What Is Internal Validity? | Definition, Example & Threats
Internal validity refers to the extent to which a research study’s design and methods minimize the likelihood of alternative explanations for the observed effect between variables.
In other words, internal validity addresses the question: “Is the observed effect or relationship likely due to the independent variable (the variable being manipulated) and not due to other factors?”
A high level of internal validity means that the study’s conclusions are likely to be reliable. It’s one of the most important types of validity in research.
What is internal validity?
Internal validity refers to the extent to which a study’s design and methodology ensure that any observed relationships or effects between variables are due to the independent variable being manipulated and not to other factors.
A study with high internal validity is considered to be strong because it’s able to isolate the effect of the independent variable and rule out alternative explanations. It increases the chance that the study is free from flaws and biases that could affect the results.
If your study has low internal validity, you can’t claim a cause-and-effect relationship between your variables because you can’t rule out other explanations.
Internal vs external validity
The degree of both internal and external validity influence the interpretation of study findings:
- Internal validity refers to the extent to which a study’s design and methods allow researchers to infer causality between variables.
- External validity, on the other hand, refers to the extent to which the results of a study can be generalized to other populations, settings, and situations beyond the specific study.
Internal validity | External validity | |
---|---|---|
Goal | Establish causality between variables | Generalize results to other contexts |
Main threat | Extraneous variables | Sampling bias |
Measures to increase validity | Randomization, control groups, and blinding | Representative probability sampling, larger sample size, and naturalistic settings |
Trade-off between internal and external validity
There’s always a trade-off between internal and external validity. The more you control for extraneous variables (internal validity), the less you can generalize your results to different populations or contexts (external validity).
Internal validity example
There are several factors that can harm or strengthen a study’s internal validity.
Low internal validity example
High internal validity example
Internal validity threats and solutions
It’s essential to recognize and solve threats to internal validity in any research design. There are different threats for single-group and multi-group studies.
Internal validity threats for single-group studies
A single-group study involves a single group of participants who are assessed at one point in time and then again at a later point in time (pre-post design). The goal is to examine the change within the same group over time.
Internal validity threat | Explanation | Example |
---|---|---|
Maturation | The research findings vary as a natural result of time. | Most students just started college at the time of the pre-test. At the time of the post-test, they’re much more familiar with college life and feel less anxious. |
Instrumentation | Outcomes differ because of the use of different measures in pre- and post-test. | In the pre-test, the anxiety was measured for 45 minutes. In the post-test, the anxiety was measured for only 15 minutes. This may have affected the outcomes. |
Testing | Results of the post-test are influenced by the pre-test. | Students show decreased anxiety at the end of the research because the same test was administered. Students achieved different results because they became aware of the study’s aim. |
History | Outcomes are affected by an unrelated event. | Two weeks before the end of the research, the students are told the rules for graduation have become more strict. The students’ anxiety has increased, which might influence the outcomes during the post-test evaluation. |
Solving threats in single-group studies
Making changes to the research design can counter threats to internal validity.
- Use of a control group. Single-group studies are often used in a quasi-experimental design when it’s not feasible or ethical to conduct randomized controlled trials. Although you might not be able to randomly assign participants to a control group, you can use existing groups as controls. For example, you can compare the outcomes of participants who have received an intervention or treatment with those of people who haven’t.
- Use of filler-tasks. You can use filler-tasks (i.e., random other questions that have nothing to do with the purpose of your study) to obscure the goal of your research. This approach counters demand characteristics, which inadvertently prompt participants to respond in a certain way, and testing threats.
- Use a larger sample size. A larger sample provides more precise estimates of the effect, even if assumptions about the data are not fully met (e.g., homoscedasticity, which is the assumption of similar variances in different groups, and normal distribution). This can help reduce the impact of random error and measurement error. It also makes it less likely that outliers will have a disproportionate impact on the results, which is particularly important in single-group studies with no control group.
Internal validity threats for multi-group studies
A multi-group study involves multiple groups or populations that are compared to each other. This design allows researchers to examine differences between groups, identify potential sources of variation, and control for extraneous factors.
Internal validity threat | Explanation | Example |
---|---|---|
Regression to the mean | The statistical tendency for participants with extremely low or high scores on a test to score closer to the mean the next time. | The students were assigned to a group based on their pre-test score. It’s difficult to attribute any change in outcomes to the teaching method instead of statistical norms. |
Social desirability and social interaction | Participants from different groups can interact with each other or researchers and either figure out the goal of the study, feel resentful of or better than others, or feel pressured to perform a certain way. | Group A gets to play a fun math game, whereas the other groups don’t. Students from Group B or Group C might resent the students from Group A for getting to play a game. This might be demotivating, which in turn could lead to poor performance. |
Attrition bias | Bias caused by dropout of participants | Many of the Group A students provided unusable data. This makes it hard to compare the results of the two treatment groups to the control group. |
Selection bias | Groups differ at the beginning of the study | Students with high math skills were placed in Group A, whereas students with low math scores were placed in Group B. This caused the groups to be systematically different from the start. Because of these differences, any observed change in test scores might be due to other reasons than the manipulation of the independent variable (teaching method). |
Solving threats in multi-group studies
Making changes to the research design can counter threats to internal validity.
- Blinding. A single-blind, double-blind, or even triple-blind research design can help you counter the effects of social interaction and desirability. If participants, researchers, and data collectors are unaware of the treatment a participant has received, they’re less likely to influence the behavior and outcomes.
- Random assignment (randomization). By randomly assigning participants to groups, you counter the effects of selection bias and regression to the mean because you’re making sure groups are comparable at the beginning of the study.
How to assess internal validity
There are three conditions that need to be met for internal validity for you to establish a cause-and-effect relationship between an independent variable (the one you manipulate) and the dependent variable (the one you measure).
- Your manipulation of the independent variable (treatment variable) precedes any changes in your dependent variable (outcome variable or response variable).
- Your independent variable and dependent variable change together.
- There are no extraneous variables or confounding variables that can explain the research findings.
Frequently asked questions about internal validity
- What is the difference between construct validity and internal validity?
-
Construct validity refers to the extent to which a study measures the underlying concept or construct that it is supposed to measure.
Internal validity refers to the extent to which observed changes in the dependent variable are caused by the manipulation of the independent variable rather than other factors, such as extraneous variables or research biases.
- What are the 12 threats to internal validity?
-
The 12 main threats to internal validity are:
- History: Changes in the environment or events that occur outside of the study can affect the outcome.
- Maturation: Changes in the participants over time (e.g., age, skill level) can affect the outcome.
- Testing: The act of testing or measurement itself can affect the outcome (testing effect, practice effect, or carryover effect).
- Instrumentation: Changes in the measuring instrument or tool used to collect data can affect the outcome.
- Statistical regression to the mean: The tendency of extreme scores to regress towards the mean, which can lead to a loss of statistical significance.
- Selection: The selection of participants for the study can affect the outcome (selection bias), especially in the case of non-probability sampling.
- Experimental mortality or attrition bias: The loss of participants or dropouts during the study can affect the outcome.
- Multiple-treatment interference: The interaction between different treatments or conditions can affect the outcome.
- Social desirability bias: The participants’ awareness of being in a study and their desire to be well-liked by researchers can affect the outcome.
- Social interaction: The participants’ awareness of being treated differently than people in other groups can affect the outcome.
- Residual confounding: The presence of unmeasured or uncontrolled extraneous or confounding variables that affect the outcome and are not accounted for in the analysis.
- Order effect: The order of the independent variable levels affects the dependent variable.
There are several ways to counter these threats to internal validity, for example, through randomization, the addition of control groups, and blinding.