Illustration of a simple confounding case: in this graphical model, given
Z, there is no association between
X and
Y. However, not observing
Z will create spurious association between
X and
Y. In the latter case,
Z is called a confounding factor.
"Confounding factor" redirects here. For other uses, see Confounding factor (disambiguation).
|
This documentation needs attention from an expert in statistics. See the talk page for details. WikiProject Statistics (or its Portal) may be able to help recruit an expert. (October 2010) |
In statistics, a confounding variable (also confounding factor, a confound, or confounder) is an extraneous variable in a statistical model that correlates (directly or inversely) with both the dependent variable and the independent variable. A perceived relationship between an independent variable and a dependent variable that has been misestimated due to the failure to account for a confounding factor is termed a spurious relationship, and the presence of misestimation for this reason is termed omitted-variable bias. In the case of risk assessments evaluating the magnitude and nature of risk to human health, it is important to control for confounding to isolate the effect of a particular hazard such as a food additive, pesticide, or new drug. For prospective studies, it is difficult to recruit and screen for volunteers with the same background (age, diet, education, geography, etc.), and in historical studies, there can be similar variability. Due to the inability to control for variability of volunteers and human studies, confounding is a particular challenge. For these reasons, experiments offer a way to avoid most forms of confounding.
As an example, suppose that there is a statistical relationship between ice-cream consumption and number of drowning deaths for a given period. These two variables have a positive correlation with each other. An evaluator might attempt to explain this correlation by inferring a causal relationship between the two variables (either that ice-cream causes drowning, or that drowning causes ice-cream consumption). However, a more likely explanation is that the relationship between ice-cream consumption and drowning is spurious and that a third, confounding, variable (the season) influences both variables: during the summer, warmer temperatures lead to increased ice-cream consumption as well as more people swimming and thus more drowning deaths.
While specific definitions may vary, in essence a confounding variable fits the following four criteria, here given in a hypothetical situation with variable of interest "V", confounding variable "C" and outcome of interest "O":
1. C is associated (inversely or directly) with O
2. C is associated with O, independent of V
3. C is associated (inversely or directly) with V
4. C is not in the causal pathway of V to O (C is not a direct consequence of V, not a way by which V produces O)
In a more concrete example, say one is studying the relation between birth order (1st child, 2nd child, etc.) and the presence of Down's Syndrome in the child. In this scenario, maternal age would be a confounding variable:
1. Higher maternal age is directly associated with Down's Syndrome in the child
2. Higher maternal age is directly associated with Down's Syndrome, regardless of birth order (a mother having her 1st vs 3rd child at age 50 confers the same risk)
3. Maternal age is directly associated with birth order (the 2nd child, except in the case of twins, is born when the mother is older than she was for the birth of the 1st child)
4. Maternal age is not a consequence of birth order (having a 2nd child does not change the mother's age)
Contents
- 1 Types of confounding
- 2 Examples
- 3 Decreasing the potential for confounding to occur
- 4 See also
- 5 References
- 6 Further reading
- 7 External links
Types of confounding
In some disciplines, confounding is categorized into different types. In epidemiology, one type is "confounding by indication",[1] which relates to confounding from observational studies. Because prognostic factors may influence treatment decisions (and bias estimates of treatment effects), controlling for known prognostic factors may reduce this problem, but it is always possible that a forgotten or unknown factor was not included or that factors interact complexly. Confounding by indication has been described as the most important limitation of observational studies. Randomized trials are not affected by confounding by indication due to random assignment.
Confounding variables may also be categorised according to their source. The choice of measurement instrument (operational confound), situational characteristics (procedural confound), or inter-individual differences (person confound).
- An operational confounding can occur in both experimental and non-experimental research designs. This type of confounding occurs when a measure designed to assess a particular construct inadvertently measures something else as well.[2]
- A procedural confounding can occur in a laboratory experiment or a quasi-experiment. This type of confound occurs when the researcher mistakenly allows another variable to change along with the manipulated independent variable.[2]
- A person confounding occurs when two or more groups of units are analyzed together (e.g., workers from different occupations), despite varying according to one or more other (observed or unobserved) characteristics (e.g., gender).[3]
Examples
In risk assessments, factors such as age, gender, and educational levels often have an impact on health status and so should be controlled. Beyond these factors, researchers may not consider or have access to data on other causal factors. An example is on the study of smoking tobacco on human health. Smoking, drinking alcohol, and diet are lifestyle activities that are related. A risk assessment that looks at the effects of smoking but does not control for alcohol consumption or diet may overestimate the risk of smoking.[4] Smoking and confounding are reviewed in occupational risk assessments such as the safety of coal mining.[5] When there is not a large sample population of non-smokers or non-drinkers in a particular occupation, the risk assessment may be biased towards finding a negative effect on health.
Decreasing the potential for confounding to occur
A reduction in the potential for the occurrence and effect of confounding factors can be obtained by increasing the types and numbers of comparisons performed in an analysis.[citation needed] If a relationship holds among different subgroups of analyzed units, confounding may be less likely. That said, if measures or manipulations of core constructs are confounded (i.e., operational or procedural confounds exist), subgroup analysis may not reveal problems in the analysis. Additionally, increasing the number of comparisons can create other problems (see multiple comparisons).
Peer review is a process that can assist in reducing instances of confounding, either before study implementation or after analysis has occurred. Peer review relies on collective expertise within a discipline to identify potential weaknesses in study design and analysis, including ways in which results may depend on confounding. Similarly, replication can test for the robustness of findings from one study under alternative study conditions or alternative analyses (e.g., controlling for potential confounds not identified in the initial study).
Confounding effects may be less likely to occur and act similarly at multiple times and locations.[citation needed] In selecting study sites, the environment can be characterized in detail at the study sites to ensure sites are ecologically similar and therefore less likely to have confounding variables. Lastly, the relationship between the environmental variables that possibly confound the analysis and the measured parameters can be studied. The information pertaining to environmental variables can then be used in site-specific models to identify residual variance that may be due to real effects.[6]
Depending on the type of study design in place, there are various ways to modify that design to actively exclude or control confounding variables:[7]
- Case-control studies assign confounders to both groups, cases and controls, equally. For example if somebody wanted to study the cause of myocardial infarct and thinks that the age is a probable confounding variable, each 67 years old infarct patient will be matched with a healthy 67 year old "control" person. In case-control studies, matched variables most often are the age and sex. Drawback: Case-control studies are feasible only when it is easy to find controls, i.e., persons whose status vis-à-vis all known potential confounding factors is the same as that of the case's patient: Suppose a case-control study attempts to find the cause of a given disease in a person who is 1) 45 years old, 2) African-American, 3) from Alaska, 4) an avid football player, 5) vegetarian, and 6) working in education. A theoretically perfect control would be a person who, in addition to not having the disease being investigated, matches all these characteristics and has no diseases that the patient does not also have — but finding such a control would be an enormous task.
- Cohort studies: A degree of matching is also possible and it is often done by only admitting certain age groups or a certain sex into the study population, creating a cohort of people who share similar characteristics and thus all cohorts are comparable in regard to the possible confounding variable. For example, if age and sex are thought to be confounders, only 40 to 50 years old males would be involved in a cohort study that would assess the myocardial infarct risk in cohorts that either are physically active or inactive. Drawback: In cohort studies, the overexclusion of input data may lead researchers to define too narrowly the set of similarly situated persons for whom they claim the study to be useful, such that other persons to whom the causal relationship does in fact apply may lose the opportunity to benefit from the study's recommendations. Similarly, "over-stratification" of input data within a study may reduce the sample size in a given stratum to the point where generalizations drawn by observing the members of that stratum alone are not statistically significant.
- Double blinding: conceals from the trial population and the observers the experiment group membership of the participants. By preventing the participants from knowing if they are receiving treatment or not, the placebo effect should be the same for the control and treatment groups. By preventing the observers from knowing of their membership, there should be no bias from researchers treating the groups differently or from interpreting the outcomes differently.
- Randomized controlled trial: A method where the study population is divided randomly in order to mitigate the chances of self-selection by participants or bias by the study designers. Before the experiment begins, the testers will assign the members of the participant pool to their groups (control, intervention, parallel), using a randomization process such as the use of a random number generator. For example, in a study on the effects of exercise, the conclusions would be less valid if participants were given a choice if they wanted to belong to the control group which would not exercise or the intervention group which would be willing to take part in an exercise program. The study would then capture other variables besides exercise, such as pre-experiment health levels and motivation to adopt healthy activities. From the observer’s side, the experimenter may choose candidates who are more likely to show the results the study wants to see or may interpret subjective results (more energetic, positive attitude) in a way favorable to their desires.
- Stratification: As in the example above, physical activity is thought to be a behaviour that protects from myocardial infarct; and age is assumed to be a possible confounder. The data sampled is then stratified by age group – this means, the association between activity and infarct would be analyzed per each age group. If the different age groups (or age strata) yield much different risk ratios, age must be viewed as a confounding variable. There exist statistical tools, among them Mantel–Haenszel methods, that account for stratification of data sets.
- Controlling for confounding by measuring the known confounders and including them as covariates in multivariate analyses such as regression analysis; however, multivariate analyses reveal much less information about the strength of the confounding variable than do stratification methods.[dubious – discuss]
All these methods have their drawbacks:
- The best available defense against the possibility of spurious results due to confounding is often to dispense with efforts at stratification and instead conduct a randomized study of a sufficiently large sample taken as a whole, such that all potential confounding variables (known and unknown) will be distributed by chance across all study groups and hence will be uncorrelated with the binary variable for inclusion/exclusion in any group.
- Ethical considerations: In double blind and randomized controlled trials, participants are not aware that they are recipients of sham treatments and may be denied effective treatments.[8] There is resistance to randomized controlled trials in surgery because patients would agree to invasive surgery which carry risks under the understanding that they were receiving treatment.
See also
- Anecdotal evidence
- Causal inference
- Epidemiological method
- Epidemiology
- Joint effect
- Simpson's paradox
References
- ^ Johnston, S. C. (2001). "Identifying Confounding by Indication through Blinded Prospective Review". Am J Epidemiol 154 (3). pp. 276–284. doi:10.1093/aje/154.3.276.
- ^ a b Pelham, Brett (2006). Conducting Research in Psychology. Belmont: Wadsworth. ISBN 0-534-53294-2.
- ^ Steg, L.; Buunk, A. P.; Rothengatter, T. (2008). "Chapter 4". Applied Social Psychology: Understanding and managing social problems. Cambridge, UK: Cambridge University Press.
- ^ Tjønneland, Anne; Morten Grønbæk, Connie Stripp and Kim Overvad (January 1999). American Society for Nutrition American Journal of Clinical Nutrition 69 (1): 49–54.
- ^ Axelson, O (1989). "Confounding from smoking in occupational epidemiology". British Journal of Industrial Medicine 46: 505–07.
- ^ Calow, Peter P. (2009) Handbook of Environmental Risk Assessment and Management, Wiley
- ^ Mayrent, Sherry L (1987). Epidemiology in Medicine. Lippincott Williams & Wilkins. ISBN 0-316-35636-0.
- ^ Emanuel, Ezekiel J; Miller, Franklin G (Sep 20, 2001). "he ethics of placebo-controlled trials--a middle ground". The New England Journal of Medicine 345 (12): 915–9.
Further reading
- Pearl, J. (1998) "Why there is no statistical test for confounding, why many think there is, and why they are almost right" UCLA Computer Science Department, Technical Report R-256, January 1998
This textbook has a nice overview of confounding factors and how to account for them in design of experiments:
- D. C. Montgomery, D.C. (2005) Design and Analysis of Experiments, 6th edition, Wiley. (Section 7-3)
External links
These sites contain descriptions or examples of confounding variables:
- Linear Regression (Yale University)
- Scatterplots (Simon Fraser University)
- Tutorial by University of New England
Statistics
|
|
Descriptive statistics
|
|
Continuous data |
Location |
- Mean (Arithmetic, Geometric, Harmonic)
- Median
- Mode
|
|
Dispersion |
- Range
- Standard deviation
- Coefficient of variation
- Percentile
- Interquartile range
|
|
Shape |
- Variance
- Skewness
- Kurtosis
- Moments
- L-moments
|
|
|
Count data |
|
|
Summary tables |
- Grouped data
- Frequency distribution
- Contingency table
|
|
Dependence |
- Pearson product-moment correlation
- Rank correlation (Spearman's rho, Kendall's tau)
- Partial correlation
- Scatter plot
|
|
Statistical graphics |
- Bar chart
- Biplot
- Box plot
- Control chart
- Correlogram
- Fan chart
- Forest plot
- Histogram
- Pie chart
- Q–Q plot
- Run chart
- Scatter plot
- Stemplot
- Radar chart
|
|
|
|
Data collection
|
|
Designing studies |
- Effect size
- Standard error
- Statistical power
- Sample size determination
|
|
Survey methodology |
- Sampling
- Stratified sampling
- Cluster sampling
- Opinion poll
- Questionnaire
|
|
Controlled experiment |
- Design of experiments
- Randomized experiment
- Random assignment
- Replication
- Blocking
- Factorial experiment
- Optimal design
|
|
Uncontrolled studies |
- Natural experiment
- Quasi-experiment
- Observational study
|
|
|
|
Statistical inference
|
|
Statistical theory |
- Sampling distribution
- Order statistic
- Scan statistic
- Record value
- Sufficiency
- Completeness
- Exponential family
- Permutation test (Randomization test)
- Empirical distribution
- Bootstrap
- U statistic
- Efficiency
- Asymptotics
- Robustness
|
|
Frequentist inference |
- Unbiased estimator (Mean unbiased minimum variance, Median unbiased)
- Biased estimators (Maximum likelihood, Method of moments, Minimum distance, Density estimation)
- Confidence interval
- Testing hypotheses
- Power
- Parametric tests (Likelihood-ratio, Wald, Score)
|
|
Specific tests |
- Z (normal)
- Student's t-test
- F
- Goodness of fit (Chi-squared, G, Sample source, sample normality, Skewness & kurtosis Normality, Model comparison, Model quality)
- Signed-rank (1-sample, 2-sample, 1-way anova)
- Shapiro–Wilk
- Kolmogorov–Smirnov
|
|
Bayesian inference |
- Bayesian probability
- Prior
- Posterior
- Credible interval
- Bayes factor
- Bayesian estimator
- Maximum posterior estimator
|
|
|
|
Correlation and regression analysis
|
|
Correlation |
- Pearson product–moment correlation
- Partial correlation
- Confounding variable
- Coefficient of determination
|
|
Regression analysis |
- Errors and residuals
- Regression model validation
- Mixed effects models
- Simultaneous equations models
- MARS
|
|
Linear regression |
- Simple linear regression
- Ordinary least squares
- General linear model
- Bayesian regression
|
|
Non-standard predictors |
- Nonlinear regression
- Nonparametric
- Semiparametric
- Isotonic
- Robust
- Heteroscedasticity
- Homoscedasticity
|
|
Generalized linear model |
- Exponential families
- Logistic (Bernoulli)
- Binomial
- Poisson
|
|
Partition of variance |
- Analysis of variance (ANOVA)
- Analysis of covariance
- Multivariate ANOVA
- Degrees of freedom
|
|
|
|
Categorical, multivariate, time-series, or survival analysis
|
|
Categorical data |
- Cohen's kappa
- Contingency table
- Graphical model
- Log-linear model
- McNemar's test
|
|
Multivariate statistics |
- Multivariate regression
- Principal components
- Factor analysis
- Cluster analysis
- Classification
- Copulas
|
|
Time series analysis |
General |
- Decomposition
- Trend
- Stationarity
- Seasonal adjustment
- Exponential smoothing
- Cointegration
|
|
Specific tests |
- Granger causality
- Q-Statistic
- Durbin–Watson
|
|
Time domain |
- ACF
- PACF
- XCF
- ARMA model
- ARIMA model
- ARCH
- Vector autoregression
|
|
Frequency domain |
- Spectral density estimation
- Fourier analysis
- Wavelet
|
|
|
Survival analysis |
- Survival function
- Kaplan–Meier
- Logrank test
- Failure rate
- Proportional hazards models
- Accelerated failure time model
|
|
|
|
Applications
|
|
Biostatistics |
- Bioinformatics
- Clinical trials & studies
- Epidemiology
- Medical statistics
|
|
Engineering statistics |
- Chemometrics
- Methods engineering
- Probabilistic design
- Process & Quality control
- Reliability
- System identification
|
|
Social statistics |
- Actuarial science
- Census
- Crime statistics
- Demography
- Econometrics
- National accounts
- Official statistics
- Population
- Psychometrics
|
|
Spatial statistics |
- Cartography
- Environmental statistics
- Geographic information system
- Geostatistics
- Kriging
|
|
|
|
- Category
- Portal
- Outline
- Index
|
|