Homepage Epidemiology Biostatistics BIAS & CONFOUNDING M.Tevfik DORAK Glossary of Bias: Delgado-Rodriguez & Llorca, 2004 Introduction In epidemiologic research, it is essential to avoid bias, to control confounding and to undertake accurate replication. Bias, confounding and random variation/chance are the non-causal reasons for an association between an exposure and outcome. These major threats to internal validity of a study should always be considered as alternative explanations in the interpretation. Bias is a mistake of the researcher; confounding is not a mistake and when obvious, it can be controlled; replication is up to the researcher. Bias is defined as 'any systematic error in an epidemiologic study that results in an incorrect estimate of the association between exposure and risk of disease.' Two major classes of bias are: 1. Selection bias 2. Observation/information (misclassification) bias Selection bias is an important problem in case-control and retrospective cohort studies while it is unlikely to occur in a prospective cohort study. The observation bias group include recall bias, interviewer bias, follow-up bias and misclassification bias. Confounding is sometimes referred to as the third major class of bias. It is a function of the complex interrelationships between various exposures and disease. Confounding can be controlled in the design (randomisation, restriction and matching) and in the analysis (stratification, multivariate analysis and matching). The best that can be done about unknown confounders is to use a randomised design. Bias and confounding are not affected by sample size but chance effect (random variation) diminishes as sample size gets larger. A small P value and a narrow odds ratio/relative risk are reassuring signs against chance effect but the same cannot be said for bias and confounding. Types of bias Chance findings are caused by random variation but bias is caused by systematic variation. Potential sources of bias should be eliminated or minimised through rigorous design and meticulous conduct of a study. Each analytic study design has particular types of bias to which it is most vulnerable. In casecontrol studies, selection bias (knowledge of exposure status influences the identification of diseased and nondiseased study subjects) and recall bias (knowledge of disease status influences the determination of exposure status) are most important. In cohort studies, bias due to loss to follow-up (attrition) would be the greatest danger (and selection bias in retrospective studies). The potential for misclassification is present in all types of epidemiologic studies. Despite all preventive efforts, bias should always be considered among alternative explanations of a finding. It has to be remembered that bias may mask an association or cause a spurious one; and it may cause over or underestimation of the effect size. A study that suffers from bias lacks internal validity. The list presented below may help to consider potential sources of bias in planning, executing, and interpreting a study. Following the classic paper by Sackett (1979), biases are classified according to stages of research: * Literature Review - Foreign language exclusion bias - Literature search bias - One-sided reference bias - Rhetoric bias * Study Design - Selection bias - Sampling frame bias Berkson (admission rate) bias Centripetal bias Diagnostic access bias Diagnostic purity bias Hospital access bias Migrator bias Prevalence-incidence (Neyman / selective survival; attrition) bias Telephone sampling bias - Nonrandom sampling bias Autopsy series bias Detection bias Diagnostic work-up bias Door-to-door solicitation bias Previous opinion bias Referral filter bias Sampling bias Self-selection bias Unmasking bias - Noncoverage bias Early-comer bias Illegal immigrant bias Loss to follow-up (attrition) bias Response bias Withdrawal bias - Noncomparability bias Ecological (aggregation) bias Healthy worker effect (HWE) Lead-time bias Length bias Membership bias Mimicry bias Nonsimultaneous comparison bias Sample size bias * Study Execution Bogus control bias Contamination bias Compliance bias * Data Collection - Instrument bias Case definition bias Diagnostic vogue bias Forced choice bias Framing bias Insensitive measure bias Juxtaposed scale bias Laboratory data bias Questionnaire bias Scale format bias Sensitive question bias Stage bias Unacceptability bias Underlying/contributing cause of death bias Voluntary reporting bias - Data source bias Competing death bias Family history bias Hospital discharge bias Spatial bias - Observer bias Diagnostic suspicion bias Exposure suspicion bias Expectation bias Interviewer bias Therapeutic personality bias - Subject bias Apprehension bias Attention bias (Hawthorne effect) Culture bias End-aversion bias (end-of-scale or central tendency bias) Faking bad bias Faking good bias Family information bias Interview setting bias Obsequiousness bias Positive satisfaction bias Proxy respondent bias - Recall bias Reporting bias Response fatigue bias Unacceptable disease bias Unacceptable exposure bias Underlying cause (rumination bias) Yes-saying bias - Data handling bias Data capture error Data entry bias Data merging error Digit preference bias Record linkage bias * Analysis - Confounding bias Latency bias Multiple exposure bias Nonrandom sampling bias Standard population bias Spectrum bias - Analysis strategy bias Distribution assumption bias Enquiry unit bias Estimator bias Missing data handling bias Outlier handling bias Overmatching bias Scale degradation bias - Post hoc analysis bias Data dredging bias Post hoc significance bias Repeated peeks bias * Interpretation of Results Assumption bias Cognitive dissonance bias Correlation bias Generalisation bias Magnitude bias Significance bias Underexhaustion bias * Publication All's well literature bias Positive result bias Hot topic bias Potential effects of sources of bias may be: 1. Positive bias: The observed measure of effect (eg, odds ratio) is larger than the true measure of effect (applies to both protective and risk associations) 2. Negative bias: The observed measure of effect is smaller than the true measure of effect 3. Toward the null: The observed measure of effect is closer to 1.0 than the true measure of effect 4. Away from the null: The observed measure of effect is farther from 1.0 than the true measure of effect Most important biases: Prevalence-incidence bias: Selective survival may be important in some conditions. For these diseases (such as cancer, HIV infection) the use of prevalent instead of incident cases usually distorts the measure of effect. The frequency of glutathione S-transferase class mu (GSTM) deletion is for example different in incident (newly and sequentially diagnosed) cases and prevalent (all patients at a point in time) cases (Kelsey, 1997). Berkson (admission rate) bias: Where cases/controls are recruited from among hospital patients, the characteristics of these groups will be influenced by hospital admission rates. Healthy Worker Effect (HWE): The overall mortality experience of an employed population is typically better than that of the general population (in Western countries at least). Use of blood donors as controls is a kind of HWE. Blood donors are self-selected on the basis of better life styles. Detection bias: The risk factor investigated itself may lead to increased diagnostic investigations and increase the probability that the disease is identified in that subset of persons. An example is women with benign breast diseases who undergo detailed follow-up programs which would detect cancer at early stages (Silber & Horwitz, 1986). Recall bias: Recall bias is caused by differences in accuracy of recalling past events by cases and controls. There is a tendency for diseased people (or their relatives) to recall past exposures more efficiently than healthy people (selective recall). For example, because women with breast cancer are more likely to remember a positive family history than control subjects, retrospective study designs are likely to overestimate the effect size of family history as a risk factor. This bias is avoided by prospective studies, and indeed the risk estimates from prospective cohorts are less than those for other types of study. Another example is the mothers of leukaemic children who would remember even trivial exposures. Family information bias: Within a family, the flow of information about exposures and diseases is stimulated by a family member who develops the disease. A person who develops a disease is more likely than his or her unaffected siblings to know that a parent has a history of the disease. Non-respondent bias: Non-respondents to a survey often differ from respondents. Volunteers also differ from non-volunteers, late respondents from early respondents, and study dropouts from those who complete the study. Misclassification bias: This is systematic distortion of estimates resulting from inaccuracy in measurement of classification of study variables. Non-differential misclassification generally dilutes the exposure effect (toward to null effect) (Copeland, 1977). It is worse when the proportions of subjects misclassified differ between the study groups (differential misclassification). Such a differential between cases and controls may mask an association or cause one when there is none. This type of misclassification is rare when exposures are recorded before the outcome is known. This bias usually results from deeper investigation or surveillance of cases. Typical sources of misclassification/information bias are: - variation among observers and among instruments - variation in the underlying characteristic - misunderstanding of questions by study subjects (interview or questionnaire) - incomplete or inaccurate record data Selection bias due to missing data: When there are a large number of variables, the regression procedure excludes an entire observation if it is missing a value for any of the variables (listwise deletion). This may result in exclusion of a considerable percentage of observations and induce selection bias. Regression to mean: This is an example of how random variability can lead to systematic error (Davis, 1976). An example would be a follow-up study of people with highly elevated cholesterol levels. During follow-up, part of reduction in cholesterol levels would be due to regression to the mean rather than drug of life modification effects. Because the initial very high level was partly because of a large positive random component (and of course, some were truly very high). This is an information bias mainly concerning cohort studies. End-aversion bias (end-of-scale or central tendency bias): In questionnaire-based surveys, respondents usually avoid ends of scales in their answers. They tend to try to be conservative and wish to be in the middle. Overmatching bias: When cases and controls are matched by a non-confounding variable (associated to the exposure but not to the disease), this is called overmatching. Overmatching can underestimate an association. Competing death bias: As each person will only die once, if there are mutually exclusive causes of death, they compete with each other in the same subject (Chiang, 1991). For example, in parenteral drug users, liver failure and AIDS are competing causes of death and may influence any research on either subject. Likewise, the apolipoprotein E genotype is associated with cardiovascular disease mortality and Alzheimer's disease; AD and death are competing risks involving apolipoprotein E genotype frequency changes with old age (Corder, 1995). Publication bias: Editors and authors tend to publish articles containing positive findings as opposed to negative result papers. This results in a belief that there is a consistent association while this may not be the case. Plots of relative risks by study may be used to check publication bias in meta-analyses. If publication bias is operating, one would expect that, of published studies, the larger ones report the smaller effects, as small positive trials are more likely to be published than negative ones. This can be examined using the funnel plot in which the effect size is plotted against sample size (Sterne & Egger, 2001). If this is done, the plot resembles an inverted funnel, with the results of the smaller studies being more widely scattered than those of the larger studies, as would be expected if there is no publication bias. Confounding Bias involves error in the measurement of a variable; confounding involves error in the interpretation of what may be an accurate measurement. Confounding in epidemiology is mixing of the effect of the exposure under study on the disease (outcome) with that of a third factor that is associated with the exposure and an independent risk factor for the disease (even among individuals nonexposed to the exposure factor under study). The consequence of confounding is that the estimated association is not the same as true effect. In contrast to other forms of bias, in confounding bias the actual data collected may be correct but the subsequent effect attributed to the exposure of interest is actually caused by something else. Classic example of confounding is the initial association between alcohol consumption and lung cancer (confounded by smoking, which is associated with alcohol use and an independent risk factor for lung cancer). Likewise, an association between gambling and cancer is confounded by at least smoking and alcohol. Confounding can cause overestimation or underestimation of the true association and may even change the direction of the observed effect. An example is the confounding by age of an inverse association between level of exercise and heart attacks (younger people having more rigorous exercise) causing overestimation. The same association can also be confounded by sex (men having more rigorous exercise) causing underestimation of the association. For a variable to confound an association, it must be associated both with the exposure and outcome, and its relation to the outcome should be independent of its association with the exposure (i.e., not through its association with the exposure). Confounding factor should not be an intermediate link in the causal chain between the exposure and disease under study. Age and sex are associated with virtually all diseases and are related to the presence or level of many exposures. Even though they act as surrogates for etiologic factors most of the time, age and sex should always be considered as potential confounders of an association. Confounders can be positive or negative. Positive confounders cause overestimation of an association (which may be an inverse association), and negative confounders cause underestimation of an association. It is not easy to recognise confounders. A practical way to achieve this is to analyse the data with and without controlling for the potential confounders. If the estimate of the association differs remarkably when controlled for the variable, it is a confounder and should be controlled for (by stratification or multivariate analysis). To be able to do this, investigators should make every effort to obtain data on all available risk factors for the disease under study. A factor can confound an association only if it differs between the study groups. Therefore, in a casecontrol study, for age and sex to be confounders, their representation should sufficiently differ between cases and controls. This is the basis of methods to control confounding in the design. Randomisation (ensures that potential confounding factors, known or unknown, are evenly distributed among the study groups), restriction (restricts admission to the study to a certain category of a confounder) and matching (equal representation of subjects with certain confounders among study groups) can overcome a great deal of confounding. The protection against confounders obtained by randomisation will only be maintained if all participants remain in the group to which they were allocated and no systematic loss to follow-up occurs. To avoid this possibility, it is best to try to maximise follow up and carry out an intention to treat analysis (Hollis & Campbell, 1999). In practice, restriction ensures comparisons to be performed only between observations that have the same value of the confounder (for example only white men); and matching ensures comparisons between groups that have the same distribution of the confounder (frequency matching or one-to-one matching). In addition to the extra effort, time, money and loss of potential study subjects, the more important disadvantage of restriction and matching is the inability to evaluate the effect of the variable used for restriction or matching. One other problem with matching is that in the analysis, the effective sample size is reduced because the analyses are based on only discordant pairs. Despite the use of these methods, at the analysis stage of a study, it may still be necessary to control for residual confounding. For example if a study of heart attacks restricted the entry to 40-65 year-old subjects, there may still be an age effect (residual confounding) within this range. Methods used to control for confounding at the analysis stage include stratified analysis (unable to control simultaneously for even a moderate number of potential confounders) and multivariable analysis (can control for a number of confounding factors simultaneously as long as there are at least ten subjects for every variable investigated -in a logistic regression situation-). If matching was done, then the most common analysis methods would be McNemar test or conditional logistic regression. Glossary of Bias: Delgado-Rodriguez & Llorca, 2004 References - Armitage P & Colton T. Encyclopedia of Biostatistics. Volumes 1-6. John Wiley & Sons, 1998 (Bias in Observational Studies section by HA Hill & DG Kleinbaum) - Bandolier Bias Guide - Bias in Clinical Research (exercises) - Bias & Effect Modification Presentation - Boffetta P. Molecular epidemiology. J Intern Med 2000;248:447-54 - Bradford Hill A & Hill ID: Bradford Hill's Principles of Medical Statistics. London: Edward Arnold. 12th Edition, 1991 - Davey Smith G & Ebrahim S. Data dredging, bias, or confounding. BMJ 2002;325: 1437–8 - Epidemiology for the Uninitiated - Greenland S & Morgenstern H. Confounding in health research. Annu Rev Public Health 2001;22:189212 (PDF) - Hennekens CH, Buring JE, Mayrent SL (Eds). Epidemiology in Medicine. Boston: Little, Brown and Company. 1987 (Amazon) - Kaptchuk TJ. Effect of interpretive bias on research evidence. BMJ 2003;326:1453-5 - Katz MH. Multivariable Analysis. Ann Intern Med 2003;138:644-50 - Potter JD. Epidemiology, cancer genetics and microarrays: making correct inferences, using appropriate designs. Trends Genet 2003;19:690-4 - Sackett DL. Bias in analytic research. J Chronic Dis 1979;32:51-63 - Schoenbach VJ & Rosamond WD. Understanding the Fundamentals of Epidemiology - Taioli E & Garte S. Covariates and confounding in epidemiologic studies using metabolic gene polymorphisms. Int J Cancer 2002;100:97-100 - Vineis P & McMichael AJ. Bias and confounding in molecular epidemiological studies. Carcinogenesis 1998;19:2063-7 See also: Choi & Noseworthy, 1992; Grimez & Schulz, 2002 Epidemiology Links Extensive Epidemiology and Biostatistics Links Online Dictionary of Epidemiology Epidemiology & BioStatistics Super Lectures Epidemiolog.Net Online Epidemiology Textbook Basic Statistics and Trial Design for Clinicians Bias & Confounding in Clinical Trials Radiology Statistical Concepts Series Article: Bias A Catalog of Biases in Questionnaires (CDC) Error, Bias & Confounding Presentation (WHO) M.Tevfik Dorak, M.D., Ph.D. Last updated on 18 April 2005 Homepage Epidemiology Biostatistics