757
Views
55
CrossRef citations to date
0
Altmetric
Methodology

Control of confounding in the analysis phase – an overview for clinicians

, , , &
Pages 195-204 | Published online: 31 Mar 2017

Abstract

In observational studies, control of confounding can be done in the design and analysis phases. Using examples from large health care database studies, this article provides the clinicians with an overview of standard methods in the analysis phase, such as stratification, standardization, multivariable regression analysis and propensity score (PS) methods, together with the more advanced high-dimensional propensity score (HD-PS) method. We describe the progression from simple stratification confined to the inclusion of a few potential confounders to complex modeling procedures such as the HD-PS approach by which hundreds of potential confounders are extracted from large health care databases. Stratification and standardization assist in the understanding of the data at a detailed level, while accounting for potential confounders. Incorporating several potential confounders in the analysis typically implies the choice between multivariable analysis and PS methods. Although PS methods have gained remarkable popularity in recent years, there is an ongoing discussion on the advantages and disadvantages of PS methods as compared to those of multivariable analysis. Furthermore, the HD-PS method, despite its generous inclusion of potential confounders, is also associated with potential pitfalls. All methods are dependent on the assumption of no unknown, unmeasured and residual confounding and suffer from the difficulty of identifying true confounders. Even in large health care databases, insufficient or poor data may contribute to these challenges. The trend in data collection is to compile more fine-grained data on lifestyle and severity of diseases, based on self-reporting and modern technologies. This will surely improve our ability to incorporate relevant confounders or their proxies. However, despite a remarkable development of methods that account for confounding and new data opportunities, confounding will remain a serious issue. Considering the advantages and disadvantages of different methods, we emphasize the importance of the clinical input and of the interplay between clinicians and analysts to ensure a proper analysis.

Introduction

During the era of modern epidemiology, we have seen large health care databases and registries emerging, contemporary with technological achievements in computing, which has paved the way for a remarkable increase in observational studies. Confounding is the concept of comparability in observational studies, which hampers causal inference.Citation1Citation3 Confounding arises when a factor is associated with both the exposure (or treatment) and the outcome, eg, a disease or death, and is not part of the causal pathway from exposure to outcome. Hence, if we study the effect of hypertension on the risk of stroke, we cannot just compare hypertensive people against people without hypertension. The reason is that we may obtain spurious results, if we do not consider confounding factors, such as smoking, diabetes, alcohol intake and cardiovascular diseases, that are likely associated with both stroke and hypertension and are not on the causal pathway from hypertension to stroke. The effect on odds ratio (OR) estimates when controlling for these confounding factors was illustrated in a UK registry-based case–control study that among other things examined the association between stroke and untreated hypertension. The estimated OR for the association of interest increased from 2.9 to 3.5 after controlling for confounding.Citation4 In this example, the magnitude of the association was underestimated, if not adjusting for confounding; however, confounding may also result in an overestimated effect, if not accounted for.

Once a potential confounding problem has been recognized, it may be dealt with in the design or the analysis phase.Citation5 Standard methods used in the design phase involve randomization, restriction and matching. In randomized studies, patients are assigned randomly to exposure categories. Restriction means that only subjects with certain values for the potential confounders are selected (eg, certain sex and age groups), while matching involves the selection of the groups to be compared (exposed vs not exposed or cases vs controls) to be comparable with respect to the distribution of potential confounders. In registry-based observational studies, it is often insufficient to control for confounding only during the design phase of the study. Usually, we wish to account for several potential confounders, which may not be possible by either restriction or matching. For example, by restriction, we may end up with a very small cohort, limiting both the precision and the generalizability of the results of the analysis. Likewise, matching on several potential confounders may reduce the likelihood of finding comparison people for the people in the patient cohort. The two approaches are not mutually exclusive. In the UK example with stroke, diagnosed cases were matched with a group of controls with the same sex and age in the design phase, and then, a multivariable regression analysis was performed in which hypertension status and the potential confounders were incorporated in the analysis phase.Citation4 In general, the control of confounding may involve design, analytical, and statistical concepts by which we can perform statistical adjustment, restructure data, remove certain observations or add comparison groups with certain characteristics (negative controls) to deal with confounding.Citation6,Citation7

In the history of control of confounding during the analysis, we have advanced from simple stratification with only a few potential confounders collated from small manageable hospital files to complex modeling procedures, using high-dimensional propensity scores (HD-PSs) by which hundreds of potential confounders are extracted from large health care databases.

The questions are, however, what have we achieved by this change in the setting of epidemiological research? Did we lose important aspects in the analysis? Are novel analysis methods to control for confounding that have become widely used in recent years, such as PS methods, our main response to the confounding issue in large and complex data sets? In the present article, we attempted to answer these questions, focusing on a registry-based setting. We considered the topic from a hands-on perspective and tried to demystify the control for confounding during analysis by explaining and discussing the nature of the various methods and referring to examples from epidemiological studies.

From the simple to the complex – stratification, standardization and multivariable analysis

Stratification

Stratification is the starting point in many textbooks dealing with confounding in the analysis phase.Citation8,Citation9 This is probably due to the simplicity of this method in which a data set is broken into a manageable number of subsets, called strata, corresponding to the levels of potential confounders (eg, age groups and sex). By comparing the overall cross-tabulation for the association between an exposure and an outcome (eg, a 2×2 table for alcohol consumption and myocardial infarction [MI]) with stratum-specific (eg, age group) cross-tabulations, it becomes evident whether a factor introduces confounding in the analysis. Thus, the stratum-specific associations (eg, measured as ORs) would deviate markedly from the overall association – refer the example in the study by MannocciCitation10 on the confounding effect of age on the association between alcohol consumption and MI. Age was a confounder since it was associated with alcohol consumption (alcohol consumption was most frequent among younger people) and with MI (MI was most common among the middle-aged people). The Mantel–Haenszel methodCitation11 is commonly used to deal with confounding using stratification. The method summarizes the stratum-specific ORs by using a weighted average of them. This approach is generally attractive because of its applicability to a number of epidemiological measures such as OR, risk difference, risk ratio and incidence rate difference.Citation9,Citation10

Stratification is an attractive method because of its simplicity; however, there are limitations to the number of factors that can be stratified, so that information can be extracted from the analysis.Citation9 For example, 10 dichotomous factors would result in 2Citation10=1,024 strata, and some strata may contain little or no data. In epidemiological research, we are expected to build on the current knowledge base and select numerous potential confounders, previously recognized, from the wealth of data that are potentially accessible from registries. Hence, when we attempt to control for confounding in the analysis, we will soon face the limitations of the stratification method regarding the number of potential confounders that are practically manageable. Stratification is therefore rarely used exclusively to control for confounding in studies that emanate from large health care databases. These days, it is used as an assisting tool in combination with other methods, and stratification may be used to identify effect measure modifications, ie, to demonstrate that the strength of the association between an exposure and an outcome depends on the value of another factor.

Standardization

Standardization provides another tool that can cope with confounding, although hampered by some of the same constraints as in stratification. Typically, disease or death rates are only standardized to age, and perhaps to sex and race, even in large registry-based studies. If more factors are considered, then separate analyses must be undertaken for specific subgroups. While stratification of confounders relies on information at the individual level in a study population, standardization involves the use of a reference population, obtained either from the data set or from an external source, such as data from a larger geographical scale. As an example, in a study based on the Korean Stroke Registry, age- and sex-standardized mortality ratios in stroke patients were calculated and compared across reasons for stroke, using the overall Korean population in 2003 as the reference population.Citation12

There are two main approaches that handle confounding by standardization: direct and indirect standardization, resulting in adjusted rates and standardized ratios. Detailed descriptions of the two methods can be found in most introductory textbooks to epidemiology (eg, Kirkwood and SternCitation13). In general, direct standardization is recommended, because the consistency of comparisons is maintained, ie, a higher rate in one study population compared to another will be preserved also after direct standardization. That said, the very rate is dependent on characteristics of the selected reference population.Citation14 When unstable rates are encountered across strata, eg, because of small numbers of patients in each stratum, indirect standardization should also be considered.Citation13 In the example from Korea, indirect standardization was used to show that the standardized mortality ratios were higher among patients with unknown stroke etiology compared to patients with known etiology.Citation12

Multivariable analysis

Multivariable regression analysis has been one of the most frequently used methods to control for confounding, and the use of this approach was particularly enhanced at a time when modeling tools were made readily available. With multivariable analysis, we get around the main limitation of stratification, as we obtain the possibility to adjust for many confounding variables in just one (assumed true) model.Citation15 Thus, we can take advantage of more of the information available in a registry than when we use stratification. In epidemiology, multivariable analysis is typically seen in analyses in which ORs or hazard ratios (HRs) are estimated. Control for confounding by multivariable analysis relies on the same principles as stratification, ie, the factors of interest (eg, a risk factor, treatment or exposure) are investigated while the potential confounders are held constant. In multivariable analysis, this is done mathematically in one integrated process, however, under certain assumptions () – here as an example of linearity for linear models. This assumption may be compromised when confounders with nonlinear effects are incorporated in a linear model as continuous variables. This leads to residual confounding (confounding remains despite controlled for in the analysis) unless other measures are taken (refer the study by Groenwold et alCitation16 for examples and solutions).

Selection of potential confounders for multivariable models has been the subject of controversy.Citation17 Confounder selection would typically rely on prior knowledge,Citation18 possibly supported by a directed acyclic graph (DAG), that is a graphical depiction of the causal relationship between, eg, an exposure and an outcome together with potential confounders.Citation6 In large study populations, the researcher would in many cases include all known measured potential confounders in the regression model. In a registry-based German study, 16 potential confounders were included in the analysis of the effect of treatment with tissue plasminogen activator (t-PA) on death (361 cases) among 6,269 ischemic stroke patients.Citation19 There was indeed a remarkable drop in the OR between t-PA and death derived from a multivariable model, when adjusting for confounding (OR=1.93 compared to OR=3.35 in the crude, unadjusted analysis). Such generous inclusion of potential confounding factors in the multivariable model is unlikely to be a problem in this example, given that there are >20 outcome events (deaths) per factor included in the model.Citation20 Factors may be omitted from a multivariable model based on preliminary data-driven procedures, such as stepwise selection, change-in-estimate procedure, least absolute shrinkage and selection operator (LASSO),Citation21 and model selection based on information criteria (eg, Akaike information criterion).Citation22 It is important to recognize that data-driven variable selection is not related to the presence of confounding factors in the data set, and hence, there is a possibility that important confounders are discarded during such procedures.Citation23

Modifications of the multivariable model have been developed to better comply with the underlying assumptions or to avoid discarding variables. These include transformations of variables,Citation16 shrinkage of parameter estimatesCitation23 and random coefficient regression models.Citation24 Despite great flexibility when exploring associations between an exposure and an outcome while controlling for potential confounders, multivariable analysis does not directly identify whether a factor is a true confounder. Therefore, it is not clear whether residual confounding remains in the model.Citation25

PS – our main response to confounding?

In recent years, PS methods have become very popular as an approach to deal with confounding in observational studies. The idea of this method is to modify the study so that exposure or treatment groups that we want to compare become comparable without influence from confounding factors.Citation26 In a cohort study, we want to get rid of confounding due to factors measured at baseline – typically defined as the period before a drug use or treatment of interest. Already in the early history of modern epidemiology, stratification by a multivariate confounder score was recognized as an attractive approach.Citation27 This is comparable to the PS approach, as it combines information on a number of variables (potential confounders) into a single score for each individual person in a data set. This score is equivalent to the probability of an exposure, given the characteristics measured at baseline. There are four conceptual steps in the PS methods: 1) selection of potential confounders; 2) estimation of the PS; 3) use of the PS to make treatment/exposure groups comparable (covariate balance) and assessment of group comparability and 4) estimation of the association between treatment/exposure and outcome.

In the German study on stroke patients mentioned earlier, PS methods were applied in addition to a multivariable analysis.Citation19 We will use the setting from this example to outline the principles of the underlying conceptual steps; further details on the methods can be found elsewhere.Citation26 In the example, we would start estimating the probability of the treatment with t-PA as a function of a number of baseline characteristics, such as the presence or absence of comorbidities (hypertension, diabetes, etc.), or person characteristics (age and sex).

Based on the PS values, we can now group individuals according to baseline characteristics – here, untreated patients and patients treated with t-PA. This can be done in several ways: matching, stratification, covariate adjustment and inverse probability (of treatment) weighting. There is an extensive literature on the different variants and the associated pros and consCitation26,Citation28Citation30 (). Different variants were applied in the example, and eventually, affected the results, ie, the ORs between t-PA and death ranged from 1.17 to 1.96,Citation19 potentially leading to different conclusions, if considered separately. However, it is important to recognize that different variants may imply answering different research questions.Citation26,Citation31 We discriminate between approaches that estimate the average effect of a treatment on the population (both treated and untreated individuals) and the average effect of treatment on those individuals who actually received the treatment. In the example with t-PA, the authors mentioned that differences in ORs between two weighting variants (inverse probability of treatment weighting and standardized mortality ratio weighting) likely derive from the fact that the two approaches are associated with different research questions,Citation19 and this may also apply to other methods that are evaluated in the study.

An important step is to evaluate whether the treated and untreated groups are comparable. The evaluation cannot be offset by a goodness-of-fit (GOF) test, which is a general approach that provides a measure of how well a statistical model fits the data. However, this approach is usually meaningless in the large data sets that are typically extracted from health care databases. Furthermore, the GOF test may not tell the researcher whether important confounders were excluded from the analysis, neither in multivariable analysis nor in PS modeling.Citation26,Citation32

The evaluation of the comparability of the groups of interest may involve measures of difference, testing or visual inspection of the PS distributions of the two groups – refer the study by Franklin et alCitation33 for a discussion under what circumstances the different approaches are considered useful. Imbalances between the two groups may necessitate that the estimation of the PS is reconsidered, meaning that the specification of the model that provides the PS is changed, another PS variant is applied or the data set is trimmed.Citation34 By trimming, a subset of data is extracted according to certain rules, and thus, the sample size is reduced, which in some cases may hamper the feasibility and interpretability of the results obtained by the PS method.

It may be difficult to balance the treatment groups in small samples or if the comparison groups are very different. Hence, the evaluation of balance represents an assurance that eventually we analyze comparable groups in the final analysis of the possible association between treatment and outcome, adjusted for (measured) confounding.

In the German study on stroke patients, there was an imbalance between the t-PA-treated and -untreated groups with a limited overlap of PSs among the two groups due to an exceptionally high proportion of untreated patients with low PS. The authors then restricted the study population to patients with a PS ≥0.05, which increased the comparability of the groups. In this setting, the results were also less sensitive to the choice of PS variant (matching and several regression adjustments) compared to the unrestricted approach.Citation19

As with multivariable analysis, there is a possibility that unknown, unmeasured and residual confounding still exists after having applied the PS approach. In order to attempt to reduce this drawback, the HD-PS approach was developed.Citation35 The HD-PS method involves a series of conceptual steps,Citation35 which in essence can be condensed to: 1) specification of data source; 2) data-driven selection of potential confounders; 3) estimation of PS; 4) use of the PS to make groups of interest comparable and assessment of group comparability and 5) estimation of the association between treatment/exposure and outcome. Essentially, it is the selection process of confounders that makes the HD-PS method differ from the conventional PS methods. For the HD-PS method, large numbers of variables (often hundreds) are selected as potential confounders. As an example in a nationwide study in Taiwan, the HD-PS method was used to adjust for confounding.Citation36 Well-known prespecified confounders, eg, sex, age and comorbidities related to lifestyle, were incorporated in the analysis together with 500 additional potential confounders. The rationale is that some of these many variables are likely proxies for unmeasured confounders that are not available in the database or the researcher is not aware of. Accordingly, we may be able to deal with at least some of the unmeasured confounding that would not be considered in a conventional PS approach. However, there is little empirical evidence that the HD-PS method is better at controlling for unmeasured confounding than other methods, and adding several hundred empirically identified factors in an HD-PS setting may lead to comparable results to those that could also be obtained from a conventional PS setting.Citation37 In addition, despite examples of HD-PS analyses that provided estimates closer to the estimates obtained in randomized trials,Citation38 we cannot conclude that HD-PS is almost as good a tool as randomization.

Given the data greediness of the HD-PS method, its application is dependent on access to large databases, although it has also been demonstrated to be quite robust in a small sample setting (down to 50 exposed patients with an event).Citation38 It is important to be aware that variable selection in the HD-PS method is mainly data driven and in principle associated with the risk of omitting important confounders. That said, the benefit of including an excessive number of proxies for potential unmeasured confounders possibly outweighs the risk of discarding important confounders. In multivariable analysis and conventional PS analysis, we select the potential confounders to adjust for from a pool of variables that are thought to be possible true confounders. Despite measures taken during the variable selection process in the HD-PS method,Citation38 the generous inclusion of variables from databases may increase the likelihood that variables are not confounders but mediator, collider or instrumental variables – see definitions elsewhere.Citation39Citation41 This may lead to inappropriate adjustment that potentially provides spurious results. However, we are limited in our understanding of all the prospects and pitfalls of the HD-PS method, given its relatively recent origin, although exploration and refinements of the approach have already emerged.Citation42Citation44

Overall, we can conclude that the PS methods have several attractive characteristics in a registry-based setting. For example, PS seems more robust in situations with rare outcomes and common exposures than multivariable analysis.Citation45,Citation46 However, even in a large sample setting, we may face the challenge of rare exposure (or treatment). The disease risk score (DRS) method is suitable to use under these circumstances, such as in the early market phase of a drug when reduction in confounder dimensions is likely important.Citation47Citation49 DRS is comparable to PS in so far that information from several variables is summarized in one single score.

The PS method cannot handle treatment defined as a continuous variable (eg, drug dosage), unless dosage is categorized, typically dichotomized into the presence or absence of treatment, associated with the risk of losing important information on the association between an exposure and baseline characteristics. DRS may again be an alternative to PS. That said, methods that are based on the inverse probability weighting (IPW) principle represent alternatives with a wide range of applications, because IPW may be generalized to a suite of different circumstances also including dichotomous and non-dichotomous exposure.Citation50 The German study of stroke comprised an additional analysis, which controlled for confounding by using the IPW principle.Citation19 Time-varying exposure and thus time-dependent confounding may also be dealt with by methods based on IPW in the form of marginal structural modelsCitation51 or structural nested models based on G-estimation.Citation52

What did we achieve and what have we lost?

It is important to stress that during selection of a method, there is no book of answers, and in many cases, simple methods may be equally valid as the complex methods. In addition to all the pros and cons of the different methods (), we may face an unusual setting or a data set with an odd structure that necessitates further consideration of the method that controls for confounding. Moreover, the specific research question that we wish to answer may determine the method selected to control for confounding ().

Both stratification and standardization represent ways of learning about the data, as we look at smaller units of the data set, and we may use these methods as preliminary analysis, before we use other methods such as multivariable analysis or PS methods to adjust for confounding. Thus, applying stratification or standardization assists in the understanding of the data at a detailed level, and we may become aware of associations in specific strata, otherwise overlooked. By the era of multivariable analysis, we may have lost some of this basic understanding of the data, because of the complexity introduced by incorporating numerous potential confounders in models. Nevertheless, we are still capable of understanding which factors substantially confound an association, and we can directly explore interactions between an exposure and other factors. After the introduction of the PS method, there has been an ongoing discussion on the advantages and disadvantages of this method as compared to multivariable analysis. Glynn et alCitation53 noticed that in the majority of studies that used both multivariable analysis and PS methods, there were no important differences in the results, and this was further confirmed by simulation studies. However, comparable results across different methods do not prove that proper adjustment of confounding was undertaken, eg, if the data quality of important confounders is poor or unmeasured confounding exists. The trend in analysis methods has dictated that we extract more and more information from databases, when attempting to account for confounding. This could potentially entail that we reduce unmeasured confounding just by chance, most notably in the HD-PS approach with the inclusion of hundreds of variables. However, there is no evidence that this method is superior to others, and even the HD-PS method would be flawed in the case that data on important confounders or their proxies are not available or if variables that are not true confounders are adjusted for.

Given the complexities of registries and data analysis, we wish to emphasize the critical importance of clinical input and of the interplay between clinicians and analysts (statisticians) during the statistical analysis. Clinicians may contribute with important scientific input regarding the initial list of potential confounders that should be considered and their availability in health care databases; if potential con-founders are missing, which surrogate factors could then be used as a replacement? Clinicians may also provide essential information on technical elements of the statistical analysis such as how variables should be categorized, the functional forms of continuous variables (eg, linear vs nonlinear) and temporal aspects (eg, the relative importance of an event of MI 1 week vs 1 year ago). Finally, clinicians have expert knowledge on the nature of treatments and treatment allocation that can guide the analyst.

Requirements to the analysis in the future

In the future, it would be an achievement if we were better at identifying confounding factors. At present, the selection of potential confounders in models largely relies on assessments, eg, prior knowledge, DAGs and arbitrarily defined differences between crude and adjusted results, or on data-driven procedures decoupled from the confounder issue.Citation17,Citation54 It would be desirable to minimize the uncertainty of a factor being a true confounder. In addition, confounder selection is commonly compromised by limited access to appropriate variables in registries.Citation18 In this respect, it would also be desirable that information on lifestyle and severity of diseases were more widespread in registries, as these elements likely represent important confounding factors.Citation55Citation57 Software that extracts information from medical records and translates it into analyzable data has already been developedCitation58 and may likely assist in compiling data on lifestyle and severity of diseases. Furthermore, self-reporting systems, remote sensing technologies and automated data logging already exist to accumulate data on, eg, blood pressure, physical activity and heart rate. This could potentially develop into health care monitoring in a citizen science setting, just as this concept has evolved in other science disciplines.Citation59,Citation60 The use of such less aggregated data than those present in the registries of today will surely improve our ability to incorporate relevant con-founders or their proxies; however, most likely it necessitates novel and innovatory methods to deal with the confounder issue in the analysis phase. However, despite a remarkable development of methods that control for confounding and new data opportunities, it is unlikely that we will be able to account completely for confounding in the data collection process in a foreseeable future. Hence, confounding will remain a serious issue that needs to be acknowledged in the interpretation of our analyses.

Acknowledgments

A hearty thank you to the editor and three anonymous reviewers for very useful comments, Lars Pedersen and Jan Vanden-broucke for stimulating discussions, Thomas B. Rasmussen for input to the PS methods and Kasper Adelborg and Troels Munch for their great hands-on knowledge on data collection in hospitals. This article was funded by the Program for Clinical Research Infrastructure (PROCRIN) established by the Lundbeck Foundation and the Novo Nordisk Foundation and administered by the Danish Regions.

Supplementary materials

Table S1 Summary of the pros and cons of five methods used to control confounding in observational studies

Table S2 Summary of the methodological pros and cons of four different types of PS methods

References

  • McNameeRRegression modelling and other methods to control confoundingOccup Environ Med200562750050615961628
  • PeduzziPConcatoJKemperEHolfordTRFeinsteinARA simulation study of the number of events per variable in logistic regression analysisJ Clin Epidemiol19964912137313798970487
  • ArbogastPGRayWAUse of disease risk scores in pharmacoepidemiologic studiesStat Methods Med Res2009181678018562398
  • CepedaMSBostonRFarrarJTStromBLComparison of logistic regression versus propensity score when the number of events is low and there are multiple confoundersAm J Epidemiol2003158328028712882951
  • AustinPCAn introduction to propensity score methods for reducing the effects of confounding in observational studiesMultivar Behav Res2011463399424
  • GlynnRJGagneJJSchneeweissSRole of disease risk scores in comparative effectiveness research with emerging therapiesPharmacoepidem Dr S201221138147
  • WilliamsonEMorleyRLucasACarpenterJPropensity scores: from naive enthusiasm to intuitive understandingStat Methods Med Res201221327329321262780
  • RobinsJMHernanMABrumbackBMarginal structural models and causal inference in epidemiologyEpidemiology200011555056010955408

Disclosure

The authors report no conflicts of interest in this work.

References

  • VandenbrouckeJPThe history of confoundingSoz Praventiv Med2002474216224
  • MorabiaAHistory of the modern epidemiological concept of confoundingJ Epidemiol Commun H2011654297300
  • GreenlandSRobinsJMIdentifiability, exchangeability, and epidemiological confoundingInt J Epidemiol19861534134193771081
  • DuXCruickshankKMcNameeRCase-control study of stroke and the quality of hypertension control in North West EnglandBMJ199731470762722769022492
  • GreenlandSMorgensternHConfounding in health researchAnnu Rev Publ Health200122189212
  • GreenlandSPearlJRobinsJMCausal diagrams for epidemiologic researchEpidemiology199910137489888278
  • SchneeweissSSuissaSAdvanced approaches to controlling confounding in pharmacoepidemiologic studiesStromBLKimmelSEHennesySTextbook of Pharmacoepidemiology2nd edUKJohn Wiley & Sons Ltd2013324336
  • KestenbaumBEpidemiology and Biostatistics – An Introduction to Clinical ResearchNew YorkSpringer2009
  • RothmanKEpidemiology – An IntroductionUKOxford University Press2002
  • MannocciAThe Mantel-Haenszel procedure. 50 years of the statistical method for confounders controlItal J Public Health200964338340
  • MantelNHaenszelWStatistical aspects of the analysis of data from retrospective studies of diseaseJ Natl Cancer I1959224719748
  • NamHSKimHCKimYDLong-term mortality in patients with stroke of undetermined etiologyStroke201243112948295622933583
  • KirkwoodBRSterneJACMedical StatisticsUKBlackwell2003
  • ChoiBCKde GuiaNAWalshPLook before you leap: stratify before you standardizeAm J Epidemiol1999149121087109610369502
  • VandenbrouckeJPShould we abandon statistical modeling altogetherAm J Epidemiol1987126110133296742
  • GroenwoldRHHKlungelOHAltmanDGAdjustment for continuous confounders: an example of how to prevent residual confoundingCMAJ2013185540140623401401
  • WalterSTiemeierHVariable selection: current practice in epidemiological studiesEur J Epidemiol2009241273373619967429
  • BrookhartMASturmerTGlynnRJRassenJSchneeweissSConfounding control in healthcare database research: challenges and potential approachesMed Care2010486 supplS114S12020473199
  • KurthTWalkerAMGlynnRJResults of multivariable logistic regression, propensity matching, propensity adjustment, and propensity-based weighting under conditions of nonuniform effectAm J Epidemiol2006163326227016371515
  • van SmedenMde GrootJAHMoonsKGMNo rationale for 1 variable per 10 events criterion for binary logistic regression analysisBMC Med Res Methodol201616116327881078
  • TibshiraniRThe lasso method for variable selection in the Cox modelStat Med19971643853959044528
  • AkaikeHNew look at statistical-model identificationIEEE T Automat Contr1974Ac196716723
  • GreenlandSInvited commentary: variable selection versus shrinkage in the control of multiple confoundersAm J Epidemiol2008167552352918227100
  • GreenlandSWhen should epidemiologic regressions use random coefficients?Biometrics200056391592110985237
  • McNameeRRegression modelling and other methods to control confoundingOccup Environ Med200562750050615961628
  • AustinPCAn introduction to propensity score methods for reducing the effects of confounding in observational studiesMultivar Behav Res2011463399424
  • MiettinenOSStratification by a multivariate confounder scoreAm J Epidemiol19761046609620998608
  • PirracchioRResche-RigonMChevretSEvaluation of the propensity score methods for estimating marginal odds ratios in case of small sample sizeBmc Med Res Methodol2012127022646911
  • AustinPCMamdaniMMA comparison of propensity score methods: a case-study estimating the effectiveness of post-AMI statin useStat Med200625122084210616220490
  • D’AgostinoRBPropensity scores in cardiovascular researchCirculation2007115172340234317470708
  • WilliamsonEMorleyRLucasACarpenterJPropensity scores: from naive enthusiasm to intuitive understandingStat Methods Med Res201221327329321262780
  • WeitzenSLapaneKLToledanoAYHumeALMorVWeaknesses of goodness-of-fit tests for evaluating propensity score models: the case of the omitted confounderPharmacoepidemiol Drug Saf200514422723815386700
  • FranklinJMRassenJAAckermannDBartelsDBSchneeweissSMetrics for covariate balance in cohort studies of causal effectsStat Med201433101685169924323618
  • SturmerTRothmanKJAvornJGlynnRJTreatment effects in the presence of unmeasured confounding: dealing with observations in the tails of the propensity score distribution – a simulation studyAm J Epidemiol2010172784385420716704
  • SchneeweissSRassenJAGlynnRJAvornJMogunHBrookhartMAHigh-dimensional propensity score adjustment in studies of treatment effects using health care claims dataEpidemiology200920451252219487948
  • HungT-YLeeY-KHuangM-YHsuC-YSuY-CIncreased risk of ischemic stroke in patients with burn injury: a nationwide cohort study in TaiwanScand J Trauma Resusc Emerg Med20162414427052491
  • TohSRodriguezLAGHernanMAConfounding adjustment via a semi-automated high-dimensional propensity score algorithm: an application to electronic medical recordsPharmacoepidem Dr S2011208849857
  • RassenJAGlynnRJBrookhartMASchneeweissSCovariate selection in high-dimensional propensity score analyses of treatment effects in small samplesAm J Epidemiol2011173121404141321602301
  • HernanMAHernandez-DiazSWerlerMMMitchellAACausal knowledge as a prerequisite for confounding evaluation: an application to birth defects epidemiologyAm J Epidemiol2002155217618411790682
  • VanderWeeleTJVansteelandtSConceptual issues concerning mediation, interventions and compositionStat Interface200924457468
  • GreenlandSAn introduction to instrumental variables for epidemiologistsInt J Epidemiol2000296110211101554
  • SchneeweissSEddingsWGlynRJFranklinJMImproving empirical variable selection in propensity-score models with high-dimensional covariate space using healthcare databasesPharmacoepidem Dr S2014231111
  • LeHVPooleCBrookhartAMSchoenbachVJBeachKJSturmerTEffects of aggregation of medical codes on the performance of the high-dimensional propensity score (hd-PS) algorithmPharmacoepidem Dr S201322181181
  • FranklinJMEddingsWGlynnRJSchneeweissSRegularized regression versus the high-dimensional propensity score for confounding adjustment in secondary database analysesAm J Epidemiol2015182765165926233956
  • WinkelmayerWCKurthTPropensity scores: help or hype?Nephrol Dial Transplant20041971671167315069182
  • CepedaMSBostonRFarrarJTStromBLComparison of logistic regression versus propensity score when the number of events is low and there are multiple confoundersAm J Epidemiol2003158328028712882951
  • GlynnRJGagneJJSchneeweissSRole of disease risk scores in comparative effectiveness research with emerging therapiesPharmacoepidem Dr S201221138147
  • ArbogastPGRayWAUse of disease risk scores in pharmacoepidemiologic studiesStat Methods Med Res2009181678018562398
  • SchmidtAFKlungelOHGroenwoldRHHConsortiumGAdjusting for confounding in early postlaunch settings going beyond logistic regression modelsEpidemiology201627113314226436519
  • HernánMRobinsJMEstimating causal effects from epidemiological dataJ Epidemiol Community Health20066057858616790829
  • RobinsJMHernanMABrumbackBMarginal structural models and causal inference in epidemiologyEpidemiology200011555056010955408
  • RobinsJA new approach to causal inference in mortality studies with a sustained exposure period – application to control of the healthy worker survivor effectMath Modelling198679–1213931512
  • GlynnRJSchneeweissSSturmerTIndications for propensity scores and review of their use in pharmacoepidemiologyBasic Clin Pharmacol2006983253259
  • HoffmannKPischonTSchulzMSchulzeMBRayJBoeingHA statistical test for the equality of differently adjusted incidence rate ratiosAm J Epidemiol2008167551752218230678
  • JiaoLSilvermanDTSchairerCAlcohol use and risk of pancreatic cancerAm J Epidemiol200916991043105119299403
  • Di MiliaLVandelanotteCDuncanMJThe association between short sleep and obesity after controlling for demographic, lifestyle, work and health related factorsSleep Med201314431932323419528
  • WeinhandlPengYGilbertsonDTBradburyBDCollinsAJHemoglobin variability and mortality: confounding by disease severityAm J Kidney Dis201157225526520801571
  • HinchcliffMJustEPodluskySVargaJChangRWKibbeWAText data extraction for a prospective, research-focused data mart: implementation and validationBmc Med Inform Decis Mak20121210622970696
  • ResnikDBElliottKCMillerAKA framework for addressing ethical issues in citizen scienceEnviron Sci Policy201554475481
  • KellingSFinkDLa SorteFAJohnstonABrunsNEHochachkaWMTaking a ‘Big Data’ approach to data quality in a citizen science projectAmbio201544S601S611