Annual Reviews home
0
Skip to content
  • For Librarians & Agents
  • For Authors
  • Knowable Magazine
  • Institutional Login
  • Login
  • Register
  • Activate
  • 0 Cart
  • Help
Annual Reviews home
  • JOURNALS A-Z
    • Analytical Chemistry
    • Animal Biosciences
    • Anthropology
    • Astronomy and Astrophysics
    • Biochemistry
    • Biomedical Data Science
    • Biomedical Engineering
    • Biophysics
    • Cancer Biology
    • Cell and Developmental Biology
    • Chemical and Biomolecular Engineering
    • Clinical Psychology
    • Computer Science
    • Condensed Matter Physics
    • Control, Robotics, and Autonomous Systems
    • Criminology
    • Developmental Psychology (new in 2019)
    • Earth and Planetary Sciences
    • Ecology, Evolution, and Systematics
    • Economics
    • Entomology
    • Environment and Resources
    • Financial Economics
    • Fluid Mechanics
    • Food Science and Technology
    • Genetics
    • Genomics and Human Genetics
    • Immunology
    • Law and Social Science
    • Linguistics
    • Marine Science
    • Materials Research
    • Medicine
    • Microbiology
    • Neuroscience
    • Nuclear and Particle Science
    • Nutrition
    • Organizational Psychology and Organizational Behavior
    • Pathology: Mechanisms of Disease
    • Pharmacology and Toxicology
    • Physical Chemistry
    • Physiology
    • Phytopathology
    • Plant Biology
    • Political Science
    • Psychology
    • Public Health
    • Resource Economics
    • Sociology
    • Statistics and Its Application
    • Virology
    • Vision Science
    • Article Collections
    • Shot of Science
  • JOURNAL INFO
    • Copyright & Permissions
    • Add To Your Course Reader
    • Expected Publication Dates
    • Impact Factor Rankings
    • Access Metadata
    • RSS Feeds
  • PRICING & SUBSCRIPTIONS
    • General Ordering Info
    • Online Activation Instructions
    • Personal Pricing
    • Institutional Pricing
    • Society Partnerships
  •     GIVE    
  • ABOUT
    • What We Do
    • Founder & History
    • Subscribe to Open
    • Our Team
    • Careers
    • Press Center
    • News
    • Global Access
    • Directory
    • Help/FAQs
    • Contact Us
  • Home >
  • Annual Review of Public Health >
  • Volume 36, 2015 >
  • Greenland, pp 89-108
  • Save
  • Email
  • Share

Statistical Foundations for Model-Based Adjustments

  • Home
  • Annual Review of Public Health
  • Volume 36, 2015
  • Greenland, pp 89-108
  • Facebook
  • Twitter
  • LinkedIn
Download PDF

Statistical Foundations for Model-Based Adjustments

Annual Review of Public Health

Vol. 36:89-108 (Volume publication date March 2015)
https://doi.org/10.1146/annurev-publhealth-031914-122559

Sander Greenland1 and Neil Pearce2,3

1Department of Epidemiology and Department of Statistics, University of California, Los Angeles, California 90095-1772; email: [email protected]

2Departments of Medical Statistics and Non-Communicable Disease Epidemiology, Faculty of Epidemiology and Population Health, London School of Hygiene and Tropical Medicine, London WC1E 7HT, United Kingdom; email: [email protected]

3Centre for Public Health Research, Massey University, Wellington 6140, New Zealand

Download PDF Article Metrics
  • Permissions
  • Reprints

  • Download Citation
  • Citation Alerts
Sections
  • Abstract
  • Keywords
  • INTRODUCTION
  • OUR FUNDAMENTAL STATISTICAL PRECEPTS
  • SOME PRACTICAL PRELIMINARIES FOR SOUND MODELING
  • CURRENT MODELING STRATEGIES FOR CONFOUNDER CONTROL
  • SUMMARY AND CONCLUSIONS
  • SUMMARY POINTS
  • disclosure statement
  • acknowledgments
  • literature cited

Abstract

Most epidemiology textbooks that discuss models are vague on details of model selection. This lack of detail may be understandable since selection should be strongly influenced by features of the particular study, including contextual (prior) information about covariates that may confound, modify, or mediate the effect under study. It is thus important that authors document their modeling goals and strategies and understand the contextual interpretation of model parameters and model selection criteria. To illustrate this point, we review several established strategies for selecting model covariates, describe their shortcomings, and point to refinements, assuming that the main goal is to derive the most accurate effect estimates obtainable from the data and available resources. This goal shifts the focus to prediction of exposure or potential outcomes (or both) to adjust for confounding; it thus differs from the goal of ordinary statistical modeling, which is to passively predict outcomes. Nonetheless, methods and software for passive prediction can be used for causal inference as well, provided that the target parameters are shifted appropriately.

Keywords

causal inference, confounding, modeling, variable selection

INTRODUCTION

In our experience, few topics cause as much consternation, both in students and in researchers, as modeling strategies do. It is straightforward to do a course on the basic principles of confounder control and assessment of interaction and mediation and to learn how to run regression models. It can be far more daunting to be confronted with a new data set that includes many potentially important covariates and have to decide what to do with it.

Most epidemiologic textbooks are vague on the practicalities of model selection, and understandably so. Arguably, model selection should be strongly influenced by factors that are specific to the particular study, including possibly controversial prior information about which variables are important potential confounders, modifiers, or mediators. It can thus be counterproductive to impose rigid modeling rules or recipes. Nonetheless, lack of guidelines can leave in place bad practices, such as choosing models based on naïve significance testing. General recommendations may thus be useful even if exceptions are common. Similarly, critical assessment of analyses requires knowledge of which strategy or guidelines were used to select analysis models.

The present article updates older commentaries on modeling guidelines (46, 47, 116, 139). Our motivation stems from noting that few subject-matter studies employ newer modeling methods (despite many sophisticated papers on these methods in leading epidemiology and statistics journals) and that simple but problematic methods remain common defaults in basic teaching, commercial software, and the clinical literature. Although simple methods may often suffice, it is important to understand their limitations and to recognize when better alternatives are required.

Scope, Aims, and Assumptions

We focus entirely on methods for observational studies of causal effects, there being some excellent texts on purely predictive modeling (79, 80, 94, 122, 132). We begin by outlining the precepts we take as fundamental to sound use of statistical methods in epidemiology, emphasizing the importance of understanding the contextual interpretation of model parameters, selection criteria, and estimates. We then critically review some established modeling approaches based either on passive predictive models or on changes in the estimate (CIE) of exposure effect and some variations of these approaches available to researchers with more technical resources. Our coverage is not intended to be a comprehensive review for highly skilled practitioners; rather, we target teachers, students, and working epidemiologists who want an accurate data analysis, but who lack resources such as R programming skills or a bona fide expert both in their field and in statistical modeling committed to their project. Elsewhere (59, 134) we discuss how simple approaches can be recast and upgraded, with little effort and without special software, to minimize harmful practices and follow more sound methodologic principles.

Throughout, we assume that we are applying a conventional risk or rate regression model (e.g., logistic, Cox, or Poisson regression) to estimate the effects of an exposure variable on the distribution of an outcome variable, while controlling for other variables (“covariates”), or that we are applying an exposure model for adjustment (as in propensity score methods). The covariates may be forced variables, which we always want to control (typically for age and sex), or unforced variables for which the decision to control may be data based. Our focus is on strategies for decisions about unforced variables.

The strategies we discuss apply to exposures and covariates of any form. We assume that data checking, description, and summarization have been done for quality control (73, 150); thus, we do not address problems arising from preliminary data examinations (e.g., from influence on subsequent analyses through collapsing away small categories) (18, 19). We advise using univariate distributions and background (contextual) information to select categories or an appropriately flexible form (e.g., splines or fractional polynomials) for detailed modeling of quantitative variables (5, 61, 67), but we must leave many difficult issues about model specification and diagnostics to more detailed discussions (3, 60, 61, 67, 79, 80, 94, 122, 132, 140). Finally, we do not consider special models and problems that arise in the contexts of ecologic analysis (55, 151), time-varying exposures (83, 140), mediation analysis (83, 115, 117, 138, 140, 142), measurement-error adjustment (17), or bias analysis (64, 68, 69, 77, 78, 92, 93, 137), although our general comments apply to all modeling applications.

OUR FUNDAMENTAL STATISTICAL PRECEPTS

Why Contextual Interpretation and Sensibility Should Dominate Statistical Theory

We agree with other authors that scientific theories, whether concerning substance or concerning methodology, are not uniformly reliable and are often filled with fictions and falsehoods (39, 113, 159). This caution applies to statistical theories of inference; in particular, strict adherence to a statistical philosophy or mathematical theory can harm scientific inference unless such adherence is moderated by contextual sensibility (13, 43, 88, 107, 111, 128, 135). In messy observational sciences such as epidemiology, this harm is often traceable to reification (157) of probability distributions and their properties (that is, treating hypothetical distributions as if real). Damage to scientific inference is especially acute when reification leads to accepting conclusions from methods whose properties are defined only within idealized sequences or models. The most common examples involve treating nominal error rates of tests as if they were real (e.g., taking a 0.05 α level as guaranteeing no more than 5% of rejections are incorrect when the tested hypothesis is correct); these error rates are in fact derived from distributions representing hypothetical infinite sequences of valid studies (which do not exist among real observations in health and medical sciences) and usually assume that there is no data-based model selection.

Observational epidemiologic studies have unique source populations, protocols, and unanticipated problems that exemplify the tenuous nature of frequency models. Consider the Nurses Health Study I (NHS I) (23): There is no similar cohort of nurses observed in such detail during the final quarter of the twentieth century, a period when low-fat, high-carbohydrate diets remained a common recommendation for weight-control and health, when statins were not yet widespread, and so on. The study was thus conducted on a population inaccessible to future studies. Available comparison cohorts do not and never will resemble the NHS-I cohort with respect to distributions of all important health predictors and measurement errors, so there is no guarantee that fully effective adjustments can be made for differences. Consequently, the number of cohorts credibly resembling NHS I after adjustment will be so few that analyses assuming replication with only random differences will be purely hypothetical.

In sum, standard statistical theories model only random uncertainty, i.e., uncertainty about what would emerge from attempts to replicate the study methods down to the last known detail predictive of outcome, successfully controlling for all other uncertainty sources (biases). Such detailed valid replication is impossible outside of a perfect parallel randomized trial. Health, medical, and social sciences rarely have such replication, and their total uncertainty often greatly exceeds uncertainty from random variation [as illustrated in bias analyses (68, 64, 69, 77, 78, 92, 93, 137)].

Case Studies in Reification: Statistical Unbiasedness and Consistency

Properties defined only over long runs must appeal to hypothetical sequences under models that are grossly oversimplified relative to the forces determining background risks, participation, and measurement errors (57). Although statistical significance and claims of objectivity are the usual targets of this criticism (9, 48, 58), concepts of unbiased and consistent estimation are also vulnerable. These concepts refer to estimators whose mean and large-sample limit equal the targeted parameter under an assumed repeated-sampling model.

A major problem is that the models used to show unbiasedness or consistency are known to be false: Uncontrolled epidemiological biases are always present and render inoperative theoretical guarantees of unbiasedness and consistency for the actual target parameters (57). Thus, no available estimator can be shown to be unbiased or consistent under realistic epidemiologic conditions. Even when only random error is present, however, unbiasedness is neither necessary nor sufficient for contextually sensible estimators (16, 33, 52, 97). We thus reject unbiasedness as an absolute requirement for point estimators of target parameters.

The hypothetical nature of distributions does not make statistical theory or methods useless, but it does greatly reduce the decisiveness and accuracy of the methods to below that indicated by p-values or confidence intervals. Properties under hypothetical distributions may serve as useful guides for determining when the distributions appear contextually reasonable. Nonetheless, any restriction of judgments by theoretical rules risks enormous error that might have been obvious from a contextual viewpoint [e.g., failure to adjust a covariate simply because it is not significant in a risk model despite its known role in etiology (71)]. Thus, statistical theories must be used with the understanding that they are theories of unattainable ideals. If a researcher has a sound basis for the assumptions of the theory and a sound mapping of these assumptions into the research context, deductions within the theory (its theorems and methods) can serve as analysis directives (87); for example, if only random error is present, statistical consistency is a compelling requirement for an estimator. But given uncertainties about the assumptions and the mapping, the use of a statistical theory as if it were certainly correct cannot be justified, and another rationale for its use should be sought.

Uses of Statistical Theory and Modeling

One rationale for statistical theories is that they can be used to conduct valuable (albeit simplistic) thought experiments within their narrow confines (77). We can analyze data using different statistical theories and methods and compare the inferences so obtained. Similarly, we can evaluate and compare methods within theories, and we can compare these evaluations across theories. For example, we can compare modeling strategies within and across frequentist theories (whether Neyman–Pearson or Fisher–Cox) and Bayesian theories (whether likelihood-based or personal betting), regardless of which theory originated them. We can also compare their performance in passive prediction versus causal (interventional) prediction. Doing so, we educate our intuitions regarding when criteria and methods will be misleading or useful, on the basis of how well the scientific problem can be mapped into the theory.

As an illustration, the unbiasedness criterion can be criticized even within the confines of classical frequentist criteria for comparing inferential methods. The criterion assumes that systematic error in a coefficient estimator is always far more costly than random (mean-zero) error. Consequently, classical unbiased estimators often perform much worse than do biased estimators according to practically relevant valuations, such as when error costs are proportional to total error or when out-of-sample prediction is at stake (which is often evaluated in terms of mean squared error or mean absolute error) (16, 25, 33, 35, 36, 52).

More generally, different strategies can be justified by different assumptions, and no strategy can be proven best or optimal or even sound when its assumptions are not assuredly true. Furthermore, criteria for optimality vary with goals and valuations. Some strategies may be preferred over others based on theory; however, these preferences arise only in special cases in which the discrepancy between statistical and contextual theory is judged minor. Thus in practical terms there are no best or optimal strategies, and the reader should be suspicious of any method or model promoted as optimal or correct. Every model and every strategy has limitations and defects that depend on the context in which it is applied; thus, given enough analysis time, every model would only contribute one result among many in a sensitivity analysis over models.

Both frequentist and Bayesian methods are subject to these cautions, because both frequency (error) calibration and Bayesian (betting) coherency are unattainable ideals in health and social research. In practice, the data models used by both methodologies fall far short of capturing all sources of error and uncertainty. Thus frequentist and Bayesian methods are simply alternative ways of looking at models and data, and are more complementary than competitive (8, 34, 43, 58, 77).

In summary, although mathematical investigations and simulations are invaluable for formulating guidelines for judgment, they rely on simplifying assumptions that are almost never fully satisfied and thus can be misleading if their results are taken with the deductive certainty obtained under their assumptions (66). On the other hand, counterexamples can show how guidelines can fail but must be critically evaluated with respect to their contextual realism and importance (66). Beyond theoretical investigation, it is equally important to assess and compare modeling strategies in real case studies involving various conditions to see when in practice they yield contextually sensible results.

SOME PRACTICAL PRELIMINARIES FOR SOUND MODELING

Unfortunately, analysis resources are usually severely limited, and many principles will be compromised to deliver the results on schedule. As a consequence, model checking may be absent or limited in scope, which makes essential a good grasp of the contextual meaning of a model and its fitting method in spotting model defects. In particular, a good modeler will recognize each assumption and valuation enforced by the model and judge these according to the degree they are contextually supportable (corroborated by available information), uncertain (plausible but not well corroborated), or implausible (contradicted by available information). The modeler may decide to use uncertain or even partially implausible models, which may be defensible if one is aware of the misfits between the model and the contextual information. The most thorough analyses will, however, probe uncertain assumptions with sensitivity analysis, discussing the contextual status of each model (e.g., supportable, uncertain, or implausible).

Any apparent conflict between formal methods and informal assessments may stem from faults in our expectations, faults in the formal methods, or both. No one doubts that prior expectations can be quite faulty, distorted as they are by biases in the methodology, interpretation, and reporting of previous studies (63, 90–93, 112). But formal statistical inferences can be quite faulty as well, insofar as they fail to account for important background information or uncertainty (error) sources such as cognitive biases built into the methodology (90–93). Conflicts will require resolution because they indicate an error in one or both of the statistical and contextual theories.

Crucial Data Preprocessing: Meaningful Centering and Scaling

There are two data-processing steps that are important to contextually sensible modeling but are usually neglected: Quantitative variables should be recentered to ensure that zero is a meaningful reference value present in the data, and they should be rescaled so that they are measured in general scientific units that represent meaningful differences spanning a range present in the data (61). Consider diastolic blood pressure (DBP) recorded as millimeters above zero (no pressure). This is a clinically unintelligible scale: A 1-mm change in pressure is below measurement noise and clinical significance, and the reference (zero) point corresponds to death. Any effect estimate (such as a regression coefficient) presented in these terms would thus be difficult to interpret because no one has a good intuition for the effect of a 1-mm change. To remove these difficulties, DBP could be recentered so that 0 represents 80 mm (often taken as clinically normal) and then rescaled to centimeters instead of millimeters so that 100 mm would become (100–80)/10 = 2 cm above reference level; now the coefficient would refer to a 1-cm (10-mm) increase in DBP from a recognizable clinical reference point. Similarly, in a study of older adults, age could be recentered so that 0 represents age 60, and rescaled to decades instead of years, so that 80 years would become (80–60)/10 = 2 decades past 60.

Unfortunately, the statistical literature is replete with bad scaling recommendations and practices, using arbitrary study-specific quantities such as sample standard deviations or interquartile ranges, despite lengthy critiques of these practices (70, 75). To illustrate, suppose the rate ratio relating packs/day smoking to mortality was 3 in every study. Using standard deviation (SD or standardized) units, each study would use a different (and strange) unit to measure smoking: A study in a population with high smoking prevalence and a smoking SD of 0.67 packs/day would report an exponentiated coefficient of 30.67 = 2.1, whereas a study in a low-prevalence population and a smoking SD of 0.25 would report an exponentiated coefficient of 30.25 = 1.3; thus the effect in the latter study would appear much weaker, even though the individual effect of smoking is identical in both populations. Thus, converting to SD units does exactly the opposite of standardization in ordinary English terms because it removes each study from a contextually understood standard (packs/day) and instead produces different units in each study. SD units will often not even apply to the source population of the study itself; for example, matching alters the sample SD of any variable associated with matching factors.

Special Cautions Regarding Product Terms

Recentering and rescaling are especially important for coefficient interpretability when examining effects of exposure combinations (interactions) (61, 100) and also when using prior distributions or penalty functions (59, 134). To illustrate, suppose we observed a cohort in which smoking conferred a mortality rate ratio (RR) of 3 when expressed in packs/day, but smoking was fitted as cigarettes/day; then, with 20 cigarettes/pack, the exponentiated smoking coefficient (output RR) in cigarettes/day would be 31/20 = 1.056. Suppose also that systolic blood pressure (SBP) had a mortality RR of 4 for a 40-mm range but is fitted in millimeters; then the exponentiated SBP coefficient would be 41/40 = 1.035. In a multiplicative model with these conditional effects, the combined effect would be 31/2041/40 = 1.094 and thus appear unimportant, despite the huge RR of 3(4) = 12 when comparing pack-a-day smokers to nonsmokers with 40-mm lower SBP.

In nearly all contexts, models containing a product term should also contain the factors in the product as main effects (the hierarchy principle) (61, 100). Without a component main effect, the product-term coefficient will depend on the center (reference or zero point) chosen for the other factor in the product, which complicates correct interpretation.

Preliminary Screening and Posterior Diagnostics Based on Background Information

Perhaps the most well-known and accepted way that contextual information enters a causal analysis is through preliminary identification of confounders based on established causal relations (44, 121). For example, estimation of net (total) effects typically requires exclusion of intermediates and their effects (44, 71, 72, 108) and any other variable influenced by the exposure or outcome (24, 72, 108, 121, 141), as well as other variables whose control may increase bias without reducing total error (24, 44, 72, 103, 108, 109, 121, 141). We assume that these variables have been identified and eliminated, leaving us with potential adjustment covariates (potential confounders). These include covariates considered essential to control (e.g., age, sex), which must be forced into models along with the study exposures. They also include covariates we are confident would reduce bias if controlled properly if only our study size were unlimited, as well as covariates of uncertain adjustment value.

In this planning stage of analysis, it is valuable to record what one expects to see for each coefficient (which is another reason why meaningful scaling is important). After the data are analyzed, one can compare these expectations to the estimates. Seemingly large discrepancies may indicate error in one's expectation, error in the analysis model, error in the data, or some combination. This contrast of residual expectations (those not used in construction of the analysis model) against model-fitted expectations (those deduced from the data model and the data) is thus an error diagnostic. The contrast can play this role even without formal testing, but formal tests can be constructed if our expectations can be assigned some degree of certainty in the form of a prior distribution or external data (12, 58).

As an example, take age in relation to typical carcinomas. From vital statistics, one expects cohort curves for these outcomes to be very steeply positive. If one then obtains a cohort coefficient of age (in decades) that is near or below zero, this apparent anomaly requires an explanation. Of course, in a very small study this difference might be within conventional random-error bounds; but if random error is not a plausible explanation, a causal explanation for the anomaly will be needed. Such anomalies may warn of problems with the study or the data model or may reflect errors in our background information.

Expectations can also be mistaken if they fail to account for features of study design and execution. For example, in an age-matched cancer case-control study, the relation of age to cancer will become sawtoothed, with jumps at the ends of each matching category, rendering invalid any expectation or data model that imposes a monotone relation between age and cancer (51, 45). As another example, if a trial was to be blinded but we knew the blinding was hopelessly compromised (e.g., by side effects), then we must adjust our expectations to account for the blinding failure. To get a sensitive diagnostic, however, we must not let our initial expectations be altered by the study results, because such alterations would bias the diagnostic toward detecting no discrepancy between expectations and observations.

Finally, expectations and interpretations need to take account of other variables in the model (155). For example, smoking is a well-known risk factor for heart disease; however, if the model includes major mediators of this smoking effect (such as blood pressure), the smoking coefficient may be diminished considerably relative to the expected total effect of smoking.

CURRENT MODELING STRATEGIES FOR CONFOUNDER CONTROL

Articles are usually clear about whether they modeled outcomes (as in risk or rate regression), exposure [as in propensity scoring (60, 119), E-estimation (118), and inverse-probability weighting (60, 115)], or both [as in doubly robust methods (7, 60, 83, 99, 136, 140, 147)]. Some are also careful to exclude variables on causal grounds, such as instrumental variables and variables affected by exposure or disease, leaving only potential confounders for selection (24, 44, 72, 109, 121).

Beyond these basics, it may be difficult to tell if a prespecified modeling strategy was used. Often, however, we see variants of the following strategies, each of which may be applied to create outcome or exposure models:

1.

Adjust all: Enter all the potential confounders in the model (only one set of covariates is considered, although the form of the model may be varied).

2.

Predictor selection: Select covariates on the basis of some measure of their ability to predict outcome or exposure (or both) given other covariates in the model.

3.

Change in estimate (CIE) selection: Select covariates on the basis of the change in the exposure effect estimate upon excluding them, given the other covariates in the model (14, 89, 125).

A problem with all these strategies is that they are not based on maximizing accuracy (minimizing bias and variance) in estimating target effects defined in an explicit causal model, and indeed they have no formal justification when there are many covariates to consider. As we now describe, each strategy has further problems.

Why Not Adjust for Every Available Covariate?

One study proposed that a relevant criterion for estimating causal effects is to adjust for all covariates known to cause exposure, disease, or both (145). This criterion can identify covariate subsets sufficient for confounding control, but it has some practical drawbacks. The set of covariates it identifies can be far larger than needed for adequate confounding control (far from minimally sufficient) and may be clumsy for subsequent analyses. Furthermore, even the largest studies can be small relative to the number of covariates potentially fulfilling this criterion, resulting in the breakdown of conventional fitting methods such as maximum likelihood (including its conditional and partial versions) (59, 76, 134). Finally, there may be many covariates whose causal status (and thus their fulfillment of the criterion) is uncertain.

Although sample-size adequacy for fitting methods is often judged by rules similar to “at least 10 subjects per regression coefficient,” such rules take no account of exposure and disease frequencies and so can be too pessimistic when only an exposure coefficient is targeted (149) and too optimistic when some exposure–outcome combination is rare (76). In the latter case, controlling too many variables by conventional means can lead to or aggravate two closely related problems: (a) data sparsity, in which full control results in too few subjects at crucial combinations of the variables, with consequent inflation of estimates (59, 76, 116, 134), and (b) multicollinearity, by which we mean high multiple correlation (or more generally, high association) of the controlled variables with study exposures (116). In particular, if we include covariates that together are highly predictive of an exposure but are not all necessary to control confounding, the resulting effect estimate may be inflated or have unnecessarily wide confidence intervals (15, 26, 116). These problems increase as the ratio of number of covariates to sample size increases, motivating strategies to reduce the number of modeled covariates (116; S. Greenland & N. Pearce, unpublished manuscript, “Modeling Strategies for Observational Epidemiology”).

Strategies Based on Predictive Modeling

Traditionally, variable selection is based solely on predicting observed outcomes under the observed distribution of exposure and covariates. The selection criterion is usually significance testing of coefficients, as in conventional stepwise regression (37) [even though all modern software allows use of better criteria (60, 79, 80, 132)]. Stepwise regression (37) attempts to achieve parsimonious noncausal prediction, searching for a model that explains the most outcome variation with the fewest variables. The goal itself is reasonable for clinical prediction whenever obtaining variables incurs notable costs; for example, there are clear practical benefits if we can predict cardiovascular disease risk with negligible loss in accuracy by collecting information on 5 variables instead of 30. Even with this goal, however, ordinary stepwise regression has many flaws and has several more when used for estimating effects (6, 32, 38, 40–42, 47, 50, 53, 54, 79, 84, 127, 132, 133, 148, 152).

For covariates subject to selection, decisions about adding (or deleting) a covariate are made according to whether adding (or deleting) the covariate significantly improves (or reduces) the fit of the model, whereby “significantly” is usually defined by some arbitrary cutoff (α-level) for the coefficient p-value, usually 0.05, but preferably much higher (e.g., 0.20) for confounder selection (27, 47). Such criteria are equivalent to assessing whether the covariate explains a significant proportion of the residual variation (the outcome variation that remains given the preexisting variables in the model). They are also equivalent to using the p-value for testing the covariate when it is in the model.

Even for pure clinical prediction, ordinary stepwise regression and other significance-based selection procedures can give very distorted p-values and confidence intervals (2, 6, 21, 22, 32, 38, 40–42, 46, 79, 84, 86, 127, 132, 133). The general problem is that these traditional algorithms use no cross-validation and so do not account for preliminary testing (the same data being used to both fit and test the model). As a result, in outcome models, they produce p-values for the exposure effect that are too small (i.e., overstate significance) and confidence intervals that are too narrow (6, 32, 38, 40–42, 47, 50, 53, 54, 79, 84, 127, 132, 133, 148, 152). In addition, the resulting model often yields much poorer predictions than can be obtained with modern techniques (79, 80, 132). These defects are especially bad for fields (such as epidemiology) that are plagued by charges of generating too many false positives and inaccurate predictions. Defects can be corrected by using advanced resampling and cross-validation methods (80, 122, 132, 140), but these corrections remain uncommon in commercial software and thus are rare in published studies. Some of these problems can be moderated by replacing significance tests with selection criteria that penalize for model complexity, such as the Akaike information criterion (AIC) or Bayesian information criterion (BIC) (79, 80, 132); nonetheless, the confidence intervals produced by these methods will remain too narrow (86).

Another persistent objection to outcome modeling is that it will select weak confounders or nonconfounders preferentially over strong confounders if the weak confounders are better than strong confounders at predicting the outcome (46). A parallel objection to exposure modeling is that selection based solely on predicting exposure will select weak confounders preferentially over strong confounders if the weak confounders are better than the strong confounders at predicting exposure (15, 31, 146, 154). Thus, covariates that statistically explain (are associated with) the most observed disease or exposure variation, or are most statistically significant (have the smallest p-value), need not be the same as the covariates that are most important in terms of confounding control (47, 71, 102) or public health importance (106).

These seemingly paradoxical facts arise because outcome and exposure models consider only one of the several parameters that determine confounding. In the extreme, an outcome risk factor may be strongly associated with the outcome, with a small p-value, and yet not be associated with exposure, whereas an exposure risk factor may be strongly associated with the exposure, with a small p-value, and yet not be associated with the outcome. Neither factor will be a confounder, yet both could be selected for adjustment in preference to actual confounders. For example, age is a strong risk factor for most diseases, but it will not be a confounder if it is independent of exposure (as with outdoor air pollution within small enough areas). Conversely, a covariate may have a nonsignificant (p > 0.05) association with outcome or exposure owing to sample-size limitations, but it may still be an important confounder (71).

The problems just described often arise in pursuit of model goodness-of-fit; in particular, some variables may not be included in the model because they do not significantly improve the fit, even though they are important confounders. Global or omnibus tests of fit are especially inadequate for confounder selection because there can be many models that fit equally well but correspond to very different confounder effects and exposure effect estimates (116).

The preceding problems explain why ordinary model-fitting criteria such as significance testing of predictors have long been considered inappropriate for the assessment of confounding (14, 47, 71, 74, 89) [indeed, there are questions about whether tests can be justified for any epidemiologic purpose (48, 131)]. The same problems apply to exposure modeling, as in propensity scoring (PS) or inverse probability of treatment weighting (IPTW), but they manifest somewhat differently than they do in outcome modeling. In exposure modeling, the only association being tested is that of the potential confounder with exposure. This reorientation eliminates the downward bias of exposure–outcome p-values and standard errors produced by covariate deletion using the covariate–outcome p-values. Nonetheless, the preferential selection of variables that predict exposure will worsen multicollinearity problems, often increasing variance without sufficient compensation via bias reduction (15, 30, 31, 109, 110, 146). Furthermore, in case-control studies, exposure modeling can lead to bias (98, 136), which can be avoided by using special algorithms that use the outcome in the exposure model (136). Standard advice for mitigating this problem is to include only factors predictive of the outcome (as well as exposure) in the exposure model (4, 15, 109); this is sound advice but may still include far more covariates than desirable, especially in terms of precision loss relative to bias reduction.

Another objection to ordinary covariate selection criteria is that, when deciding whether to control for a variable, the null hypothesis is arguably not the correct one to test. If the goal is to avoid bias, it may be more accurate to start with the hypothesis that a particular variable is an important confounder and to decide to exclude the variable from the analysis only if the data indicate that it can be ignored (46, 47). For example, instead of testing whether a confounder coefficient is zero, one could test whether its absolute magnitude is larger than a given important size (a coefficient equivalence test) (81). More directly, one could test whether the change in the exposure effect estimate from controlling the covariate is larger than a certain size (a collapsibility equivalence test) (96, 101) as discussed below.

Change-in-Estimate Strategies

In contrast with traditional predictive modeling, CIE strategies select covariates on the basis of how much their control changes exposure effect estimates; this observed change is presumed to measure confounding by the covariate. Since the late 1970s, epidemiologic textbooks and articles have recommended CIE rather than significance testing of the covariate coefficient (14, 89, 125). Later versions (47, 74) suggest using change in the confidence limits instead because those are usually the final analysis product.

One caution to these approaches is that an accurate assessment of confounding may require examining changes from removing entire sets of covariates. Another caution, which arises if the disease frequency is high and measured by odds or rates, is that the change may partly reflect noncollapsibility of the effect measure rather than confounding (72, 104). Nonetheless, CIE methods have an advantage over selection based only on outcome or exposure prediction insofar as the selection criterion is on the scale used for contextual interpretation.

A complication of CIE methods is that one must choose an effect measure to judge change importance, where “importance” needs to be evaluated along a contextually meaningful scale. Except for linear models, this scale will not be that of model coefficients, but rather will be a nonlinear transformation of the coefficients. If the outcome is rare enough to ignore distinctions among risk, rate, and odds ratios, and the exposure effect is represented by a single coefficient β in a multiplicative model, one can simply compare the estimates of the constant (homogeneous) ratio effect exp(β) from models with and without the covariate at issue. Otherwise it may be necessary to compare standardized (marginal) risk ratios estimated from the models (104).

Suppose RRa and RRu denote the estimated risk ratio with and without adjustment for the covariate; then RRa/RRu is traditionally used to judge change importance. However, in public health applications in which total caseload is of primary concern, arguably the exposed attributable fraction (AF) = (RR − 1)/RR is more relevant, in which case change could be measured by AFa − AFu, perhaps multiplied by 100 to express change in the percent caseload attributed to exposure. Many variations could arise depending on the ultimate target parameter.

Outcome Modeling, Exposure Modeling, or Both?

Outcome modeling is usually the simplest approach, especially when there is only one outcome but multiple exposures or an exposure with multiple levels. Nonetheless, exposure modeling may provide more valid effect estimates when there is more information for modeling exposures than for modeling outcomes. This advantage would usually arise when exposure is a treatment (and thus much is known about reasons for its use) or when the exposure is common but the outcome is not (and thus there are relatively few cases available for outcome modeling). Nonetheless, exposure modeling risks greater imprecision than outcome modeling does if the exposure is predicted better than necessary for confounding control (4).

Methods that model both outcome and exposure (including doubly robust methods) (7, 60, 83, 99, 136, 140, 147) avoid having to make a choice, but at the cost of more modeling effort. They have the option of using more data information with potential accuracy gains as a result. They may also use different covariates in the two models: Only confounders requiring the most accurate control (e.g., age) need appear in both models; minor or doubtful covariates may be limited to the model in which their role is better understood, knowing that their adequacy of control depends entirely on the accuracy of their specification in that model.

Parsimony versus Confounding

Both predictive modeling and CIE lack strong theoretical foundation for confounder selection and share several weaknesses. As commonly implemented, both assess confounding solely on the basis of the analysis data, ignoring the earlier data that led to consideration of certain variables as potential confounders. Both have parsimony as a key goal: to find a model that maximizes prediction or confounding control using few covariates. Nonetheless, parsimony is not a worthwhile goal in itself; its benefits must be demonstrated to justify its use as a criterion. Although parsimony can improve predictive accuracy and simplify analysis and presentation, typical arguments offered for parsimony are fallacious (87).

Some have argued that apparently weak confounders should be deleted from the model because “the use of a reduced model…can sometimes lead to a gain in precision” (89), pointing to the smaller estimated standard errors from smaller (simpler) fitted models as exhibiting that benefit. Unfortunately, this apparent variance reduction in single data sets is largely illusory because it ignores the component of variance due to model selection. Again, there are ways to account for this problem (79, 80, 132, 140), but we know of none that are easy to implement with popular software.

An objection to all variable selection is that if our goal is to estimate the average effect of a particular exposure rather than all the model coefficients, then there is no direct reason for concern about accuracy or parsimony in estimating individual confounder effects. In that case, all that matters is whether the resulting model, taken as a whole, successfully removes confounding by the covariates. This argument has often been taken to imply that we should adjust for all measured potential confounders or at least a maximal number (47, 124, 145, 146), but again this view takes no account of the problems that can arise from doing so (62, 109, 110, 129, 130, 153).

Studies of Simple Strategies

There are many possible ways to use significance testing and CIE separately or together to reduce the number of model covariates. For the simplest methods, simulation studies (96, 101, 103) appear to confirm earlier suggestions (27) that false negatives (incorrect exclusion of confounders) are a greater threat to effect estimation accuracy than are false positives (incorrect inclusion of nonconfounders), consonant with theoretical criteria (145) and supporting weak exclusion criteria. As an illustration, consider one simple simulation study (96), which compared the following strategies:

I.

Significance test the covariate coefficient in the outcome model; e.g., delete if the coefficient's p-value is under 0.05.

II.

See whether the change in the exposure effect estimate from adjusting for the covariate falls outside an interval of practical equivalence; e.g., delete if 0.91 < RRa/RRu < 1.1 (which is the 10%-change rule for the risk ratio modified to be proportionally symmetric).

III.

Significance test the change from adjusting for the covariate (collapsibility testing); e.g., delete if a test of RRa/RRu = 1 yields p < 0.05.

IV.

Test whether the change from adjusting for the covariate (noncollapsibility testing) is important (falls outside an interval of practical equivalence); e.g., delete if the 95% confidence interval for RRa/RRu (not just the point estimate) falls between 0.80 and 1.25 [this is a 0.05-level equivalence test (11, 81)].

V.

A hybrid strategy uses a weighted average of ln(RRa) and ln(RRu) as the estimated exposure effect to reduce mean-squared error (49); this is not a selection strategy but instead a method for partial adjustment of all candidate covariates.

Strategies based on equivalence criteria (II, IV) performed best when the equivalence interval was narrow [e.g., 0.91 < RRa/RRc < 1.1 for (II)], whereas significance-test strategies (I, III) performed best when the α-level was very high (e.g., using p < 0.20 instead of p < 0.05); thus, for all approaches, it appeared more important to avoid excluding possible important confounders than to avoid including weak confounders or nonconfounders.

None of the above selection strategies have a good grounding in theory, in that they are not derived to minimize error in any dimension; they are merely heuristics for finding simpler models that are not misleading for subsequent analysis and presentation. They thus leave considerable room for improvement. Noteworthy in this regard is the substantial literature in algorithmic modeling (machine learning), which has found that the performance of most simple predictive-modeling algorithms can be boosted considerably via computer-intensive methods, such as cross-validation or bootstrapping, to a level that rivals much more sophisticated algorithms (80). Such algorithms have been used to improve performance of exposure-modeling methods (95, 99, 105, 114, 126, 156) and can be applied similarly to outcome modeling (80) as long as exposure is forced into the model.

Further Considerations for Effect Estimation (Causal Inference)

All procedures require a target against which error (whether systematic or random) will be measured and a scale or loss function for measuring error and performance. Although there have been several applications of algorithmic exposure modeling for causal targets (140), the machine-learning algorithms in common software assume that the target is passive prediction to populations identical to that observed, apart from random variation. In outcome modeling, this amounts to success in passively predicting the outcome over the observed joint distribution of the exposure and covariates; in exposure modeling, this amounts to success in predicting the exposure over the observed joint distribution of covariates.

Effect estimation involves a different target: prediction of the outcome variable under at least two different exposure regimes and possibly under different covariate distributions than observed (65, 61, 83, 108, 140). Algorithms for passive prediction can be exploited for such causal prediction if care is taken to shift the final prediction target from passive to potential outcomes. Specifically, an individual's outcome under a treatment regime is only a potential outcome until that regime is chosen; after it is chosen, the other regimes become counterfactual for that individual, and outcomes under those regimes thus become unobservable. With this conceptualization, the problem of causal inference can be recast as a missing-data problem, opening the way for use of concepts and tools for missing-data analysis in causal inference (123, 140).

There are, however, elements of potential-outcome modeling that are not identified by statistical experiments, which have led to objections and alternatives (28, 29). Nonetheless, both potential-outcome models and their competitors must be elaborated considerably when exposure is ambiguously defined or is not an intervention or decision (e.g., age, sex) (56, 82, 144, 143). As argued by the discussants of Dawid (28), potential-outcome models are valuable if not indispensable for providing insights needed to develop and criticize causal questions and inference methods in these cases (56, 82, 83, 108).

SUMMARY AND CONCLUSIONS

Students are often taught that statistical significance testing is inappropriate for confounder evaluation and that CIE methods are preferable (14, 20, 60, 89, 125). Nonetheless, they are also often taught that the goal of modeling is to produce a model that is as simple as possible (parsimonious) while providing an adequate fit to the data, without recognizing that this goal is closer to that of passive prediction (as in ordinary stepwise regression) than it is to that of accurate estimation or prediction of effects. Confusion is exacerbated when the different goals of noncausal prediction and effect estimation are not made explicit because these differences lead to different modeling strategies. Further confusion arises from rigid application of idealized estimation criteria to messy observational settings in which the supporting theory is at best a thin and possibly misleading sketch of the real context.

We have emphasized that parsimony and goodness-of-fit are inappropriate end goals for modeling, as indicated by simulation studies in which full-model analysis sometimes outperforms conventional selection strategies (96, 101, 153). Conversely, however, rejecting any use of data-based model selection for causal inference is also inappropriate because it ignores the harsh reality that even databases of studies with hundreds of thousands of individuals often have limited numbers of pivotal observations (such as exposed cases). Coupled with the availability of what may be dozens or even hundreds of variables, some kind of targeted approach to error control is essential, accounting for both systematic and random errors. Taking these precautions will lead to restoration of parsimony as a heuristic, formalized in model-dimension (degree-of-freedom) reduction strategies such as shrinkage, algorithmic modeling, or combinations thereof (62, 85, 99, 158).

The simple strategies we have discussed are relevant to common borderline situations in which control of all potential confounders may be possible, yet benefits are expected from eliminating some or all variables whose inclusion is of uncertain value. The benefits may be theoretical, such as accuracy improvement, or practical, such as simplification of analysis or presentation. While we would prefer to see a shift toward methods with sound and relevant theoretical and practical foundations, the gap between state-of-the-art methodology and what is done in most publications has only grown despite early criticisms of common modeling strategies (94). This lag may be attributable in part to methodologic conservatism, which serves a purpose insofar as it takes time to evaluate new methodology in enough contexts to consider it adequately field tested. The accelerating pace of methodologic development might also be partly responsible.

Nonetheless, we also see major obstacles in the time and resource limits that constrain typical research teams. Our experience suggests that considerable software development and training will be needed to facilitate use of better methods before simple strategies can be retired. We discuss elsewhere some easily implemented methods with stronger theoretical justification and broader capabilities than the simple strategies reviewed above (59, 134). Regardless of modeling approach, however, we caution that no methodology is foolproof (especially when faced with uncontrolled confounding, selection bias, or measurement error) and that modeling methods should be documented in enough detail so that readers can interpret results in light of the strengths and weaknesses of those methods.

SUMMARY POINTS

1.

Because models always fall far short of the complex reality under study, there are no best or optimal strategies for modeling. Strategies promoted as optimal or correct have these properties only under highly simplified models. When many covariates are available, however, traditional modeling strategies are demonstrably inferior to more modern techniques.

2.

When we can discern a modeling strategy in epidemiological and medical studies, it is usually a variant of one of these approaches: (a) use of a model with all measured covariates a priori identified as potential confounders; (b) testing of covariate coefficients to eliminate some variables from a model predicting outcome or exposure (as in stepwise regression); or (c) using a change-in-estimate criterion to eliminate variables.

3.

The goal of testing covariate coefficients is to obtain a model that predicts observed outcomes or exposures well with a minimal number of variables, whereas the goal of change-in-estimate criteria is to obtain a model that controls most or all confounding with a minimal number of variables.

4.

The main goal of a statistical analysis of effects should be the production of the most accurate (least erroneous) effect estimates obtainable from the data and available software. None of the common strategies are targeted toward that goal, and better strategies are not yet integrated into popular commercial software.

5.

Regardless of the modeling strategy chosen, it is important that authors document the strategy used so that readers can interpret the results in light of the strategy's strengths and weaknesses.

disclosure statement

The authors are not aware of any affiliations, memberships, funding, or financial holdings that might be perceived as affecting the objectivity of this review.

acknowledgments

We thank Steven Cole, Katherine J. Hoggatt, Mohammad A. Mansournia, and Charles Poole for helpful comments on earlier drafts of this paper. Neil Pearce also thanks Tony Blakely, Naomi Brewer, Tim Clayton, Simon Cousens, Rhian Daniel, Shah Ebrahim, Chris Frost, Michael Hills, Nancy Krieger, Dorothea Nitsch, George Davey Smith, Lorenzo Richiardi, Bianca de Stavola, and Jan Vandenbroucke for their valuable comments on earlier work leading to the present paper. The Centre for Public Health Research is supported by a program grant from the Health Research Council of New Zealand.

literature cited

  • 1. 
    Ahrens W, Pigeot I, eds. 2014. Handbook of Epidemiology. New York: Springer. 2nd ed.
    • Crossref
    • Google Scholar
  • 2. 
    Arabatzis AA, Gregoire TG, Reynolds MR Jr. 1989. Conditional interval estimation of the mean following rejection of a two-sided test. Commun. Stat.-Theory Methods 18:4359–73
    • Crossref
    • Web of Science ®
    • Google Scholar
    Article Location
  • 3. 
    Austin PC. 2009. Balance diagnostics for comparing the distribution of baseline covariates between treatment groups in propensity-score matched samples. Stat. Med. 28:3083–107
    • Crossref
    • Medline
    • Web of Science ®
    • Google Scholar
    Article Location
  • 4. 
    Austin PC. 2014. A tutorial on the use of propensity score methods with survival time or time-to-event outcomes: reporting measures of effect similar to those used in randomized experiments. Stat. Med. 33:1242–58
    • Crossref
    • Medline
    • Web of Science ®
    • Google Scholar
    Article Locations:
    • Article Location
    • Article Location
  • 5. 
    Austin PC, Brunner LJ. 2004. Inflation of the type I error rate when a continuous confounding variable is categorized in logistic regression analyses. Stat. Med. 23:1159–78
    • Crossref
    • Medline
    • Web of Science ®
    • Google Scholar
    Article Location
  • 6. 
    Bancroft TA, Han CP. 1977. Inference based on conditional specification - note and a bibliography. Int. Stat. Rev. 45:117–27
    • Web of Science ®
    • Google Scholar
    Article Locations:
    • Article Location
    • Article Location
    • Article Location
  • 7. 
    Bang H, Robins JM. 2005. Doubly robust estimation in missing data and causal inference models. Biometrics 61:962–72. Erratum. 2008. Biometrics 64(2):650
    • Crossref
    • Medline
    • Web of Science ®
    • Google Scholar
    Article Locations:
    • Article Location
    • Article Location
    More AR articles citing this reference

    • Econometric Methods for Program Evaluation

      Alberto Abadie1 and Matias D. Cattaneo21Department of Economics, Massachusetts Institute of Technology, Cambridge, Massachusetts 02142, USA; email: [email protected]2Department of Economics and Department of Statistics, University of Michigan, Ann Arbor, Michigan 48109, USA; email: [email protected]
      Annual Review of Economics Vol. 10: 465 - 503
      • ...locally robust estimators) of Van der Laan & Robins (2003), Bang & Robins (2005), Cattaneo (2010), Farrell (2015), Chernozhukov et al. (2016), ...

  • 8. 
    Bayarri MJ, Berger JO. 2004. The interplay of Bayesian and frequentist analysis. Stat. Sci. 19:58–80
    • Crossref
    • Web of Science ®
    • Google Scholar
    Article Location
    More AR articles citing this reference

    • Evaluating Model Performance in Evolutionary Biology

      Jeremy M. Brown1 and Robert C. Thomson21Department of Biological Sciences and Museum of Natural Science, Louisiana State University, Baton Rouge, Louisiana 70803, USA; email: [email protected]2Department of Biology, University of Hawai'i, Honolulu, Hawai'i 96822, USA; email: [email protected]
      Annual Review of Ecology, Evolution, and Systematics Vol. 49: 95 - 114
      • ... and Bayarri & Berger 2004 for modern perspectives on their interplay), ...
    • There Is Individualized Treatment. Why Not Individualized Inference?

      Keli Liu1 and Xiao-Li Meng21Department of Statistics, Stanford University, Stanford, California 94305; email: [email protected]2Department of Statistics, Harvard University, Cambridge, Massachusetts 02138; email: [email protected]
      Annual Review of Statistics and Its Application Vol. 3: 79 - 111
      • Systematic Errors

        Joel Heinrich1 and Louis Lyons21Department of Physics and Astronomy, University of Pennsylvania, Philadelphia, Pennsylvania 19104; email: [email protected]2Department of Physics, University of Oxford, Oxford OX1 3RH, United Kingdom; email: [email protected]
        Annual Review of Nuclear and Particle Science Vol. 57: 145 - 169
        • ...Reference 31 advocates the use of coverage as a diagnostic for the selection of priors—one approach within a larger objective Bayesian methodology—and seems a practical way to reduce the ambiguity associated with the choice of prior....
      • The Posterior and the Prior in Bayesian Phylogenetics

        Michael E. Alfaro1 and Mark T. Holder21School of Biological Sciences, Washington State University, Pullman, Washington 99164-4236; email: [email protected]2School of Computational Science, Florida State University, Tallahassee, Florida 32306-4120; email: [email protected]
        Annual Review of Ecology, Evolution, and Systematics Vol. 37: 19 - 42
        • ...and they often result in analyses that behave well from a frequentist viewpoint (Bayarri & Berger 2004)....

    • 9. 
      Berger JO, Berry DA. 1988. Statistical analysis and the illusion of objectivity. Am. Sci. 76:159–65
      • Web of Science ®
      • Google Scholar
      Article Location
    • 10. 
      Berzuini C, Dawid AP, Bernardinelli L, eds. 2012. Causal Inference: Statistical Perspectives and Applications. Chichester, UK: Wiley
      • Crossref
      • Google Scholar
      More AR articles citing this reference

      • Statistical Causality from a Decision-Theoretic Perspective

        A. Philip DawidCentre for Mathematical Sciences, University of Cambridge, Cambridge CB3 0WB, United Kingdom; email: [email protected]
        Annual Review of Statistics and Its Application Vol. 2: 273 - 303

      • 11. 
        Blackwelder WC. 1998. Equivalence trials. In Encyclopedia of Biostatistics, ed. P Armitage, T Colton, pp. 1367–72. New York: Wiley
        • Google Scholar
        Article Location
      • 12. 
        Box GEP. 1980. Sampling and Bayes inference in scientific modeling and robustness. J. R. Stat. Soc. Ser. A 143:383–430
        • Crossref
        • Google Scholar
        Article Location
        More AR articles citing this reference

        • There Is Individualized Treatment. Why Not Individualized Inference?

          Keli Liu1 and Xiao-Li Meng21Department of Statistics, Stanford University, Stanford, California 94305; email: [email protected]2Department of Statistics, Harvard University, Cambridge, Massachusetts 02138; email: [email protected]
          Annual Review of Statistics and Its Application Vol. 3: 79 - 111
          • Climate Simulators and Climate Projections

            Jonathan Rougier1 and Michael Goldstein21Department of Mathematics, University of Bristol, Bristol, BS8 1TW, United Kingdom; email: [email protected]2Department of Mathematical Sciences, University of Durham, Durham, DH1 3LE, United Kingdom; email: [email protected]
            Annual Review of Statistics and Its Application Vol. 1: 103 - 123
            • ...(Box 1980)....
            • ...15Boundary values such as ρ=1 and σ2=0 give the appearance of objectivity but are of course just as subjective as any other values and less defensible than many. Box (1980, ...
          • Build, Compute, Critique, Repeat: Data Analysis with Latent Variable Models

            David M. BleiComputer Science Department, Princeton University, Princeton, New Jersey 08540; email: [email protected]
            Annual Review of Statistics and Its Application Vol. 1: 203 - 232
            • ...really—of the ideas of George Box and collaborators beginning in the 1960s (Box & Hunter 1962, 1965; Box & Hill 1967; Box 1976, 1980)....
            • ... and posterior predictive checks (PPCs) (Box 1980, Rubin 1984, Meng 1994, Gelman et al. 1996)....
            • ... and posterior predictive checks (Box 1980, Rubin 1984, Meng 1994, Gelman et al. 1996)....
            • ...we step out of the formal Bayesian framework to ask how well the model (and inference) captures the underlying data distribution, a perspective that is beautifully laid out in Box (1980)...
            • ...note that model criticism can be a philosophically controversial idea because it is incoherent under certain theoretical frameworks [see the discussion of Box (1980)...
            • ...Posterior predictive checks (PPCs) (Guttman 1967, Box 1980, Rubin 1984, Meng 1994, Gelman et al. 1996) help answer the important question, ...

        • 13. 
          Box GEP. 1990. Comment. Stat. Sci. 5:390–91
          • Crossref
          • Google Scholar
          Article Location
        • 14. 
          Breslow NE, Day NE. 1980. Statistical Methods in Cancer Research. Volume 1: The Analysis of Case-Control Studies. Lyon: IARC
          • Google Scholar
          Article Locations:
          • Article Location
          • Article Location
          • Article Location
          • Article Location
        • 15. 
          Brookhart MA, Schneeweiss S, Rothman KJ, Glynn RJ, Avorn J, Sturmer T. 2006. Variable selection for propensity score models. Am. J. Epidemiol. 163:1149–56
          • Crossref
          • Medline
          • Web of Science ®
          • Google Scholar
          Article Locations:
          • Article Location
          • Article Location
          • Article Location
          • Article Location
        • 16. 
          Carlin BP, Louis TA. 2000. Bayes and Empirical Bayes Methods of Data Analysis. New York: Chapman & Hall. 2nd ed.
          • Crossref
          • Google Scholar
          Article Locations:
          • Article Location
          • Article Location
          More AR articles citing this reference

          • There Is Individualized Treatment. Why Not Individualized Inference?

            Keli Liu1 and Xiao-Li Meng21Department of Statistics, Stanford University, Stanford, California 94305; email: [email protected]2Department of Statistics, Harvard University, Cambridge, Massachusetts 02138; email: [email protected]
            Annual Review of Statistics and Its Application Vol. 3: 79 - 111
            • Confronting Prior Convictions: On Issues of Prior Sensitivity and Likelihood Robustness in Bayesian Analysis

              Hedibert F. Lopes1 and Justin L. Tobias21Booth School of Business, University of Chicago, Chicago, Illinois 60637; email: [email protected]2Economics Department, Purdue University, Lafayette, Indiana 47907; email: [email protected]
              Annual Review of Economics Vol. 3: 107 - 131
              • ...including Poirier (1995), Carlin & Louis (2000), Koop (2003), Gelman et al. (2004), Lancaster (2004), Geweke (2005), Gamerman & Lopes (2006), Koop et al. (2007), ...
            • BAYESIAN ANALYSIS FOR POLITICAL RESEARCH

              Simon JackmanDepartment of Political Science, Stanford University, Stanford, California 94305-6044; email: [email protected]
              Annual Review of Political Science Vol. 7: 483 - 505
              • ...Detailed treatments of these topics and others appear in the book-length treatments of Carlin & Louis (2000)...

          • 17. 
            Carroll RJ, Ruppert D, Stefanski LA, Craineceanu C. 2006. Measurement Error in Non-Linear Models. Boca Raton, FL: Chapman & Hall
            • Crossref
            • Google Scholar
            Article Location
            More AR articles citing this reference

            • Agricultural Crop Forecasting for Large Geographical Areas

              Linda J. YoungResearch and Development Division, USDA National Agricultural Statistics Service, Washington, DC 20250, USA; email: [email protected]
              Annual Review of Statistics and Its Application Vol. 6: 173 - 196
              • ...and thus is measured with error (Carroll et al. 2006, Fuller 2006)....
            • Handling Missing Data in Instrumental Variable Methods for Causal Inference

              Edward H. Kennedy,1 Jacqueline A. Mauro,1 Michael J. Daniels,2 Natalie Burns,2 Dylan S. Small31Department of Statistics and Data Science, Carnegie Mellon University, Pittsburgh, Pennsylvania 15213, USA2Department of Statistics, University of Florida, Gainesville, Florida 32611, USA3Department of Statistics, The Wharton School, University of Pennsylvania, Philadelphia, Pennsylvania 19104, USA; email: [email protected]
              Annual Review of Statistics and Its Application Vol. 6: 125 - 148
              • ...the presence of unmeasured pretreatment differences between the treated and control group (another motivation for using IV methods that we do not focus on in this review is for handling measurement error; see Carroll et al. 2006, ...
            • Cure Models in Survival Analysis

              Maïlis Amico and Ingrid Van KeilegomResearch Centre for Operations Research and Business Statistics (ORSTAT), KU Leuven, 3000 Leuven, Belgium; email: [email protected]
              Annual Review of Statistics and Its Application Vol. 5: 311 - 342
              • ...Ignoring this measurement error can lead to wrong conclusions since the presence of measurement error leads to biased estimators (see, e.g., Carroll et al. 2006)....
            • Informatics and Data Analytics to Support Exposome-Based Discovery for Public Health

              Arjun K. Manrai,1 Yuxia Cui,2 Pierre R. Bushel,2 Molly Hall,3 Spyros Karakitsios,4 Carolyn J. Mattingly,5 Marylyn Ritchie,3,6 Charles Schmitt,7 Denis A. Sarigiannis,4 Duncan C. Thomas,8 David Wishart,9 David M. Balshaw,2 and Chirag J. Patel1,101Department of Biomedical Informatics, Harvard Medical School, Boston, Massachusetts 02115; email: [email protected]2National Institute of Environmental Health Sciences, National Institutes of Health, Research Triangle Park, North Carolina 27709; email: [email protected]3Center for Systems Genomics, The Pennsylvania State University, College Station, Pennsylvania 168024Environmental Engineering Laboratory, Department of Chemical Engineering, Aristotle University of Thessaloniki, Thessaloniki 54124, Greece5Department of Biological Sciences, College of Sciences, North Carolina State University, Raleigh, North Carolina 276956Geisinger Health System, Danville, Pennsylvania 178217Renaissance Computing Institute, Chapel Hill, North Carolina 275178Division of Biostatistics, Department of Preventive Medicine, Keck School of Medicine, University of Southern California, Los Angeles, California 90089-90119Departments of Biological Sciences and Computing Science, University of Alberta, Edmonton, Alberta T6G 2E8, Canada10Center for Assessment Technology and Continuous Health, Massachusetts General Hospital, Boston, Massachusetts 02114
              Annual Review of Public Health Vol. 38: 279 - 294
              • ...the problem of exposure measurement error has been widely discussed in the statistical and epidemiological literature (5)....
            • Understanding and Assessing Nutrition

              Alicia L. CarriquiryDepartment of Statistics, Iowa State University, Ames, Iowa 50011; email: [email protected]
              Annual Review of Statistics and Its Application Vol. 4: 123 - 146
              • ...If ρ denotes the correlation between X and W (Carroll et al. 2006), ...
            • Recent Advances in the Measurement Error Literature

              Susanne M. SchennachDepartment of Economics, Brown University, Providence, Rhode Island 02912; email: [email protected]
              Annual Review of Economics Vol. 8: 341 - 377
              • ...The popular book by Carroll et al. (2006) covers nonlinear measurement error models, ...
            • Probabilistic Record Linkage in Astronomy: Directional Cross-Identification and Beyond

              Tamás Budavári1 and Thomas J. Loredo21Department of Applied Mathematics and Statistics, The Johns Hopkins University, Baltimore, Maryland 21218; email: [email protected]2Center for Radiophysics and Space Research, Cornell University, Ithaca, New York 14853; email: [email protected]
              Annual Review of Statistics and Its Application Vol. 2: 113 - 139
              • ...Our framework corresponds to a Bayesian measurement error model (Carroll et al. 2006, ...
            • Approaches to Uncertainty in Exposure Assessment in Environmental Epidemiology

              Donna SpiegelmanDepartments of Epidemiology and Biostatistics, Harvard School of Public Health, Boston, Massachusetts 02115; email: [email protected]
              Annual Review of Public Health Vol. 31: 149 - 163
              • ...exposure misclassification probably occurs in nearly every epidemiologic study.” Reviews of the effects of exposure measurement error and the numerous methods that have been proposed to correct for biases that result when exposure measurement error is present have been published (1, 8, 10, 26, 46)....
              • ...using a second-order Taylor series expansion around the conditional mean of E given X, has been given previously (8, 9), ...

          • 18. 
            Chatfield C. 1995. Model uncertainty, data mining and statistical inference. J. R. Stat. Soc. Ser. A 158:419–66
            • Crossref
            • Google Scholar
            Article Location
          • 19. 
            Chatfield C. 2002. Confessions of a pragmatic statistician. J. R. Stat. Soc. Ser. D 51:1–20
            • Crossref
            • Google Scholar
            Article Location
          • 20. 
            Checkoway H, Pearce N, Kriebel D. 2004. Research Methods in Occupational Epidemiology. New York: Oxford Univ. Press
            • Crossref
            • Google Scholar
            Article Location
          • 21. 
            Chiou P. 1997. Interval estimation of scale parameters following a pre-test for two exponential distributions. Comput. Stat. Data Anal. 23:477–89
            • Crossref
            • Web of Science ®
            • Google Scholar
            Article Location
          • 22. 
            Chiou P, Han CP. 1995. Conditional interval estimation of the exponential location parameter following rejection of a pretest. Commun. Stat.-Theory Methods 24:1481–92
            • Crossref
            • Web of Science ®
            • Google Scholar
            Article Location
          • 23. 
            Colditz GA, Manson JE, Hankinson SE. 1997. The Nurses' Health Study: 20-year contribution to the understanding of health among women. J. Womens Health 6:49–62
            • Crossref
            • Medline
            • Web of Science ®
            • Google Scholar
            Article Location
          • 24. 
            Cole SR, Platt RW, Schisterman EF, Chu HT, Westreich D, et al. 2010. Illustrating bias due to conditioning on a collider. Int. J. Epidemiol. 39:417–20
            • Crossref
            • Medline
            • Web of Science ®
            • Google Scholar
            Article Locations:
            • Article Location
            • Article Location
            • Article Location
            More AR articles citing this reference

            • Accounting for Confounding in Observational Studies

              Brian M. D'Onofrio,1,2 Arvid Sjölander,2 Benjamin B. Lahey,3 Paul Lichtenstein,2 and A. Sara Öberg2,41Department of Psychological and Brain Sciences, Indiana University, Bloomington, Indiana 47405, USA; email: [email protected]2Department of Medical Epidemiology and Biostatistics, Karolinska Institutet, SE-171 77 Stockholm, Sweden; email: [email protected], [email protected], [email protected]3Departments of Health Studies and Psychiatry and Behavioral Neuroscience, University of Chicago, Chicago, Illinois 60637, USA; email: [email protected]4Department of Epidemiology, Harvard T.H. Chan School of Public Health, Boston, Massachusetts 02115, USA
              Annual Review of Clinical Psychology Vol. 16: 25 - 48
              • ...but there are several helpful published reviews (Cole et al. 2009, Elwert & Winship 2014)....
            • Inferring Causal Relationships Between Risk Factors and Outcomes from Genome-Wide Association Study Data

              Stephen Burgess,1,2 Christopher N. Foley,1 and Verena Zuber11MRC Biostatistics Unit, University of Cambridge, Cambridge CB2 0SR, United Kingdom; email: [email protected]2Cardiovascular Epidemiology Unit, University of Cambridge, Cambridge CB1 8RN, United Kingdom
              Annual Review of Genomics and Human Genetics Vol. 19: 303 - 327
              • ...A collider is a variable that is a common effect of two variables (that is, it is causally downstream of the two variables) (35)....
            • Causal Inference in Developmental Origins of Health and Disease (DOHaD) Research

              Suzanne H. Gage,1,2 Marcus R. Munafò,1,2 and George Davey Smith11MRC Integrative Epidemiology Unit (IEU) at the University of Bristol, Bristol BS8 2BN, United Kingdom; email: [email protected]2UK Center for Tobacco and Alcohol Studies, School of Experimental Psychology, University of Bristol, Bristol BS8 1TU, United Kingdom
              Annual Review of Psychology Vol. 67: 567 - 585
              • ...This is a form of collider bias (Cole et al. 2010)—a family of biases that can distort observational estimates of exposure effects—that has perhaps been underappreciated in the literature until recently....
            • A Systematic Statistical Approach to Evaluating Evidence from Observational Studies

              David Madigan,1,2 Paul E. Stang,2,3 Jesse A. Berlin,4 Martijn Schuemie,2,3 J. Marc Overhage,2,5 Marc A. Suchard,2,6,7,8 Bill Dumouchel,2,9 Abraham G. Hartzema,2,10 and Patrick B. Ryan2,31Department of Statistics, Columbia University, New York, New York 10027; email: [email protected]2Observational Medical Outcomes Partnership, Foundation for the National Institutes of Health, Bethesda, Maryland 208103Janssen Research and Development LLC, Titusville, New Jersey, 085604Johnson & Johnson, New Brunswick, New Jersey, 08901; email: [email protected], [email protected], [email protected], [email protected]5Siemens Health Services, Malvern, Pennsylvania, 19355; email: [email protected]6Department of Biomathematics, David Geffen School of Medicine, University of California, Los Angeles, California, 90095; email: [email protected]7Department of Human Genetics, David Geffen School of Medicine, University of California, Los Angeles, California, 900958Department of Biostatistics, Fielding School of Public Health, University of California, Los Angeles, California, 900959Oracle Health Sciences, Burlington, Massachusetts, 01803; email: [email protected]10College of Pharmacy, University of Florida, Gainesville, Florida, 32610; email: [email protected]
              Annual Review of Statistics and Its Application Vol. 1: 11 - 39
              • ...bias can also be induced if an analysis improperly stratifies on a collider variable (Cole et al. 2010), ...

          • 25. 
            Copas JB. 1983. Regression, prediction and shrinkage. J. R. Stat. Soc. Ser. B 45:311–54
            • Google Scholar
            Article Location
          • 26. 
            Courvoisier DS, Combescure C, Agoritsas T, Gayet-Ageron A, Perneger TV. 2011. Performance of logistic regression modeling: beyond the number of events per variable, the role of data structure. J. Clin. Epidemiol. 64:1463–64
            • Crossref
            • Web of Science ®
            • Google Scholar
            Article Location
          • 27. 
            Dales LD, Ury HK. 1978. An improper use of statistical significance testing in studying covariables. Int. J. Epidemiol. 4:373–75
            • Crossref
            • Web of Science ®
            • Google Scholar
            Article Locations:
            • Article Location
            • Article Location
          • 28. 
            Dawid AP. 2000. Causal inference without counterfactuals. J. Am. Stat. Assoc. 95:407–48
            • Crossref
            • Web of Science ®
            • Google Scholar
            Article Locations:
            • Article Location
            • Article Location
            More AR articles citing this reference

            • Bayesian Additive Regression Trees: A Review and Look Forward

              Jennifer Hill,1 Antonio Linero,2 and Jared Murray31Department of Applied Statistics, Social Science, and Humanities, New York University, New York, NY 10003, USA; email: [email protected]2Department of Statistics and Data Sciences, University of Texas, Austin, Texas 78712, USA3Department of Information, Risk, and Operations Management, McCombs School of Business, University of Texas, Austin, Texas 78712, USA
              Annual Review of Statistics and Its Application Vol. 7: 251 - 278
              • ...2While this is the dominant perspective among methodological researchers in causal inference we note that dissenting opinions exist (for example Dawid 2000)....
            • Causal Structure Learning

              Christina Heinze-Deml, Marloes H. Maathuis, and Nicolai MeinshausenSeminar for Statistics, Department of Mathematics, ETH Zurich, CH-8092 Zurich, Switzerland; email: [email protected], [email protected], [email protected]
              Annual Review of Statistics and Its Application Vol. 5: 371 - 391
              • ...and any statement about the joint distribution of the noise under different interventions is thus in principle unfalsifiable and untestable (Dawid 2000)....
            • Statistical Causality from a Decision-Theoretic Perspective

              A. Philip DawidCentre for Mathematical Sciences, University of Cambridge, Cambridge CB3 0WB, United Kingdom; email: [email protected]
              Annual Review of Statistics and Its Application Vol. 2: 273 - 303
              • ...For further details and developments of the material in this article, the reader is referred to Berzuini et al. 2012b,c; Dawid 2000, 2002, 2003, 2007a, 2010a,b, 2011, 2012...
              • ...as well as their requisite infrastructures, in some detail in another article (Dawid 2000), ...
              • ...is as follows [see section 8.1 of Dawid (2000) for a more detailed account]: Because of different selection criteria, ...
            • Causal Inference in Public Health

              Thomas A. Glass,1 Steven N. Goodman,2 Miguel A. Hernán,3,4 and Jonathan M. Samet51Department of Epidemiology, Bloomberg School of Public Health, Johns Hopkins University, Baltimore, Maryland 21205; email: [email protected]2Department of Medicine, Stanford University, Palo Alto, California 94305; email: [email protected]3Departments of Epidemiology and Biostatistics, School of Public Health, Harvard University, Boston, Massachusetts 021154Harvard-MIT Division of Health Sciences and Technology, Cambridge, Massachusetts 02115; email: [email protected]5Department of Preventive Medicine, Keck School of Medicine, and USC Institute for Global Health, University of Southern California, Los Angeles, California 90089; email: [email protected]
              Annual Review of Public Health Vol. 34: 61 - 75
              • ...or age) that cannot be reasonably translated into hypothetical interventions (4, 41, 42, 47, 66)? And how should investigators address individual (e.g., ...

          • 29. 
            Dawid AP. 2012. The decision-theoretic approach to causal inference. See Ref. 10, pp. 25–42
            • Google Scholar
            Article Location
          • 30. 
            Day NE, Byar DP, Green SB. 1980. Overadjustment in case-control studies. Am. J. Epidemiol. 112:696–706
            • Crossref
            • Medline
            • Web of Science ®
            • Google Scholar
            Article Location
          • 31. 
            De Luna X, Waernbaum I, Richardson TS. 2011. Covariate selection for the nonparametric estimation of an average treatment effect. Biometrika 98:861–75
            • Crossref
            • Web of Science ®
            • Google Scholar
            Article Locations:
            • Article Location
            • Article Location
            More AR articles citing this reference

            • Bias Analysis for Uncontrolled Confounding in the Health Sciences

              Onyebuchi A. ArahDepartment of Epidemiology, Fielding School of Public Health; UCLA Center for Health Policy Research; and California Center for Population Research, University of California, Los Angeles, California 90095; email: [email protected]
              Annual Review of Public Health Vol. 38: 23 - 38
              • ...These criteria subsume a scenario in which the unmeasured variable is a common cause of the exposure and outcome (35, 42, 65)....

          • 32. 
            Draper NR, Guttman I, Lapczak L. 1979. Actual rejection levels in a certain stepwise test. Commun. Stat. A 8:99–105
            • Crossref
            • Web of Science ®
            • Google Scholar
            Article Locations:
            • Article Location
            • Article Location
            • Article Location
          • 33. 
            Efron B. 1975. Biased versus unbiased estimation. Adv. Math. 16:259–77
            • Crossref
            • Web of Science ®
            • Google Scholar
            Article Locations:
            • Article Location
            • Article Location
          • 34. 
            Efron B. 2005. Bayesians, frequentists, and scientists. J. Am. Stat. Assoc. 100:1–5
            • Crossref
            • Web of Science ®
            • Google Scholar
            Article Location
          • 35. 
            Efron B, Morris C. 1973. Stein's estimation rule and its competitors—empirical Bayes approach. J. Am. Stat. Assoc. 68:117–30
            • Web of Science ®
            • Google Scholar
            Article Location
          • 36. 
            Efron B, Morris C. 1975. Data analysis using Stein's estimator and its generalizations. J. Am. Stat. Assoc. 70:311–19
            • Crossref
            • Web of Science ®
            • Google Scholar
            Article Location
            More AR articles citing this reference

            • There Is Individualized Treatment. Why Not Individualized Inference?

              Keli Liu1 and Xiao-Li Meng21Department of Statistics, Stanford University, Stanford, California 94305; email: [email protected]2Department of Statistics, Harvard University, Cambridge, Massachusetts 02138; email: [email protected]
              Annual Review of Statistics and Its Application Vol. 3: 79 - 111
              • Statistical and Substantive Inferences in Public Health: Issues in the Application of Multilevel Models

                Jeffrey B. Bingenheimer1 and Stephen W. Raudenbush2 1 School of Public Health, University of Michigan,
                Ann Arbor, Michigan 48109
                ; email: [email protected] 2 and 2School of Education, Survey Research Center, and Departments of Statistics and Sociology, University of Michigan,
                Ann Arbor, Michigan 48109
                ; email: [email protected]
                Annual Review of Public Health Vol. 25: 53 - 77
                • ...have been used to characterize toxoplasmosis rates in cities in El Salvador (39, 104); to obtain age- and sex-specific stomach, ...

            • 37. 
              Efroymson MA. 1960. Multiple regression analysis. In Mathematical Methods for Digital Computers, ed. A Ralston, HS Wilf, pp. 191–203. Hoboken, NJ: Wiley
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              More AR articles citing this reference

              • Big Data in Public Health: Terminology, Machine Learning, and Privacy

                Stephen J. Mooney1 and Vikas Pejaver21Harborview Injury Prevention and Research Center, University of Washington, Seattle, Washington 98122, USA; email: [email protected]2Department of Biomedical Informatics and Medical Education and the eScience Institute, University of Washington, Seattle, Washington 98109, USA; email: [email protected]
                Annual Review of Public Health Vol. 39: 95 - 112
                • ...the familiar stepwise regression technique is one such wrapper method (30, 127)....

            • 38. 
              Faraway JJ. 1992. On the cost of data analysis. J. Comput. Graph. Stat. 1:213–19
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              • Article Location
            • 39. 
              Feyerabend P. 1993. Against Method. New York: Verso
              • Google Scholar
              Article Location
            • 40. 
              Flack VF, Chang PC. 1987. Frequency of selecting noise variables in subset regression analysis—a simulation study. Am. Stat. 41:84–86
              • Web of Science ®
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              • Article Location
            • 41. 
              Freedman DA. 1983. A note on screening regression equations. Am. Stat. 37:152–55
              • Web of Science ®
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              • Article Location
            • 42. 
              Freedman DA, Navidi W, Peters SC. 1988. On the impact of variable selection in fitting regression equations. In On Model Uncertainty and Its Statistical Implications, ed. TK Dijlestra, pp. 1–16. Berlin: Springer-Verlag
              • Crossref
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              • Article Location
              More AR articles citing this reference

              • Forecasting Methods in Crime and Justice

                Richard BerkDepartment of Statistics, Department of Criminology, University of Pennsylvania, Philadelphia, Pennsylvania 19104; email: [email protected]
                Annual Review of Law and Social Science Vol. 4: 219 - 238
                • ...there are now several important papers showing that no such tests can be sensibly constructed. Freedman and colleagues (1988) made an early contribution by showing that resampling procedures such as the bootstrap do not solve the problem....

            • 43. 
              Gelman A, Shalizi CR. 2013. Philosophy and the practice of Bayesian statistics. Br. J. Math. Stat. Psychol. 66:8–38
              • Crossref
              • Medline
              • Web of Science ®
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              More AR articles citing this reference

              • Correlational Studies in Typological and Historical Linguistics

                D. Robert Ladd,1 Seán G. Roberts,2 and Dan Dediu3,4 1Department of Linguistics and English Language, University of Edinburgh, Edinburgh EH8 9AD, United Kingdom; email: [email protected] 2Language and Cognition Department, 3Language and Genetics Department, Max Planck Institute for Psycholinguistics, Nijmegen 6500 AH, Netherlands 4The Donders Institute for Brain, Cognition and Behaviour, Radboud University, Nijmegen 6525 EN, Netherlands
                Annual Review of Linguistics Vol. 1: 221 - 241
                • ...In this Bayesian view of science (see, e.g., Easwaran 2013, Talbott 2013, but also see Gelman & Shalizi 2013), ...
              • Build, Compute, Critique, Repeat: Data Analysis with Latent Variable Models

                David M. BleiComputer Science Department, Princeton University, Princeton, New Jersey 08540; email: [email protected]
                Annual Review of Statistics and Its Application Vol. 1: 203 - 232
                • ...these methods connect to ideas of falsification in the philosophy of science (Popper 1959, Gelman & Shalizi 2012)....

            • 44. 
              Glymour MM, Greenland S. 2008. Causal diagrams. See Ref. 120, pp. 183–209
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              • Article Location
              • Article Location
            • 45. 
              Greenland S. 1986. Partial and marginal matching in case-control studies. In Modern Statistical Methods in Chronic Disease Epidemiology, ed. SH Moolgavkar, RL Prentice, pp. 35–49. New York: Wiley
              • Google Scholar
              Article Location
            • 46. 
              Greenland S. 1989. Comment: cautions in the use of preliminary test estimators. Stat. Med. 8:669–73
              • Crossref
              • Web of Science ®
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              • Article Location
              • Article Location
            • 47. 
              Greenland S. 1989. Modeling and variable selection in epidemiologic analysis. Am. J. Public Health 79:340–49
              • Crossref
              • Medline
              • Web of Science ®
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              • Article Location
              • Article Location
              • Article Location
              • Article Location
              • Article Location
              • Article Location
              • Article Location
            • 48. 
              Greenland S. 1990. Randomization, statistics, and causal inference. Epidemiology 1:421–29
              • Crossref
              • Medline
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              More AR articles citing this reference

              • Commentary: Causal Inference for Social Exposures

                Jay S. KaufmanDepartment of Epidemiology, Biostatistics, and Occupational Health, McGill University, Montreal, Quebec H3A 1A2, Canada; email: [email protected]
                Annual Review of Public Health Vol. 40: 7 - 21
                • ...Randomization at least provides the assurance of comparability, in expectation, across the treatment groups (17)....
              • Confounding in Health Research

                Sander Greenland1,2 and Hal Morgenstern11Department of Epidemiology, Los Angeles School of Public Health, University of California, Los Angeles, California 90095-1772; e-mail: [email protected] 2Department of Statistics, Los Angeles College of Letters and Science, University of California, Los Angeles, California 90095
                Annual Review of Public Health Vol. 22: 189 - 212
                • ...Counterfactual approaches to causal inference emphasize the importance of randomization in assuring identifiability of causal effects (30, 36, 38, 62, 70, 86, 91, 92, 93)....
                • ...and compliance and because of random differences between comparison groups (22, 30, 36, 38, 87)....

            • 49. 
              Greenland S. 1991. Reducing mean squared error in the analysis of stratified epidemiologic studies. Biometrics 47:773–75
              • Crossref
              • Medline
              • Web of Science ®
              • Google Scholar
              Article Location
            • 50. 
              Greenland S. 1993. Methods for epidemiologic analyses of multiple exposures: a review and a comparative study of maximum-likelihood, preliminary testing, and empirical-Bayes regression. Stat. Med. 12:717–36
              • Crossref
              • Medline
              • Web of Science ®
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              More AR articles citing this reference

              • Statistical and Substantive Inferences in Public Health: Issues in the Application of Multilevel Models

                Jeffrey B. Bingenheimer1 and Stephen W. Raudenbush2 1 School of Public Health, University of Michigan,
                Ann Arbor, Michigan 48109
                ; email: [email protected] 2 and 2School of Education, Survey Research Center, and Departments of Statistics and Sociology, University of Michigan,
                Ann Arbor, Michigan 48109
                ; email: [email protected]
                Annual Review of Public Health Vol. 25: 53 - 77
                • ...such as the simultaneous estimation of effects of multiple exposures on health (57, 154)...

            • 51. 
              Greenland S. 1997. Estimating relative risk functions in case-control studies using a nonparametric logistic regression. Am. J. Epidemiol. 146:883–84
              • Crossref
              • Medline
              • Web of Science ®
              • Google Scholar
              Article Location
            • 52. 
              Greenland S. 2000. Principles of multilevel modelling. Int. J. Epidemiol. 29:158–67
              • Crossref
              • Medline
              • Web of Science ®
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              More AR articles citing this reference

              • Confounding in Health Research

                Sander Greenland1,2 and Hal Morgenstern11Department of Epidemiology, Los Angeles School of Public Health, University of California, Los Angeles, California 90095-1772; e-mail: [email protected] 2Department of Statistics, Los Angeles College of Letters and Science, University of California, Los Angeles, California 90095
                Annual Review of Public Health Vol. 22: 189 - 212
                • ...although flexibility in using these degrees of freedom can be greatly improved via hierarchical regression (mixed or multilevel modeling) (32, 32A)....

            • 53. 
              Greenland S. 2000. When should epidemiologic regressions use random coefficients? Biometrics 56:915–21
              • Crossref
              • Medline
              • Web of Science ®
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              More AR articles citing this reference

              • Confounding in Health Research

                Sander Greenland1,2 and Hal Morgenstern11Department of Epidemiology, Los Angeles School of Public Health, University of California, Los Angeles, California 90095-1772; e-mail: [email protected] 2Department of Statistics, Los Angeles College of Letters and Science, University of California, Los Angeles, California 90095
                Annual Review of Public Health Vol. 22: 189 - 212
                • ...although flexibility in using these degrees of freedom can be greatly improved via hierarchical regression (mixed or multilevel modeling) (32, 32A)....

            • 54. 
              Greenland S. 2001. Putting background information about relative risks into conjugate prior distributions. Biometrics 57:663–70
              • Crossref
              • Medline
              • Web of Science ®
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
            • 55. 
              Greenland S. 2002. A review of multilevel theory for ecologic analyses. Stat. Med. 21:389–95
              • Crossref
              • Medline
              • Web of Science ®
              • Google Scholar
              Article Location
              More AR articles citing this reference

              • Ecologic Studies Revisited

                Jonathan WakefieldDepartments of Statistics and Biostatistics, University of Washington, Seattle, Washington 98195; email: [email protected]
                Annual Review of Public Health Vol. 29: 75 - 90
                • ...Further discussion may be found in Reference 27, and interpretation of more complex models is provided in References 2...

            • 56. 
              Greenland S. 2005. Epidemiologic measures and policy formulation: lessons from potential outcomes. Emerg. Themes Epidemiol. 2:5
              • Crossref
              • Medline
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              More AR articles citing this reference

              • Commentary: Causal Inference for Social Exposures

                Jay S. KaufmanDepartment of Epidemiology, Biostatistics, and Occupational Health, McGill University, Montreal, Quebec H3A 1A2, Canada; email: [email protected]
                Annual Review of Public Health Vol. 40: 7 - 21
                • ...The second common form of verification is evaluation of policy changes (19)....

            • 57. 
              Greenland S. 2005. Multiple-bias modeling for observational studies (with discussion). J. R. Stat. Soc. Ser. A 168:267–308
              • Crossref
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              More AR articles citing this reference

              • Bias Analysis for Uncontrolled Confounding in the Health Sciences

                Onyebuchi A. ArahDepartment of Epidemiology, Fielding School of Public Health; UCLA Center for Health Policy Research; and California Center for Population Research, University of California, Los Angeles, California 90095; email: [email protected]
                Annual Review of Public Health Vol. 38: 23 - 38
                • ...One of the oldest and most commonly used methods for adjusting an association estimate for an unmeasured confounder is the use of a bias formula to calculate a bias factor (3–11, 14, 17, 19, 20, 23, 32, 44, 53, 68)....
                • ...Several techniques based on the idea of bias formulas have been in use for more than half a century (3–11, 14, 17, 19, 20, 23, 32, 44, 53, 68)...
              • Incorporating Both Randomized and Observational Data into a Single Analysis

                Eloise E. KaizarDepartment of Statistics, The Ohio State University, Columbus, Ohio 43210; email: [email protected]
                Annual Review of Statistics and Its Application Vol. 2: 49 - 72
                • ...For example, Greenland (2005, 2009) suggests multiple bias models, focusing on methods for estimation based on sensitivity analyses and discussing technical issues that arise in models with high-dimensional bias parameters....
              • EMF AND HEALTH

                Maria Feychting,1 Anders Ahlbom,1,2 and Leeka Kheifets31Institute of Environmental Medicine, Karolinska Institutet, S-171 77 Stockholm, Sweden; email: [email protected] 2Division of Epidemiology, Stockholm Center of Public Health, 171 76 Stockholm, Sweden; email: Anders.[email protected] 3Department of Epidemiology, School of Public Health, University of California, Los Angeles, California 90095-1772; email: [email protected]
                Annual Review of Public Health Vol. 26: 165 - 189
                • ...a case-control study in Italy, an evaluation of selection bias (24) (G....

            • 58. 
              Greenland S. 2006. Bayesian perspectives for epidemiological research: I. Foundations and basic methods. Int. J. Epidemiol. 35:765–75
              • Crossref
              • Medline
              • Web of Science ®
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              • Article Location
            • 59. 
              Greenland S. 2007. Bayesian perspectives for epidemiological research. II. Regression analysis. Int. J. Epidemiol. 36:195–202
              • Crossref
              • Medline
              • Web of Science ®
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              • Article Location
              • Article Location
              • Article Location
            • 60. 
              Greenland S. 2008. Introduction to regression modeling. See Ref. 120, pp. 418–55
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              • Article Location
              • Article Location
              • Article Location
              • Article Location
              • Article Location
            • 61. 
              Greenland S. 2008. Introduction to regression models. See Ref. 120, pp. 381–417
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              • Article Location
              • Article Location
              • Article Location
              • Article Location
            • 62. 
              Greenland S. 2008. Invited commentary: Variable selection versus shrinkage in the control of multiple confounders. Am. J. Epidemiol. 167:523–29
              • Crossref
              • Medline
              • Web of Science ®
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
            • 63. 
              Greenland S. 2009. Dealing with uncertainty about investigator bias: Disclosure is informative. J. Epidemiol. Community Health 63:593–98
              • Crossref
              • Medline
              • Web of Science ®
              • Google Scholar
              Article Location
            • 64. 
              Greenland S. 2009. Relaxation penalties and priors for plausible modeling of nonidentified bias sources. Stat. Sci. 24:195–210
              • Crossref
              • Web of Science ®
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              More AR articles citing this reference

              • Incorporating Both Randomized and Observational Data into a Single Analysis

                Eloise E. KaizarDepartment of Statistics, The Ohio State University, Columbus, Ohio 43210; email: [email protected]
                Annual Review of Statistics and Its Application Vol. 2: 49 - 72
                • ...For example, Greenland (2005, 2009) suggests multiple bias models, focusing on methods for estimation based on sensitivity analyses and discussing technical issues that arise in models with high-dimensional bias parameters....

            • 65. 
              Greenland S. 2012. Causal inference as a prediction problem: assumptions, identification, and evidence synthesis. See Ref. 10, pp. 43–58
              • Google Scholar
              Article Location
            • 66. 
              Greenland S. 2012. Intuitions, simulations, theorems: the role and limits of methodology. Epidemiology 23:440–42
              • Crossref
              • Medline
              • Web of Science ®
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
            • 67. 
              Greenland S. 2014. Regression methods for epidemiological analysis. See Ref. 1, pp. 1087–159
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
            • 68. 
              Greenland S. 2014. Sensitivity analysis and bias analysis. See Ref. 1, pp. 685–706
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
            • 69. 
              Greenland S, Lash TL. 2008. Bias analysis. See Ref. 120, pp. 345–80
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
            • 70. 
              Greenland S, Maclure M, Schlesselman JJ, Poole C. 1991. Standardized regression coefficients: a further critique and review of some alternatives. Epidemiology 2:387–92
              • Crossref
              • Medline
              • Web of Science ®
              • Google Scholar
              Article Location
            • 71. 
              Greenland S, Neutra RR. 1980. Control of confounding in the assessment of medical technology. Int. J. Epidemiol. 9:361–67
              • Crossref
              • Medline
              • Web of Science ®
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              • Article Location
              • Article Location
              • Article Location
            • 72. 
              Greenland S, Pearl J. 2011. Adjustments and their consequences—collapsibility analysis using graphical models. Int. Stat. Rev. 79:401–26
              • Crossref
              • Web of Science ®
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              • Article Location
              • Article Location
              • Article Location
              More AR articles citing this reference

              • Evaluation of Causal Effects and Local Structure Learning of Causal Networks

                Zhi Geng,1 Yue Liu,1 Chunchen Liu,2 and Wang Miao3 1School of Mathematical Sciences, Peking University, Beijing 100871, China; email: [email protected], [email protected]2Department of Data Mining, NEC Laboratories China, Beijing 100600, China; email: [email protected]3Guanghua School of Management, Peking University, Beijing 100871, China; email: [email protected]
                Annual Review of Statistics and Its Application Vol. 6: 103 - 124
                • ...The covariates leading to the confounding bias are called confounders. Greenland et al. (1999b) and Greenland & Pearl (2011) gave overviews of these concepts....

            • 73. 
              Greenland S, Rothman KJ. 2008. Fundamentals of epidemiologic data analysis. See Ref. 120, pp. 213–37
              • Google Scholar
              Article Location
            • 74. 
              Greenland S, Rothman KJ. 2008. Introduction to stratified analysis. See Ref. 120, pp. 258–82
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
            • 75. 
              Greenland S, Schlesselman JJ, Criqui MH. 1986. The fallacy of employing standardized regression coefficients and correlations as measures of effect. Am. J. Epidemiol. 123:203–8
              • Crossref
              • Medline
              • Web of Science ®
              • Google Scholar
              Article Location
            • 76. 
              Greenland S, Schwartzbaum JA, Finkle WD. 2000. Problems due to small samples and sparse data in conditional logistic regression analysis. Am. J. Epidemiol. 151:531–39
              • Crossref
              • Medline
              • Web of Science ®
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              • Article Location
              More AR articles citing this reference

              • Confounding in Health Research

                Sander Greenland1,2 and Hal Morgenstern11Department of Epidemiology, Los Angeles School of Public Health, University of California, Los Angeles, California 90095-1772; e-mail: [email protected] 2Department of Statistics, Los Angeles College of Letters and Science, University of California, Los Angeles, California 90095
                Annual Review of Public Health Vol. 22: 189 - 212
                • ...few if any strata would have subjects in both treatment groups, thereby making comparisons biased, inefficient, or impossible (38A, 79)....

            • 77. 
              Gustafson P, Greenland S. 2009. Interval estimation for messy observational data. Stat. Sci. 24:328–42
              • Crossref
              • Web of Science ®
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              • Article Location
              • Article Location
            • 78. 
              Gustafson P, McCandless LC. 2010. Probabilistic approaches to better quantifying the results of epidemiologic studies. Int. J. Environ. Res. Public Health 7:1520–39
              • Crossref
              • Medline
              • Web of Science ®
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
            • 79. 
              Harrell F. 2001. Regression Modeling Strategies. New York: Springer
              • Crossref
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              • Article Location
              • Article Location
              • Article Location
              • Article Location
              • Article Location
              • Article Location
              • Article Location
              More AR articles citing this reference

              • Data Visualization in Sociology

                Kieran Healy and James MoodySociology Department, Duke University, Durham, North Carolina 27708; email: [email protected], [email protected]
                Annual Review of Sociology Vol. 40: 105 - 128
                • ...but to open things up for further exploration. Harrell (2001) remains an exemplary book-length demonstration of the virtues of integrating graphical methods with the process of data exploration (including exploring patterns of missingness in the data) right across the process of model building, ...

            • 80. 
              Hastie T, Tibshirani R, Friendman J. 2009. The Elements of Statistical Learning: Data Mining, Inference and Prediction. New York: Springer. 2nd ed.
              • Crossref
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              • Article Location
              • Article Location
              • Article Location
              • Article Location
              • Article Location
              • Article Location
              • Article Location
              More AR articles citing this reference

              • Machine Learning Methods That Economists Should Know About

                Susan Athey1,2,3 and Guido W. Imbens1,2,3,41Graduate School of Business, Stanford University, Stanford, California 94305, USA; email: [email protected], [email protected]2Stanford Institute for Economic Policy Research, Stanford University, Stanford, California 94305, USA3National Bureau of Economic Research, Cambridge, Massachusetts 02138, USA4Department of Economics, Stanford University, Stanford, California 94305, USA
                Annual Review of Economics Vol. 11: 685 - 725
                • ...and many textbooks discuss ML methods alongside more traditional statistical methods (e.g., Hastie et al. 2009, Efron & Hastie 2016)....
                • ...including the work of Efron & Hastie (2016); Hastie et al. (2009), ...
                • ...Hastie et al. (2009, 2015) discuss what they call the sparsity principle: ...
              • Machine Learning for Sociology

                Mario Molina and Filiz GaripDepartment of Sociology, Cornell University, Ithaca, New York 14853, USA; email: [email protected], [email protected]
                Annual Review of Sociology Vol. 45: 27 - 45
                • ...This error comprises two components: bias and variance (Hastie et al. 2009)....
                • ...estimating its generalization (prediction) error on new data (Hastie et al. 2009)....
                • ...and a quarter each for validation and testing (Hastie et al. 2009)....
                • ...Boosting involves giving more weight to misclassified observations over repeated estimation (Hastie et al. 2009)....
                • ...4Hastie et al. (2009, table 10.1) compare different methods on several criteria (e.g., ...
              • Precision Medicine

                Michael R. Kosorok1 and Eric B. Laber2 1Department of Biostatistics and Department of Statistics and Operations Research, University of North Carolina at Chapel Hill, Chapel Hill, North Carolina 27599, USA; email: [email protected]2Department of Statistics, North Carolina State University, Raleigh, North Carolina 27695, USA; email: [email protected]
                Annual Review of Statistics and Its Application Vol. 6: 263 - 286
                • ... uses this framework with support vector machines (see, e.g., chapter 12 of Hastie et al. 2009)...
              • Forecasting Methods in Finance

                Allan TimmermannRady School of Management, University of California, San Diego, La Jolla, California 92093, USA; email: [email protected]
                Annual Review of Financial Economics Vol. 10: 449 - 479
                • ...have been developed in recent years (for an excellent introduction, see Hastie, Tibshirani & Friedman 2009)....
              • Big Data Approaches for Modeling Response and Resistance to Cancer Drugs

                Peng Jiang,1 William R. Sellers,2 and X. Shirley Liu11Dana–Farber Cancer Institute and Harvard T.H. Chan School of Public Health, Boston, Massachusetts 02215, USA; email: [email protected]2Broad Institute of MIT and Harvard, Cambridge, Massachusetts 02142, USA; email: [email protected]
                Annual Review of Biomedical Data Science Vol. 1: 1 - 27
                • ...These penalties help find coefficients of the optimal solution in high-dimensional settings while preventing the regression procedure from overfitting the training data (86)....
              • Machine Learning Approaches for Clinical Psychology and Psychiatry

                Dominic B. Dwyer, Peter Falkai, and Nikolaos KoutsoulerisDepartment of Psychiatry and Psychotherapy, Section for Neurodiagnostic Applications, Ludwig-Maximilian University, Munich 80638, Germany; email: [email protected], [email protected], [email protected]
                Annual Review of Clinical Psychology Vol. 14: 91 - 118
                • ...there are several well-established and highly regarded comprehensive methodological guides to machine learning that include formal statistical nomenclature (Bishop 2006, Hastie et al. 2009, James et al. 2015)....
                • ...and the long computational time (Hastie et al. 2009, Varoquaux et al. 2017)....
                • ...This process is then repeated for a prespecified number of k folds and results in more stable estimates of generalizability outside the sample because the training groups are more variable and there are more individuals in the left-out test sets (Hastie et al. 2009)....
                • ...and while authors recommend 5- or 10-fold CV (Breiman & Spector 1992) or statistical criteria (Hastie et al. 2009, James et al. 2015), ...
                • ...Aspects of the SVM algorithm have developed over time to include the ability to characterize nonlinear hyperplanes by using a data transformation implemented by a kernel function (Hastie et al. 2009, James et al. 2015)....
              • Treatment Selection in Depression

                Zachary D. Cohen and Robert J. DeRubeisDepartment of Psychology, University of Pennsylvania, Philadelphia, Pennsylvania 19104, USA; email: [email protected]
                Annual Review of Clinical Psychology Vol. 14: 209 - 236
                • ...researchers must weigh the increased flexibility and predictive power of such approaches against the interpretability (Hastie et al. 2009, James et al. 2013)...
                • ...data from the to-be-predicted patient cannot be included in the course of development of the algorithm (Hastie et al. 2009)....
              • Forecasting in Economics and Finance

                Graham Elliott1 and Allan Timmermann1,21Department of Economics, University of California, San Diego, La Jolla, California 92093; email: [email protected]2Center for Research in Econometric Analysis of Time Series, Aarhus University, DK-8210 Aarhus, Denmark
                Annual Review of Economics Vol. 8: 81 - 110
                • ...these methods will receive further consideration in future work. Hastie et al. (2009) provide a terrific introduction to statistical learning methods, ...
              • League Tables for Hospital Comparisons

                Sharon-Lise T. Normand,1 Arlene S. Ash,2 Stephen E. Fienberg,3 Thérèse A. Stukel,4 Jessica Utts,5 and Thomas A. Louis61Department of Health Care Policy, Harvard Medical School, and Department of Biostatistics, Harvard T.H. Chan School of Public Health, Boston, Massachusetts 02115; email: [email protected]2Department of Quantitative Health Sciences, University of Massachusetts Medical School, Worcester, Massachusetts 01605; email: [email protected]3Department of Statistics, Carnegie Mellon University, Pittsburgh, Pennsylvania 15213; email: [email protected]4Institute for Clinical Evaluative Sciences, Toronto, Ontario M4N 3M5, Canada, and the Institute of Health Policy, Management & Evaluation, University of Toronto, Toronto, Ontario M5T 3M6, Canada, and Dartmouth Institute for Health Policy and Clinical Practice, Hanover, New Hampshire 03766; email: [email protected]5Department of Statistics, University of California, Irvine, California 92697; email: [email protected]6Department of Biostatistics, Johns Hopkins Bloomberg School of Public Health, Baltimore, Maryland 21205; email: [email protected]
                Annual Review of Statistics and Its Application Vol. 3: 21 - 50
                • ...and boosting (Berk 2008, Breiman 2001, Hastie et al. 2009, McCaffrey et al. 2004), ...
              • Far Right Parties in Europe

                Matt GolderDepartment of Political Science, Pennsylvania State University, University Park, Pennsylvania 16802; email: [email protected]
                Annual Review of Political Science Vol. 19: 477 - 497
                • ...They have yet to exploit recent developments in data mining, particularly with respect to cluster analysis (Hastie et al. 2009)....
              • Computerized Adaptive Diagnosis and Testing of Mental Health Disorders

                Robert D. Gibbons,1 David J. Weiss,2 Ellen Frank,3 and David Kupfer31Center for Health Statistics and Departments of Medicine and Public Health Sciences, University of Chicago, Chicago, Illinois 60612; email: [email protected]2Department of Psychology, University of Minnesota, Minneapolis, Minnesota 554553Department of Psychiatry and Western Psychiatric Institute and Clinic, University of Pittsburgh, Pittsburgh, Pennsylvania 15213
                Annual Review of Clinical Psychology Vol. 12: 83 - 104
                • ...have received considerable attention in statistics and machine learning (Brieman 1996, Hastie et al. 2009)....
                • ...decision trees have frequently suffered from poor performance (Hastie et al. 2009) because algorithms used to build trees from data can exhibit sensitivity to small changes in the data sets that are provided....
                • ...Random forests require minimal human intervention and have historically exhibited good performance across a wide range of domains (Brieman 2001, Hastie et al. 2009)....
              • Modular Brain Networks

                Olaf Sporns1,2 and Richard F. Betzel11Department of Psychological and Brain Sciences, Indiana University, Bloomington, Indiana 47405; email: [email protected]2Indiana University Network Science Institute, Indiana University, Bloomington, Indiana 47405
                Annual Review of Psychology Vol. 67: 613 - 640
                • ...Distance-based modules.One of the simplest methods for detecting modules in complex networks is to extend distance-based clustering techniques to be compatible with network data (Hastie et al. 2009)....
              • Analytics of Insurance Markets

                Edward W. FreesWisconsin School of Business, University of Wisconsin–Madison, Madison, Wisconsin 53706; email: [email protected]
                Annual Review of Financial Economics Vol. 7: 253 - 277
                • ...predictive analytics means advanced data-mining tools such as described in The Elements of Statistical Learning: Data Mining, Inference, and Prediction (Hastie, Tibshirani & Friedman 2009)....
              • Empirical Comparative Law

                Holger SpamannHarvard Law School, Harvard University, Cambridge, Massachusetts 02138; email: [email protected]
                Annual Review of Law and Social Science Vol. 11: 131 - 153
                • ...25This is the “bet on sparsity” principle coined by Hastie et al. (2009, ...
              • Valid Post-Selection and Post-Regularization Inference: An Elementary, General Approach

                Victor Chernozhukov,1 Christian Hansen,2 and Martin Spindler3 1Department of Economics, Massachusetts Institute of Technology, Cambridge, Massachusetts 02142; email: [email protected] 2University of Chicago Booth School of Business, Chicago, Illinois 60637; email: [email protected] 3Munich Center for the Economics of Aging, 80799 Munich, Germany; email: [email protected]
                Annual Review of Economics Vol. 7: 649 - 688
                • ...Many other interesting procedures beyond those mentioned in this review have been developed for estimating high-dimensional models (see, e.g., Hastie et al. 2009 for a textbook review)....
              • High-Dimensional Statistics with a View Toward Applications in Biology

                Peter Bühlmann, Markus Kalisch, and Lukas MeierSeminar for Statistics, ETH Zürich, CH-8092 Zürich, Switzerland; email: [email protected], [email protected], [email protected]
                Annual Review of Statistics and Its Application Vol. 1: 255 - 278
                • ...Assessing the accuracy of prediction is relatively straightforward using the tool of cross-validation (cf. Hastie et al. 2009)....
              • Sensors and Decoding for Intracortical Brain Computer Interfaces

                Mark L. Homer,1 Arto V. Nurmikko,2 John P. Donoghue,4,3 and Leigh R. Hochberg4,2,51Biomedical Engineering,2School of Engineering,3Department of Neuroscience, Brown University, Providence, Rhode Island 02912; email: [email protected]4Center for Neurorestoration and Neurotechnology, Veterans Affairs Medical Center, Providence, Rhode Island 029085Neurology, Massachusetts General Hospital, Harvard Medical School, Boston, Massachusetts 02114
                Annual Review of Biomedical Engineering Vol. 15: 383 - 405
                • ...akin to forward stepwise regression, can find promising, though typically not optimal, subsets (83)....
              • Sparse High-Dimensional Models in Economics

                Jianqing Fan,1,2 Jinchi Lv,3 and Lei Qi1,21Bendheim Center for Finance and 2Department of Operations Research and Financial Engineering, Princeton University, Princeton, New Jersey 08544; email: [email protected], [email protected]3Information and Operations Management Department, Marshall School of Business, University of Southern California, Los Angeles, California 90089; email: [email protected]
                Annual Review of Economics Vol. 3: 291 - 317
                • ...such as text and document classification and computer vision (see Hastie et al. 2009 for more examples)....
              • Species Distribution Models: Ecological Explanation and Prediction Across Space and Time

                Jane Elith1 and John R. Leathwick21School of Botany, The University of Melbourne, Victoria 3010, Australia; email: [email protected]2National Institute of Water and Atmospheric Research, Hamilton, New Zealand; email: [email protected]
                Annual Review of Ecology, Evolution, and Systematics Vol. 40: 677 - 697
                • ...both within the model-fitting process, and for model evaluation (Hastie et al. 2009)....
                • ...Other model averaging techniques from computer science use a range of approaches to concurrently develop a set of models that together predict well (Hastie et al. 2009)....
                • ...The different information criteria provide a range of trade-offs between model complexity and predictive performance and can be used within cross-validation to select a model (Hastie et al. 2009)....

            • 81. 
              Hauck WW, Anderson S. 1986. A proposal for interpreting and reporting negative studies. Stat. Med. 5:203–9
              • Crossref
              • Medline
              • Web of Science ®
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
            • 82. 
              Hernán MA. 2005. Invited commentary: Hypothetical interventions to define causal effects—afterthought or prerequisite? Am. J. Epidemiol. 162:618–20
              • Crossref
              • Medline
              • Web of Science ®
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              More AR articles citing this reference

              • Causal Inference in Public Health

                Thomas A. Glass,1 Steven N. Goodman,2 Miguel A. Hernán,3,4 and Jonathan M. Samet51Department of Epidemiology, Bloomberg School of Public Health, Johns Hopkins University, Baltimore, Maryland 21205; email: [email protected]2Department of Medicine, Stanford University, Palo Alto, California 94305; email: [email protected]3Departments of Epidemiology and Biostatistics, School of Public Health, Harvard University, Boston, Massachusetts 021154Harvard-MIT Division of Health Sciences and Technology, Cambridge, Massachusetts 02115; email: [email protected]5Department of Preventive Medicine, Keck School of Medicine, and USC Institute for Global Health, University of Southern California, Los Angeles, California 90089; email: [email protected]
                Annual Review of Public Health Vol. 34: 61 - 75
                • ...liposuction (26, 32)? Although obesity may meet criteria for a causal factor in the classic framework, ...

            • 83. 
              Hernán MA, Robins JM. 2015. Causal Inference. New York: Chapman & Hall/CRC. In press
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              • Article Location
              • Article Location
              • Article Location
              • Article Location
            • 84. 
              Hurvich CM, Tsai CL. 1990. The impact of model selection on inference in linear regression. Am. Stat. 44:214–17
              • Web of Science ®
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              • Article Location
            • 85. 
              Joffe MM. 2009. Exhaustion, automation, theory, and confounding. Epidemiology 20:523–24
              • Crossref
              • Medline
              • Web of Science ®
              • Google Scholar
              Article Location
            • 86. 
              Kabaila P. 2009. The coverage properties of confidence regions after model selection. Int. Stat. Rev. 77:405–14
              • Crossref
              • Web of Science ®
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
            • 87. 
              Kelly KT. 2011. Simplicity, truth, and probability. In Handbook of the Philosophy of Science: Philosophy of Statistics, ed. PS Bandyopadhyay, MR Forster, pp. 983–1026. North Holland: Elsevier
              • Crossref
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
            • 88. 
              Kendall MG. 1963. Comment on JO Irwin. The place of mathematics in medical and biological statistics. J. R. Stat. Soc. Ser. A 126:1–45
              • Crossref
              • Google Scholar
              Article Location
            • 89. 
              Kleinbaum D, Kupper LL, Morgenstern H. 1982. Epidemiologic Research: Principles and Quantitative Methods. Belmont, CA: Lifetime Learn.
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              • Article Location
              • Article Location
              • Article Location
            • 90. 
              Krieger N. 2011. Epidemiology and the People's Health: Theory and Context. New York: Oxford Univ. Press
              • Crossref
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
            • 91. 
              Lash TL. 2007. Heuristic thinking and inference from observational epidemiology. Epidemiology 18:67–72
              • Crossref
              • Medline
              • Web of Science ®
              • Google Scholar
            • 92. 
              Lash TL, Fox MP, Fink AK. 2009. Applying Quantitative Bias Analysis to Epidemiologic Data. Boston: Springer
              • Crossref
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              More AR articles citing this reference

              • Approaches to Uncertainty in Exposure Assessment in Environmental Epidemiology

                Donna SpiegelmanDepartments of Epidemiology and Biostatistics, Harvard School of Public Health, Boston, Massachusetts 02115; email: [email protected]
                Annual Review of Public Health Vol. 31: 149 - 163
                • ...Uncertainty analysis (5, 29), bias analysis (25), sensitivity analysis (13) (see sidebar), measurement error correction, ...
                • ...A more sophisticated approach, Bayesian (13) or probabilistic (25) sensitivity analysis, generalizes the point-wise sensitivity analysis previously described by using a weighted distribution of possible measurement error or misclassification parameter values in place of a single point....

            • 93. 
              Lash TL, Fox MP, MacLehose RF, Maldonado G, McCandless LC, Greenland S. 2014. Good practices for quantitative bias analysis. Int. J. Epidemiol. 43:1969–85
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              • Article Location
              • Article Location
            • 94. 
              Leamer E. 1978. Specification Searches. New York: Wiley
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              • Article Location
              More AR articles citing this reference

              • Credible Causal Inference for Empirical Legal Studies

                Daniel E. Ho1 and Donald B. Rubin21Stanford Law School, Stanford University, Stanford, California 94305; email: [email protected]2Department of Statistics, Harvard University, Cambridge, Massachusetts 02138; email: [email protected]
                Annual Review of Law and Social Science Vol. 7: 17 - 40
                • ...; see also Berk 2004; Donohue & Wolfers 2006; Gelman & Meng 2004; Leamer 1978, 1983...
                • ...; Dehejia & Wahba 1999; Ho et al. 2007; King & Zeng 2007; Lalonde 1986; Leamer 1978, 1983...

            • 95. 
              Lee BK, Lessler J, Stuart EA. 2010. Improving propensity score weighting using machine learning. Stat. Med. 29:337–46
              • Medline
              • Web of Science ®
              • Google Scholar
              Article Location
              More AR articles citing this reference

              • Machine Learning for Sociology

                Mario Molina and Filiz GaripDepartment of Sociology, Cornell University, Ithaca, New York 14853, USA; email: [email protected], [email protected]
                Annual Review of Sociology Vol. 45: 27 - 45
                • ...and regression trees for this task (Diamond & Sekhon 2013, Hill 2011, Lee et al. 2010, Wyss et al. 2014) as alternatives to traditional logistic regression....
              • Nonprobability Sampling and Causal Analysis

                Ulrich Kohler,1 Frauke Kreuter,2,3,4 and Elizabeth A. Stuart51Faculty of Economics and Social Sciences, University of Potsdam, 14482 Potsdam, Germany; email: [email protected]2Joint Program in Survey Methodology, University of Maryland, College Park, Maryland 20742, USA; email: [email protected]3School of Social Sciences, University of Mannheim, 68131 Mannheim, Germany4Statistical Methods Research Department, Institute for Employment Research (IAB), 90478 Nuremberg, Germany5Department of Mental Health, Department of Biostatistics, and Department of Health Policy and Management, Bloomberg School of Public Health, Johns Hopkins University, Baltimore, Maryland 21205, USA; email: [email protected]
                Annual Review of Statistics and Its Application Vol. 6: 149 - 172
                • ...in particular under conditions of both moderate nonadditivity and moderate nonlinearity (Lee et al. 2010)....
              • Evaluation of Causal Effects and Local Structure Learning of Causal Networks

                Zhi Geng,1 Yue Liu,1 Chunchen Liu,2 and Wang Miao3 1School of Mathematical Sciences, Peking University, Beijing 100871, China; email: [email protected], [email protected]2Department of Data Mining, NEC Laboratories China, Beijing 100600, China; email: [email protected]3Guanghua School of Management, Peking University, Beijing 100871, China; email: [email protected]
                Annual Review of Statistics and Its Application Vol. 6: 103 - 124
                • ...semiparametric models and flexible machine learning methods can be applied (Lee et al. 2010)....

            • 96. 
              Maldonado G, Greenland S. 1993. Simulation study of confounder-selection strategies. Am. J. Epidemiol. 138:923–36
              • Crossref
              • Medline
              • Web of Science ®
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              • Article Location
              • Article Location
            • 97. 
              Mallows CL, Nair VN. 1987. A unique unbiased estimator with an interesting property. Am. Stat. 41:205–6
              • Web of Science ®
              • Google Scholar
              Article Location
            • 98. 
              Mansson R, Joffe MM, Sun WG, Hennessy S. 2007. On the estimation and use of propensity scores in case-control and case-cohort studies. Am. J. Epidemiol. 166:332–39
              • Crossref
              • Medline
              • Web of Science ®
              • Google Scholar
              Article Location
            • 99. 
              McCaffrey DF, Ridgeway G, Morral AR. 2004. Propensity score estimation with boosted regression for evaluating causal effects in observational studies. Psychol. Methods 9:403–25
              • Crossref
              • Medline
              • Web of Science ®
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              • Article Location
              • Article Location
              More AR articles citing this reference

              • Machine Learning for Sociology

                Mario Molina and Filiz GaripDepartment of Sociology, Cornell University, Ithaca, New York 14853, USA; email: [email protected], [email protected]
                Annual Review of Sociology Vol. 45: 27 - 45
                • ...Recent work uses boosting (McCaffrey et al. 2004), neural networks (Setoguchi et al. 2008, Westreich et al. 2010)...
              • Experiments in Criminology: Improving Our Understanding of Crime and the Criminal Justice System

                Greg RidgewayDepartment of Criminology and Department of Statistics, University of Pennsylvania, Philadelphia, Pennsylvania 19104, USA; email: [email protected]
                Annual Review of Statistics and Its Application Vol. 6: 37 - 61
                • ... use boosted logistic regression to estimate the propensity scores (McCaffrey et al. 2004), ...
              • League Tables for Hospital Comparisons

                Sharon-Lise T. Normand,1 Arlene S. Ash,2 Stephen E. Fienberg,3 Thérèse A. Stukel,4 Jessica Utts,5 and Thomas A. Louis61Department of Health Care Policy, Harvard Medical School, and Department of Biostatistics, Harvard T.H. Chan School of Public Health, Boston, Massachusetts 02115; email: [email protected]2Department of Quantitative Health Sciences, University of Massachusetts Medical School, Worcester, Massachusetts 01605; email: [email protected]3Department of Statistics, Carnegie Mellon University, Pittsburgh, Pennsylvania 15213; email: [email protected]4Institute for Clinical Evaluative Sciences, Toronto, Ontario M4N 3M5, Canada, and the Institute of Health Policy, Management & Evaluation, University of Toronto, Toronto, Ontario M5T 3M6, Canada, and Dartmouth Institute for Health Policy and Clinical Practice, Hanover, New Hampshire 03766; email: [email protected]5Department of Statistics, University of California, Irvine, California 92697; email: [email protected]6Department of Biostatistics, Johns Hopkins Bloomberg School of Public Health, Baltimore, Maryland 21205; email: [email protected]
                Annual Review of Statistics and Its Application Vol. 3: 21 - 50
                • ...and boosting (Berk 2008, Breiman 2001, Hastie et al. 2009, McCaffrey et al. 2004), ...

            • 100. 
              McCullagh P, Nelder JA. 1989. Generalized Linear Models. New York: Chapman & Hall
              • Crossref
              • Google Scholar
              Article Locations:
              • Article Location
              • Article Location
              More AR articles citing this reference

              • Behavioral Assays for Studies of Host Plant Choice and Adaptation in Herbivorous Insects

                Lisa M. Knolhoff1 and David G. Heckel2,1Genective, c/o AgReliant Genetics, Champaign, Illinois 61801; email: [email protected]2Max Planck Institute for Chemical Ecology, Jena 07745, Germany; email: [email protected]
                Annual Review of Entomology Vol. 59: 263 - 278
                • ...The generalized linear model (55) is well-suited to count data and obviates the need for less satisfactory approaches....
              • High-Dimensional Statistics with a View Toward Applications in Biology

                Peter Bühlmann, Markus Kalisch, and Lukas MeierSeminar for Statistics, ETH Zürich, CH-8092 Zürich, Switzerland; email: [email protected], [email protected], [email protected]
                Annual Review of Statistics and Its Application Vol. 1: 255 - 278
                • ...Generalized linear models (McCullagh & Nelder 1989) are very popular for extending the linear model in a unified way....
                • ...if the (conditional) distribution of Y|X = x is from a subclass of the exponential family model (see McCullagh & Nelder 1989, ...
                • ...replace the t-test in step 3 of Algorithm 1 with the log-likelihood ratio test (McCullagh & Nelder 1989), ...
              • Event History Analysis

                Niels KeidingDepartment of Biostatistics, University of Copenhagen, Copenhagen DK-1014, Denmark; email: [email protected]
                Annual Review of Statistics and Its Application Vol. 1: 333 - 360
                • ...the statistical model is essentially a generalized linear model in the sense of McCullagh & Nelder (1989), ...
                • ...One is based on a postulation of an overdispersion parameter to capture at least some of the unaccounted excess variation (e.g., Breslow 1984, 1990; McCullagh & Nelder 1989, ...
              • Probabilistic Forecasting

                Tilmann Gneiting1 and Matthias Katzfuss21Institut für Angewandte Mathematik, Universität Heidelberg, 69120 Heidelberg, Germany; email: [email protected]2Department of Statistics, Texas A&M University, College Station, TX 77843; email: [email protected]
                Annual Review of Statistics and Its Application Vol. 1: 125 - 151
                • ...generalized linear models (McCullagh & Nelder 1989) relate the mean of a certain specified distribution to a linear function in the explanatory variables via a suitably chosen link function....
              • Build, Compute, Critique, Repeat: Data Analysis with Latent Variable Models

                David M. BleiComputer Science Department, Princeton University, Princeton, New Jersey 08540; email: [email protected]
                Annual Review of Statistics and Its Application Vol. 1: 203 - 232
                • ...such as generalized linear models (Nelder & Wedderburn 1972, McCullagh & Nelder 1989), ...
              • Using Geographic Information Systems and Decision Support Systems for the Prediction, Prevention, and Control of Vector-Borne Diseases

                Lars Eisen1 and Rebecca J. Eisen21Department of Microbiology, Immunology, and Pathology, Colorado State University, Fort Collins, Colorado 80523; email: [email protected]2Division of Vector-Borne Infectious Diseases, Coordinating Center for Infectious Diseases, Centers for Disease Control and Prevention, Fort Collins, Colorado 80522; email: [email protected]
                Annual Review of Entomology Vol. 56: 41 - 61
                • ...and the statistical methods have been described in detail elsewhere (34, 35, 44, 52, 58, 73, 82, 84, 89, 95, 115)....
              • Nonparametric Methods for Modeling Nonlinearity in Regression Analysis

                Robert AndersenDepartment of Sociology, University of Toronto, Toronto, Ontario M5S 2J4, Canada; email: [email protected]
                Annual Review of Sociology Vol. 35: 67 - 85
                • ...I start with a brief description of the GLM (for a more extensive treatment, see McCullagh & Nelder 1989)....
                • ...Maximum likelihood estimates for GLMs can be obtained using the Newton-Raphson method or iteratively reweighted least squares (IRLS) (Nelder & Wedderburn 1972, McCullagh & Nelder 1989)....
              • Forecasting Methods in Crime and Justice

                Richard BerkDepartment of Statistics, Department of Criminology, University of Pennsylvania, Philadelphia, Pennsylvania 19104; email: [email protected]
                Annual Review of Law and Social Science Vol. 4: 219 - 238
                • ...The classic work on the generalized linear model is by McCullagh & Nelder (1989)....
              • Methods for Improving Regression Analysis for Skewed Continuous or Counted Responses

                Abdelmonem A. Afifi,1 Jenny B. Kotlerman,2 Susan L. Ettner,1,3 and Marie Cowan21School of Public Health, 2School of Nursing, 3School of Medicine, University of California, Los Angeles, California 90095-1772; email: [email protected], [email protected], [email protected], [email protected]
                Annual Review of Public Health Vol. 28: 95 - 111
                • ...Such programs rely on the concept of generalized linear models (GLM) [Dobson (16), McCullagh & Nelder (43)]....
              • On Time Series Analysis of Public Health and Biomedical Data

                Scott L. Zeger, Rafael Irizarry, and Roger D. PengDepartment of Biostatistics, The Johns Hopkins University Bloomberg School of Public Health, Baltimore, Maryland 21205; email: [email protected], [email protected], [email protected]
                Annual Review of Public Health Vol. 27: 57 - 79
                • ...generalized linear models (GLMs; 31) have unified regression analysis for binary, ...
                • ...We have already mentioned the generalized linear model (31) extensions of autoregressive models for binary, ...
              • Time-Series Studies of Particulate Matter

                Michelle L. Bell,1 Jonathan M. Samet,1 and Francesca Dominici2 1 Department of Epidemiology, Johns Hopkins Bloomberg School of Public Health,
                Baltimore, Maryland 21205
                ; email: [email protected], [email protected] 2 and 2Department of Biostatistics, Johns Hopkins Bloomberg School of Public Health,
                Baltimore, Maryland 21205
                ; email: [email protected]
                Annual Review of Public Health Vol. 25: 247 - 280
                • ...Currently used statistical approaches for time-series analysis are generalized linear models (GLM) with parametric splines (101)...
              • Assessing Change with Longitudinal and Clustered Binary Data

                John M NeuhausDepartment of Epidemiology and Biostatistics, University of California, San Francisco, California 94143-0560; e-mail: [email protected]
                Annual Review of Public Health Vol. 22: 115 - 128
                • ...Many of these approaches are extensions of generalized linear models (17) to accommodate intracluster correlation....
              • Selected Statistical Issues in Group Randomized Trials

                Ziding Feng,1 Paula Diehr,1,2 Arthur Peterson,1 and Dale McLerran11Cancer Prevention Research Program, Fred Hutchinson Cancer Research Center, 1100 Fairview Ave.N. MP-702, Seattle, Washington 98109-1024; e-mail: [email protected] ,[email protected] ,[email protected] [email protected] ,2Department of Biostatistics, University of Washington, Seattle, Washington 98195 [email protected] ,
                Annual Review of Public Health Vol. 22: 167 - 187
                • Comparing Personal Trajectories and Drawing Causal Inferences from Longitudinal Data

                  Stephen W. RaudenbushSchool of Education and Institute for Social Research, University of Michigan, 610 East University Avenue, Ann Arbor, Michigan 48109; e-mail: [email protected]
                  Annual Review of Psychology Vol. 52: 501 - 525
                  • ...The generalized linear model of McCullagh & Nelder (1989) provides a far more general class of level-1 models....
                  • ...In the language of the generalized linear model (McCullagh & Nelder 1989), ...
                • CATEGORICAL DATA ANALYSIS

                  Thomas D. WickensDepartment of Psychology, University of California, Los Angeles, California 90095; e-mail: [email protected]
                  Annual Review of Psychology Vol. 49: 537 - 558
                  • ...This approach is most strongly represented by the generalized linear models (McCullagh & Nelder, 1989)....
                  • ...which can be estimated in a variety of ways (Collett, 1991;, McCullagh & Nelder, 1989; for examples, ...
                • CATEGORICAL DATA ANALYSIS IN PUBLIC HEALTH

                  John S. PreisserSection on Biostatistics, Department of Public Health Sciences, Bowman Gray School of Medicine, Winston-Salem, North Carolina 27157 Gary G. KochDepartment of Biostatistics, School of Public Health, University of North Carolina, Chapel Hill, North Carolina 27599-7400
                  Annual Review of Public Health Vol. 18: 51 - 82
                  • ...marginal modeling has been unified for different types of responses into the theory of generalized linear models (22, 62, 68)....
                  • ...A book by McCullagh & Nelder (62) is considered the primary reference....
                  • ...Without further distributional assumptions it would be called a quasi-score and its integral a quasi-likelihood (62, 75, 88)....

              • 101. 
                Mickey RM, Greenland S. 1989. The impact of confounder selection criteria on effect estimation. Am. J. Epidemiol. 129:125–37
                • Crossref
                • Medline
                • Web of Science ®
                • Google Scholar
                Article Locations:
                • Article Location
                • Article Location
                • Article Location
                More AR articles citing this reference

                • Treatment Selection in Depression

                  Zachary D. Cohen and Robert J. DeRubeisDepartment of Psychology, University of Pennsylvania, Philadelphia, Pennsylvania 19104, USA; email: [email protected]
                  Annual Review of Clinical Psychology Vol. 14: 209 - 236
                  • ...p = 0.04) is, of course, trivial (Mickey & Greenland 1989)....

              • 102. 
                Miettinen OS. 1976. Stratification by a multivariate confounder score. Am. J. Epidemiol. 104:609–20
                • Crossref
                • Medline
                • Web of Science ®
                • Google Scholar
                Article Location
              • 103. 
                Myers JA, Rassen JA, Gagne JJ, Huybrechts KF, Schneeweiss S, et al. 2011. Effects of adjusting for instrumental variables on bias and precision of effect estimates. Am. J. Epidemiol. 174:1223–27
                • Crossref
                • Medline
                • Web of Science ®
                • Google Scholar
                Article Locations:
                • Article Location
                • Article Location
                More AR articles citing this reference

                • Bias Analysis for Uncontrolled Confounding in the Health Sciences

                  Onyebuchi A. ArahDepartment of Epidemiology, Fielding School of Public Health; UCLA Center for Health Policy Research; and California Center for Population Research, University of California, Los Angeles, California 90095; email: [email protected]
                  Annual Review of Public Health Vol. 38: 23 - 38
                  • ...There is a third type of bias that could also result from not controlling for U: This is bias amplification of the uncontrolled confounding in the X→Y effect if the investigator adds to the model an instrumental variable—a preexposure variable that is only a cause of or is associated with the exposure X but not with Y except through X (40, 43)....

              • 104. 
                Pang M, Kaufman JS, Platt RW. 2014. Studying noncollapsibility of the odds ratio with marginal structural and logistic regression models. Stat. Methods Med. Res. In press
                • Google Scholar
                Article Locations:
                • Article Location
                • Article Location
              • 105. 
                Patorno E, Glynn RJ, Hernandez-Diaz S, Liu J, Schneeweiss S. 2014. Studies with many covariates and few outcomes selecting covariates and implementing propensity-score-based confounding adjustments. Epidemiology 25:268–78
                • Crossref
                • Medline
                • Web of Science ®
                • Google Scholar
                Article Location
              • 106. 
                Pearce N. 2011. Epidemiology in a changing world: variation, causation and ubiquitous risk factors. Int. J. Epidemiol. 40:503–12
                • Crossref
                • Medline
                • Web of Science ®
                • Google Scholar
                Article Location
              • 107. 
                Pearce SC. 1983. The monstrous regiment of mathematicians. Statistician 32:375–78
                • Crossref
                • Google Scholar
                Article Location
              • 108. 
                Pearl J. 2009. Causality: Models, Reasoning, and Inference. New York: Cambridge Univ. Press. 2nd ed.
                • Crossref
                • Google Scholar
                Article Locations:
                • Article Location
                • Article Location
                • Article Location
                • Article Location
                • Article Location
                More AR articles citing this reference

                • Analyzing Age-Period-Cohort Data: A Review and Critique

                  Ethan Fosse1 and Christopher Winship21Department of Sociology, University of Toronto, Toronto, Ontario M5S 2J4, Canada; email: [email protected]2Department of Sociology, Harvard University, Cambridge, Massachusetts 02138, USA
                  Annual Review of Sociology Vol. 45: 467 - 492
                  • Commentary: Causal Inference for Social Exposures

                    Jay S. KaufmanDepartment of Epidemiology, Biostatistics, and Occupational Health, McGill University, Montreal, Quebec H3A 1A2, Canada; email: [email protected]
                    Annual Review of Public Health Vol. 40: 7 - 21
                    • ...where the subscript in Ya = 1 is analogous to Pearl's Y|do(a = 1) or Y|set(a = 1) (71)....
                  • Evaluation of Causal Effects and Local Structure Learning of Causal Networks

                    Zhi Geng,1 Yue Liu,1 Chunchen Liu,2 and Wang Miao3 1School of Mathematical Sciences, Peking University, Beijing 100871, China; email: [email protected], [email protected]2Department of Data Mining, NEC Laboratories China, Beijing 100600, China; email: [email protected]3Guanghua School of Management, Peking University, Beijing 100871, China; email: [email protected]
                    Annual Review of Statistics and Its Application Vol. 6: 103 - 124
                    • ...Causal networks are used to depict the causal relationships among multiple variables (Pearl 2009)....
                    • ...Pearl (1988, 2009) presented causal networks depicted by directed acyclic graphs (DAGs)....
                    • ...the joint distribution of variables can be factorized into We make the faithfulness assumption, which is used for structure learning of DAGs (Pearl 2009)....
                  • Nonprobability Sampling and Causal Analysis

                    Ulrich Kohler,1 Frauke Kreuter,2,3,4 and Elizabeth A. Stuart51Faculty of Economics and Social Sciences, University of Potsdam, 14482 Potsdam, Germany; email: [email protected]2Joint Program in Survey Methodology, University of Maryland, College Park, Maryland 20742, USA; email: [email protected]3School of Social Sciences, University of Mannheim, 68131 Mannheim, Germany4Statistical Methods Research Department, Institute for Employment Research (IAB), 90478 Nuremberg, Germany5Department of Mental Health, Department of Biostatistics, and Department of Health Policy and Management, Bloomberg School of Public Health, Johns Hopkins University, Baltimore, Maryland 21205, USA; email: [email protected]
                    Annual Review of Statistics and Its Application Vol. 6: 149 - 172
                    • ...the inferential question is to infer from the data at hand to the unknown causal effect (see Mill 1843, Neyman et al. 1935, Rubin 1974, and Pearl 2009...
                    • ...Such adjustment sets have been developed in the growing body of literature on graphical causal models (Pearl 2009)...
                    • ...Graphical rules then make it relatively easy to identify adjustment sets that satisfy, for example, the backdoor criterion (Pearl 1993, 2009), ...
                  • Econometric Methods for Program Evaluation

                    Alberto Abadie1 and Matias D. Cattaneo21Department of Economics, Massachusetts Institute of Technology, Cambridge, Massachusetts 02142, USA; email: [email protected]2Department of Economics and Department of Statistics, University of Michigan, Ann Arbor, Michigan 48109, USA; email: [email protected]
                    Annual Review of Economics Vol. 10: 465 - 503
                    • ...Blundell & Costa Dias (2009), DiNardo & Lee (2011), Heckman & Vytlacil (2007), Hernán & Robins (2018), Imbens & Rubin (2015), Imbens & Wooldridge (2009), Lee (2016), Manski (2008), Pearl (2009), Rosenbaum (2002, 2010), ...
                    • ...Directed acyclic graphs (DAGs; see Pearl 2009) provide graphical representations of causal relationships and confounding....
                    • ...as it depends on potential outcomes that are not always observed. Pearl (2009) and others have investigated graphical causal structures that allow identification by conditioning on covariates....
                    • ...6Pearl (2009) and Morgan & Winship (2015) provide detailed introductions to identification in DAGs, ...
                  • Causal Structure Learning

                    Christina Heinze-Deml, Marloes H. Maathuis, and Nicolai MeinshausenSeminar for Statistics, Department of Mathematics, ETH Zurich, CH-8092 Zurich, Switzerland; email: [email protected], [email protected], [email protected]
                    Annual Review of Statistics and Its Application Vol. 5: 371 - 391
                    • ...and hence the estimation of causal effects (e.g., Wright 1934, Spirtes et al. 2000, Pearl 2009)....
                    • ...We formulate the model as a structural causal model (Pearl 2009)....
                    • ...the following holds: If A and B are separated by S in G according to a graphical criterion called d-separation (Pearl 2009), ...
                    • ...the manipulated density (Spirtes et al. 2000), or the truncated factorization formula (Pearl 2009)....
                    • ...the set of DAGs that encode the same set of d-separation relationships (Pearl 2009)....
                  • Web-Based Enrollment and Other Types of Self-Selection in Surveys and Studies: Consequences for Generalizability

                    Niels Keiding1 and Thomas A. Louis21Department of Biostatistics, University of Copenhagen, Copenhagen DK-1014, Denmark; email: [email protected]2Department of Biostatistics, Johns Hopkins Bloomberg School of Public Health, Baltimore, Maryland 21205, USA; email: [email protected]
                    Annual Review of Statistics and Its Application Vol. 5: 25 - 47
                    • ...The authors assumed the potential outcomes framework for causality analysis, as described in detail in the monograph by Pearl (2009)....
                  • Bias Analysis for Uncontrolled Confounding in the Health Sciences

                    Onyebuchi A. ArahDepartment of Epidemiology, Fielding School of Public Health; UCLA Center for Health Policy Research; and California Center for Population Research, University of California, Los Angeles, California 90095; email: [email protected]
                    Annual Review of Public Health Vol. 38: 23 - 38
                    • ...are not measured and controlled for during study design or analysis (42, 49)....
                    • ...These criteria subsume a scenario in which the unmeasured variable is a common cause of the exposure and outcome (35, 42, 65)....
                    • ...See Figures 1–3 for directed acyclic graphs (DAGs) that depict data-generating processes or causal structures whereby U and C confound the effect of X on Y. [There are now accessible introductions and detailed resources on DAGs (22, 42, 49).] Figures 1–3 report information regarding U in relation to the observed X, ...
                    • ...conditioning on U would have controlled for confounding by U (via the path X←U→Y) but would have introduced a new bias (called collider-stratification bias) by opening up the colliding arrowheads at U (X←→[U] ←→Y) (22, 42)....
                    • ...Concerns about uncontrolled confounding should accompany any covariate selection issue for confounding control in empirical quantitative analysis (2, 4, 15, 17, 23, 42, 58)....
                  • Structure Learning in Graphical Modeling

                    Mathias Drton1 and Marloes H. Maathuis21Department of Statistics, University of Washington, Seattle 98195; email: [email protected]2Seminar für Statistik, ETH Zürich, 8092 Zürich, Switzerland; email: [email protected]
                    Annual Review of Statistics and Its Application Vol. 4: 365 - 393
                    • ...This is the starting point for a connection to causal modeling (Spirtes et al. 2000, Pearl 2009)....
                    • ...we review some essential concepts for undirected and directed graphical models (e.g., Lauritzen 1996, Studený 2005, Pearl 2009)....
                    • ...interventions to the system can be modeled by changing the structural equations for precisely those variables that are affected by the intervention (Pearl 2009). ...
                    • ...Mixed graphs provide a useful approach to address these problems without explicit modeling of latent variables (e.g., Spirtes et al. 2000, Pearl 2009, Wermuth 2011)....
                  • Information Recovery and Causality: A Tribute to George Judge

                    Gordon Rausser1 and David A. Bessler21Department of Agricultural and Resource Economics, University of California, Berkeley, California 94720; email: [email protected]2Department of Agricultural Economics, Texas A&M University, College Station, Texas 77843; email: [email protected]
                    Annual Review of Resource Economics Vol. 8: 7 - 23
                    • ...including the Computer Science Laboratory at the University of California, Los Angeles (UCLA) under the direction of Judea Pearl (2009)...
                    • ... and IC (Pearl 2009) algorithms are machine learning programs that sort through data sets, ...
                    • ...18Bayes' nets (or Bayesian networks) are directed acyclic graphs that maintain the probability updating rule of Thomas Bayes (see Pearl 2009, ...
                  • Statistical Causality from a Decision-Theoretic Perspective

                    A. Philip DawidCentre for Mathematical Sciences, University of Cambridge, Cambridge CB3 0WB, United Kingdom; email: [email protected]
                    Annual Review of Statistics and Its Application Vol. 2: 273 - 303
                    • ...structural equation modeling and path analysis (Wright 1921), potential response models (Rubin 1978), functional models (Pearl 2009), ...
                    • ...The importance of making a clear distinction between intervening and merely observing has been stressed by numerous authors, including Meek & Glymour (1994), Pearl (2009), Rubin (1978), ...
                    • ...Most of the effort to date in statistical causality has focused on CoE problems, an important exception being Pearl (2009), ...
                    • ...I introduce influence diagrams as useful graphical representations of causal problems and relate these diagrams to the use of directed acyclic graph (DAG) representations as described by Pearl (2009)....
                    • ...In the approach of Pearl (2009), a DAG such as that in Figure 6 is taken to represent causal properties....
                    • ...We here make use of the notation of Pearl (2009), in which, ...
                    • ...; the reader is also referred to the book chapter by Lauritzen (2000) and the book by Pearl (2009)....
                  • Incorporating Both Randomized and Observational Data into a Single Analysis

                    Eloise E. KaizarDepartment of Statistics, The Ohio State University, Columbus, Ohio 43210; email: [email protected]
                    Annual Review of Statistics and Its Application Vol. 2: 49 - 72
                    • ...Pearl (2009) also suggests using CPM ideas in a hypothetical estimation of a lower bound for the probability of causation for a treatment....
                    • ...As Pearl (2009, p. 303) notes, the combination of estimates he suggests is valid only when the participants in both studies “were sampled properly from the population at large.” That is, ...
                  • Advances in Mediation Analysis: A Survey and Synthesis of New Developments

                    Kristopher J. PreacherDepartment of Psychology and Human Development, Vanderbilt University, Nashville, Tennessee 37203-5721; email: [email protected]
                    Annual Review of Psychology Vol. 66: 825 - 852
                    • ...A distinct yet compatible perspective is offered by Judea Pearl (see Pearl 2009, 2010...
                    • ...Also, although space does not permit elaboration here, Pearl (2001, 2009, 2010, 2012, 2014; Shpitser 2013)...
                  • Advances in Measuring the Environmental and Social Impacts of Environmental Programs

                    Paul J. Ferraro1 and Merlin M. Hanauer21Department of Economics, Andrew Young School of Policy Studies, Georgia State University, Atlanta, Georgia 30302; email: [email protected]2Department of Economics, Sonoma State University, Rohnert Park, California 94928; email: [email protected]
                    Annual Review of Environment and Resources Vol. 39: 495 - 517
                    • ...scholars have made substantial advances in the empirical study of causal relationships—the effect of one variable on another (1...
                    • ...one can think about two potential pollution outcomes: the outcome that would be observed in the presence of the program, Yi(1), ...
                    • ...one eliminates them as rival explanations [called the “back-door criterion” by Pearl (1)]....
                    • ...and then estimates the effect of M on Y without bias [called the “front-door criterion” by Pearl (1)]....
                  • Endogenous Selection Bias: The Problem of Conditioning on a Collider Variable

                    Felix Elwert1 and Christopher Winship21Department of Sociology, University of Wisconsin, Madison, Wisconsin 53706; email: [email protected]2Department of Sociology, Harvard University, Cambridge, Massachusetts 02138; email: [email protected]
                    Annual Review of Sociology Vol. 40: 31 - 53
                    • ...our presentation draws on graphical approaches from computer science (Pearl 1995, 2009, 2010)...
                    • ...Specifically, our presentation relies on direct acyclic graphs (DAGs) (Pearl 1995, 2009; Elwert 2013)...
                    • ...DAGs encode the analyst's qualitative causal assumptions about the data-generating process in the population (thus abstracting from sampling variability; see Pearl 2009 for technical details)....
                    • ...The power of DAGs lies in their ability to reveal all marginal and conditional associations and independences implied by a qualitative causal model (Pearl 2009)....
                    • ...all observable associations originate from just three elementary configurations: chains (A→B, A→C→B, etc.), forks (A←C→B), and inverted forks (A→C←B) (Pearl 1988, 2009...
                    • ...Adding arrows to a given set of variables never helps nonparametric identification (Pearl 2009)....
                    • ...The proper course of action differs sharply depending on how the data were generated (Greenland & Robins 1986, Weinberg 1993, Cole & Hernán 2002; Hernán et al. 2002, Pearl 2009)....
                    • ...Drawing on recent work in theoretical computer science (Pearl 1995, 2009) and epidemiology (Hernán et al. 2004)...
                    • ...Methodological warnings are most useful if they are phrased in an accessible language (Pearl 2009, ...
                    • ...2The equivalence of DAGs and nonparametric structural equation models is discussed in Pearl (2009, 2010)....
                    • ...See Pearl (2009) and Spirtes et al. (2000) for details....
                  • Event History Analysis

                    Niels KeidingDepartment of Biostatistics, University of Copenhagen, Copenhagen DK-1014, Denmark; email: [email protected]
                    Annual Review of Statistics and Its Application Vol. 1: 333 - 360
                    • ...The relation of local independence to the recent development in statistical causality based on hypothetical interventions (cf. Pearl 2009)...
                  • How to Conduct a Mixed Methods Study: Recent Trends in a Rapidly Growing Literature

                    Mario Luis SmallDepartment of Sociology, University of Chicago, Chicago, Illinois 60637; email: [email protected]
                    Annual Review of Sociology Vol. 37: 57 - 86
                    • ...Examples are “back-door path,” “axial coding,” and “structural equivalence” (see Pearl 2009, ...
                  • Using Marketing Muscle to Sell Fat: The Rise of Obesity in the Modern Economy

                    Frederick J. ZimmermanDepartment of Health Services, School of Public Health, University of California, Los Angeles, California 90095-1772; email: [email protected]
                    Annual Review of Public Health Vol. 32: 285 - 306
                    • ...One scholar argues for a “sharp distinction between statistical and causal concepts” (78)....

                • 109. 
                  Pearl J. 2010. On a class of bias-amplifying variables that endanger effect estimates. In Proc. 26th Conf. Uncertain. Artificial Intell., ed. P Grunwald, P Spirtes, pp. 417–24. Corvallis, OR: AUAI
                  • Google Scholar
                  Article Locations:
                  • Article Location
                  • Article Location
                  • Article Location
                  • Article Location
                  • Article Location
                • 110. 
                  Pearl J. 2012. Understanding bias amplification. Am. J. Epidemiol. 174:1223–27
                  • Crossref
                  • Web of Science ®
                  • Google Scholar
                  Article Locations:
                  • Article Location
                  • Article Location
                  More AR articles citing this reference

                  • Bias Analysis for Uncontrolled Confounding in the Health Sciences

                    Onyebuchi A. ArahDepartment of Epidemiology, Fielding School of Public Health; UCLA Center for Health Policy Research; and California Center for Population Research, University of California, Los Angeles, California 90095; email: [email protected]
                    Annual Review of Public Health Vol. 38: 23 - 38
                    • ...There is a third type of bias that could also result from not controlling for U: This is bias amplification of the uncontrolled confounding in the X→Y effect if the investigator adds to the model an instrumental variable—a preexposure variable that is only a cause of or is associated with the exposure X but not with Y except through X (40, 43)....

                • 111. 
                  Perlman MD, Wu L. 1999. The emperor's new tests (with discussion). Stat. Sci. 14:355–81
                  • Crossref
                  • Web of Science ®
                  • Google Scholar
                  Article Location
                • 112. 
                  Phillips CV. 2004. Publication bias in situ. BMC Med. Res. Methodol. 4:20
                  • Crossref
                  • Medline
                  • Google Scholar
                  Article Location
                  More AR articles citing this reference

                  • Calibrating the Scientific Ecosystem Through Meta-Research

                    Tom E. Hardwicke,1,2 Stylianos Serghiou,2,3 Perrine Janiaud,2 Valentin Danchev,2 Sophia Crüwell,1,4 Steven N. Goodman,2,3,5 and John P.A. Ioannidis1,2,3,5,61Meta-Research Innovation Center Berlin (METRIC-B), QUEST Center for Transforming Biomedical Research, Berlin Institute of Health, Charité–Universitätsmedizin Berlin, 10178 Berlin, Germany; email: [email protected]2Meta-Research Innovation Center at Stanford (METRICS), Stanford University, Stanford, California 94305, USA3Department of Health Research and Policy, Stanford University, Stanford, California 94305, USA4Department of Psychological Methods, University of Amsterdam, 1018 WS Amsterdam, Netherlands5Department of Medicine, Stanford University, Stanford, California 94305, USA6Departments of Biomedical Data Science, and of Statistics, Stanford University, California, USA
                    Annual Review of Statistics and Its Application Vol. 7: 11 - 37
                    • ...specific measured outcomes, and outcomes arising from particular analyses (Phillips 2004)....
                  • Psychology's Renaissance

                    Leif D. Nelson,1 Joseph Simmons,2 and Uri Simonsohn21Haas School of Business, University of California, Berkeley, California 94720; email: [email protected]2The Wharton School, University of Pennsylvania, Philadelphia, Pennsylvania 19104; email: [email protected], [email protected]
                    Annual Review of Psychology Vol. 69: 511 - 534
                    • ...4Methodologists in other fields had brought up the problem we now know as p-hacking (Cole 1957, Ioannidis 2005, Leamer 1983, Phillips 2004)....

                • 113. 
                  Popper KR. 1959. The Logic of Scientific Discovery. New York: Basic Books
                  • Crossref
                  • Google Scholar
                  Article Location
                  More AR articles citing this reference

                  • The Personal Journey of a Resource Economist

                    Gardner M. BrownDepartment of Economics, University of Washington, Seattle, Washington 98195, USA; email: [email protected]

                    Annual Review of Resource Economics Vol. 10: 1 - 18
                    • ...As I would much later learn, this is a crucial approach that Karl Popper (1959), ...
                    • ...I remember one quote from Popper (1959, p. 59): “Scientific theories … are nets cast to catch what we call ‘the world’: to explain, ...
                  • Replication in Social Science

                    Jeremy Freese1 and David Peterson21Department of Sociology, Stanford University, Stanford, California 94305; email: [email protected]2Department of Sociology, Northwestern University, Evanston, Illinois 60208; email: [email protected]
                    Annual Review of Sociology Vol. 43: 147 - 165
                    • ...replication was considered essential to science because theories could be “falsified only if we discover a reproducible effect which refutes the theory” (Popper 1959 (1992), ...
                  • Information Recovery and Causality: A Tribute to George Judge

                    Gordon Rausser1 and David A. Bessler21Department of Agricultural and Resource Economics, University of California, Berkeley, California 94720; email: [email protected]2Department of Agricultural Economics, Texas A&M University, College Station, Texas 77843; email: [email protected]
                    Annual Review of Resource Economics Vol. 8: 7 - 23
                    • ...This Newton-style research was consistent with the logic of science as suggested by Popper (1959)....
                  • Clashing Diagnostic Approaches: DSM-ICD Versus RDoC

                    Scott O. Lilienfeld and Michael T. TreadwayDepartment of Psychology, Emory University, Atlanta, Georgia 30322; email: [email protected]
                    Annual Review of Clinical Psychology Vol. 12: 435 - 463
                    • ...Popper (1959) famously argued that for a theoretical model to be scientific, ...
                  • Build, Compute, Critique, Repeat: Data Analysis with Latent Variable Models

                    David M. BleiComputer Science Department, Princeton University, Princeton, New Jersey 08540; email: [email protected]
                    Annual Review of Statistics and Its Application Vol. 1: 203 - 232
                    • ...these methods connect to ideas of falsification in the philosophy of science (Popper 1959, Gelman & Shalizi 2012)....
                  • Rational Choice Theory and Empirical Research: Methodological and Theoretical Contributions in Europe

                    Clemens Kroneberg and Frank KalterSchool of Social Sciences, University of Mannheim, 68131 Mannheim, Germany; email: [email protected], [email protected]
                    Annual Review of Sociology Vol. 38: 73 - 92
                    • ...The main roots of this mission lie in Popper's (1959) critical rationalism and the deductive-nomological model of explanation (Hempel & Oppenheim 1948)...
                    • ...Although these practical limitations of falsifiability are a known general feature of empirical testing (Popper 1959), ...
                  • Working Memory: Theories, Models, and Controversies

                    Alan BaddeleyDepartment of Psychology, University of York, York YO10 5DD, United Kingdom; email: [email protected]

                    Annual Review of Psychology Vol. 63: 1 - 29
                    • ...in the United Kingdom at least, was provided by Karl Popper (1959), ...
                  • Failed Forensics: How Forensic Science Lost Its Way and How It Might Yet Find It

                    Michael J. Saks1 and David L. Faigman21College of Law and Department of Psychology, Arizona State University, Tempe, Arizona 85287; email: [email protected]2Hastings College of the Law, University of California, San Francisco, California 94102; email: [email protected]
                    Annual Review of Law and Social Science Vol. 4: 149 - 171
                    • ...expertise based on simple experience or observation). Popper (1959) asked his readers to consider the hypothesis that “all swans are white.” This hypothesis might be tested by the experience of observing 1000 swans, ...

                • 114. 
                  Rassen JA, Glynn RJ, Brookhart MA, Schneeweiss S. 2011. Covariate selection in high-dimensional propensity score analyses of treatment effects in small samples. Am. J. Epidemiol. 173:1404–13
                  • Crossref
                  • Medline
                  • Web of Science ®
                  • Google Scholar
                  Article Location
                • 115. 
                  Robins JM. 1999. Marginal structural models versus structural nested models as tools for causal inference. In Statistical Models in Epidemiology: The Environment and Clinical Trials, ed. ME Halloran, D Berry, 116:95–134. New York: Springer
                  • Google Scholar
                  Article Locations:
                  • Article Location
                  • Article Location
                • 116. 
                  Robins JM, Greenland S. 1986. The role of model selection in causal inference from nonexperimental data. Am. J. Epidemiol. 123:392–402
                  • Crossref
                  • Medline
                  • Web of Science ®
                  • Google Scholar
                  Article Locations:
                  • Article Location
                  • Article Location
                  • Article Location
                  • Article Location
                  • Article Location
                  • Article Location
                  More AR articles citing this reference

                  • Confounding in Health Research

                    Sander Greenland1,2 and Hal Morgenstern11Department of Epidemiology, Los Angeles School of Public Health, University of California, Los Angeles, California 90095-1772; e-mail: [email protected] 2Department of Statistics, Los Angeles College of Letters and Science, University of California, Los Angeles, California 90095
                    Annual Review of Public Health Vol. 22: 189 - 212
                    • ...few if any strata would have subjects in both treatment groups, thereby making comparisons biased, inefficient, or impossible (38A, 79)....

                • 117. 
                  Robins JM, Hernán MA. 2009. Estimation of the causal effects of time-varying exposures. In Longitudinal Data Analysis, ed. G Fitzmaurice, M Davidian, G Verbeke, G Molenberghs, pp. 553–600. New York: Chapman & Hall/CRC Press
                  • Google Scholar
                  Article Location
                  More AR articles citing this reference

                  • Event History Analysis

                    Niels KeidingDepartment of Biostatistics, University of Copenhagen, Copenhagen DK-1014, Denmark; email: [email protected]
                    Annual Review of Statistics and Its Application Vol. 1: 333 - 360
                    • ...one that Robins and coworkers have carried out since 1986 (for recent surveys of time-dependent confounding, see Robins & Hernán 2009...

                • 118. 
                  Robins JM, Mark SD, Newey WK. 1992. Estimating exposure effects by modeling the expectation of exposure conditional on confounders. Biometrics 48:479–95
                  • Crossref
                  • Medline
                  • Web of Science ®
                  • Google Scholar
                  Article Location
                  More AR articles citing this reference

                  • Evaluation of Causal Effects and Local Structure Learning of Causal Networks

                    Zhi Geng,1 Yue Liu,1 Chunchen Liu,2 and Wang Miao3 1School of Mathematical Sciences, Peking University, Beijing 100871, China; email: [email protected], [email protected]2Department of Data Mining, NEC Laboratories China, Beijing 100600, China; email: [email protected]3Guanghua School of Management, Peking University, Beijing 100871, China; email: [email protected]
                    Annual Review of Statistics and Its Application Vol. 6: 103 - 124
                    • ...such as doubly robust estimation (Scharfstein et al. 1999) and g-estimation (Robins et al. 1992), ...
                  • Causal Inference in Sociological Research

                    Markus GanglDepartment of Sociology, University of Wisconsin, Madison, Wisconsin 53706-1320; email: [email protected]
                    Annual Review of Sociology Vol. 36: 21 - 47
                    • ...matching and regression can even be seen as members of a larger class of inverse probability of treatment weighted (IPW; see Robins et al. 1992, Hirano et al. 2003, Wooldridge 2007) estimators of the form...

                • 119. 
                  Rosenbaum PR. 2002. Observational Studies. New York: Springer
                  • Crossref
                  • Google Scholar
                  Article Location
                  More AR articles citing this reference

                  • Evaluation of Causal Effects and Local Structure Learning of Causal Networks

                    Zhi Geng,1 Yue Liu,1 Chunchen Liu,2 and Wang Miao3 1School of Mathematical Sciences, Peking University, Beijing 100871, China; email: [email protected], [email protected]2Department of Data Mining, NEC Laboratories China, Beijing 100600, China; email: [email protected]3Guanghua School of Management, Peking University, Beijing 100871, China; email: [email protected]
                    Annual Review of Statistics and Its Application Vol. 6: 103 - 124
                    • ...sensitivity analysis (Cornfield et al. 1959, Rosenbaum & Rubin 1983a, Rosenbaum 2002) is a powerful tool to assess the impact of confounding and untestable assumptions on causal conclusions....
                  • Methodological Challenges and Opportunities in Testing for Racial Discrimination in Policing

                    Roland Neil and Christopher WinshipDepartment of Sociology, Harvard University, Cambridge, Massachusetts, 02138, USA; email: [email protected]
                    Annual Review of Criminology Vol. 2: 73 - 98
                    • ...Sensitivity analysis is a means of testing how robust conclusions are to processes that may matter but are not explicitly modeled (Lin et al. 1998, Morgan & Winship 2015, Rosenbaum 2002, VanderWeele 2008)....
                  • Econometric Methods for Program Evaluation

                    Alberto Abadie1 and Matias D. Cattaneo21Department of Economics, Massachusetts Institute of Technology, Cambridge, Massachusetts 02142, USA; email: [email protected]2Department of Economics and Department of Statistics, University of Michigan, Ann Arbor, Michigan 48109, USA; email: [email protected]
                    Annual Review of Economics Vol. 10: 465 - 503
                    • ...Blundell & Costa Dias (2009), DiNardo & Lee (2011), Heckman & Vytlacil (2007), Hernán & Robins (2018), Imbens & Rubin (2015), Imbens & Wooldridge (2009), Lee (2016), Manski (2008), Pearl (2009), Rosenbaum (2002, 2010), ...
                    • ...a large literature on sensitivity to unobserved confounders analyzes the impact of departures from the unconfoundedness assumption in Equation 12 (see, e.g., Altonji et al. 2005, Imbens 2003, Rosenbaum 2002)....
                  • On the Measurement and Identification of Turning Points in Criminology

                    Holly Nguyen1 and Thomas A. Loughran21Department of Sociology and Criminology, Pennsylvania State University, University Park, Pennsylvania 16802, USA2Department of Criminology and Criminal Justice, University of Maryland, College Park, Maryland 20742, USA; email: [email protected]
                    Annual Review of Criminology Vol. 1: 335 - 358
                    • ...One common strategy is to create a suitable counterfactual through the use of a balancing score, such as a propensity score (Rosenbaum 2002, Rosenbaum & Rubin 1984)....
                  • Bias Analysis for Uncontrolled Confounding in the Health Sciences

                    Onyebuchi A. ArahDepartment of Epidemiology, Fielding School of Public Health; UCLA Center for Health Policy Research; and California Center for Population Research, University of California, Los Angeles, California 90095; email: [email protected]
                    Annual Review of Public Health Vol. 38: 23 - 38
                    • ...Observational studies play a central role in the health sciences (47...
                    • ...the numerical value from the bias formula without specifying the underlying bias parameters relating U to X given C and relating U to Y given X and C) (21) or related methods (47, 48)...
                  • Race as a Bundle of Sticks: Designs that Estimate Effects of Seemingly Immutable Characteristics

                    Maya Sen1 and Omar Wasow21Harvard Kennedy School, Cambridge, Massachusetts 02138; email: [email protected]2Department of Politics, Princeton University, Princeton, New Jersey 08544; email: [email protected]
                    Annual Review of Political Science Vol. 19: 499 - 522
                    • ...a pervasive problem within observational social science research (King et al. 1994, Rosenbaum 2002)....
                    • ... provide some useful examples, and Keele (2010) and Rosenbaum (2002) discuss the methodology....
                  • Transparency, Replication, and Cumulative Learning: What Experiments Alone Cannot Achieve

                    Thad DunningTravers Department of Political Science, University of California, Berkeley, California 94720-1950; email: [email protected]
                    Annual Review of Political Science Vol. 19: 541 - 563
                    • ...as have I in the case of natural experiments (see, e.g., Brady & Collier 2010, Rosenbaum 2002, Dunning 2012)....
                  • Transparency, Replication, and Cumulative Learning: What Experiments Alone Cannot Achieve

                    Thad DunningTravers Department of Political Science, University of California, Berkeley, California 94720-1950; email: [email protected]
                    Annual Review of Political Science Vol. 19: S1 - S23
                    • ...as have I in the case of natural experiments (see, e.g., Brady & Collier 2010, Rosenbaum 2002, Dunning 2012)....
                  • How to See More in Observational Studies: Some New Quasi-Experimental Devices

                    Paul R. RosenbaumDepartment of Statistics, Wharton School, University of Pennsylvania, Philadelphia, Pennsylvania 19104; email: [email protected]
                    Annual Review of Statistics and Its Application Vol. 2: 21 - 48
                    • ...some form of sensitivity analysis is required (e.g., Cornfield et al. 1959; Rosenbaum & Rubin 1983; Rosenbaum 1988, 1991, 2002, 2007, 2010a...
                    • ...and a bias of this magnitude would suffice to explain away the association between heavy smoking and lung cancer in Hammond's (1964) study (see Rosenbaum 2002, ...
                    • ...Hammond's (1964) study of smoking and lung cancer yields a maximum p-value of 0.10 at Γ=6 (see Rosenbaum 2002, ...
                    • ...for nonrandom assignment to rows of Table 1 to spuriously produce a p-value less than or equal to 0.05 in the absence of an effect of dose difference on toxicity difference would require a bias greater than Γ=15 by the method proposed by Rosenbaum (2002, ...
                  • Advances in Measuring the Environmental and Social Impacts of Environmental Programs

                    Paul J. Ferraro1 and Merlin M. Hanauer21Department of Economics, Andrew Young School of Policy Studies, Georgia State University, Atlanta, Georgia 30302; email: [email protected]2Department of Economics, Sonoma State University, Rohnert Park, California 94928; email: [email protected]
                    Annual Review of Environment and Resources Vol. 39: 495 - 517
                    • ...scholars have made substantial advances in the empirical study of causal relationships—the effect of one variable on another (1...
                    • ...analysts can use tests of known effects [also called placebo or falsification tests (2)]....
                    • ...one should posit that the red lines in Figure 3 may exist and then ask how the conclusions would change (2, 68...
                    • ...they can be used both to test for hidden bias and to bound the range of plausible impact estimates (see 2, 71)....
                  • A Systematic Statistical Approach to Evaluating Evidence from Observational Studies

                    David Madigan,1,2 Paul E. Stang,2,3 Jesse A. Berlin,4 Martijn Schuemie,2,3 J. Marc Overhage,2,5 Marc A. Suchard,2,6,7,8 Bill Dumouchel,2,9 Abraham G. Hartzema,2,10 and Patrick B. Ryan2,31Department of Statistics, Columbia University, New York, New York 10027; email: [email protected]2Observational Medical Outcomes Partnership, Foundation for the National Institutes of Health, Bethesda, Maryland 208103Janssen Research and Development LLC, Titusville, New Jersey, 085604Johnson & Johnson, New Brunswick, New Jersey, 08901; email: [email protected], [email protected], [email protected], [email protected]5Siemens Health Services, Malvern, Pennsylvania, 19355; email: [email protected]6Department of Biomathematics, David Geffen School of Medicine, University of California, Los Angeles, California, 90095; email: [email protected]7Department of Human Genetics, David Geffen School of Medicine, University of California, Los Angeles, California, 900958Department of Biostatistics, Fielding School of Public Health, University of California, Los Angeles, California, 900959Oracle Health Sciences, Burlington, Massachusetts, 01803; email: [email protected]10College of Pharmacy, University of Florida, Gainesville, Florida, 32610; email: [email protected]
                    Annual Review of Statistics and Its Application Vol. 1: 11 - 39
                    • ...but differ from experiments in that assignment of treatment to subjects is not controlled (Rosenbaum 2002)....
                    • ...An observational study is biased if the treated and control groups differ prior to treatment in ways that can influence the outcome under study (Rosenbaum 2002)....
                    • ...A popular tool to overcome this limitation is propensity score analysis (Rosenbaum 2002, Rubin 1997)....
                    • ...Rosenbaum (2002) posits the existence of a latent confounder and explores the magnitude of the confounding that would be required to explain away the observed effect....
                  • Credible Causal Inference for Empirical Legal Studies

                    Daniel E. Ho1 and Donald B. Rubin21Stanford Law School, Stanford University, Stanford, California 94305; email: [email protected]2Department of Statistics, Harvard University, Cambridge, Massachusetts 02138; email: [email protected]
                    Annual Review of Law and Social Science Vol. 7: 17 - 40
                    • ...and Rosenbaum (2002) provide general overviews of program evaluation and causal inference....
                    • ...For alternative matching approaches, see Iacus et al. (2009a), Rosenbaum (2002), Abadie & Gardeazabal (2003), ...
                  • Causal Inference in Sociological Research

                    Markus GanglDepartment of Sociology, University of Wisconsin, Madison, Wisconsin 53706-1320; email: [email protected]
                    Annual Review of Sociology Vol. 36: 21 - 47
                    • ...Originating in the statistical (e.g., Holland 1986; Rosenbaum 2002; Rubin 2005, 2006) and econometrics literature (see Heckman 2000, 2001, 2005...
                    • ...The propensity score thus reduces a high-dimensional to a one-dimensional matching problem and has become the main vehicle to implement matching estimators. Rosenbaum (2002), Rubin (2006), ...
                    • ...Originally developed for matching estimators by Rosenbaum & Rubin (1983a) and Rosenbaum (2002), ...
                  • Opiates for the Matches: Matching Methods for Causal Inference

                    Jasjeet S. SekhonTravers Department of Political Science, Survey Research Center, University of California, Berkeley, California 94720; email: [email protected]
                    Annual Review of Political Science Vol. 12: 487 - 508
                    • ...for example, conditioning on post-treatment variables (Cox 1958, Section 4.2; Rosenbaum 2002, ...
                    • ...The Neyman-Rubin framework has become increasingly popular in many fields, including statistics (Holland 1986; Rosenbaum 2002...
                    • ...Extensions to the case of multiple discrete treatment are straightforward (e.g., Imbens 2000; Rosenbaum 2002, ...
                  • Methods for Improving Regression Analysis for Skewed Continuous or Counted Responses

                    Abdelmonem A. Afifi,1 Jenny B. Kotlerman,2 Susan L. Ettner,1,3 and Marie Cowan21School of Public Health, 2School of Nursing, 3School of Medicine, University of California, Los Angeles, California 90095-1772; email: [email protected], [email protected], [email protected], [email protected]
                    Annual Review of Public Health Vol. 28: 95 - 111
                    • ...Readers interested in that topic are referred to the following articles: D'Agostino (13), Rosenbaum (51), ...

                • 120. 
                  Rothman KJ, Greenland S, Lash TL, eds. 2008. Modern Epidemiology. Philadelphia, PA: Lippincott Williams & Wilkins. 3rd ed.
                  • Google Scholar
                • 121. 
                  Rothman KJ, Greenland S, Lash TL. 2008. Validity in epidemiologic studies. See Ref. 120, pp. 128–47
                  • Google Scholar
                  Article Locations:
                  • Article Location
                  • Article Location
                  • Article Location
                  • Article Location
                • 122. 
                  Royston P, Sauerbrei W. 2008. Multivariable Model-Building: A Pragmatic Approach to Regression Analysis Based on Fractional Polynomials for Modelling Continuous Variables. Chichester, UK: Wiley
                  • Crossref
                  • Google Scholar
                  Article Locations:
                  • Article Location
                  • Article Location
                  • Article Location
                • 123. 
                  Rubin DB. 1991. Practical implications of modes of statistical inference for causal effects and the critical role of the assignment mechanism. Biometrics 47:1213–34
                  • Crossref
                  • Medline
                  • Web of Science ®
                  • Google Scholar
                  Article Location
                  More AR articles citing this reference

                  • Confounding in Health Research

                    Sander Greenland1,2 and Hal Morgenstern11Department of Epidemiology, Los Angeles School of Public Health, University of California, Los Angeles, California 90095-1772; e-mail: [email protected] 2Department of Statistics, Los Angeles College of Letters and Science, University of California, Los Angeles, California 90095
                    Annual Review of Public Health Vol. 22: 189 - 212
                    • ...Counterfactual approaches to causal inference emphasize the importance of randomization in assuring identifiability of causal effects (30, 36, 38, 62, 70, 86, 91, 92, 93)....
                  • THE ESTIMATION OF CAUSAL EFFECTS FROM OBSERVATIONAL DATA

                    Christopher Winship and Stephen L. MorganDepartment of Sociology, Harvard University, William James Hall, 33 Kirkland Street, Cambridge, Massachusetts 02138; e-mail: [email protected] ;[email protected]
                    Annual Review of Sociology Vol. 25: 659 - 706
                    • ...The propensity score method (Rosenbaum & Rubin 1983, 1984, 1985;, Rosenbaum 1984a, b, 1995;, Rubin 1991;, Rubin & Thomas 1996) provides a much more general approach that is nonetheless based on the same strategy as the regression discontinuity design....

                • 124. 
                  Rubin DB. 2007. The design versus the analysis of observational studies for causal effects: parallels with the design of randomized trials. Stat. Med. 26:20–36
                  • Crossref
                  • Medline
                  • Web of Science ®
                  • Google Scholar
                  Article Location
                  More AR articles citing this reference

                  • Modern Algorithms for Matching in Observational Studies

                    Paul R. RosenbaumDepartment of Statistics, Wharton School, University of Pennsylvania, Philadelphia, Pennsylvania 19104-6340, USA; email: [email protected]
                    Annual Review of Statistics and Its Application Vol. 7: 143 - 176
                    • ...Developing these thoughts, Rubin (2007, pp. 20, 26) wrote: ...
                  • Causal Modeling in Environmental Health

                    Marie-Abèle BindDepartment of Statistics, Faculty of Arts and Sciences, Harvard University, Cambridge, Massachusetts 02138, USA; email: [email protected]
                    Annual Review of Public Health Vol. 40: 23 - 43
                    • ...Rubin (88) has promoted for decades a separation of the design stage from the analysis stage in observational studies and illustrated his strategy in the context of the US tobacco litigation....
                    • ...he created two comparable groups of male current smokers and of male never smokers, blindly of any outcome data (88)....
                    • ...Matching units with respect to their estimated propensity score has been a successful strategy to obtain balanced groups (88)....
                  • Econometric Methods for Program Evaluation

                    Alberto Abadie1 and Matias D. Cattaneo21Department of Economics, Massachusetts Institute of Technology, Cambridge, Massachusetts 02142, USA; email: [email protected]2Department of Economics and Department of Statistics, University of Michigan, Ann Arbor, Michigan 48109, USA; email: [email protected]
                    Annual Review of Economics Vol. 10: 465 - 503
                    • ...This feature of matching and IPW estimators provides a potential safeguard against specification searches and p-hacking, a point forcefully made by Rubin (2007)....
                  • Advances in Measuring the Environmental and Social Impacts of Environmental Programs

                    Paul J. Ferraro1 and Merlin M. Hanauer21Department of Economics, Andrew Young School of Policy Studies, Georgia State University, Atlanta, Georgia 30302; email: [email protected]2Department of Economics, Sonoma State University, Rohnert Park, California 94928; email: [email protected]
                    Annual Review of Environment and Resources Vol. 39: 495 - 517
                    • ...scholars have made substantial advances in the empirical study of causal relationships—the effect of one variable on another (1...
                  • A "SMART" Design for Building Individualized Treatment Sequences

                    H. Lei,1 I. Nahum-Shani,2 K. Lynch,3 D. Oslin,4 and S.A. Murphy51Department of Statistics, University of Michigan, Ann Arbor, Michigan 48109; email: [email protected]2Institute for Social Research, University of Michigan, Ann Arbor, Michigan 48106; email: [email protected]3Treatment Research Center and Center for Studies of Addictions, Department of Psychiatry, University of Pennsylvania, Philadelphia, Pennsylvania 19104; email: [email protected]4Philadelphia Veterans Administration Medical Center, Philadelphia, Pennsylvania 19104, and Treatment Research Center and Center for Studies of Addictions, Department of Psychiatry, University of Pennsylvania, Philadelphia, Pennsylvania 19104; email: [email protected]5Department of Statistics, Institute for Social Research, and Department of Psychiatry, University of Michigan, Ann Arbor, Michigan 48109; email: [email protected]
                    Annual Review of Clinical Psychology Vol. 8: 21 - 48
                    • ...the randomization allows scientists to make valid causal inferences concerning the relative effectiveness of the intervention options without having to make unverifiable assumptions (Rubin 1974, 2007)....

                • 125. 
                  Schlesselman JJ. 1982. Case-Control Studies: Design, Conduct, Analysis. New York: Oxford Univ. Press
                  • Google Scholar
                  Article Locations:
                  • Article Location
                  • Article Location
                  • Article Location
                • 126. 
                  Schneeweiss S, Rassen JA, Glynn RJ, Avorn J, Mogun H, Brookhart MA. 2009. High-dimensional propensity score adjustment in studies of treatment effects using health care claims data. Epidemiology 20:512–22
                  • Crossref
                  • Medline
                  • Web of Science ®
                  • Google Scholar
                  Article Location
                  More AR articles citing this reference

                  • Accounting for Confounding in Observational Studies

                    Brian M. D'Onofrio,1,2 Arvid Sjölander,2 Benjamin B. Lahey,3 Paul Lichtenstein,2 and A. Sara Öberg2,41Department of Psychological and Brain Sciences, Indiana University, Bloomington, Indiana 47405, USA; email: [email protected]2Department of Medical Epidemiology and Biostatistics, Karolinska Institutet, SE-171 77 Stockholm, Sweden; email: [email protected], [email protected], [email protected]3Departments of Health Studies and Psychiatry and Behavioral Neuroscience, University of Chicago, Chicago, Illinois 60637, USA; email: [email protected]4Department of Epidemiology, Harvard T.H. Chan School of Public Health, Boston, Massachusetts 02115, USA
                    Annual Review of Clinical Psychology Vol. 16: 25 - 48
                    • ...hundreds of covariates) increases the likelihood that such PSs serve as proxies for relevant unobserved confounders (Schneeweiss et al. 2009)....
                  • A Systematic Statistical Approach to Evaluating Evidence from Observational Studies

                    David Madigan,1,2 Paul E. Stang,2,3 Jesse A. Berlin,4 Martijn Schuemie,2,3 J. Marc Overhage,2,5 Marc A. Suchard,2,6,7,8 Bill Dumouchel,2,9 Abraham G. Hartzema,2,10 and Patrick B. Ryan2,31Department of Statistics, Columbia University, New York, New York 10027; email: [email protected]2Observational Medical Outcomes Partnership, Foundation for the National Institutes of Health, Bethesda, Maryland 208103Janssen Research and Development LLC, Titusville, New Jersey, 085604Johnson & Johnson, New Brunswick, New Jersey, 08901; email: [email protected], [email protected], [email protected], [email protected]5Siemens Health Services, Malvern, Pennsylvania, 19355; email: [email protected]6Department of Biomathematics, David Geffen School of Medicine, University of California, Los Angeles, California, 90095; email: [email protected]7Department of Human Genetics, David Geffen School of Medicine, University of California, Los Angeles, California, 900958Department of Biostatistics, Fielding School of Public Health, University of California, Los Angeles, California, 900959Oracle Health Sciences, Burlington, Massachusetts, 01803; email: [email protected]10College of Pharmacy, University of Florida, Gainesville, Florida, 32610; email: [email protected]
                    Annual Review of Statistics and Its Application Vol. 1: 11 - 39
                    • ...Strategies for automated selection of large sets of covariates have been proposed as potential solutions to reduce the possibility of missing an empiric confounder (Schneeweiss et al. 2009)....

                • 127. 
                  Sclove SL, Morris C, Radhakrishnan R. 1972. Non-optimality of preliminary-test estimators for mean of a multivariate normal distribution. Ann. Math. Stat. 43:1481–90
                  • Crossref
                  • Google Scholar
                  Article Locations:
                  • Article Location
                  • Article Location
                  • Article Location
                • 128. 
                  Senn SJ. 1998. Mathematics: governess or handmaiden? Statistician 47:251–59
                  • Google Scholar
                  Article Location
                • 129. 
                  Shrier I. 2009. Propensity scores. Stat. Med. 28:1317–18
                  • Crossref
                  • Medline
                  • Web of Science ®
                  • Google Scholar
                  Article Location
                • 130. 
                  Sjölander A. 2009. Propensity scores and M-structures. Stat. Med. 28:1416–20
                  • Crossref
                  • Medline
                  • Web of Science ®
                  • Google Scholar
                  Article Location
                • 131. 
                  Sterne JAC, Smith GD. 2001. Sifting the evidence—what's wrong with significance tests? BMJ 322:226–31
                  • Crossref
                  • Medline
                  • Web of Science ®
                  • Google Scholar
                  Article Location
                • 132. 
                  Steyerberg EW. 2008. Clinical Prediction Models: A Practical Approach to Development, Validation, and Updating. New York: Springer
                  • Google Scholar
                  Article Locations:
                  • Article Location
                  • Article Location
                  • Article Location
                  • Article Location
                  • Article Location
                  • Article Location
                  • Article Location
                  • Article Location
                  • Article Location
                  • Article Location
                  More AR articles citing this reference

                  • Treatment Selection in Depression

                    Zachary D. Cohen and Robert J. DeRubeisDepartment of Psychology, University of Pennsylvania, Philadelphia, Pennsylvania 19104, USA; email: [email protected]
                    Annual Review of Clinical Psychology Vol. 14: 209 - 236
                    • ... provide an in-depth review of the merits of machine learning in mental health; interested readers can also consult books focused on applied clinical predictive modeling (Chakraborty & Moodie 2013, Parmigiani 2002, Steyerberg 2008)....

                • 133. 
                  Steyerberg EW, Eijkemans MJC, Habbema JDF. 1999. Stepwise selection in small data sets: a simulation study of bias in logistic regression analysis. J. Clin. Epidemiol. 52:935–42
                  • Crossref
                  • Medline
                  • Web of Science ®
                  • Google Scholar
                  Article Locations:
                  • Article Location
                  • Article Location
                  • Article Location
                  More AR articles citing this reference

                  • Species Distribution Models: Ecological Explanation and Prediction Across Space and Time

                    Jane Elith1 and John R. Leathwick21School of Botany, The University of Melbourne, Victoria 3010, Australia; email: [email protected]2National Institute of Water and Atmospheric Research, Hamilton, New Zealand; email: [email protected]
                    Annual Review of Ecology, Evolution, and Systematics Vol. 40: 677 - 697
                    • ...and small data sets and model selection difficulties mean that models can select irrelevant variables (Mac Nally 2000, Steyerberg et al. 1999)....

                • 134. 
                  Sullivan S, Greenland S. 2013. Bayesian regression in SAS software. Int. J. Epidemiol. 42:308–17
                  • Crossref
                  • Medline
                  • Web of Science ®
                  • Google Scholar
                  Article Locations:
                  • Article Location
                  • Article Location
                  • Article Location
                  • Article Location
                  • Article Location
                • 135. 
                  Susser M. 1977. Judgement and causal inference. Am. J. Epidemiol. 105:1–15
                  • Crossref
                  • Medline
                  • Web of Science ®
                  • Google Scholar
                  Article Location
                  More AR articles citing this reference

                  • Looking Back on “Causal Thinking in the Health Sciences”

                    J. S. Kaufman1,2 1Department of Epidemiology, University of North Carolina School of Public Health, Chapel Hill, North Carolina 27599-7400; e-mail: [email protected] C. Poole1 2Carolina Population Center, University of North Carolina at Chapel Hill, Chapel Hill, North Carolina 27516-3997
                    Annual Review of Public Health Vol. 21: 101 - 119
                    • ...which has persisted unabated to the present day, including substantial refinements by Susser himself (52, 56, 57)....
                    • ...Susser's elaboration and expansion of this list over the ensuing years (52, 53, 54, 57) forms the most detailed and prolonged attempt to develop criteria for causality in the field of epidemiology....
                    • ...But in his 1977 case study of the debate on smoking and lung cancer (52), ...

                • 136. 
                  Tchetgen Tchetgen EJ. 2013. On a closed-form doubly robust estimator of the adjusted odds ratio for a binary exposure. Am. J. Epidemiol. 177:1314–16
                  • Crossref
                  • Medline
                  • Web of Science ®
                  • Google Scholar
                  Article Locations:
                  • Article Location
                  • Article Location
                  • Article Location
                  • Article Location
                • 137. 
                  Turner RM, Spiegelhalter DJ, Smith GCS, Thompson SG. 2009. Bias modelling in evidence synthesis. J. R. Stat. Soc. Ser. A 172:21–47
                  • Crossref
                  • Google Scholar
                  Article Locations:
                  • Article Location
                  • Article Location
                  More AR articles citing this reference

                  • Incorporating Both Randomized and Observational Data into a Single Analysis

                    Eloise E. KaizarDepartment of Statistics, The Ohio State University, Columbus, Ohio 43210; email: [email protected]
                    Annual Review of Statistics and Its Application Vol. 2: 49 - 72
                    • ...by assuming that indicators of high risk of bias in multiple domains linearly contribute to the mean; also presented in articles by Siersma et al. (2007) and Turner et al. (2009)]....

                • 138. 
                  Valeri L, VanderWeele TJ. 2013. Mediation analysis allowing for exposure-mediator interactions and causal interpretation: theoretical assumptions and implementation with SAS and SPSS macros. Psychol. Methods 18:137–50
                  • Crossref
                  • Medline
                  • Web of Science ®
                  • Google Scholar
                  Article Location
                  More AR articles citing this reference

                  • Mediation Analysis: A Practitioner's Guide

                    Tyler J. VanderWeeleT.H. Chan School of Public Health, Harvard University, Boston, Massachusetts 02115; email: [email protected]
                    Annual Review of Public Health Vol. 37: 17 - 32
                    • ...and software to compute these effects and their standard errors and confidence intervals automatically are available in SAS (Statistical Analysis System), Stata, and SPSS (Statistical Package for the Social Sciences) (39)....
                    • ...and software to compute these effects and their standard errors and confidence intervals automatically are available in SAS, Stata, and SPSS (39)....
                    • ...Software for estimating direct and indirect effects with log-binomial models is also available (39)....
                    • ...Software for estimating direct and indirect effects for a case-control design is also available (39)....
                    • ...Software to estimate direct and indirect effects with binary mediators is also available (39)....
                    • ...Stata, and SPSS macros described above (39) consider numerous different scenarios....
                  • Advances in Mediation Analysis: A Survey and Synthesis of New Developments

                    Kristopher J. PreacherDepartment of Psychology and Human Development, Vanderbilt University, Nashville, Tennessee 37203-5721; email: [email protected]
                    Annual Review of Psychology Vol. 66: 825 - 852
                    • ...inferring that an effect is causal requires more than establishing the correct temporal ordering of one's variables (Valeri & VanderWeele 2013)....
                    • ...Other assumptions are necessary but implicit (data are longitudinal; Valeri & VanderWeele 2013), ...
                    • ..., Stata (Hicks & Tingley 2011), and SAS (Valeri & VanderWeele 2013)....
                    • ...(Coxe & MacKinnon 2010, Imai et al. 2010a, Valeri & VanderWeele 2013)....

                • 139. 
                  Vandenbroucke JP. 1987. Should we abandon statistical modeling altogether? Am. J. Epidemiol. 126:10–13
                  • Crossref
                  • Medline
                  • Web of Science ®
                  • Google Scholar
                  Article Location
                • 140. 
                  van der Laan MJ, Rose R. 2011. Targeted Learning: Causal Inference for Observational and Experimental Data. New York: Springer
                  • Crossref
                  • Google Scholar
                  Article Locations:
                  • Article Location
                  • Article Location
                  • Article Location
                  • Article Location
                  • Article Location
                  • Article Location
                  • Article Location
                  • Article Location
                  • Article Location
                  • Article Location
                  More AR articles citing this reference

                  • The Digitization of Patient Care: A Review of the Effects of Electronic Health Records on Health Care Quality and Utilization

                    Hilal Atasoy,1 Brad N. Greenwood,2 and Jeffrey Scott McCullough31Department of Accounting, Temple University, Philadelphia, Pennsylvania 19122, USA; email: [email protected]2Carlson School of Management, University of Minnesota, Minneapolis, Minnesota 55455, USA; email: [email protected]3Department of Health Management and Policy, University of Michigan, Ann Arbor, Michigan 48109-2029, USA; email: [email protected]
                    Annual Review of Public Health Vol. 40: 487 - 500
                    • ...there exists a large and growing literature on causality for machine learning (14, 104); although these approaches will never supplant clinical trials, ...
                  • Field Experimentation and the Study of Law and Policy

                    Donald P. Green1 and Dane R. Thorley1,21Department of Political Science,2Columbia Law School, Columbia University, New York, NY 10027; email: [email protected], [email protected]
                    Annual Review of Law and Social Science Vol. 10: 53 - 72
                    • ... and machine learning (Chipman et al. 2010, Imai & Ratkovic 2013, van der Laan & Rose 2011) that facilitate experimental design and data analysis....

                • 141. 
                  VanderWeele TJ. 2009. On the relative nature of overadjustment and unnecessary adjustment. Epidemiology 20:496–99
                  • Crossref
                  • Medline
                  • Web of Science ®
                  • Google Scholar
                  Article Locations:
                  • Article Location
                  • Article Location
                • 142. 
                  VanderWeele TJ. 2011. Subtleties of explanatory language: What is meant by “mediation”? Eur. J. Epidemiol. 26:343–46
                  • Crossref
                  • Medline
                  • Web of Science ®
                  • Google Scholar
                  Article Location
                • 143. 
                  VanderWeele TJ, Hernán MA. 2012. Causal effects and natural laws: towards a conceptualization of causal counterfactuals for nonmanipulable exposures with application to the effects of race and sex. See Ref. 10, pp. 101–13
                  • Google Scholar
                  Article Location
                • 144. 
                  VanderWeele TJ, Hernán M. 2013. Causal inference under multiple versions of treatment. J. Causal Inference 1:1–20
                  • Crossref
                  • Medline
                  • Google Scholar
                  Article Location
                • 145. 
                  VanderWeele TJ, Shpitser I. 2011. A new criterion for confounder selection. Biometrics 67:1406–13
                  • Crossref
                  • Medline
                  • Web of Science ®
                  • Google Scholar
                  Article Locations:
                  • Article Location
                  • Article Location
                  • Article Location
                  More AR articles citing this reference

                  • Nonprobability Sampling and Causal Analysis

                    Ulrich Kohler,1 Frauke Kreuter,2,3,4 and Elizabeth A. Stuart51Faculty of Economics and Social Sciences, University of Potsdam, 14482 Potsdam, Germany; email: [email protected]2Joint Program in Survey Methodology, University of Maryland, College Park, Maryland 20742, USA; email: [email protected]3School of Social Sciences, University of Mannheim, 68131 Mannheim, Germany4Statistical Methods Research Department, Institute for Employment Research (IAB), 90478 Nuremberg, Germany5Department of Mental Health, Department of Biostatistics, and Department of Health Policy and Management, Bloomberg School of Public Health, Johns Hopkins University, Baltimore, Maryland 21205, USA; email: [email protected]
                    Annual Review of Statistics and Its Application Vol. 6: 149 - 172
                    • ...the parents of treatment criterion (Pearl 1995), the confounder selection criterion (VanderWeele & Shpister 2011), ...
                  • Evaluation of Causal Effects and Local Structure Learning of Causal Networks

                    Zhi Geng,1 Yue Liu,1 Chunchen Liu,2 and Wang Miao3 1School of Mathematical Sciences, Peking University, Beijing 100871, China; email: [email protected], [email protected]2Department of Data Mining, NEC Laboratories China, Beijing 100600, China; email: [email protected]3Guanghua School of Management, Peking University, Beijing 100871, China; email: [email protected]
                    Annual Review of Statistics and Its Application Vol. 6: 103 - 124
                    • ...it is difficult to have the knowledge of a complete causal diagram. Geng & Li (2002) and VanderWeele & Shpitser (2011)...
                  • Bias Analysis for Uncontrolled Confounding in the Health Sciences

                    Onyebuchi A. ArahDepartment of Epidemiology, Fielding School of Public Health; UCLA Center for Health Policy Research; and California Center for Population Research, University of California, Los Angeles, California 90095; email: [email protected]
                    Annual Review of Public Health Vol. 38: 23 - 38
                    • ...These criteria subsume a scenario in which the unmeasured variable is a common cause of the exposure and outcome (35, 42, 65)....
                  • Endogenous Selection Bias: The Problem of Conditioning on a Collider Variable

                    Felix Elwert1 and Christopher Winship21Department of Sociology, University of Wisconsin, Madison, Wisconsin 53706; email: [email protected]2Department of Sociology, Harvard University, Cambridge, Massachusetts 02138; email: [email protected]
                    Annual Review of Sociology Vol. 40: 31 - 53
                    • ...then it is sufficient to control for all variables that (directly or indirectly) cause treatment or outcome or both (VanderWeele & Shpitser 2011)....
                    • ...and hence the VanderWeele & Shpitser (2011) rule does not apply....

                • 146. 
                  Vansteelandt S, Bekaert M, Claeskens G. 2012. On model selection and model misspecification in causal inference. Stat. Methods Med. Res. 21:7–30
                  • Crossref
                  • Medline
                  • Web of Science ®
                  • Google Scholar
                  Article Locations:
                  • Article Location
                  • Article Location
                  • Article Location
                  More AR articles citing this reference

                  • Statistical Model Choice

                    Gerda ClaeskensResearch Center ORSTAT and Leuven Statistics Research Center, KU Leuven, B-3000 Leuven, Belgium; email: [email protected]
                    Annual Review of Statistics and Its Application Vol. 3: 233 - 256
                    • ...Vansteelandt et al. (2012) use the FIC to estimate the marginal treatment effect with focus on a marginal causal odds ratio, ...

                • 147. 
                  Vansteelandt S, Daniel RM. 2014. On regression adjustment for the propensity score. Stat. Med. 33:4053–72
                  • Crossref
                  • Medline
                  • Web of Science ®
                  • Google Scholar
                  Article Locations:
                  • Article Location
                  • Article Location
                • 148. 
                  Viallefont V, Raftery AE, Richardson S. 2001. Variable selection and Bayesian model averaging in case-control studies. Stat. Med. 20:3215–30
                  • Crossref
                  • Medline
                  • Web of Science ®
                  • Google Scholar
                  Article Locations:
                  • Article Location
                  • Article Location
                • 149. 
                  Vittinghoff E, McCulloch CE. 2007. Relaxing the rule of ten events per variable in logistic and Cox regression. Am. J. Epidemiol. 165:710–18
                  • Crossref
                  • Medline
                  • Web of Science ®
                  • Google Scholar
                  Article Location
                • 150. 
                  von Elm E, Altman DG, Egger M, Pocock SJ, Gøtzsche PC, et al. 2007. The Strengthening the Reporting of Observational Studies in Epidemiology (STROBE) statement: guidelines for reporting observational studies. Lancet 370:1453–57
                  • Crossref
                  • Medline
                  • Web of Science ®
                  • Google Scholar
                  Article Location
                • 151. 
                  Wakefield J, Salway R. 2001. A statistical framework for ecological and aggregate studies. J. R. Stat. Soc. Ser. A 164:119–37
                  • Crossref
                  • Google Scholar
                  Article Location
                  More AR articles citing this reference

                  • Ecologic Studies Revisited

                    Jonathan WakefieldDepartments of Statistics and Biostatistics, University of Washington, Seattle, Washington 98195; email: [email protected]
                    Annual Review of Public Health Vol. 29: 75 - 90
                    • ...When trying to understand ecologic bias it is beneficial to specify an individual-level model and aggregate to determine the consequences (64, 74, 76)....
                    • ...An alternative approach is to assume a parametric distribution for the within-area distribution of exposures and confounders (57, 76), ...

                • 152. 
                  Weiss RE. 1995. The influence of variable selection—a Bayesian diagnostic perspective. J. Am. Stat. Assoc. 90:619–25
                  • Crossref
                  • Web of Science ®
                  • Google Scholar
                  Article Locations:
                  • Article Location
                  • Article Location
                • 153. 
                  Weng HY, Hsueh YH, Messam LLM, Hertz-Picciotto I. 2009. Methods of covariate selection: directed acyclic graphs and the change-in-estimate procedure. Am. J. Epidemiol. 169:1182–90
                  • Crossref
                  • Medline
                  • Web of Science ®
                  • Google Scholar
                  Article Locations:
                  • Article Location
                  • Article Location
                • 154. 
                  Westreich D, Cole SR, Funk MJ, Brookhart MA, Stuermer T. 2011. The role of the c-statistic in variable selection for propensity score models. Pharmacoepidemiol. Drug Saf. 20:317–20
                  • Crossref
                  • Medline
                  • Web of Science ®
                  • Google Scholar
                  Article Location
                • 155. 
                  Westreich D, Greenland S. 2013. The table 2 fallacy: presenting and interpreting confounder and modifier coefficients. Am. J. Epidemiol. 177:292–98
                  • Crossref
                  • Medline
                  • Web of Science ®
                  • Google Scholar
                  Article Location
                  More AR articles citing this reference

                  • Bias Analysis for Uncontrolled Confounding in the Health Sciences

                    Onyebuchi A. ArahDepartment of Epidemiology, Fielding School of Public Health; UCLA Center for Health Policy Research; and California Center for Population Research, University of California, Los Angeles, California 90095; email: [email protected]
                    Annual Review of Public Health Vol. 38: 23 - 38
                    • ...need not have a causal interpretation in the first instance (63, 67)....

                • 156. 
                  Westreich D, Lessler J, Funk MJ. 2010. Propensity score estimation: neural networks, support vector machines, decision trees (CART), and meta-classifiers as alternatives to logistic regression. J. Clin. Epidemiol. 63:826–33
                  • Crossref
                  • Medline
                  • Web of Science ®
                  • Google Scholar
                  Article Location
                  More AR articles citing this reference

                  • Machine Learning for Sociology

                    Mario Molina and Filiz GaripDepartment of Sociology, Cornell University, Ithaca, New York 14853, USA; email: [email protected], [email protected]
                    Annual Review of Sociology Vol. 45: 27 - 45
                    • ...Recent work uses boosting (McCaffrey et al. 2004), neural networks (Setoguchi et al. 2008, Westreich et al. 2010), ...
                    • ...estimation of the propensity score in matching (Westreich et al. 2010)...

                • 157. 
                  Whitehead AN. 1929. Process and Reality. New York: Harper
                  • Google Scholar
                  Article Location
                • 158. 
                  Wilson A, Reich BJ. 2014. Confounder selection via penalized credible regions. Biometrics 70:852–61
                  • Google Scholar
                  Article Location
                • 159. 
                  Ziman J. 1978. Reliable Knowledge. New York: Cambridge
                  • Google Scholar
                  Article Location

                More AR articles citing this reference

                    Previous Article Next Article
                    From KNOWABLE MAGAZINE
                    The pernicious contagion of misinformation
                    From KNOWABLE MAGAZINE
                    Recycling meets reality
                    • Related Articles
                    • Literature Cited
                    • Most Downloaded
                    Most Downloaded from this journal

                    Designing Difference in Difference Studies: Best Practices for Public Health Policy Research

                    Coady Wing, Kosali Simon, Ricardo A. Bello-Gomez
                    Vol. 39, 2018

                    AbstractPreview

                    Abstract

                    The difference in difference (DID) design is a quasi-experimental research design that researchers often use to study causal relationships in public health settings where randomized controlled trials (RCTs) are infeasible or unethical. However, causal ...Read More

                    • Full Text HTML
                    • Download PDF

                    Racism and Health: Evidence and Needed Research

                    David R. Williams, Jourdyn A. Lawrence, Brigette A. Davis
                    Vol. 40, 2019

                    AbstractPreview

                    Abstract

                    In recent decades, there has been remarkable growth in scientific research examining the multiple ways in which racism can adversely affect health. This interest has been driven in part by the striking persistence of racial/ethnic inequities in health and ...Read More

                    • Full Text HTML
                    • Download PDF

                    The Prescription Opioid and Heroin Crisis: A Public Health Approach to an Epidemic of Addiction

                    Andrew Kolodny, David T. Courtwright, Catherine S. Hwang, Peter Kreiner, John L. Eadie, Thomas W. Clark, G. Caleb Alexander
                    Vol. 36, 2015

                    Abstract - FiguresPreview

                    Abstract

                    Public health authorities have described, with growing alarm, an unprecedented increase in morbidity and mortality associated with use of opioid pain relievers (OPRs). Efforts to address the opioid crisis have focused mainly on reducing nonmedical OPR ...Read More

                    • Full Text HTML
                    • Download PDF
                    • Figures
                    image

                    Figure 1: Rates of OPR sales, OPR-related unintentional overdose deaths, and OPR addiction treatment admissions, 1999–2010. Abbreviation: OPR, opioid pain reliever. Source: 10.

                    image

                    Figure 2: Heroin admissions, by age group and race/ethnicity: 2001–2011. Source: 68.

                    image

                    Figure 3: First-time nonmedical use of pain relievers. Source: 64, 70.

                    image

                    Figure 4: (a) Past month nonmedical OPR use by age versus (b) OPR-related unintentional overdose deaths by age. Abbreviation: OPR, opioid pain reliever. Sources: 58, 68.

                    image

                    Figure 5: Rate of hospital inpatient stays related to OPR use by adult age group, 1993 and 2012. Source: 56.


                    Can We Say What Diet Is Best for Health?

                    D.L. Katz and S. Meller
                    Vol. 35, 2014

                    Abstract - FiguresPreview

                    Abstract

                    Diet is established among the most important influences on health in modern societies. Injudicious diet figures among the leading causes of premature death and chronic disease. Optimal eating is associated with increased life expectancy, dramatic ...Read More

                    • Full Text HTML
                    • Download PDF
                    • Figures
                    image

                    Figure 1: The theme of optimal eating. Diverse diets making competing claims actually emphasize key elements that are generally compatible, complementary, or even duplicative. Competition for public a...


                    The Growing Impact of Globalization for Health and Public Health Practice

                    Ronald Labonté, Katia Mohindra, and Ted Schrecker
                    Vol. 32, 2011

                    Abstract - FiguresPreview

                    Abstract

                    In recent decades, public health policy and practice have been increasingly challenged by globalization, even as global financing for health has increased dramatically. This article discusses globalization and its health challenges from a vantage of ...Read More

                    • Full Text HTML
                    • Download PDF
                    • Figures
                    image

                    Figure 1: Global poverty: World Bank $1.25/day poverty line. Source: Data from Reference 24. Note that East Asia and Pacific includes China; South Asia includes India.

                    image

                    Figure 2: Global poverty: World Bank $2.50/day poverty line. Source: Data from Reference 24. Note that East Asia and Pacific includes China; South Asia includes India.

                    image

                    Figure 3: Quadruple burden of disease in South Africa: percentage of overall years of life lost, 2000. Source: (16). “Pre-transitional causes” of death include communicable diseases, maternal and peri...


                    See More
                    • © Copyright 2020
                    • Contact Us
                    • Email Preferences
                    • Annual Reviews Directory
                    • Multimedia
                    • Supplemental Materials
                    • FAQs
                    • Privacy Policy
                    Back to Top

                    PRIVACY NOTICE

                    Accept

                    This site requires the use of cookies to function. It also uses cookies for the purposes of performance measurement. Please see our Privacy Policy.