A Theory of Statistical Inference for Ensuring the Robustness of Scientific Results
Beau Coker (),
Cynthia Rudin () and
Gary King ()
Additional contact information
Beau Coker: Department of Biostatistics, Harvard T. H. Chan School of Public Health, Boston, Massachusetts 02115
Cynthia Rudin: Department of Computer Science and Department of Electrical and Computer Engineering, Duke University, Durham, North Carolina 27708
Gary King: Institute for Quantitative Social Science, Harvard University, Cambridge, Massachusetts 02138
Management Science, 2021, vol. 67, issue 10, 6174-6197
Abstract:
Inference is the process of using facts we know to learn about facts we do not know. A theory of inference gives assumptions necessary to get from the former to the latter, along with a definition for and summary of the resulting uncertainty. Any one theory of inference is neither right nor wrong but merely an axiom that may or may not be useful. Each of the many diverse theories of inference can be valuable for certain applications. However, no existing theory of inference addresses the tendency to choose, from the range of plausible data analysis specifications consistent with prior evidence, those that inadvertently favor one’s own hypotheses. Because the biases from these choices are a growing concern across scientific fields, and in a sense the reason the scientific community was invented in the first place, we introduce a new theory of inference designed to address this critical problem. We introduce hacking intervals , which are the range of a summary statistic one may obtain given a class of possible endogenous manipulations of the data. Hacking intervals require no appeal to hypothetical data sets drawn from imaginary superpopulations. A scientific result with a small hacking interval is more robust to researcher manipulation than one with a larger interval and is often easier to interpret than a classical confidence interval. Some versions of hacking intervals turn out to be equivalent to classical confidence intervals, which means they may also provide a more intuitive and potentially more useful interpretation of classical confidence intervals.
Keywords: robustness; replicability; observational data; model dependence; causal inference; matching (search for similar items in EconPapers)
Date: 2021
References: View references in EconPapers View complete reference list from CitEc
Citations: View citations in EconPapers (4)
Downloads: (external link)
http://dx.doi.org/10.1287/mnsc.2020.3818 (application/pdf)
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:inm:ormnsc:v:67:y:2021:i:10:p:6174-6197
Access Statistics for this article
More articles in Management Science from INFORMS Contact information at EDIRC.
Bibliographic data for series maintained by Chris Asher ().