Poor practice is catching up with science,1-3 manifesting in part in the failure of results to be reproducible and replicable.4-7 Various causes have been posited,1, 8 but we believe that poor statistical education and practice are symptoms of and contributors to problems in science as a whole.
The problem is one of cargo-cult statistics – the ritualistic miming of statistics rather than conscientious practice. This has become the norm in many disciplines, reinforced and abetted by statistical education, statistical software, and editorial policies.
At the risk of oversimplifying a complex historical process, we think the strongest force pushing science (and statistics) in the wrong direction is existential: science has become a career, rather than a calling, while quality control mechanisms have not kept pace.9
Some, such as historian and sociologist of science Steven Shapin, still argue that science survives thanks to the ethical commitment of scientists,10 but others, such as philosopher of science Jerome Ravetz, find this a charitable perspective.11,12 Much of what is currently called “science” may be viewed as mechanical application of particular technologies, including statistical calculations,13 rather than adherence to shared moral norms.14, 15
We believe the root of the problem lies in the mid-twentieth century.
The bigger picture
After World War II, governments increased funding for science in response to the assessment16 that scientific progress is important for national security, prosperity, and quality of life. This increased the scale of science and the pool of scientific labour: science became “big science” conducted by career professionals. The resulting increase in scientific output17, 18 required new approaches to managing science and scientists, including new government agencies and a focus on quantifying scientific productivity.19 The norms and self-regulating aspects of “little science” – communities that valued questioning, craftsmanship, scepticism, self-doubt, critical appraisal of the quality of evidence, and the verifiable, and verifiably replicable, advancement of human knowledge – gave way to current approaches centring on metrics, funding, publication, and prestige.9 Such approaches may invite and reward “gaming” of the system.9, 20-22
When understanding, care, and honesty become valued less than novelty, visibility, scale, funding, and salary, science is at risk.3 Elements of the present crisis were anticipated1 by scholars such as Price23 and Ravetz9 in the 1960s and 1970s; a more modern explanation of science’s crisis, in terms of the prevailing economic model – and of the commodification of science – is offered by Mirowski.24 Other scholars such as John Ioannidis, Brian Nosek, Marc Edwards, et al., now study the perverse system of incentives and its consequences5, 25, 26 and how bad science outcompetes8 better science.
While some argue that there is no crisis27 (or at least not a systemic problem28), bad incentives, bad scientific practices, outdated methods of vetting and disseminating results,29 and techno-science appear to be producing misleading30 and incorrect results.31 This might produce crisis of biblical proportions: as Edwards and Roy write: “If a critical mass of scientists become untrustworthy, a tipping point is possible in which the scientific enterprise itself becomes inherently corrupt and public trust is lost, risking a new dark age with devastating consequences to humanity.”19
Scientists collectively risk losing credibility and authority in part because of prominent examples of poor practice, but also because many are guilty of ultracrepidation: acting as if their stature in one domain makes them authoritative in others. Science is “show me,” not “trust me.”32 The example of 107 Nobel laureates – mostly in areas unrelated to genetics, agriculture, ecology, or public health – endorsing one side of the genetically modified organisms in food argument as “scientific”33 is a visible example of prestige and uninformed consensus conflated with evidence. As G.K. Chesterton wrote: “Fallacies do not cease to be fallacies because they become fashions.”34
Statistics was developed to root out error, appraise evidence, quantify uncertainty, and generally to keep us from fooling ourselves. Increasingly often, it is used instead to aid and abet weak science, a role it can perform well when used mechanically or ritually.35
In his 1974 Caltech commencement speech, Nobel physicist Richard Feynman coined the label “cargo-cult science”36 for work that has some formal trappings of science but does not practise the scientific method.
Feynman’s neologism borrows from anthropological observations of Melanesian cultures that experienced a bonanza in World War II, when military cargo aircraft landed on the islands, bringing a wealth of goods. To bring back the cargo planes, islanders set up landing strips, lit fires as runway lights, and mimed communication with the oncoming planes using makeshift communication huts, wooden headsets, and the like.36 They went through the motions that had led to landings, without understanding the significance of those motions.
In our experience, many applications of statistics are cargo-cult statistics: practitioners go through the motions of fitting models, computing p-values or confidence intervals, or simulating posterior distributions. They invoke statistical terms and procedures as incantations, with scant understanding of the assumptions or relevance of the calculations, or even the meaning of the terminology. This demotes statistics from a way of thinking about evidence and avoiding self-deception to a formal “blessing” of claims. The effectiveness of cargo-cult statistics is predictably uneven. But it is effective at getting weak work published – and is even required by some journals.
The crisis in statistics is a microcosm of the crisis in science: the mechanical application of methods without understanding their assumptions, limitations, or interpretation will surely reduce scientific replicability. There are, of course, concerns about statistical practice.37 For instance, a statement on p-values38 by the American Statistical Association (ASA) was accompanied by no fewer than 21 commentaries, mostly by practitioners involved in drafting the ASA statement. Their disagreement could be misinterpreted to suggest that anything goes in statistics, but diversity of opinion within statistics is not as broad as it may appear to outsiders.39
The largest divide is between frequentist and Bayesian philosophies, which differ fundamentally in how they conceive of and quantify uncertainty, and even in their ontologies. But no good statistician, Bayesian or frequentist, would ignore how the data were generated in assessing statistical evidence nor would claim that a p-value is the probability that the null hypothesis is true – two common abuses.
The misuse of p-values, hypothesis tests, and confidence intervals might be deemed frequentist cargo-cult statistics. There is also Bayesian cargo-cult statistics. While a great deal of thought has been given to methods for eliciting priors,40, 41 in practice, priors are often chosen for convenience or out of habit; perhaps worse, some practitioners choose the prior after looking at the data, trying several priors, and looking at the results – in which case Bayes’ rule no longer applies! Such practices make Bayesian data analysis a rote, conventional calculation rather than a circumspect application of probability theory and Bayesian philosophy.
Some scientists use an incoherent hotchpotch of Bayesian methods and frequentist measures of uncertainty in the same analysis, with no apparent understanding of the fundamental mathematical and philosophical incommensurability of Bayesian and frequentist measures of uncertainty. Some add the lengths of confidence intervals to the lengths of credible intervals or add systematic uncertainties in quadrature, as if they were independent random errors. Some conflate confidence levels with credible levels. We have seen examples in a number of disciplines, notably high-energy particle physics and cosmology.42, 43
To paraphrase David Freedman, much frequentist statistics is about what you would do if you had a model, and much Bayesian statistics is about what you would do if you had a prior.44 To proceed with a model or prior that is not chosen carefully and well grounded in disciplinary knowledge, to mix frequentist and Bayesian methods obliviously, to select the prior after looking at the data to get a result one likes, and to combine systematic and stochastic errors as if they were independent random errors are all forms of cargo-cult statistics. The calculations are as likely to produce valid inferences as cargo cults were to summon cargo planes.
Statistics education: contributory negligence
While statistical education has started a sea change for the better,45, 46 in our experience, many statistics courses – especially “service” courses for non-specialists – teach cargo-cult statistics: mechanical calculations with little attention to scientific context, experimental design, assumptions and limitations of methods, or the interpretation of results.
This should not be surprising. These courses are often taught outside statistics departments by faculty whose own understanding of foundational issues is limited, having possibly taken similarly shallow courses that emphasise technique and calculation over understanding and evidence.
Service courses taught in statistics departments often have high enrolments, which help justify departmental budgets and staffing levels. Statistics departments may be under administrative, social, and financial pressure to cater to the disciplinary “consumers” of the courses. Consumers may not care whether methods are used appropriately, in part because, in their fields, the norm (including the expectations of editors and referees) is cargo-cult statistics. The bad incentives for individuals, departments, and disciplines are clear; negative consequences for science and society are expected.
Statistical software: power without wisdom
Statistical software enables and promotes cargo-cult statistics. Marketing and adoption of statistical software are driven by ease of use and the range of statistical routines the software implements. Offering complex and “modern” methods provides a competitive advantage. And some disciplines have in effect standardised on particular statistical software, often proprietary software.
Statistical software does not help you know what to compute, nor how to interpret the result. It does not offer to explain the assumptions behind methods, nor does it flag delicate or dubious assumptions. It does not warn you about multiplicity or p-hacking. It does not check whether you picked the hypothesis or analysis after looking at the data,47 nor track the number of analyses you tried before arriving at the one you sought to publish – another form of multiplicity.48-50 The more “powerful” and “user-friendly” the software is, the more it invites cargo-cult statistics.
This is hard to fix. Checks of residuals and similar tests cannot yield evidence that modelling assumptions are true – and running such checks makes the estimates and inferences conditional, which software generally does not take into account. In-built warnings could be used to remind the user of the assumptions, but these are unlikely to have much effect without serious changes to incentives. Indeed, if software offered such warnings, it might be seen as an irritant, and hence a competitive disadvantage to the vendor and the user, rather than an aid.
Scientific publishing and open science
Peer review can reinforce bad scientific and statistical practice. Indeed, journals may reject papers that use more reliable or more rigorous methods than the discipline is accustomed to, simply because the methods are unfamiliar. Conversely, some disciplines become enthralled with methodology du jour without careful vetting. Even the increased volume of research suggests that quality must suffer.51
There is structural moral hazard in the current scientific publishing system. Many turf battles are fought at the editorial level. Our own experience suggests that journals are reluctant to publish papers critical of work the journal published previously, or of work by scientists who are referees or editors for the journal.
Editorial control of prestigious journals confers indirect but substantial control of employment, research directions, research funding, and professional recognition. Editors and referees can keep competitors from being heard, funded, and hired. Nobel biologist Randy Shekman reports: “Young people tell me all the time, ‘If I don’t publish in CNS [a common acronym for Cell/Nature/Science, the most prestigious journals in biology], I won’t get a job’ … [Those journals] have a very big influence on where science goes.”52
Editorial policies may preclude authors from providing enough information for a reviewer (or reader) to check whether the results are correct, or even to check whether the figures and tables accurately reflect the underlying data. As a result, the editorial process simply cannot perform its intended quality-control function.
Academic research is often funded, at least in part, by taxes. Yet many scientists try to become rent seekers, keeping data and code resulting from public funding to themselves indefinitely, or until they feel they have exhausted its main value. This is morally murky. To “publish” the resulting research behind a paywall, inaccessible to the general public, is even more troubling. Scientific publishing is big business,52 and its interests are not those of science or scientists. Open data, open software, and open publication may provide better value for society53 and a better ethical foundation for science.
What can statisticians do?
Statisticians can help with important, controversial issues with immediate consequences for society.35 We can help fight power asymmetries in the use of evidence.54 We can stand up for the responsible use of statistics, even when that means taking personal risks.
We should be vocally critical of cargo-cult statistics, including where study design is ignored, where p-values, confidence intervals and posterior distributions are misused, and where probabilities are calculated under irrelevant, misleading assumptions. We should be critical even when the abuses involve politically charged issues, such as the social cost of climate change.55 If an authority treats estimates based on an ad hoc collection of related numerical models with unknown, potentially large systematic errors as if they were a random sample from a distribution centred at the parameter,56 we should object – whether or not we like the conclusion.
We can insist that “service” courses foster statistical thinking, deep understanding, and appropriate scepticism, rather than promulgating cargo-cult statistics. We can help empower individuals to appraise quantitative information critically – to be informed, effective citizens of the world. We also can help educate the media, which often reduces science to “infotainment” through inaccurate, sensationalised, truncated, and uncircumspect reporting. Journalists rarely ask basic questions about experimental design or data quality, report uncertainties, ask whether clinical endpoints were changed, try to read the underlying scientific report, check the scientific literature for conflicting results, etc. We can address this by teaching courses for journalists and editors.
When we appraise each other’s work in academia, we can ignore impact factors, citation counts, and the like: they do not measure importance, correctness, or quality.57 We can pay attention to the work itself, rather than the masthead of the journal in which it appeared, the press coverage it received, or the funding that supported it. We can insist on evidence that the work is correct – on reproducibility and replicability – rather than pretend that editors and referees can reliably vet research by proxy when the requisite evidence was not even submitted for scrutiny.58
We can decline to referee manuscripts that do not include enough information to tell whether they are correct.32, 58, 59 We can commit to working reproducibly,60 to publishing code and data,61 and generally to contributing to the intellectual commons. We can point out when studies change endpoints.62 We can decline to serve as an editor or referee for journals that profiteer or that enable scientists to be rent-seekers by publishing “results” without the underlying publicly funded evidence: data and code.
And we can be of service. Direct involvement of statisticians on the side of citizens in societal and environmental problems can help earn the justified trust of society. For instance, statisticians helped show that the erroneous use of zip codes to identify the geographic area of interest in the Flint, Michigan water pollution scandal made the water contamination problem disappear.63-65
Statistical election forensics66 has revealed electoral manipulation in countries such as Russia.67 Statistical “risk-limiting” audits, endorsed by the ASA,68 can provide assurance that election outcomes are correct.69 Such methods have been tested in California, Colorado, Ohio, and Denmark, and are required by law in Colorado and Rhode Island; other states have pending legislation. Statisticians and computer scientists developed methods and software; worked with election officials, legislators, and government agencies on logistics, laws, and regulations; and advocated with the public through op-eds and media appearances. Statisticians and mathematicians can help assess and combat gerrymandering,70 the practice of redrawing electoral districts to advantage a party unfairly.
Statisticians are pointing out biases inherent in “big data” and machine-learning approaches to social issues, such as predictive policing.71 They could also work with economists to monitor new forms of exploitation of intellectual labour now that new modes of working can be exploited in old ways.72, 73
We statisticians can support initiatives such as the Reproducibility Project,74 the Meta-research Innovation Center,75 the EQUATOR network, alltrials.net, retractionwatch.com, and others that aim to improve quality and ethics in science, and hold scientists accountable for sloppy, disingenuous, or fraudulent work.
And we can change how we work. We can recognise that software engineering is as important to modern data analysis as washing test tubes is to wet chemistry: we can develop better computational hygiene.60 And we can ensure that publicly funded research is public.76
In the 1660s, radical philosophers sought to understand and master the world by becoming scientific – creating science. In the 1970s, radical scientists sought to change the world by changing science.77 Perhaps that is now needed again.78
About the authors
Philip B. Stark is professor in the Department of Statistics and Associate Dean in the Division of Mathematical and Physical Sciences at the University of California, Berkeley.
Andrea Saltelli is an adjunct professor in the Centre for the Study of the Sciences and the Humanities at the University of Bergen, and a researcher with the Open Evidence Research group at the Universitat Oberta de Catalunya, Barcelona.
This is a fully referenced version of an article published in our August 2018 print edition.
- A. Benessia et al., The Rightful Place of Science: Science on the Verge. Consortium for Science, Policy, & Outcomes, 2016.
- A. Saltelli and S. Funtowicz, “What is science’s crisis really about?,” Futures, vol. 91, pp. 5–11, 2017.
- R. F. Harris, Rigor mortis : how sloppy science creates worthless cures, crushes hope, and wastes billions. Basic Books, 2017.
- C. G. Begley and J. P. A. Ioannidis, “Reproducibility in Science: Improving the Standard for Basic and Preclinical Research,” Circ. Res., vol. 116, no. 1, pp. 116–126, Jan. 2015.
- J. P. A. Ioannidis, J. Chen, R. Kodell, C. Haug, and J. Hoey, “Why Most Published Research Findings Are False,” PLoS Med., vol. 2, no. 8, p. e124, Aug. 2005.
- J. P. A. Ioannidis et al., “Why Most Clinical Research Is Not Useful,” PLOS Med., vol. 13, no. 6, p. e1002049, Jun. 2016.
- J. P. A. Ioannidis, T. D. Stanley, and H. Doucouliagos, “The Power of Bias in Economics Research,” Econ. J., vol. 127, pp. F236–F265, 2017.
- P. E. Smaldino and R. McElreath, “The natural selection of bad science,” R. Soc. Open Sci., vol. 3, no. 160384, 2016.
- J. R. Ravetz, Scientific knowledge and its social problems. Clarendon Press, 1971.
- S. Shapin, The scientific life : a moral history of a late modern vocation. University of Chicago Press, 2008.
- J. Ravetz, “Morals and manners in modern science,” Nature, vol. 457, no. 7230, pp. 662–663, Feb. 2009.
- C. Macilwain, “The elephant in the room we can’t ignore,” Nature, vol. 531, no. 7594, pp. 277–277, Mar. 2016.
- G. Gigerenzer and J. N. Marewski, “Surrogate Science,” J. Manage., vol. 41, no. 2, pp. 421–440, Feb. 2015.
- R. Merton, The sociology of science: Theoretical and empirical investigations. 1973.
- I. I. Mitroff, “Norms and Counter-Norms in a Select Group of the Apollo Moon Scientists: A Case Study of the Ambivalence of Scientists,” Am. Sociol. Rev., vol. 39, no. 4, p. 579, Aug. 1974.
- V. Bush, “Science, The Endless Frontier, A Report to the President by Vannevar Bush, Director of the Office of Scientific Research and Development,” 1945.
- A. E. Jinha, “Article 50 million: an estimate of the number of scholarly articles in existence,” Learn. Publ., vol. 23, no. 3, pp. 258–263, Jul. 2010.
- L. Bornmann and R. Mutz, “Growth rates of modern science: A bibliometric analysis based on the number of publications and cited references,” J. Assoc. Inf. Sci. Technol., vol. 66, no. 11, pp. 2215–2222, Nov. 2015.
- M. A. Edwards and S. Roy, “Academic Research in the 21st Century: Maintaining Scientific Integrity in a Climate of Perverse Incentives and Hypercompetition,” Environ. Eng. Sci., vol. 34, no. 1, pp. 51–61, Jan. 2017.
- E. Callaway, “Beat it, impact factor! Publishing elite turns against controversial metric,” Nature, vol. 535, no. 7611, pp. 210–211, Jul. 2016.
- J. P. A. Ioannidis, B. Forstman, I. Boutron, L. Yu, and J. Cook, “How to Make More Published Research True,” PLoS Med., vol. 11, no. 10, p. e1001747, Oct. 2014.
- J. Wilsdon, The metric tide : the independent review of the role of metrics in research assessment and management. Sage Publications, Ltd., 2016.
- D. J. de Solla Price, Little science, big science. Columbia University Press, 1963.
- P. Mirowski, Science-Mart, Privatizing American Science. Harvard University Press, 2011.
- J. P. A. Ioannidis, J. Chen, R. Kodell, C. Haug, and J. Hoey, “Why Most Published Research Findings Are False,” PLoS Med., vol. 2, no. 8, p. e124, Aug. 2005.
- M. R. Munafò et al., “A manifesto for reproducible science,” Nat. Hum. Behav., vol. 1, no. January, pp. 1–9, 2017.
- K. H. Jamieson, “Crisis or self-correction: Rethinking media narratives about the well-being of science,” Proc. Natl. Acad. Sci. U. S. A., vol. 115, no. 11, pp. 2620–2627, Mar. 2018.
- D. Fanelli, “Opinion: Is science really facing a reproducibility crisis, and do we need it to?,” Proc. Natl. Acad. Sci. U. S. A., vol. 115, no. 11, pp. 2628–2631, Mar. 2018.
- P. B. Stark, “Preface: Nullius in Verba,” in The Practice of Reproducible Research, J. Kitzes, D. Turek, and F. Deniz, Eds. Gitbook, 2018.
- C. E. Kearns, L. A. Schmidt, and S. A. Glantz, “Sugar Industry and Coronary Heart Disease Research,” JAMA Intern. Med., vol. 176, no. 11, p. 1680, Nov. 2016.
- A. McCook, “A new record: Major publisher retracting more than 100 studies from cancer journal over fake peer reviews,” Retraction Watch, 2017. [Online]. Available: http://retractionwatch.com/2017/04/20/new-record-major-publisher-retracting-100-studies-cancer-journal-fake-peer-reviews/. [Accessed: 11-Aug-2017].
- P. B. Stark, “Science is ‘show me,’ not ‘trust me’ – Berkeley Initiative for Transparency in the Social Sciences,” Berkeley Initiative for Transparency in the Social Sciences, 2015. [Online]. Available: http://www.bitss.org/2015/12/31/science-is-show-me-not-trust-me/. [Accessed: 11-Aug-2017].
- A. Hilbeck and H. Herren, “Millions Spent, No One Served: Who Is to Blame for the Failure of GMO Golden Rice?,” Independent Science News, 10-Aug-2016.
- G. K. Chesterton, Napoleon of Notting Hill. Open Road Media, 2015.
- A. Saltelli, “Young statistician, you shall live in adventurous times,” Significance, vol. 13, no. 6, pp. 38–41, Dec. 2016.
- R. P. (Richard P. Feynman, R. Leighton, and E. Hutchings, Surely you’re joking, Mr. Feynman! W.W. Norton, 1985.
- J. T. Leek and R. D. Peng, “P values are just the tip of the iceberg,” Nature, vol. 520, p. 612, 2015.
- R. L. Wasserstein and N. A. Lazar, “The ASA’s Statement on p -Values: Context, Process, and Purpose,” Am. Stat., vol. 70, no. 2, pp. 129–133, Apr. 2016.
- J. Leek, B. B. McShane, A. Gelman, D. Colquhoun, M. B. Nuijten, and S. N. Goodman, “Five ways to fix statistics,” Nature, vol. 551, pp. 557–559, 2017.
- J. Kadane and L. J.Wolfson, “Experiences in elicitation,” Journal of the Royal Statistical Society: Series D (The Statistician), vol. 47, no. 1, pp. 3-19, 1998.
- P. H. Garthwaite, J. B. Kadane, and A. O’Hagan, “Statistical Methods for Eliciting Probability Distributions,” J. Am. Stat. Assoc., vol. 100, no. 470, pp. 680–701, Jun. 2005.
- A. E. Lange et al., “Cosmological parameters from the first results of Boomerang,” Phys. Rev. D, vol. 63, no. 4, p. 042001, Jan. 2001.
- The ATLAS Collaboration, “Improved Luminosity Determination in pp Collisions at √ s = 7 TeV using the ATLAS Detector at the LHC,” 2012.
- D. Freedman, “Some issues in the foundation of statistics,” Found. Sci., vol. 1, no. 1, pp. 19–39, Mar. 1995.
- UC Berkeley Foundations of Data Science, “data-8.github.io.” [Online]. Available: http://data8.org/. [Accessed: 04-Apr-2018].
- A. Adhikari and J. DeNero, “Introduction · Computational and Inferential Thinking,” 2017. [Online]. Available: https://www.inferentialthinking.com/. [Accessed: 04-Apr-2018].
- N. L. Kerr, “HARKing: Hypothesizing After the Results are Known,” Personal. Soc. Psychol. Rev., vol. 2, no. 3, pp. 196–217, Aug. 1998.
- E. E. Leamer, “Sensitivity Analyses Would Help,” Am. Econ. Rev., vol. 75, no. 3, pp. 308–313, 1985.
- E. E. Leamer, “Tantalus on the Road to Asymptopia,” J. Econ. Perspect., vol. 24, no. 2, pp. 31–46, May 2010.
- P. Kennedy, A guide to econometrics. Blackwell Pub, 2008.
- H. Devlin, “‘Exaggerations’ threaten public trust in science, says leading statistician,” The Guardian, 28-Jun-2017.
- S. Buranyi, “Is the staggeringly profitable business of scientific publishing bad for science?,” The Guardian, 27-Jun-2017.
- M. Eisen, “The Past, Present and Future of Scholarly Publishing,” Blog: it is NOT junk, 28-Mar-2013.
- A. Saltelli and M. Giampietro, “What is wrong with evidence based policy, and how can it be improved?,” Futures, vol. 91, pp. 62–71, Feb. 2017.
- A. Saltelli, P. B. Stark, W. Becker, and P. Stano, “Climate Models as Economic Gudes – Scientific Challenge of Quixotic Quest?,” no. Issues in Science and Technology, pp. 1–8, 2015.
- R. Knutti et al., “Challenges in Combining Projections from Multiple Climate Models,” J. Clim., vol. 23, no. 10, pp. 2739–2758, May 2010.
- D. Colquhoun and A. Plested, “Scientists don’t count: why you should ignore altmetrics and other bibliometric nightmares,” www.dcscience.net, 16-Jan-2014.
- P. B. Stark, “Before reproducibility must come preproducibility,” Nature, vol. 557, no. 7707, pp. 613–613, May 2018.
- R. D. Morey et al., “The Peer Reviewers’ Openness Initiative: incentivizing open research practices through peer review,” R. Soc. Open Sci., vol. 3, no. 1, 2016.
- J. Kitzes, D. Turek, and F. Deniz, The Practice of Reproducible Research Case Studies and Lessons from the Data-Intensive Sciences. University of California Press, 2018.
- “Open Science Framework.” [Online]. Available: https://osf.io/.
- A. Matthees, T. Kindlon, C. Maryhew, P. Stark, and B. Levin, “A preliminary analysis of ‘recovery’ from chronic fatigue syndrome in the PACE trial using individual participant data,” Virology Blog, 21-Sep-2016.
- “flintwaterstudy.org/.” [Online]. Available: http://flintwaterstudy.org/.
- D. Hohn, “Flint’s Water Crisis and the ‘Troublemaker’ Scientist,” New York Times Magazine, 16-Aug-2016.
- R. C. Sadler, “How ZIP codes nearly masked the lead problem in Flint,” The Conversation, 2016.
- A. Hicken and W. R. Mebane, “A Guide to Election Forensics,” 2015.
- K. Kalinin and W. R. Mebane, “When the Russians fake their election results, they may be giving us the statistical finger,” The Washington Post, 11-Jan-2017.
- ASA, “American Statistical Association Statement on Risk-Limiting Post-Election Audits, 4/17/10,” 2010. [Online]. Available: http://www.amstat.org/asa/files/pdfs/POL-Risk-Limiting_Endorsement.pdf.
- M. Lindeman and P. B. Stark, “A Gentle Introduction to Risk-Limiting Audits,” IEEE Secur. Priv., vol. 10, no. 5, pp. 42–49, Sep. 2012.
- C. Arnold, “The mathematicians who want to save democracy,” Nature, vol. 546, no. 7657, pp. 200–202, Jun. 2017.
- K. Lum and W. Isaac, “To predict and serve?,” Significance, vol. 13, no. 5, pp. 14–19, Oct. 2016.
- T. Scholz, Digital labor : the Internet as playground and factory. Routledge, 2013.
- N. Heath, “Inside Amazon’s clickworker platform: How half a million people are being paid pennies to train AI,” Dec-2016.
- Open Science Collaboration (OSC), “Estimating the reproducibility of psychological science,” Science (80-. ), vol. 349, no. 6251, aac4716, 2015.
- “Meta-Research Innovation Center at Stanford (METRICS).” [Online]. Available: https://metrics.stanford.edu/.
- M. Taylor, “Hiding your research behind a paywall is immoral,” The Guardian, 17-Jan-2013.
- Alice Bell, “How Radical 70s Scientists Tried to Change the World,” GIZMODO, 2015. [Online]. Available: http://gizmodo.com/how-radical-70s-scientists-tried-to-change-the-world-1681987399. [Accessed: 25-Aug-2017].
- A. Saltelli and S. O. Funtowicz, “To tackle the post-truth world, science must reform itself,” The Conversation, 27-Jan-2017.