Library / Confusion Over Measures of Evidence (p's) Versus Errors (alpha's) In Classical Statistical Testing


Raymond Hubbard, Mar'\ia Jes'us Bayarri “Confusion over measures of evidence (p’s) versus errors (alpha’s) in classical statistical testing” (2003) // The American Statistician. Publisher: Taylor & Francis. Vol. 57. No 3. Pp. 171–178. DOI: 10.1198/0003130031856


  title = {Confusion over measures of evidence (p's) versus errors (alpha's) in classical statistical testing},
  author = {Hubbard, Raymond and Bayarri, Mar\'\ia Jes\'us},
  journal = {The American Statistician},
  volume = {57},
  number = {3},
  pages = {171--178},
  year = {2003},
  publisher = {Taylor \& Francis},
  doi = {10.1198/0003130031856}

Quotes (1)

Neyman–pearson vs. Fisher

Confusion surrounding the reporting and interpretation of results of classical statistical tests is widespread among applied researchers, most of whom erroneously believe that such tests are prescribed by a single coherent theory of statistical inference. This is not the case: Classical statistical testing is an anonymous hybrid of the competing and frequently contradictory approaches formulated by R. A. Fisher on the one hand, and Jerzy Neyman and Egon Pearson on the other. In particular, there is a widespread failure to appreciate the incompatibility of Fisher’s evidential p value with the Type I error rate, α, of Neyman–Pearson statistical orthodoxy. The distinction between evidence (p’s) and error (α ’s) is not trivial. Instead, it reflects the fundamental differences between Fisher’s ideas on significance testing and inductive inference, and Neyman–Pearson’s views on hypothesis testing and inductive behavior. The emphasis of the article is to expose this incompatibility, but we also briefly note a possible reconciliation.