False positive and false negative rates The
false positive rate (FPR) is the proportion of all negatives that still yield positive test outcomes, i.e., the
conditional probability of a positive test result given an event that was not present. The false positive rate is equal to the
significance level. The
specificity of the test is equal to
1 minus the false positive rate. In
statistical hypothesis testing, this fraction is given the Greek letter
α, and 1 −
α is defined as the specificity of the test. Increasing the specificity of the test lowers the probability of type I errors, but may raise the probability of type II errors (false negatives that reject the
alternative hypothesis when it is true). Complementarily, the '''''' (FNR) is the proportion of positives which yield negative test outcomes with the test, i.e., the conditional probability of a negative test result given that the condition being looked for is present. In
statistical hypothesis testing, this fraction is given the letter
β. The "
power" (or the "
sensitivity") of the test is equal to 1 −
β.
Ambiguity in the definition of false positive rate The term false discovery rate (FDR) was used by Colquhoun (2014) to mean the probability that a "significant" result was a false positive. Later Colquhoun (2017) used the term false positive risk (FPR) for the same quantity, to avoid confusion with the term FDR as used by people who work on
multiple comparisons. Corrections for multiple comparisons aim only to correct the type I error rate, so the result is a (corrected)
p-value. Thus they are susceptible to the same misinterpretation as any other
p-value. The false positive risk is always higher, often much higher, than the
p-value. Because of the ambiguity of notation in this field, it is essential to look at the definition in every paper. The hazards of reliance on
p-values was emphasized in Colquhoun (2017) that every
p-value should be accompanied by the prior probability of there being a real effect that it would be necessary to assume in order to achieve a false positive risk of 5%. For example, if we observe
p = 0.05 in a single experiment, we would have to be 87% certain that there as a real effect before the experiment was done to achieve a false positive risk of 5%.
Receiver operating characteristic The article "
Receiver operating characteristic" discusses parameters in statistical signal processing based on ratios of errors of various types. == See also ==