Some widely used methods of conventional statistics, for example many
significance tests, are not consistent with the likelihood principle. Let us briefly consider some of the arguments for and against the likelihood principle.
The original Birnbaum argument According to Giere (1977), Birnbaum rejected p. 263). Birnbaum's original argument for the likelihood principle has also been disputed by other statisticians including
Akaike, Evans and philosophers of science, including
Deborah Mayo.
Dawid points out fundamental differences between Mayo's and Birnbaum's definitions of the conditionality principle, arguing Birnbaum's argument cannot be so readily dismissed. A new proof of the likelihood principle has been provided by Gandenberger that addresses some of the counterarguments to the original proof.
Experimental design arguments on the likelihood principle Unrealized events play a role in some common statistical methods. For example, the result of a
significance test depends on the
-value, the probability of a result as extreme or more extreme than the observation, and that probability may depend on the design of the experiment. To the extent that the likelihood principle is accepted, such methods are therefore denied. Some classical significance tests are not based on the likelihood. The following are a simple and more complicated example of those, using a commonly cited example called
the optional stopping problem. ;Example 1 – simple version: Suppose I tell you that I tossed a coin 12 times and in the process observed 3 heads. You might make some inference about the probability of heads and whether the coin was fair. Suppose now I tell that I tossed the coin
until I observed 3 heads, and I tossed it 12 times. Will you now make some different inference? The likelihood function is the same in both cases: It is proportional to :p^3 (1-p)^9 ~. So according to the
likelihood principle, in either case the inference should be the same. ;Example 2 – a more elaborated version of the same statistics: Suppose a number of scientists are assessing the probability of a certain outcome (which we shall call 'success') in experimental trials. Conventional wisdom suggests that if there is no bias towards success or failure then the success probability would be one half. Adam, a scientist, conducted 12 trials and obtains 3 successes and 9 failures.
One of those successes was the 12th and last observation. Then Adam left the lab. Bill, a colleague in the same lab, continued Adam's work and published Adam's results, along with a significance test. He tested the
null hypothesis that , the success probability, is equal to a half, versus . If we ignore the information that the third success was the 12th and last observation, the probability of the observed result that out of 12 trials 3 or something fewer (i.e. more extreme) were successes, if is true, is :\left[{12 \choose 3}+{12 \choose 2}+{12 \choose 1}+{12 \choose 0}\right]\left({1 \over 2}\right)^{12} ~ , which is . Thus the null hypothesis is not rejected at the 5% significance level if we ignore the knowledge that the third success was the 12th result. However observe that this first calculation also includes 12 token long sequences that end in tails contrary to the problem statement! If we redo this calculation we realize the likelihood according to the null hypothesis must be the probability of a fair coin landing 2 or fewer heads on 11 trials multiplied with the probability of the fair coin landing a head for the 12th trial: :\left[{11 \choose 2}+{11 \choose 1}+{11 \choose 0}\right]\left({1 \over 2}\right)^{11}{1 \over 2} ~ , which is . Now the result
is statistically significant at the level. Charlotte, another scientist, reads Bill's paper and writes a letter, saying that it is possible that Adam kept trying until he obtained 3 successes, in which case the probability of needing to conduct 12 or more experiments is given by :\left[{11 \choose 2}+{11 \choose 1}+{11 \choose 0}\right]\left({1 \over 2}\right)^{11}{1 \over 2} ~ , which is . Now the result
is statistically significant at the level. Note that there is no contradiction between the latter two correct analyses; both computations are correct, and result in the same p-value. To these scientists, whether a result is significant or not does not depend on the design of the experiment, but does on the likelihood (in the sense of the likelihood function) of the parameter value being . ;Summary of the illustrated issues: Results of this kind are considered by some as arguments against the likelihood principle. For others it exemplifies the value of the likelihood principle and is an argument against significance tests. Similar themes appear when comparing
Fisher's exact test with
Pearson's chi-squared test.
The voltmeter story An argument in favor of the likelihood principle is given by Edwards in his book
Likelihood. He cites the following story from J.W. Pratt, slightly condensed here. Note that the likelihood function depends only on what actually happened, and not on what
could have happened. : An engineer draws a random sample of electron tubes and measures their voltages. The measurements range from 75 to 99 Volts. A statistician computes the sample mean and a confidence interval for the true mean. Later the statistician discovers that the voltmeter reads only as far as 100 Volts, so technically, the population appears to be “
censored”. If the statistician is orthodox this necessitates a new analysis. : However, the engineer says he has another meter reading to 1000 Volts, which he would have used if any voltage had been over 100. This is a relief to the statistician, because it means the population was effectively uncensored after all. But later, the statistician discovers that the second meter had not been working when the measurements were taken. The engineer informs the statistician that he would not have held up the original measurements until the second meter was fixed, and the statistician informs him that new measurements are required. The engineer is astounded. “''Next you'll be asking about my oscilloscope!''” ;Throwback to
Example 2 in the prior section: This story can be translated to Adam's stopping rule above, as follows: Adam stopped immediately after 3 successes, because his boss Bill had instructed him to do so. After the publication of the statistical analysis by Bill, Adam realizes that he has missed a later instruction from Bill to instead conduct 12 trials, and that Bill's paper is based on this second instruction. Adam is very glad that he got his 3 successes after exactly 12 trials, and explains to his friend Charlotte that by coincidence he executed the second instruction. Later, Adam is astonished to hear about Charlotte's letter, explaining that
now the result is significant. == See also ==