HOME

TheInfoList



OR:

In null-hypothesis significance testing, the ''p''-value is the probability of obtaining test results at least as extreme as the result actually observed, under the assumption that the
null hypothesis In scientific research, the null hypothesis (often denoted ''H''0) is the claim that no difference or relationship exists between two sets of data or variables being analyzed. The null hypothesis is that any experimentally observed difference is ...
is correct. A very small ''p''-value means that such an extreme observed outcome would be very unlikely under the null hypothesis. Reporting ''p''-values of statistical tests is common practice in
academic publications Academic publishing is the subfield of publishing which distributes academic research and scholarship. Most academic work is published in academic journal articles, books or theses. The part of academic written output that is not formally publ ...
of many quantitative fields. Since the precise meaning of ''p''-value is hard to grasp, misuse is widespread and has been a major topic in
metascience Metascience (also known as meta-research) is the use of scientific methodology to study science itself. Metascience seeks to increase the quality of scientific research while reducing inefficiency. It is also known as "''research on research''" ...
.


Basic concepts

In statistics, every conjecture concerning the unknown probability distribution of a collection of random variables representing the observed data X in some study is called a ''statistical hypothesis''. If we state one hypothesis only and the aim of the statistical test is to see whether this hypothesis is tenable, but not to investigate other specific hypotheses, then such a test is called a null hypothesis test. As our statistical hypothesis will, by definition, state some property of the distribution, the
null hypothesis In scientific research, the null hypothesis (often denoted ''H''0) is the claim that no difference or relationship exists between two sets of data or variables being analyzed. The null hypothesis is that any experimentally observed difference is ...
is the default hypothesis under which that property does not exist. The null hypothesis is typically that some parameter (such as a correlation or a difference between means) in the populations of interest is zero. Note that our hypothesis might specify the probability distribution of X precisely, or it might only specify that it belongs to some class of distributions. Often, we reduce the data to a single numerical statistic, e.g., T, whose marginal probability distribution is closely connected to a main question of interest in the study. The ''p''-value is used in the context of null hypothesis testing in order to quantify the
statistical significance In statistical hypothesis testing, a result has statistical significance when it is very unlikely to have occurred given the null hypothesis (simply by chance alone). More precisely, a study's defined significance level, denoted by \alpha, is the p ...
of a result, the result being the observed value of the chosen statistic T. The lower the ''p''-value is, the lower the probability of getting that result if the null hypothesis were true. A result is said to be ''statistically significant'' if it allows us to reject the null hypothesis. All other things being equal, smaller p-values are taken as stronger evidence against the null hypothesis. Loosely speaking, rejection of the null hypothesis implies that there is sufficient evidence against it. As a particular example, if a null hypothesis states that a certain summary statistic T follows the standard
normal distribution In statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is : f(x) = \frac e^ The parameter \mu ...
N(0,1), then the rejection of this null hypothesis could mean that (i) the mean of T is not 0, or (ii) the
variance In probability theory and statistics, variance is the expectation of the squared deviation of a random variable from its population mean or sample mean. Variance is a measure of dispersion, meaning it is a measure of how far a set of numbe ...
of T is not 1, or (iii) T is not normally distributed. Different tests of the same null hypothesis would be more or less sensitive to different alternatives. However, even if we do manage to reject the null hypothesis for all 3 alternatives, and even if we know the distribution is normal and variance is 1, the null hypothesis test does not tell us which non-zero values of the mean are now most plausible. The more independent observations from the same probability distribution one has, the more accurate the test will be, and the higher the precision with which one will be able to determine the mean value and show that it is not equal to zero; but this will also increase the importance of evaluating the real-world or scientific relevance of this deviation.


Definition and interpretation


Definition


Probability of obtaining a real-valued test statistic at least as extreme as the one actually obtained

Consider an observed test-statistic t from unknown distribution T. Then the ''p''-value p is what the prior probability would be of observing a test-statistic value at least as "extreme" as t if null hypothesis H_0 were true. That is: * p = \Pr(T \geq t \mid H_0) for a one-sided right-tail test, * p = \Pr(T \leq t \mid H_0) for a one-sided left-tail test, * p = 2\min\ for a two-sided test. If the distribution of T is symmetric about zero, then p =\Pr(, T, \geq , t, \mid H_0)


Interpretations


p-value as the statistic for performing significance tests

In a significance test, the null hypothesis H_0 is rejected if the ''p''-value is less than or equal to a predefined threshold value \alpha, which is referred to as the alpha level or
significance level In statistical hypothesis testing, a result has statistical significance when it is very unlikely to have occurred given the null hypothesis (simply by chance alone). More precisely, a study's defined significance level, denoted by \alpha, is the p ...
. \alpha is not derived from the data, but rather is set by the researcher before examining the data. \alpha is commonly set to 0.05, though lower alpha levels are sometimes used. The ''p''-value is a function of the chosen test statistic T and is therefore a
random variable A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. It is a mapping or a function from possible outcomes (e.g., the po ...
. If the null hypothesis fixes the probability distribution of T precisely, and if that distribution is continuous, then when the null-hypothesis is true, the p-value is uniformly distributed between 0 and 1. Thus, the ''p''-value is not fixed. If the same test is repeated independently with fresh data, one will typically obtain a different ''p''-value in each iteration. If the null-hypothesis is composite, or the distribution of the statistic is discrete, the probability of obtaining a ''p''-value less than or equal to any number between 0 and 1 is less than or equal to that number, if the null-hypothesis is true. It remains the case that very small values are relatively unlikely if the null-hypothesis is true, and that a significance test at level \alpha is obtained by rejecting the null-hypothesis if the significance level is less than or equal to \alpha. Different ''p''-values based on independent sets of data can be combined, for instance using Fisher's combined probability test.


Distribution

When the null hypothesis is true, if it takes the form H_0: \theta = \theta_0, and the underlying random variable is continuous, then the
probability distribution In probability theory and statistics, a probability distribution is the mathematical function that gives the probabilities of occurrence of different possible outcomes for an experiment. It is a mathematical description of a random phenomenon ...
of the ''p''-value is
uniform A uniform is a variety of clothing worn by members of an organization while participating in that organization's activity. Modern uniforms are most often worn by armed forces and paramilitary organizations such as police, emergency services, ...
on the interval ,1 By contrast, if the alternative hypothesis is true, the distribution is dependent on sample size and the true value of the parameter being studied. The distribution of ''p''-values for a group of studies is sometimes called a ''p''-curve. A ''p''-curve can be used to assess the reliability of scientific literature, such as by detecting publication bias or ''p''-hacking.


For composite hypothesis

In parametric hypothesis testing problems, a ''simple or point hypothesis'' refers to a hypothesis where the parameter's value is assumed to be a single number. In contrast, in a ''composite hypothesis'' the parameter's value is given by a set of numbers. For example, when testing the null hypothesis that a distribution is normal with a mean less than or equal to zero against the alternative that the mean is greater than zero (variance known), the null hypothesis does not specify the probability distribution of the appropriate test statistic. In the just mentioned example that would be the ''Z''-statistic belonging to the one-sided one-sample ''Z''-test. For each possible value of the theoretical mean, the ''Z''-test statistic has a different probability distribution. In these circumstances (the case of a so-called composite null hypothesis) the ''p''-value is defined by taking the least favourable null-hypothesis case, which is typically on the border between null and alternative. This definition ensures the complementarity of p-values and alpha-levels. If we set the significance level alpha to 0.05, and only reject the null hypothesis if the p-value is less than or equal to 0.05, then our hypothesis test will indeed have significance level (maximal type 1 error rate) 0.05. As Neyman wrote: “The error that a practising statistician would consider the more important to avoid (which is a subjective judgment) is called the error of the first kind. The first demand of the mathematical theory is to deduce such test criteria as would ensure that the probability of committing an error of the first kind would equal (or approximately equal, or not exceed) a preassigned number α, such as α = 0.05 or 0.01, etc. This number is called the level of significance”; Neyman 1976, p. 161 in "The Emergence of Mathematical Statistics: A Historical Sketch with Particular Reference to the United States","On the History of Statistics and Probability", ed. D.B. Owen, New York: Marcel Dekker, pp. 149-193. See also "Confusion Over Measures of Evidence (p's) Versus Errors (a's) in Classical Statistical Testing", Raymond Hubbard and M. J. Bayarri, The American Statistician, August 2003, Vol. 57, No 3, 171--182 (with discussion). For a concise modern statement see Chapter 10 of "All of Statistics: A Concise Course in Statistical Inference", Springer; 1st Corrected ed. 20 edition (September 17, 2004). Larry Wasserman.


Usage

The ''p''-value is widely used in
statistical hypothesis testing A statistical hypothesis test is a method of statistical inference used to decide whether the data at hand sufficiently support a particular hypothesis. Hypothesis testing allows us to make probabilistic statements about population parameters. ...
, specifically in null hypothesis significance testing. In this method, before conducting the study, one first chooses a model (the
null hypothesis In scientific research, the null hypothesis (often denoted ''H''0) is the claim that no difference or relationship exists between two sets of data or variables being analyzed. The null hypothesis is that any experimentally observed difference is ...
) and the alpha level ''α'' (most commonly .05). After analyzing the data, if the ''p''-value is less than ''α'', that is taken to mean that the observed data is sufficiently inconsistent with the
null hypothesis In scientific research, the null hypothesis (often denoted ''H''0) is the claim that no difference or relationship exists between two sets of data or variables being analyzed. The null hypothesis is that any experimentally observed difference is ...
for the null hypothesis to be rejected. However, that does not prove that the null hypothesis is false. The ''p''-value does not, in itself, establish probabilities of hypotheses. Rather, it is a tool for deciding whether to reject the null hypothesis.


Misuse

According to the ASA, there is widespread agreement that ''p''-values are often misused and misinterpreted. One practice that has been particularly criticized is accepting the alternative hypothesis for any ''p''-value nominally less than .05 without other supporting evidence. Although ''p''-values are helpful in assessing how incompatible the data are with a specified statistical model, contextual factors must also be considered, such as "the design of a study, the quality of the measurements, the external evidence for the phenomenon under study, and the validity of assumptions that underlie the data analysis". Another concern is that the ''p''-value is often misunderstood as being the probability that the null hypothesis is true. Some statisticians have proposed abandoning ''p''-values and focusing more on other inferential statistics, such as
confidence intervals In frequentist statistics, a confidence interval (CI) is a range of estimates for an unknown parameter. A confidence interval is computed at a designated ''confidence level''; the 95% confidence level is most common, but other levels, such as 9 ...
, likelihood ratios, or Bayes factors, but there is heated debate on the feasibility of these alternatives. Others have suggested to remove fixed significance thresholds and to interpret ''p''-values as continuous indices of the strength of evidence against the null hypothesis. Yet others suggested to report alongside p-values the prior probability of a real effect that would be required to obtain a false positive risk (i.e. the probability that there is no real effect) below a pre specified threshold (e.g. 5%).


Calculation

Usually, T is a
test statistic A test statistic is a statistic (a quantity derived from the sample) used in statistical hypothesis testing.Berger, R. L.; Casella, G. (2001). ''Statistical Inference'', Duxbury Press, Second Edition (p.374) A hypothesis test is typically specifi ...
. A test statistic is the output of a scalar function of all the observations. This statistic provides a single number, such as a t-statistic or an F-statistic. As such, the test statistic follows a distribution determined by the function used to define that test statistic and the distribution of the input observational data. For the important case in which the data are hypothesized to be a random sample from a normal distribution, depending on the nature of the test statistic and the hypotheses of interest about its distribution, different null hypothesis tests have been developed. Some such tests are the
z-test A ''Z''-test is any statistical test for which the distribution of the test statistic under the null hypothesis can be approximated by a normal distribution. Z-tests test the mean of a distribution. For each significance level in the confide ...
for hypotheses concerning the mean of a
normal distribution In statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is : f(x) = \frac e^ The parameter \mu ...
with known variance, the
t-test A ''t''-test is any statistical hypothesis test in which the test statistic follows a Student's ''t''-distribution under the null hypothesis. It is most commonly applied when the test statistic would follow a normal distribution if the value of ...
based on
Student's t-distribution In probability and statistics, Student's ''t''-distribution (or simply the ''t''-distribution) is any member of a family of continuous probability distributions that arise when estimating the mean of a normally distributed population in situ ...
of a suitable statistic for hypotheses concerning the mean of a normal distribution when the variance is unknown, the
F-test An ''F''-test is any statistical test in which the test statistic has an ''F''-distribution under the null hypothesis. It is most often used when comparing statistical models that have been fitted to a data set, in order to identify the model ...
based on the
F-distribution In probability theory and statistics, the ''F''-distribution or F-ratio, also known as Snedecor's ''F'' distribution or the Fisher–Snedecor distribution (after Ronald Fisher and George W. Snedecor) is a continuous probability distribution ...
of yet another statistic for hypotheses concerning the variance. For data of other nature, for instance categorical (discrete) data, test statistics might be constructed whose null hypothesis distribution is based on normal approximations to appropriate statistics obtained by invoking the
central limit theorem In probability theory, the central limit theorem (CLT) establishes that, in many situations, when independent random variables are summed up, their properly normalized sum tends toward a normal distribution even if the original variables themsel ...
for large samples, as in the case of
Pearson's chi-squared test Pearson's chi-squared test (\chi^2) is a statistical test applied to sets of categorical data to evaluate how likely it is that any observed difference between the sets arose by chance. It is the most widely used of many chi-squared tests (e.g. ...
. Thus computing a ''p''-value requires a null hypothesis, a test statistic (together with deciding whether the researcher is performing a
one-tailed test In statistical significance testing, a one-tailed test and a two-tailed test are alternative ways of computing the statistical significance of a parameter inferred from a data set, in terms of a test statistic. A two-tailed test is appropriate if ...
or a
two-tailed test In statistical significance testing, a one-tailed test and a two-tailed test are alternative ways of computing the statistical significance of a parameter inferred from a data set, in terms of a test statistic. A two-tailed test is appropriate i ...
), and data. Even though computing the test statistic on given data may be easy, computing the sampling distribution under the null hypothesis, and then computing its
cumulative distribution function In probability theory and statistics, the cumulative distribution function (CDF) of a real-valued random variable X, or just distribution function of X, evaluated at x, is the probability that X will take a value less than or equal to x. Eve ...
(CDF) is often a difficult problem. Today, this computation is done using statistical software, often via numeric methods (rather than exact formulae), but, in the early and mid 20th century, this was instead done via tables of values, and one interpolated or extrapolated ''p''-values from these discrete values. Rather than using a table of ''p''-values, Fisher instead inverted the CDF, publishing a list of values of the test statistic for given fixed ''p''-values; this corresponds to computing the
Quantile function In probability and statistics, the quantile function, associated with a probability distribution of a random variable, specifies the value of the random variable such that the probability of the variable being less than or equal to that value equ ...
(inverse CDF).


Example


Testing the fairness of coin

As an example of a statistical test, an experiment is performed to determine whether a
coin flip Coin flipping, coin tossing, or heads or tails is the practice of throwing a coin in the air and checking which side is showing when it lands, in order to choose between two alternatives, heads or tails, sometimes used to resolve a dispute betwe ...
is
fair A fair (archaic: faire or fayre) is a gathering of people for a variety of entertainment or commercial activities. Fairs are typically temporary with scheduled times lasting from an afternoon to several weeks. Types Variations of fairs incl ...
(equal chance of landing heads or tails) or unfairly biased (one outcome being more likely than the other). Suppose that the experimental results show the coin turning up heads 14 times out of 20 total flips. The full data X would be a sequence of twenty times the symbol "H" or "T". The statistic on which one might focus could be the total number T of heads. The null hypothesis is that the coin is fair, and coin tosses are independent of one another. If a right-tailed test is considered, which would be the case if one is actually interested in the possibility that the coin is biased towards falling heads, then the ''p''-value of this result is the chance of a fair coin landing on heads ''at least'' 14 times out of 20 flips. That probability can be computed from
binomial coefficient In mathematics, the binomial coefficients are the positive integers that occur as coefficients in the binomial theorem. Commonly, a binomial coefficient is indexed by a pair of integers and is written \tbinom. It is the coefficient of the t ...
s as : \begin & \Pr(14\text) + \Pr(15\text) + \cdots + \Pr(20\text) \\ & = \frac \left \binom + \binom + \cdots + \binom \right= \frac \approx 0.058 \end This probability is the ''p''-value, considering only extreme results that favor heads. This is called a
one-tailed test In statistical significance testing, a one-tailed test and a two-tailed test are alternative ways of computing the statistical significance of a parameter inferred from a data set, in terms of a test statistic. A two-tailed test is appropriate if ...
. However, one might be interested in deviations in either direction, favoring either heads or tails. The two-tailed ''p''-value, which considers deviations favoring either heads or tails, may instead be calculated. As the
binomial distribution In probability theory and statistics, the binomial distribution with parameters ''n'' and ''p'' is the discrete probability distribution of the number of successes in a sequence of ''n'' independent experiments, each asking a yes–no ques ...
is symmetrical for a fair coin, the two-sided ''p''-value is simply twice the above calculated single-sided ''p''-value: the two-sided ''p''-value is 0.115. In the above example: * Null hypothesis (H0): The coin is fair, with Pr(heads) = 0.5 * Test statistic: Number of heads * Alpha level (designated threshold of significance): 0.05 * Observation O: 14 heads out of 20 flips; and * Two-tailed ''p''-value of observation O given H0 = 2 × min(Pr(no. of heads ≥ 14 heads), Pr(no. of heads ≤ 14 heads)) = 2 × min(0.058, 0.978) = 2*0.058 = 0.115. Note that the Pr (no. of heads ≤ 14 heads) = 1 - Pr(no. of heads ≥ 14 heads) + Pr (no. of head = 14) = 1 - 0.058 + 0.036 = 0.978; however, the symmetry of this binomial distribution makes it an unnecessary computation to find the smaller of the two probabilities. Here, the calculated ''p''-value exceeds .05, meaning that the data falls within the range of what would happen 95% of the time were the coin in fact fair. Hence, the null hypothesis is not rejected at the .05 level. However, had one more head been obtained, the resulting ''p''-value (two-tailed) would have been 0.0414 (4.14%), in which case the null hypothesis would be rejected at the .05 level.


Multistage experiment design

The difference between the two meanings of "extreme" appear when we consider a multistage experiment for testing the fairness of the coin. Suppose we design the experiment as follows: * Flip the coin twice. If both comes up heads or tails, end the experiment. * Else, flip the coin 4 more times. This experiment has 7 types of outcomes: 2 heads, 2 tails, 5 heads 1 tail..., 1 head 5 tails. We now calculate the p-value of the "3 heads 3 tails" outcome . If we use the test statistic \frac, then under the null hypothesis is exactly 1 for two-sided p-value, and exactly \frac for one-sided left-tail p-value, and same for one-sided right-tail p-value. If we consider every outcome that has equal or lower probability than "3 heads 3 tails" as "at least as extreme", then the p-value is exactly \frac 1 2. However, suppose we have planned to simply flip the coin 6 times no matter what happens, then the second definition of p-value would mean that the p-value of "3 heads 3 tails" is exactly 1. Thus, the "at least as extreme" definition of p-value is deeply contextual, and depend on what the experimenter ''planned'' to do even in situations that did not occur.


History

''P''-value computations date back to the 1700s, where they were computed for the
human sex ratio In anthropology and demography, the human sex ratio is the ratio of males to females in a population. Like most sexual species, the sex ratio in humans is close to 1:1. In humans, the natural ratio at birth between males and females is sligh ...
at birth, and used to compute statistical significance compared to the null hypothesis of equal probability of male and female births.
John Arbuthnot John Arbuthnot FRS (''baptised'' 29 April 1667 – 27 February 1735), often known simply as Dr Arbuthnot, was a Scottish physician, satirist and polymath in London. He is best remembered for his contributions to mathematics, his members ...
studied this question in 1710, and examined birth records in London for each of the 82 years from 1629 to 1710. In every year, the number of males born in London exceeded the number of females. Considering more male or more female births as equally likely, the probability of the observed outcome is 1/282, or about 1 in 4,836,000,000,000,000,000,000,000; in modern terms, the ''p''-value. This is vanishingly small, leading Arbuthnot that this was not due to chance, but to divine providence: "From whence it follows, that it is Art, not Chance, that governs." In modern terms, he rejected the null hypothesis of equally likely male and female births at the ''p'' = 1/282 significance level. This and other work by Arbuthnot is credited as "… the first use of significance tests …" the first example of reasoning about statistical significance, and "… perhaps the first published report of a nonparametric test …", specifically the
sign test The sign test is a statistical method to test for consistent differences between pairs of observations, such as the weight of subjects before and after treatment. Given pairs of observations (such as weight pre- and post-treatment) for each subject ...
; see details at . The same question was later addressed by
Pierre-Simon Laplace Pierre-Simon, marquis de Laplace (; ; 23 March 1749 – 5 March 1827) was a French scholar and polymath whose work was important to the development of engineering, mathematics, statistics, physics, astronomy, and philosophy. He summarize ...
, who instead used a ''parametric'' test, modeling the number of male births with a
binomial distribution In probability theory and statistics, the binomial distribution with parameters ''n'' and ''p'' is the discrete probability distribution of the number of successes in a sequence of ''n'' independent experiments, each asking a yes–no ques ...
: The ''p''-value was first formally introduced by
Karl Pearson Karl Pearson (; born Carl Pearson; 27 March 1857 – 27 April 1936) was an English mathematician and biostatistician. He has been credited with establishing the discipline of mathematical statistics. He founded the world's first university st ...
, in his
Pearson's chi-squared test Pearson's chi-squared test (\chi^2) is a statistical test applied to sets of categorical data to evaluate how likely it is that any observed difference between the sets arose by chance. It is the most widely used of many chi-squared tests (e.g. ...
, using the
chi-squared distribution In probability theory and statistics, the chi-squared distribution (also chi-square or \chi^2-distribution) with k degrees of freedom is the distribution of a sum of the squares of k independent standard normal random variables. The chi-squar ...
and notated as capital P. The ''p''-values for the
chi-squared distribution In probability theory and statistics, the chi-squared distribution (also chi-square or \chi^2-distribution) with k degrees of freedom is the distribution of a sum of the squares of k independent standard normal random variables. The chi-squar ...
(for various values of ''χ''2 and degrees of freedom), now notated as ''P,'' were calculated in , collected in . The use of the ''p''-value in statistics was popularized by
Ronald Fisher Sir Ronald Aylmer Fisher (17 February 1890 – 29 July 1962) was a British polymath who was active as a mathematician, statistician, biologist, geneticist, and academic. For his work in statistics, he has been described as "a genius who ...
, and it plays a central role in his approach to the subject. In his influential book ''
Statistical Methods for Research Workers ''Statistical Methods for Research Workers'' is a classic book on statistics, written by the statistician R. A. Fisher. It is considered by some to be one of the 20th century's most influential books on statistical methods, together with his ''The ...
'' (1925), Fisher proposed the level ''p'' = 0.05, or a 1 in 20 chance of being exceeded by chance, as a limit for
statistical significance In statistical hypothesis testing, a result has statistical significance when it is very unlikely to have occurred given the null hypothesis (simply by chance alone). More precisely, a study's defined significance level, denoted by \alpha, is the p ...
, and applied this to a normal distribution (as a two-tailed test), thus yielding the rule of two standard deviations (on a normal distribution) for statistical significance (see
68–95–99.7 rule In statistics, the 68–95–99.7 rule, also known as the empirical rule, is a shorthand used to remember the percentage of values that lie within an interval estimate in a normal distribution: 68%, 95%, and 99.7% of the values lie within one, t ...
). He then computed a table of values, similar to Elderton but, importantly, reversed the roles of ''χ''2 and ''p.'' That is, rather than computing ''p'' for different values of ''χ''2 (and degrees of freedom ''n''), he computed values of ''χ''2 that yield specified ''p''-values, specifically 0.99, 0.98, 0.95, 0,90, 0.80, 0.70, 0.50, 0.30, 0.20, 0.10, 0.05, 0.02, and 0.01. That allowed computed values of ''χ''2 to be compared against cutoffs and encouraged the use of ''p''-values (especially 0.05, 0.02, and 0.01) as cutoffs, instead of computing and reporting ''p''-values themselves. The same type of tables were then compiled in , which cemented the approach. As an illustration of the application of ''p''-values to the design and interpretation of experiments, in his following book ''
The Design of Experiments ''The Design of Experiments'' is a 1935 book by the English statistician Ronald Fisher about the design of experiments and is considered a foundational work in experimental design. Among other contributions, the book introduced the concept of th ...
'' (1935), Fisher presented the lady tasting tea experiment, which is the archetypal example of the ''p''-value. To evaluate a lady's claim that she ( Muriel Bristol) could distinguish by taste how tea is prepared (first adding the milk to the cup, then the tea, or first tea, then milk), she was sequentially presented with 8 cups: 4 prepared one way, 4 prepared the other, and asked to determine the preparation of each cup (knowing that there were 4 of each). In that case, the null hypothesis was that she had no special ability, the test was
Fisher's exact test Fisher's exact test is a statistical significance test used in the analysis of contingency tables. Although in practice it is employed when sample sizes are small, it is valid for all sample sizes. It is named after its inventor, Ronald Fisher, a ...
, and the ''p''-value was 1/\binom = 1/70 \approx 0.014, so Fisher was willing to reject the null hypothesis (consider the outcome highly unlikely to be due to chance) if all were classified correctly. (In the actual experiment, Bristol correctly classified all 8 cups.) Fisher reiterated the ''p'' = 0.05 threshold and explained its rationale, stating: He also applies this threshold to the design of experiments, noting that had only 6 cups been presented (3 of each), a perfect classification would have only yielded a ''p''-value of 1/\binom = 1/20 = 0.05, which would not have met this level of significance. Fisher also underlined the interpretation of ''p,'' as the long-run proportion of values at least as extreme as the data, assuming the null hypothesis is true. In later editions, Fisher explicitly contrasted the use of the ''p''-value for statistical inference in science with the Neyman–Pearson method, which he terms "Acceptance Procedures". Fisher emphasizes that while fixed levels such as 5%, 2%, and 1% are convenient, the exact ''p''-value can be used, and the strength of evidence can and will be revised with further experimentation. In contrast, decision procedures require a clear-cut decision, yielding an irreversible action, and the procedure is based on costs of error, which, he argues, are inapplicable to scientific research.


Related indices

The ''E''-value corresponds to the expected number of times in multiple testing that one expects to obtain a test statistic at least as extreme as the one that was actually observed if one assumes that the null hypothesis is true. The ''E''-value is the product of the number of tests and the ''p''-value. The ''q''-value is the analog of the ''p''-value with respect to the positive false discovery rate. It is used in multiple hypothesis testing to maintain statistical power while minimizing the
false positive rate In statistics, when performing multiple comparisons, a false positive ratio (also known as fall-out or false alarm ratio) is the probability of falsely rejecting the null hypothesis for a particular test. The false positive rate is calculated as th ...
. The Probability of Direction (''pd'') is the Bayesian numerical equivalent of the ''p''-value. It corresponds to the proportion of the posterior distribution that is of the median's sign, typically varying between 50% and 100%, and representing the certainty with which an effect is positive or negative.


See also

*
Student's t-test A ''t''-test is any statistical hypothesis test in which the test statistic follows a Student's ''t''-distribution under the null hypothesis. It is most commonly applied when the test statistic would follow a normal distribution if the value of ...
*
Bonferroni correction In statistics, the Bonferroni correction is a method to counteract the multiple comparisons problem. Background The method is named for its use of the Bonferroni inequalities. An extension of the method to confidence intervals was proposed by Ol ...
* Counternull * Fisher's method of combining ''p''-values * Generalized ''p''-value * Harmonic mean ''p''-value * Holm–Bonferroni method *
Multiple comparisons problem In statistics, the multiple comparisons, multiplicity or multiple testing problem occurs when one considers a set of statistical inferences simultaneously or infers a subset of parameters selected based on the observed values. The more inferences ...
* ''p''-rep * ''p''-value fallacy


Notes


References


Further reading

* * * * * * * * * * * * *


External links


Free online ''p''-values calculators
for various specific tests (chi-square, Fisher's F-test, etc.).

including a Java applet that illustrates how the numerical values of ''p''-values can give quite misleading impressions about the truth or falsity of the hypothesis under test. * *
Science Isn’t Broken - Article on how ''p''-values can be manipulated and an interactive tool to visualize it.
{{Statistics Statistical hypothesis testing ja:有意#p値