Nonparametric Skew
   HOME

TheInfoList



OR:

In
statistics Statistics (from German language, German: ', "description of a State (polity), state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of data. In applying statistics to a s ...
and
probability theory Probability theory or probability calculus is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expre ...
, the nonparametric skew is a
statistic A statistic (singular) or sample statistic is any quantity computed from values in a sample which is considered for a statistical purpose. Statistical purposes include estimating a population parameter, describing a sample, or evaluating a hypot ...
occasionally used with
random variable A random variable (also called random quantity, aleatory variable, or stochastic variable) is a Mathematics, mathematical formalization of a quantity or object which depends on randomness, random events. The term 'random variable' in its mathema ...
s that take real values.Arnold BC, Groeneveld RA (1995) Measuring skewness with respect to the mode. The American Statistician 49 (1) 34–38 DOI:10.1080/00031305.1995.10476109Rubio F.J.; Steel M.F.J. (2012) "On the Marshall–Olkin transformation as a skewing mechanism". ''Computational Statistics & Data Analysis''
Preprint
/ref> It is a measure of the
skewness In probability theory and statistics, skewness is a measure of the asymmetry of the probability distribution of a real-valued random variable about its mean. The skewness value can be positive, zero, negative, or undefined. For a unimodal ...
of a random variable's distribution—that is, the distribution's tendency to "lean" to one side or the other of the
mean A mean is a quantity representing the "center" of a collection of numbers and is intermediate to the extreme values of the set of numbers. There are several kinds of means (or "measures of central tendency") in mathematics, especially in statist ...
. Its calculation does not require any knowledge of the form of the underlying distribution—hence the name nonparametric. It has some desirable properties: it is zero for any
symmetric distribution In statistics, a symmetric probability distribution is a probability distribution—an assignment of probabilities to possible occurrences—which is unchanged when its probability density function (for continuous probability distribution) or pro ...
; it is unaffected by a scale shift; and it reveals either left- or right-skewness equally well. In some
statistical sample In this statistics, quality assurance, and survey methodology, sampling is the selection of a subset or a statistical sample (termed sample for short) of individuals from within a statistical population to estimate characteristics of the whole ...
s it has been shown to be less powerfulTabor J (2010) Investigating the Investigative Task: Testing for skewness - An investigation of different test statistics and their power to detect skewness. J Stat Ed 18: 1–13 than the usual measures of skewness in detecting departures of the
population Population is a set of humans or other organisms in a given region or area. Governments conduct a census to quantify the resident population size within a given jurisdiction. The term is also applied to non-human animals, microorganisms, and pl ...
from normality.


Properties


Definition

The nonparametric skew is defined as : S = \frac where the
mean A mean is a quantity representing the "center" of a collection of numbers and is intermediate to the extreme values of the set of numbers. There are several kinds of means (or "measures of central tendency") in mathematics, especially in statist ...
(''μ''),
median The median of a set of numbers is the value separating the higher half from the lower half of a Sample (statistics), data sample, a statistical population, population, or a probability distribution. For a data set, it may be thought of as the “ ...
(''ν'') and
standard deviation In statistics, the standard deviation is a measure of the amount of variation of the values of a variable about its Expected value, mean. A low standard Deviation (statistics), deviation indicates that the values tend to be close to the mean ( ...
(''σ'') of the population have their usual meanings.


Properties

The nonparametric skew is one third of the Pearson 2 skewness coefficient and lies between −1 and +1 for any distribution.Hotelling H, Solomons LM (1932) The limits of a measure of skewness. Annals Math Stat 3, 141–114Garver (1932) Concerning the limits of a mesuare of skewness. Ann Math Stats 3(4) 141–142 This range is implied by the fact that the mean lies within one standard deviation of any median.O’Cinneide CA (1990) The mean is within one standard deviation of any median. Amer Statist 44, 292–293 Under an
affine transformation In Euclidean geometry, an affine transformation or affinity (from the Latin, '' affinis'', "connected with") is a geometric transformation that preserves lines and parallelism, but not necessarily Euclidean distances and angles. More general ...
of the variable (''X''), the value of ''S'' does not change except for a possible change in sign. In symbols : S( aX + b ) = \operatorname (a)\, S(X) where ''a'' ≠ 0 and ''b'' are constants and ''S''( ''X'' ) is the nonparametric skew of the variable ''X''.


Sharper bounds

The bounds of this statistic ( ±1 ) were sharpened by MajindarMajindar KN (1962) "Improved bounds on a measure of skewness". ''Annals of Mathematical Statistics'', 33, 1192–1194 who showed that its
absolute value In mathematics, the absolute value or modulus of a real number x, is the non-negative value without regard to its sign. Namely, , x, =x if x is a positive number, and , x, =-x if x is negative (in which case negating x makes -x positive), ...
is bounded by : \frac with : p = \Pr( X > \operatorname( X ) ) and : q = \Pr( X < \operatorname( X ) ) , where ''X'' is a random variable with finite
variance In probability theory and statistics, variance is the expected value of the squared deviation from the mean of a random variable. The standard deviation (SD) is obtained as the square root of the variance. Variance is a measure of dispersion ...
, ''E''() is the expectation operator and ''Pr''() is the probability of the event occurring. When ''p'' = ''q'' = 0.5 the absolute value of this statistic is bounded by 1. With ''p'' = 0.1 and ''p'' = 0.01, the statistic's absolute value is bounded by 0.6 and 0.199 respectively.


Extensions

It is also known thatMallows CCC, Richter D (1969) "Inequalities of Chebyschev type involving conditional expectations". ''Annals of Mathematical Statistics'', 40:1922–1932 : , \mu - \nu_0 , \le \operatorname ( , X - \nu_0 , ) \le \operatorname ( , X - \mu , ) \le \sigma , where ''ν''0 is any median and ''E''(.) is the expectation operator. It has been shown that : \frac \le \max\left( \sqrt, \sqrt \right) where ''x''''q'' is the ''q''th
quantile In statistics and probability, quantiles are cut points dividing the range of a probability distribution into continuous intervals with equal probabilities or dividing the observations in a sample in the same way. There is one fewer quantile t ...
. Quantiles lie between 0 and 1: the median (the 0.5 quantile) has ''q'' = 0.5. This inequality has also been used to define a measure of skewness.Dziubinska R, Szynal D (1996) On functional measures of skewness. Applicationes Mathematicae 23(4) 395–403 This latter inequality has been sharpened further.Dharmadhikari SS (1991) Bounds on quantiles: a comment on O'Cinneide. The Am Statist 45: 257-58 : \mu -\sigma \sqrt \le x_q \le \mu + \sigma \sqrt Another extension for a distribution with a finite mean has been published:Gilat D, Hill TP(1993) Quantile-locating functions and the distance between the mean and quantiles. Statistica Neerlandica 47 (4) 279–283 DOI: 10.1111/j.1467-9574.1993.tb01424.x

/ref> : \mu - \frac \operatorname, X - \mu , \le x_q \le \mu + \frac \operatorname, X - \mu , The bounds in this last pair of inequalities are attained when \Pr (X=a) = q and \Pr (X=b) = 1-q for fixed numbers ''a'' < ''b''.


Finite samples

For a finite sample with sample size ''n'' ≥ 2 with ''x''r is the ''r''th
order statistic In statistics, the ''k''th order statistic of a statistical sample is equal to its ''k''th-smallest value. Together with Ranking (statistics), rank statistics, order statistics are among the most fundamental tools in non-parametric statistics and ...
, ''m'' the sample mean and ''s'' the sample standard deviation corrected for degrees of freedom,David HA (1991) Mean minus median: A comment on O'Cinneide. The Am Statist 45: 257 \frac \le \text\left \sqrt , \sqrt \right Replacing ''r'' with ''n'' / 2 gives the result appropriate for the sample median:Joarder AH, Laradji A (2004) Some inequalities in descriptive statistics. Technical Report Series TR 321 \frac \le \sqrt = \sqrt where ''a'' is the sample median.


Statistical tests

Hotelling and Solomons considered the distribution of the test statistic : D = \frac where ''n'' is the sample size, ''m'' is the sample mean, ''a'' is the sample median and ''s'' is the sample's standard deviation. Statistical tests of ''D'' have assumed that the null hypothesis being tested is that the distribution is symmetric . Gastwirth estimated the asymptotic
variance In probability theory and statistics, variance is the expected value of the squared deviation from the mean of a random variable. The standard deviation (SD) is obtained as the square root of the variance. Variance is a measure of dispersion ...
of ''n''−1/2''D''.Gastwirth JL (1971) "On the sign test for symmetry". ''
Journal of the American Statistical Association The ''Journal of the American Statistical Association'' is a quarterly peer-reviewed scientific journal published by Taylor & Francis on behalf of the American Statistical Association. It covers work primarily focused on the application of statis ...
'' 66:821–823
If the distribution is unimodal and symmetric about 0, the asymptotic variance lies between 1/4 and 1. Assuming a conservative estimate (putting the variance equal to 1) can lead to a true level of significance well below the nominal level. Assuming that the underlying distribution is symmetric Cabilio and Masaro have shown that the distribution of ''S'' is asymptotically normal.Cabilio P, Masaro J (1996) "A simple test of symmetry about an unknown median". ''Canadian Journal of Statistics-Revue Canadienne De Statistique'', 24:349–361 The asymptotic variance depends on the underlying distribution: for the normal distribution, the asymptotic variance of ''S'' is 0.5708... Assuming that the underlying distribution is symmetric, by considering the distribution of values above and below the median Zheng and Gastwirth have argued thatZheng T, Gastwirth J (2010) "On bootstrap tests of symmetry about an unknown median". ''Journal of Data Science'', 8(3): 413–427 : \sqrt \left( \frac \right) where ''n'' is the sample size, is distributed as a t distribution.


Related statistics

Antonietta Mira studied the distribution of the difference between the mean and the median. Mira A (1999) "Distribution-free test for symmetry based on Bonferroni’s measure", ''Journal of Applied Statistics'', 26:959–972 : \gamma_1 = 2 ( m - a ) , where ''m'' is the sample mean and ''a'' is the median. If the underlying distribution is symmetrical ''γ''1 itself is asymptotically normal. This statistic had been earlier suggested by Bonferroni.Bonferroni CE (1930) ''Elementi di statistica generale''. Seeber, Firenze Assuming a symmetric underlying distribution, a modification of ''S'' was studied by Miao, Gel and Gastwirth who modified the standard deviation to create their statistic.Miao W, Gel YR, Gastwirth JL (2006) "A new test of symmetry about an unknown median". In: Hsiung A, Zhang C-H, Ying Z, eds. ''Random Walk, Sequential Analysis and Related Topics — A Festschrift in honor of Yuan-Shih Chow''. World Scientific; Singapore : J = \frac \sqrt \sum where ''X''i are the sample values, , , is the
absolute value In mathematics, the absolute value or modulus of a real number x, is the non-negative value without regard to its sign. Namely, , x, =x if x is a positive number, and , x, =-x if x is negative (in which case negating x makes -x positive), ...
and the sum is taken over all ''n'' sample values. The test statistic was : T = \frac . The scaled statistic ''T'' is asymptotically normal with a mean of zero for a symmetric distribution. Its asymptotic variance depends on the underlying distribution: the limiting values are, for the normal distribution = 0.5708... and, for the t distribution with three
degrees of freedom In many scientific fields, the degrees of freedom of a system is the number of parameters of the system that may vary independently. For example, a point in the plane has two degrees of freedom for translation: its two coordinates; a non-infinite ...
, = 0.9689...


Values for individual distributions


Symmetric distributions

For
symmetric probability distribution In statistics, a symmetric probability distribution is a probability distribution—an assignment of probabilities to possible occurrences—which is unchanged when its probability density function (for continuous probability distribution) or pro ...
s the value of the nonparametric skew is 0.


Asymmetric distributions

It is positive for right skewed distributions and negative for left skewed distributions. Absolute values ≥ 0.2 indicate marked skewness. It may be difficult to determine ''S'' for some distributions. This is usually because a closed form for the median is not known: examples of such distributions include the gamma distribution, inverse-chi-squared distribution, the inverse-gamma distribution and the scaled inverse chi-squared distribution. The following values for ''S'' are known: *
Beta distribution In probability theory and statistics, the beta distribution is a family of continuous probability distributions defined on the interval
, 1 The comma is a punctuation mark that appears in several variants in different languages. Some typefaces render it as a small line, slightly curved or straight, but inclined from the vertical; others give it the appearance of a miniature fille ...
or (0, 1) in terms of two positive Statistical parameter, parameters, denoted by ''alpha'' (''α'') an ...
: 1 < ''α'' < ''β'' where ''α'' and ''β'' are the parameters of the distribution, then to a good approximationKerman J (2011) "A closed-form approximation for the median of the beta distribution". :: S = \frac\frac : If 1 < ''β'' < ''α'' then the positions of ''α'' and ''β'' are reversed in the formula. ''S'' is always < 0. *
Binomial distribution In probability theory and statistics, the binomial distribution with parameters and is the discrete probability distribution of the number of successes in a sequence of statistical independence, independent experiment (probability theory) ...
: varies. If the mean is an
integer An integer is the number zero (0), a positive natural number (1, 2, 3, ...), or the negation of a positive natural number (−1, −2, −3, ...). The negations or additive inverses of the positive natural numbers are referred to as negative in ...
then ''S'' = 0. If the mean is not an integer ''S'' may have either sign or be zero.Kaas R, Buhrman JM (1980) Mean, median and mode in binomial distributions. Statistica Neerlandica 34 (1) 13–18 It is bounded by ±min / ''σ'' where ''σ'' is the standard deviation of the binomial distribution.Hamza K (1995) "The smallest uniform upper bound on the distance between the mean and the median of the binomial and Poisson distributions". ''Statistics and Probability Letters'', 23 (1) 21–25 *
Burr distribution In probability theory, statistics and econometrics, the Burr Type XII distribution or simply the Burr distribution is a continuous probability distribution for a non-negative random variable. It is also known as the Singh–Maddala distribution a ...
: * Birnbaum–Saunders distribution: :: S = \frac : where ''α'' is the shape parameter and ''β'' is the location parameter. * Cantor distribution: despite the distribution being symmetric about its mean of \tfrac12, the median can be any value in \left tfrac13,\tfrac23\right/math> as this central interval has zero probability :: \frac \le S \le \frac * Chi square distribution: Although ''S'' ≥ 0 its value depends on the numbers of
degrees of freedom In many scientific fields, the degrees of freedom of a system is the number of parameters of the system that may vary independently. For example, a point in the plane has two degrees of freedom for translation: its two coordinates; a non-infinite ...
(''k''). :: S \approx \frac * Dagum distribution: *
Exponential distribution In probability theory and statistics, the exponential distribution or negative exponential distribution is the probability distribution of the distance between events in a Poisson point process, i.e., a process in which events occur continuousl ...
: :: S = 1 - \log_e( 2 ) \approx 0.31 *
Exponential distribution In probability theory and statistics, the exponential distribution or negative exponential distribution is the probability distribution of the distance between events in a Poisson point process, i.e., a process in which events occur continuousl ...
with two parameters: :: S = 1 - \log_e( 2 ) \approx 0.31 * Exponential-logarithmic distribution :: S = - \frac : Here ''S'' is always > 0. * Exponentially modified Gaussian distribution: :: 0 \le S \le 1 - \log_e( 2 ) * F distribution with ''n'' and ''n''
degrees of freedom In many scientific fields, the degrees of freedom of a system is the number of parameters of the system that may vary independently. For example, a point in the plane has two degrees of freedom for translation: its two coordinates; a non-infinite ...
( ''n'' > 4 ):Terrell GR (1986) "Pearson's rule for sample medians". Technical Report 86-2 :: S = n^ \sqrt + O( n^ ) * Fréchet distribution: The variance of this distribution is defined only for ''α'' > 2. :: S = \frac \sqrt * Gamma distribution: The median can only be determined approximately for this distribution.Banneheka BMSG, Ekanayake GEMUPD (2009) A new point estimator for the median of Gamma distribution. Viyodaya J Science 14:95–103 If the shape parameter ''α'' is ≥ 1 then :: S \approx \frac : where ''β'' > 0 is the rate parameter. Here ''S'' is always > 0. * Generalized normal distribution version 2 :: S = - \frac : ''S'' is always < 0. * Generalized Pareto distribution: ''S'' is defined only when the shape parameter ( ''k'' ) is < 1/2. ''S'' is < 0 for this distribution. :: S = \left( \frac - 2^k \right)( 1 - 2k )^ * Gumbel distribution: :: \frac \approx 0.1643 : where ''γ'' is Euler's constant.Ferguson T
"Asymptotic Joint Distribution of Sample Mean and a Sample Quantile"
Unpublished
*
Half-normal distribution In probability theory and statistics, the half-normal distribution is a special case of the folded normal distribution. Let X follow an ordinary normal distribution, N(0,\sigma^2). Then, Y=, X, follows a half-normal distribution. Thus, the ha ...
: :: S \approx \frac \approx 0.36279 * Kumaraswamy distribution * Log-logistic distribution (Fisk distribution): Let ''β'' be the shape parameter. The variance and mean of this distribution are only defined when ''β'' > 2. To simplify the notation let ''b'' = ''β'' / . :: S = \frac : The standard deviation does not exist for values of ''b'' > 4.932 (approximately). For values for which the standard deviation is defined, ''S'' is > 0. *
Log-normal distribution In probability theory, a log-normal (or lognormal) distribution is a continuous probability distribution of a random variable whose logarithm is normal distribution, normally distributed. Thus, if the random variable is log-normally distributed ...
: With mean ( ''μ'' ) and variance ( ''σ''2 ) :: S = \frac * Log-Weibull distribution: :: S \approx \frac \approx -0.1643 * Lomax distribution: ''S'' is defined only for ''α'' > 2 :: S = \frac *
Maxwell–Boltzmann distribution In physics (in particular in statistical mechanics), the Maxwell–Boltzmann distribution, or Maxwell(ian) distribution, is a particular probability distribution named after James Clerk Maxwell and Ludwig Boltzmann. It was first defined and use ...
: :: S \approx \frac \approx 0.0854 * Nakagami distribution :: S = -1 *
Pareto distribution The Pareto distribution, named after the Italian civil engineer, economist, and sociologist Vilfredo Pareto, is a power-law probability distribution that is used in description of social, quality control, scientific, geophysical, actuarial scien ...
: for ''α'' > 2 where ''α'' is the shape parameter of the distribution, :: S = ( \alpha - 2^ \alpha - 1 ) ( \frac )^ , :and ''S'' is always > 0. *
Poisson distribution In probability theory and statistics, the Poisson distribution () is a discrete probability distribution that expresses the probability of a given number of events occurring in a fixed interval of time if these events occur with a known const ...
: :: \frac \le S \le \frac : where ''λ'' is the parameter of the distribution.Choi KP (1994) "On the medians of Gamma distributions and an equation of Ramanujan". ''Proc Amer Math Soc'' 121 (1) 245–251 * Rayleigh distribution: :: S = \sqrt ( \frac )^ - \log_e( 4 ) \approx 0.1251 * Weibull distribution: :: S = \frac, : where ''k'' is the shape parameter of the distribution. Here ''S'' is always > 0.


History

In 1895 Pearson first suggested measuring skewness by standardizing the difference between the mean and the mode,Pearson K (1895) Contributions to the Mathematical Theory of Evolution–II. Skew variation in homogenous material. Phil Trans Roy Soc A. 186: 343–414 giving : \frac , where ''μ'', ''θ'' and ''σ'' is the mean, mode and standard deviation of the distribution respectively. Estimates of the population mode from the sample data may be difficult but the difference between the mean and the mode for many distributions is approximately three times the difference between the mean and the medianStuart A, Ord JK (1994) ''Kendall’s advanced theory of statistics. Vol 1. Distribution theory''. 6th Edition. Edward Arnold, London which suggested to Pearson a second skewness coefficient: : \frac , where ''ν'' is the median of the distribution. Bowley dropped the factor 3 from this formula in 1901 leading to the nonparametric skew statistic. The relationship between the median, the mean and the mode was first noted by Pearson when he was investigating his type III distributions.


Relationships between the mean, median and mode

For an arbitrary distribution the mode, median and mean may appear in any order.Dharmadhikari SW, Joag-dev K (1983) Mean, Median, Mode III. Statistica Neerlandica, 33: 165–168 Analyses have been made of some of the relationships between the mean, median, mode and standard deviation. and these relationships place some restrictions on the sign and magnitude of the nonparametric skew. A simple example illustrating these relationships is the
binomial distribution In probability theory and statistics, the binomial distribution with parameters and is the discrete probability distribution of the number of successes in a sequence of statistical independence, independent experiment (probability theory) ...
with ''n'' = 10 and ''p'' = 0.09.Lesser LM (2005
"Letter to the editor"
omment on von Hippel (2005) ''Journal of Statistics Education'' 13(2).
This distribution when plotted has a long right tail. The mean (0.9) is to the left of the median (1) but the skew (0.906) as defined by the third standardized moment is positive. In contrast the nonparametric skew is -0.110.


Pearson's rule

The rule that for some distributions the difference between the mean and the mode is three times that between the mean and the median is due to Pearson who discovered it while investigating his Type 3 distributions. It is often applied to slightly asymmetric distributions that resemble a normal distribution but it is not always true. In 1895 Pearson noted that for what is now known as the gamma distribution that the relation : \nu - \theta = 2( \mu - \nu ) where ''θ'', ''ν'' and ''μ'' are the mode, median and mean of the distribution respectively was approximately true for distributions with a large shape parameter. Doodson in 1917 proved that the median lies between the mode and the mean for moderately skewed distributions with finite fourth moments.Doodson AT (1917) "Relation of the mode, median and mean in frequency functions". '' Biometrika'', 11 (4) 425–429 This relationship holds for all the Pearson distributions and all of these distributions have a positive nonparametric skew. Doodson also noted that for this family of distributions to a good approximation, : \theta = 3 \nu - 2 \mu , where ''θ'', ''ν'' and ''μ'' are the mode, median and mean of the distribution respectively. Doodson's approximation was further investigated and confirmed by Haldane.Haldane JBS (1942) "The mode and median of a nearly normal distribution with given cumulants". '' Biometrika'', 32: 294–299 Haldane noted that samples with identical and independent variates with a third
cumulant In probability theory and statistics, the cumulants of a probability distribution are a set of quantities that provide an alternative to the '' moments'' of the distribution. Any two probability distributions whose moments are identical will have ...
had sample means that obeyed Pearson's relationship for large sample sizes. Haldane required a number of conditions for this relationship to hold including the existence of an Edgeworth expansion and the uniqueness of both the median and the mode. Under these conditions he found that mode and the median converged to 1/2 and 1/6 of the third moment respectively. This result was confirmed by Hall under weaker conditions using
characteristic function In mathematics, the term "characteristic function" can refer to any of several distinct concepts: * The indicator function of a subset, that is the function \mathbf_A\colon X \to \, which for a given subset ''A'' of ''X'', has value 1 at points ...
s.Hall P (1980) "On the limiting behaviour of the mode and median of a sum of independent random variables". ''Annals of Probability'' 8: 419–430 Doodson's relationship was studied by Kendall and Stuart in the
log-normal distribution In probability theory, a log-normal (or lognormal) distribution is a continuous probability distribution of a random variable whose logarithm is normal distribution, normally distributed. Thus, if the random variable is log-normally distributed ...
for which they found an exact relationship close to it.Kendall M.G., Stuart A. (1958) ''The advanced theory of statistics''. p53 Vol 1. Griffin. London Hall also showed that for a distribution with regularly varying tails and exponent ''α'' that : \mu - \theta = \alpha ( \mu - \nu )


Unimodal distributions

Gauss showed in 1823 that for a
unimodal distribution In mathematics, unimodality means possessing a unique mode (statistics), mode. More generally, unimodality means there is only a single highest value, somehow defined, of some mathematical object. Unimodal probability distribution In statis ...
Gauss C.F. Theoria Combinationis Observationum Erroribus Minimis Obnoxiae. Pars Prior. Pars Posterior. Supplementum. Theory of the Combination of Observations Least Subject to Errors. Part One. Part Two. Supplement. 1995. Translated by G.W. Stewart. Classics in Applied Mathematics Series, Society for Industrial and Applied Mathematics, Philadelphia : \sigma \le \omega \le 2 \sigma and : , \nu - \mu , \le \sqrt \omega , where ''ω'' is the root mean square deviation from the mode. For a large class of unimodal distributions that are positively skewed the mode, median and mean fall in that order.MacGillivray HL (1981) The mean, median, mode inequality and skewness for a class of densities. Aust J Stat 23(2) 247–250 Conversely for a large class of unimodal distributions that are negatively skewed the mean is less than the median which in turn is less than the mode. In symbols for these positively skewed unimodal distributions : \theta \le \nu \le \mu and for these negatively skewed unimodal distributions : \mu \le \nu \le \theta This class includes the important F, beta and gamma distributions. This rule does not hold for the unimodal Weibull distribution.Groeneveld RA (1986) Skewness for the Weibull family. Statistica Neerlandica 40: 135–140 For a unimodal distribution the following bounds are known and are sharp:Johnson NL, Rogers CA (1951) "The moment problem for unimodal distributions". ''Annals of Mathematical Statistics'', 22 (3) 433–439 : \frac \le \sqrt , : \frac \le \sqrt , : \frac \le \sqrt , where ''μ'',''ν'' and ''θ'' are the mean, median and mode respectively. The middle bound limits the nonparametric skew of a unimodal distribution to approximately ±0.775.


van Zwet condition

The following inequality, : \theta \le \nu \le \mu , where ''θ'', ''ν'' and ''μ'' is the mode, median and mean of the distribution respectively, holds if : F( \nu - x ) + F( \nu + x ) \ge 1 \text x, where ''F'' is the
cumulative distribution function In probability theory and statistics, the cumulative distribution function (CDF) of a real-valued random variable X, or just distribution function of X, evaluated at x, is the probability that X will take a value less than or equal to x. Ever ...
of the distribution.van Zwet W.R. (1979) "Mean, median, mode II". ''Statistica Neerlandica'' 33(1) 1–5 These conditions have since been generalised and extended to discrete distributions.Abdous B, Theodorescu R (1998) Mean, median, mode IV. Statistica Neerlandica. 52 (3) 356–359 Any distribution for which this holds has either a zero or a positive nonparametric skew.


Notes


Ordering of skewness

In 1964 van Zwet proposed a series of axioms for ordering measures of skewness.van Zwet, W.R. (1964) "Convex transformations of random variables". ''Mathematics Centre Tract'', 7, Mathematisch Centrum, Amsterdam The nonparametric skew does not satisfy these axioms.


Benford's law

Benford's law is an empirical law concerning the distribution of digits in a list of numbers. It has been suggested that random variates from distributions with a positive nonparametric skew will obey this law.Durtschi C, Hillison W, Pacini C (2004) The effective use of Benford’s Law to assist in detecting fraud in accounting data. J Forensic Accounting 5: 17–34


Relation to Bowley's coefficient

This statistic is very similar to Bowley's coefficient of skewnessBowley AL (1920) Elements of statistics. New York: Charles Scribner's Sons : SK_2 = \frac where Qi is the ith quartile of the distribution. Hinkley generalised thisHinkley DV (1975) On power transformations to symmetry. Biometrika 62: 101–111 : SK = \frac where \alpha lies between 0 and 0.5. Bowley's coefficient is a special case with \alpha equal to 0.25. Groeneveld and MeedenGroeneveld RA, Meeden G (1984) Measuring skewness and kurtosis. The Statistician, 33: 391–399 removed the dependence on \alpha by integrating over it. : SK_3 = \frac The denominator is a measure of dispersion. Replacing the denominator with the standard deviation we obtain the nonparametric skew.


References

{{Statistics, descriptive Summary statistics