Jump to content
Main menu
Main menu
move to sidebar
hide
Navigation
Main page
Recent changes
Random page
Help about MediaWiki
Special pages
Niidae Wiki
Search
Search
Appearance
Create account
Log in
Personal tools
Create account
Log in
Pages for logged out editors
learn more
Contributions
Talk
Editing
Skewness
Page
Discussion
English
Read
Edit
View history
Tools
Tools
move to sidebar
hide
Actions
Read
Edit
View history
General
What links here
Related changes
Page information
Appearance
move to sidebar
hide
Warning:
You are not logged in. Your IP address will be publicly visible if you make any edits. If you
log in
or
create an account
, your edits will be attributed to your username, along with other benefits.
Anti-spam check. Do
not
fill this in!
{{Short description|Measure of the asymmetry of random variables}} {{For|the planarity measure in graph theory|Graph skewness}} {{use American English|date = January 2019}} {{Use dmy dates|date=October 2020}} [[File:SkewedDistribution.png|thumb|200px|Example distribution with positive skewness. These data are from experiments on wheat grass growth.]] In [[probability theory]] and [[statistics]], '''skewness''' is a measure of the asymmetry of the [[probability distribution]] of a [[real number|real]]-valued [[random variable]] about its mean. The skewness value can be positive, zero, negative, or undefined. For a [[unimodal]] distribution (a distribution with a single peak), negative skew commonly indicates that the ''tail'' is on the left side of the distribution, and positive skew indicates that the tail is on the right. In cases where one tail is long but the other tail is fat, skewness does not obey a simple rule. For example, a zero value in skewness means that the tails on both sides of the mean balance out overall; this is the case for a symmetric distribution but can also be true for an asymmetric distribution where one tail is long and thin, and the other is short but fat. Thus, the judgement on the symmetry of a given distribution by using only its skewness is risky; the distribution shape must be taken into account. == Introduction == Consider the two distributions in the figure just below. Within each graph, the values on the right side of the distribution taper differently from the values on the left side. These tapering sides are called ''tails'', and they provide a visual means to determine which of the two kinds of skewness a distribution has: # ''{{visible anchor|negative skew}}'': The left tail is longer; the mass of the distribution is concentrated on the right of the figure. The distribution is said to be ''left-skewed'', ''left-tailed'', or ''skewed to the left'', despite the fact that the curve itself appears to be skewed or leaning to the right; ''left'' instead refers to the left tail being drawn out and, often, the mean being skewed to the left of a typical center of the data. A left-skewed distribution usually appears as a ''right-leaning'' curve.<ref name="cnx.org">{{Cite web |last=Illowsky |first=Barbara |last2=Dean |first2=Susan |date=2020-03-27 |title=2.6 Skewness and the Mean, Median, and Mode – Statistics |url=https://openstax.org/books/statistics/pages/2-6-skewness-and-the-mean-median-and-mode |access-date=2022-12-21 |website=[[OpenStax]] |language=en}}</ref> # ''{{visible anchor|positive skew}}'': The right tail is longer; the mass of the distribution is concentrated on the left of the figure. The distribution is said to be ''right-skewed'', ''right-tailed'', or ''skewed to the right'', despite the fact that the curve itself appears to be skewed or leaning to the left; ''right'' instead refers to the right tail being drawn out and, often, the mean being skewed to the right of a typical center of the data. A right-skewed distribution usually appears as a ''left-leaning'' curve.<ref name="cnx.org" /> [[Image:Negative and positive skew diagrams (English).svg|thumb|left|upright=2]] Skewness in a data series may sometimes be observed not only graphically but by simple inspection of the values. For instance, consider the numeric sequence (49, 50, 51), whose values are evenly distributed around a central value of 50. We can transform this sequence into a negatively skewed distribution by adding a value far below the mean, which is probably a negative [[outlier]], e.g. (40, 49, 50, 51). Therefore, the mean of the sequence becomes 47.5, and the median is 49.5. Based on the formula of [[nonparametric skew]], defined as <math>(\mu - \nu)/\sigma,</math> the skew is negative. Similarly, we can make the sequence positively skewed by adding a value far above the mean, which is probably a positive outlier, e.g. (49, 50, 51, 60), where the mean is 52.5, and the median is 50.5. As mentioned earlier, a unimodal distribution with zero value of skewness does not imply that this distribution is symmetric necessarily. However, a symmetric unimodal or multimodal distribution always has zero skewness. [[File:Asymmetric Distribution with Zero Skewness.jpg|434px|thumb|right|Example of an asymmetric distribution with zero skewness. This figure serves as a counterexample that zero skewness does not imply symmetric distribution necessarily. (Skewness was calculated by Pearson's moment coefficient of skewness.)]] == Relationship of mean and median == The skewness is not directly related to the relationship between the mean and median: a distribution with negative skew can have its mean greater than or less than the median, and likewise for positive skew.<ref name="von Hippel 2005">{{cite journal |last=von Hippel |first=Paul T. |year=2005 |title=Mean, Median, and Skew: Correcting a Textbook Rule |url=http://www.amstat.org/publications/jse/v13n2/vonhippel.html |url-status=dead |journal=Journal of Statistics Education |volume=13 |issue=2 |archive-url=https://web.archive.org/web/20160220181456/http://www.amstat.org/publications/jse/v13n2/vonhippel.html |archive-date=2016-02-20}}</ref> [[File:Relationship between mean and median under different skewness.png|thumb|434x434px|A general relationship of mean and median under differently skewed unimodal distribution.]] In the older notion of [[nonparametric skew]], defined as <math>(\mu - \nu)/\sigma,</math> where <math>\mu</math> is the [[mean]], <math>\nu</math> is the [[median]], and <math>\sigma</math> is the [[standard deviation]], the skewness is defined in terms of this relationship: positive/right nonparametric skew means the mean is greater than (to the right of) the median, while negative/left nonparametric skew means the mean is less than (to the left of) the median. However, the modern definition of skewness and the traditional nonparametric definition do not always have the same sign: while they agree for some families of distributions, they differ in some of the cases, and conflating them is misleading. If the distribution is [[Symmetric probability distribution|symmetric]], then the mean is equal to the median, and the distribution has zero skewness.<ref>{{cite web|title=1.3.5.11. Measures of Skewness and Kurtosis |url=http://www.itl.nist.gov/div898/handbook/eda/section3/eda35b.htm|publisher=NIST|access-date=18 March 2012}}</ref> If the distribution is both symmetric and [[Unimodal distribution|unimodal]], then the [[mean]] = [[median]] = [[Mode (statistics)|mode]]. This is the case of a coin toss or the series 1,2,3,4,... Note, however, that the converse is not true in general, i.e. zero skewness (defined below) does not imply that the mean is equal to the median. A 2005 journal article points out:<ref name="von Hippel 2005"/><blockquote>Many textbooks teach a rule of thumb stating that the mean is right of the median under right skew, and left of the median under left skew. This rule fails with surprising frequency. It can fail in [[multimodal distribution]]s, or in distributions where one tail is [[Long tail|long]] but the other is [[Heavy-tailed distribution|heavy]]. Most commonly, though, the rule fails in discrete distributions where the areas to the left and right of the median are not equal. Such distributions not only contradict the textbook relationship between mean, median, and skew, they also contradict the textbook interpretation of the median.</blockquote> [[File:Positive skewness with mean less than median.png|thumb|288x288px|Distribution of adult residents across US households]] For example, in the distribution of adult residents across US households, the skew is to the right. However, since the majority of cases is less than or equal to the mode, which is also the median, the mean sits in the heavier left tail. As a result, the rule of thumb that the mean is right of the median under right skew failed.<ref name="von Hippel 2005" /> ==Definition== ===Fisher's moment coefficient of skewness=== The skewness <math>\gamma_1</math> of a random variable {{mvar|X}} is the third [[standardized moment]] <math>\tilde{\mu}_3</math>, defined as:<ref name="StanBrown1"/><ref name="FXSolver1"/> <math display="block"> \gamma_1 := \tilde{\mu}_3 = \operatorname{E}\left[\left(\frac{X-\mu}{\sigma}\right)^3 \right] = \frac{\mu_3}{\sigma^3} = \frac{\operatorname{E}\left[(X-\mu)^3\right]}{\left( \operatorname{E}\left[ (X-\mu)^2 \right] \right)^{3/2}} = \frac{\kappa_3}{\kappa_2^{3/2}} </math> where {{mvar|μ}} is the mean, {{mvar|σ}} is the [[standard deviation]], E is the [[expected value|expectation operator]], {{math|''μ''<sub>3</sub>}} is the third [[central moment]], and {{math|''κ''<sub>''t''</sub>}} are the {{mvar|t}}-th [[cumulant]]s. It is sometimes referred to as '''Pearson's moment coefficient of skewness''',<ref name="FXSolver1">[http://www.fxsolver.com/browse/formulas/Pearson's+moment+coefficient+of+skewness Pearson's moment coefficient of skewness], FXSolver.com</ref> or simply the '''moment coefficient of skewness''',<ref name="StanBrown1">[http://brownmath.com/stat/shape.htm "Measures of Shape: Skewness and Kurtosis"], 2008–2016 by Stan Brown, Oak Road Systems</ref> but should not be confused with Pearson's other skewness statistics (see below). The last equality expresses skewness in terms of the ratio of the third cumulant {{math|''κ''<sub>3</sub>}} to the 1.5th power of the second cumulant {{math|''κ''<sub>2</sub>}}. This is analogous to the definition of [[kurtosis]] as the fourth cumulant normalized by the square of the second cumulant. The skewness is also sometimes denoted {{math|Skew[''X'']}}. If {{mvar|σ}} is finite and {{mvar|μ}} is finite too, then skewness can be expressed in terms of the non-central moment {{math|E[''X''<sup>3</sup>]}} by expanding the previous formula: <math display="block"> \begin{align} \tilde{\mu}_3 &= \operatorname{E}\left[\left(\frac{X-\mu}{\sigma}\right)^3 \right] \\ &= \frac{\operatorname{E}[X^3] - 3\mu\operatorname E[X^2] + 3\mu^2\operatorname E[X] - \mu^3}{\sigma^3}\\ &= \frac{\operatorname{E}[X^3] - 3\mu(\operatorname E[X^2] -\mu\operatorname E[X]) - \mu^3}{\sigma^3}\\ &= \frac{\operatorname{E}[X^3] - 3\mu\sigma^2 - \mu^3}{\sigma^3}. \end{align} </math> <!-- EDITORS BEWARE: DO NOT CHANGE THIS INTO E[X^3] - 3\mu E[X^2] + 2\mu^3 /// SEE TALK PAGE --> ===Examples=== Skewness can be infinite, as when <math display="block">\Pr \left[ X > x \right]=x^{-2}\mbox{ for }x>1,\ \Pr[X<1]=0</math> where the third cumulants are infinite, or as when <math display="block">\Pr[X<x] = \begin{cases} \frac{1}{2} (1-x)^{-3} & \text{ for } x < 0, \\[2pt] \frac{1}{2} (1+x)^{-3} & \text{ for } x > 0. \end{cases}</math> where the third cumulant is undefined. Examples of distributions with finite skewness include the following. * A [[normal distribution]] and any other symmetric distribution with finite third moment has a skewness of 0 * A [[half-normal distribution]] has a skewness just below 1 * An [[exponential distribution]] has a skewness of 2 * A [[lognormal distribution]] can have a skewness of any positive value, depending on its parameters ===Sample skewness=== For a sample of ''n'' values, two natural estimators of the population skewness are<ref name=JG>{{cite journal |last=Joanes<!--sic--> |first=D. N. |last2=Gill |first2=C. A. |year=1998 |title=Comparing measures of sample skewness and kurtosis |journal=[[Journal of the Royal Statistical Society, Series D]] |volume=47 |issue=1 |pages=183–189 |doi=10.1111/1467-9884.00122}}</ref> <math display="block"> b_1 = \frac{m_3}{s^3} = \frac{\tfrac{1}{n} \sum_{i=1}^n \left(x_i-\bar{x}\right)^3}{\left[\tfrac{1}{n-1} \sum_{i=1}^n \left(x_i-\bar{x}\right)^2 \right]^{3/2}} </math> and <math display="block"> g_1 = \frac{m_3}{m_2^{3/2}} = \frac{\tfrac{1}{n} \sum_{i=1}^n (x_i-\bar{x})^3}{\left[\tfrac{1}{n} \sum_{i=1}^n \left(x_i-\bar{x}\right)^2 \right]^{3/2}}, </math> where <math>\bar{x}</math> is the [[sample mean]], {{mvar|s}} is the [[Standard deviation#Corrected sample standard deviation|sample standard deviation]], {{math|''m''<sub>2</sub>}} is the (biased) sample second central [[Moment (mathematics)|moment]], and {{math|''m''<sub>3</sub>}} is the (biased) sample third central moment.<ref name=JG/> <math>g_1</math> is a [[Method of moments (statistics)|method of moments]] estimator. Another common definition of the ''sample skewness'' is<ref name=JG/><ref name=Doane2011>Doane, David P., and Lori E. Seward. [http://jse.amstat.org/v19n2/doane.pdf "Measuring skewness: a forgotten statistic."] Journal of Statistics Education 19.2 (2011): 1-18. (Page 7)</ref> <math display="block">\begin{align} G_1 & = \frac{k_3}{k_2^{3/2}} = \frac{n^2}{(n-1)(n-2)}\; b_1 = \frac{\sqrt{n(n-1)}}{n-2}\; g_1, \\ \end{align}</math> where <math>k_3</math> is the unique symmetric unbiased estimator of the third [[cumulant]] and <math>k_2 = s^2</math> is the symmetric unbiased estimator of the second cumulant (i.e. the [[Variance#Population variance and sample variance|sample variance]]). This adjusted Fisher–Pearson standardized moment coefficient <math> G_1 </math> is the version found in [[Microsoft Excel|Excel]] and several statistical packages including [[Minitab]], [[SAS (software)|SAS]] and [[SPSS]].<ref name=Doane2011/> Under the assumption that the underlying random variable <math>X</math> is normally distributed, it can be shown that all three ratios <math>b_1</math>, <math>g_1</math> and <math>G_1</math> are unbiased and [[Consistent estimator|consistent]] estimators of the population skewness <math>\gamma_1=0</math>, with <math>\sqrt{n} b_1 \mathrel{\xrightarrow{d}} N(0, 6)</math>, i.e., their distributions converge to a normal distribution with mean 0 and variance 6 ([[Ronald Fisher|Fisher]], 1930).<ref name=JG/> The variance of the sample skewness is thus approximately <math>6/n</math> for sufficiently large samples. More precisely, in a random sample of size ''n'' from a normal distribution,<ref name=Duncan1997>Duncan Cramer (1997) Fundamental Statistics for Social Research. Routledge. {{isbn|9780415172042}} (p 85)</ref><ref>Kendall, M.G.; Stuart, A. (1969) ''The Advanced Theory of Statistics, Volume 1: Distribution Theory, 3rd Edition'', Griffin. {{isbn|0-85264-141-9}} (Ex 12.9)</ref> <math display="block"> \operatorname{var}(G_1)= \frac{6n ( n - 1 )}{ ( n - 2 )( n + 1 )( n + 3 ) } .</math> In normal samples, <math>b_1</math> has the smaller variance of the three estimators, with<ref name=JG/> <math display="block"> \operatorname{var}(b_1) < \operatorname{var} (g_1) < \operatorname{var}(G_1).</math> For non-normal distributions, <math>b_1</math>, <math>g_1</math> and <math>G_1</math> are generally [[Bias of an estimator|biased estimators]] of the population skewness <math>\gamma_1</math>; their expected values can even have the opposite sign from the true skewness. For instance, a mixed distribution consisting of very thin Gaussians centred at −99, 0.5, and 2 with weights 0.01, 0.66, and 0.33 has a skewness <math>\gamma_1</math> of about −9.77, but in a sample of 3 <math>G_1</math> has an expected value of about 0.32, since usually all three samples are in the positive-valued part of the distribution, which is skewed the other way. ==Applications== Skewness is a descriptive statistic that can be used in conjunction with the [[histogram]] and the normal [[quantile plot]] to characterize the data or distribution. Skewness indicates the direction and relative magnitude of a distribution's deviation from the normal distribution. With pronounced skewness, standard statistical inference procedures such as a [[confidence interval]] for a mean will be not only incorrect, in the sense that the true coverage level will differ from the nominal (e.g., 95%) level, but they will also result in unequal error probabilities on each side. Skewness can be used to obtain approximate probabilities and quantiles of distributions (such as [[value at risk]] in finance) via the [[Cornish–Fisher expansion]]. Many models assume normal distribution; i.e., data are symmetric about the mean. The normal distribution has a skewness of zero. But in reality, data points may not be perfectly symmetric. So, an understanding of the skewness of the dataset indicates whether deviations from the mean are going to be positive or negative. [[D'Agostino's K-squared test]] is a [[goodness-of-fit]] [[normality test]] based on sample skewness and sample kurtosis. ==Other measures of skewness== [[Image:Comparison mean median mode.svg|thumb|upright=1.35|Comparison of [[mean]], [[median]] and [[mode (statistics)|mode]] of two [[log-normal distribution]]s with the same medians and different skewnesses.]] Other measures of skewness have been used, including simpler calculations suggested by [[Karl Pearson]]<ref name="Capistrano1">{{cite web |url=http://www.stat.upd.edu.ph/s114%20cnotes%20fcapistrano/Chapter%2010.pdf |title=Archived copy |access-date=2010-04-09 |url-status=dead |archive-url=https://web.archive.org/web/20100705025706/http://www.stat.upd.edu.ph/s114%20cnotes%20fcapistrano/Chapter%2010.pdf |archive-date=5 July 2010}}</ref> (not to be confused with Pearson's moment coefficient of skewness, see above). These other measures are: ===Pearson's first skewness coefficient (mode skewness)=== The Pearson mode skewness,<ref>{{MathWorld|title=Pearson Mode Skewness|urlname=PearsonModeSkewness}}</ref> or first skewness coefficient, is defined as {{block indent | em = 1.5 | text = <span style="font-size:125%;">{{sfrac|[[mean]] − [[Mode (statistics)|mode]]|[[standard deviation]]}}</span>.}} ===Pearson's second skewness coefficient (median skewness)=== The Pearson median skewness, or second skewness coefficient,<ref name="MWPSK">{{MathWorld | title = Pearson's skewness coefficients | urlname = PearsonsSkewnessCoefficients}}</ref><ref name="amstat2011">{{cite journal | last1 = Doane | first1 = David P. | last2 = Seward | first2 = Lori E. | year = 2011 | title = Measuring Skewness: A Forgotten Statistic? | url = http://www.amstat.org/publications/jse/v19n2/doane.pdf | journal = Journal of Statistics Education | volume = 19 | issue = 2| pages = 1–18 | doi = 10.1080/10691898.2011.11889611 | doi-access = free }}</ref> is defined as {{block indent | em = 1.5 | text = <span style="font-size:125%;">{{sfrac|3 ([[mean]] − [[median]])|[[standard deviation]]}}</span>.}} Which is a simple multiple of the [[nonparametric skew]]. ===Quantile-based measures=== Bowley's measure of skewness (from 1901),<ref>Bowley, A. L. (1901). Elements of Statistics, P.S. King & Son, Laondon. Or in a later edition: BOWLEY, AL. "Elements of Statistics, 4th Edn (New York, Charles Scribner)."(1920).</ref><ref name=Kenney1962>Kenney JF and Keeping ES (1962) ''Mathematics of Statistics, Pt. 1, 3rd ed.'', Van Nostrand, (page 102).</ref> also called '''Yule's coefficient''' (from 1912)<ref>Yule, George Udny. An introduction to the theory of statistics. C. Griffin, limited, 1912.</ref><ref>{{cite journal | last1 = Groeneveld | first1 = Richard A | year = 1991 | title = An influence function approach to describing the skewness of a distribution | journal = The American Statistician | volume = 45 | issue = 2| pages = 97–102 | doi = 10.2307/2684367 | jstor = 2684367 }}</ref> is defined as: <math display="block">\frac{\frac{Q(3/4) + Q(1/4)}{2} - Q(1/2)}{\frac{Q(3/4) - Q(1/4)}{2}} = \frac{Q(3/4) + Q(1/4) - 2 Q(1/2)}{Q(3/4) - Q(1/4)},</math> where ''Q'' is the [[quantile function]] (i.e., the inverse of the [[cumulative distribution function]]). The numerator is difference between the average of the upper and lower quartiles (a measure of location) and the median (another measure of location), while the denominator is the [[semi-interquartile range]] <math>(Q(3/4)}-{Q(1/4))/2</math>, which for symmetric distributions is equal to the [[Average absolute deviation|MAD]] measure of [[statistical dispersion|dispersion]].{{citation needed|date=May 2024}} Other names for this measure are Galton's measure of skewness,<ref name=Johnson1994>{{harvp|Johnson, NL|Kotz, S|Balakrishnan, N|1994}} p. 3 and p. 40</ref> the Yule–Kendall index<ref name=Wilks1995>Wilks DS (1995) ''Statistical Methods in the Atmospheric Sciences'', p 27. Academic Press. {{isbn|0-12-751965-3}}</ref> and the quartile skewness,<ref>{{Cite web|url=http://mathworld.wolfram.com/Skewness.html|title=Skewness|last=Weisstein|first=Eric W.|website=mathworld.wolfram.com|language=en|access-date=2019-11-21}}</ref> Similarly, Kelly's measure of skewness is defined as<ref>{{cite web |url=http://www.math.ruh.ac.lk/~pubudu/app5.pdf |title=Applied Statistics I: Chapter 5: Measures of skewness |author=A.W.L. Pubudu Thilan |website=University of Ruhuna |page=21}}</ref> <math display="block">\frac{Q(9/10) + Q(1/10) - 2 Q(1/2)}{Q(9/10) - Q(1/10)}.</math> A more general formulation of a skewness function was described by Groeneveld, R. A. and Meeden, G. (1984):<ref name=Groeneveld1984>{{Cite journal | doi = 10.2307/2987742 | last1 = Groeneveld | first1 = R.A. | last2 = Meeden | first2 = G. | year = 1984 | title = Measuring Skewness and Kurtosis | jstor = 2987742| journal = The Statistician | volume = 33 | issue = 4| pages = 391–399 }}</ref><ref name=MacGillivray1992>MacGillivray (1992)</ref><ref name=Hinkley1975>Hinkley DV (1975) "On power transformations to symmetry", ''[[Biometrika]]'', 62, 101–111</ref> <math display="block"> \gamma( u ) = \frac{ Q( u ) +Q( 1 - u )-2Q( 1 / 2 ) }{Q( u ) -Q( 1 - u ) } </math> The function {{math|''γ''(''u'')}} satisfies {{math|−1 ≤ ''γ''(''u'') ≤ 1}} and is well defined without requiring the existence of any moments of the distribution.<ref name=Groeneveld1984/> Bowley's measure of skewness is {{math|''γ''(''u'')}} evaluated at {{math|1=''u'' = 3/4}} while Kelly's measure of skewness is {{math|''γ''(''u'')}} evaluated at {{math|1=''u'' = 9/10}}. This definition leads to a corresponding overall measure of skewness<ref name=MacGillivray1992>MacGillivray (1992)</ref> defined as the [[supremum]] of this over the range {{math|1/2 ≤ ''u'' < 1}}. Another measure can be obtained by integrating the numerator and denominator of this expression.<ref name=Groeneveld1984/> Quantile-based skewness measures are at first glance easy to interpret, but they often show significantly larger sample variations than moment-based methods. This means that often samples from a symmetric distribution (like the uniform distribution) have a large quantile-based skewness, just by chance. ===Groeneveld and Meeden's coefficient=== Groeneveld and Meeden have suggested, as an alternative measure of skewness,<ref name=Groeneveld1984 /> <math display="block"> \operatorname{skew}(X) = \frac{\mu - \nu}{ \operatorname E( | X - \nu | ) }, </math> where {{mvar|μ}} is the mean, {{mvar|ν}} is the median, {{math|{{abs|...}}}} is the [[absolute value]], and {{math|E()}} is the expectation operator. This is closely related in form to [[Skewness#Pearson.27s second skewness coefficient .28median skewness.29|Pearson's second skewness coefficient]]. ===L-moments=== Use of [[L-moment]]s in place of moments provides a measure of skewness known as the L-skewness.<ref name=hos:92>{{cite journal | last=Hosking | first= J.R.M. | year=1992 | title=Moments or L moments? An example comparing two measures of distributional shape | journal=The American Statistician | volume=46 | number=3 | pages=186–189 | jstor=2685210 | doi=10.2307/2685210}}</ref> ===Distance skewness=== A value of skewness equal to zero does not imply that the probability distribution is symmetric. Thus there is a need for another measure of asymmetry that has this property: such a measure was introduced in 2000.<ref>Szekely, G.J. (2000). "Pre-limit and post-limit theorems for statistics", In: ''Statistics for the 21st Century'' (eds. [[C. R. Rao]] and G. J. Szekely), Dekker, New York, pp. 411–422.</ref> It is called '''distance skewness''' and denoted by {{math|dSkew}}. If ''X'' is a random variable taking values in the {{mvar|d}}-dimensional Euclidean space, {{mvar|X}} has finite expectation, {{math|''X''{{'}}}} is an independent identically distributed copy of {{mvar|X}}, and <math>\|\cdot\|</math> denotes the norm in the Euclidean space, then a simple ''measure of asymmetry'' with respect to location parameter {{mvar|θ}} is <math display="block"> \operatorname{dSkew}(X) := 1 - \frac{\operatorname{E}\|X-X'\|}{\operatorname{E}\|X+X'-2 \theta\|} \text{ if } \Pr(X=\theta)\ne 1 </math> and {{math|1=dSkew(''X'') := 0}} for {{math|1=''X'' = ''θ''}} (with probability 1). Distance skewness is always between 0 and 1, equals 0 if and only if ''X'' is diagonally symmetric with respect to {{mvar|θ}} ({{mvar|X}} and {{math|2''θ'' − ''X''}} have the same probability distribution) and equals 1 if and only if ''X'' is a constant ''c'' (<math>c \neq \theta</math>) with probability one.<ref>Szekely, G. J. and Mori, T. F. (2001) "A characteristic measure of asymmetry and its application for testing diagonal symmetry", ''Communications in Statistics – Theory and Methods'' 30/8&9, 1633–1639.</ref> Thus there is a simple consistent [[statistical test]] of diagonal symmetry based on the '''sample distance skewness''': <math display="block"> \operatorname{dSkew}_n(X) := 1 - \frac{\sum_{i,j} \|x_i-x_j\| }{\sum_{i,j} \|x_i+x_j-2\theta \|}. </math> ===Medcouple=== The [[medcouple]] is a scale-invariant robust measure of skewness, with a [[breakdown point]] of 25%.<ref name="Brys2004">{{cite journal |author=G. Brys |author2=M. Hubert|author2-link=Mia Hubert |author3=A. Struyf |date=November 2004 |title=A Robust Measure of Skewness |journal=Journal of Computational and Graphical Statistics |volume=13 |issue=4 |pages=996–1017 |doi=10.1198/106186004X12632|s2cid=120919149 }}</ref> It is the [[median]] of the values of the kernel function <math display="block"> h(x_i, x_j) = \frac{ (x_i - x_m) - (x_m - x_j)}{x_i - x_j} </math> taken over all couples <math>(x_i, x_j)</math> such that <math>x_i \geq x_m \geq x_j</math>, where <math>x_m</math> is the median of the [[sample (statistics)|sample]] <math>\{x_1, x_2, \ldots, x_n\}</math>. It can be seen as the median of all possible quantile skewness measures. == See also == {{Portal|Mathematics}} *[[Bragg peak]] *[[Coskewness]] *[[Kurtosis]] *[[Shape parameter]]s *[[Skew normal distribution]] *[[Skewness risk]] == References == === Citations === {{Reflist}} === Sources === {{refbegin}} * {{cite book|author1=Johnson, NL|author2=Kotz, S|author3=Balakrishnan, N|year=1994|title=Continuous Univariate Distributions|volume=1|edition=2|publisher=Wiley|isbn=0-471-58495-9}} * {{cite journal | last1 = MacGillivray | first1 = HL | year = 1992 | title = Shape properties of the g- and h- and Johnson families | journal = Communications in Statistics – Theory and Methods | volume = 21 | issue = 5| pages = 1244–1250 | doi = 10.1080/03610929208830842 }} * Premaratne, G., Bera, A. K. (2001). Adjusting the Tests for Skewness and Kurtosis for Distributional Misspecifications. Working Paper Number 01-0116, University of Illinois. Forthcoming in Comm in Statistics, Simulation and Computation. 2016 1–15 * Premaratne, G., Bera, A. K. (2000). Modeling Asymmetry and Excess Kurtosis in Stock Return Data. Office of Research Working Paper Number 00-0123, University of Illinois. * [https://ssrn.com/abstract=2590356 Skewness Measures for the Weibull Distribution] {{refend}} == External links == {{Wikiversity}} {{Commons category|Skewness (statistics)}} *{{springer|title=Asymmetry coefficient|id=p/a013590}} *[http://petitjeanmichel.free.fr/itoweb.petitjean.skewness.html An Asymmetry Coefficient for Multivariate Distributions] by Michel Petitjean *[http://repositories.cdlib.org/cgi/viewcontent.cgi?article=1017&context=ucsdecon On More Robust Estimation of Skewness and Kurtosis] Comparison of skew estimators by Kim and White. *[http://dahoiv.net/master/index.html Closed-skew Distributions — Simulation, Inversion and Parameter Estimation] {{-}} {{Statistics|descriptive}} {{Theory_of_probability_distributions}} [[Category:Moments (mathematics)]] [[Category:Statistical deviation and dispersion]]
Summary:
Please note that all contributions to Niidae Wiki may be edited, altered, or removed by other contributors. If you do not want your writing to be edited mercilessly, then do not submit it here.
You are also promising us that you wrote this yourself, or copied it from a public domain or similar free resource (see
Encyclopedia:Copyrights
for details).
Do not submit copyrighted work without permission!
Cancel
Editing help
(opens in new window)
Templates used on this page:
Template:-
(
edit
)
Template:Block indent
(
edit
)
Template:Citation needed
(
edit
)
Template:Cite book
(
edit
)
Template:Cite journal
(
edit
)
Template:Cite web
(
edit
)
Template:Commons category
(
edit
)
Template:For
(
edit
)
Template:Harvp
(
edit
)
Template:Isbn
(
edit
)
Template:Math
(
edit
)
Template:MathWorld
(
edit
)
Template:Mvar
(
edit
)
Template:Portal
(
edit
)
Template:Refbegin
(
edit
)
Template:Refend
(
edit
)
Template:Reflist
(
edit
)
Template:Short description
(
edit
)
Template:Springer
(
edit
)
Template:Statistics
(
edit
)
Template:Theory of probability distributions
(
edit
)
Template:Use American English
(
edit
)
Template:Use dmy dates
(
edit
)
Template:Visible anchor
(
edit
)
Template:Wikiversity
(
edit
)
Search
Search
Editing
Skewness
Add topic