Skewness Calculator
Calculate the skewness of a dataset to measure distribution asymmetry. Includes Fisher-Pearson sample skewness, Pearson method, and Bowley quartile-based skewness.
Skewness (Fisher-Pearson)
—
Distribution Shape —
n —
Extended More scenarios, charts & detailed breakdown ▾
Pearson's Skewness 1 (mean-mode)/SD
—
Pearson's Skewness 2 (3×(mean-median)/SD) —
Mean —
Median —
Professional Full parameters & maximum detail ▾
Skewness Result
Skewness G1 —
Mean —
Median —
Statistical Implications
Mean vs Median —
Recommended Summary Statistic —
How to Use This Calculator
- Enter your data as a comma-separated list (at least 3 values).
- Read the Fisher-Pearson sample skewness G1 and distribution shape instantly.
- Use Pearson's Skewness tab for the simpler (mean−median)/SD and (mean−mode)/SD formulas.
- Use Sample Skewness tab for G1 with its standard error and Z-score for significance testing.
- Use Bowley's Skewness tab for the quartile-based skewness — more robust to outliers.
- Professional tier adds mean vs median comparison and recommended summary statistic.
Formula
G1 (Fisher-Pearson): [n/((n-1)(n-2))] × Σ[(xᵢ−x̄)³/s³]
Pearson's 2nd: 3(mean − median) / SD
Example
Data: [2, 3, 3, 4, 4, 4, 5, 6, 10]. The value 10 creates right skew. G1 ≈ 1.3 — highly right-skewed: use median instead of mean.
Frequently Asked Questions
- Skewness is a measure of the asymmetry of a probability distribution. A symmetric distribution (like the normal bell curve) has skewness = 0. A positively skewed distribution has a longer tail extending to the right — the bulk of values cluster at the lower end with a few very large outliers pulling the tail rightward. A negatively skewed distribution has a longer tail extending to the left — values cluster at the upper end with a few very small outliers pulling the tail leftward. Skewness tells you several things: first, whether your data has outliers on one side; second, whether the mean or median is a better measure of center; and third, which statistical tests and models are appropriate. Income distributions are the classic example of positive skewness — most people earn near the median, but a small number of very high earners pull the mean far above the median and create a long right tail. Exam scores in easy tests are often negatively skewed — most students score near the maximum, with a few very low scores pulling the tail left. Karl Pearson formalized skewness in 1895 as part of his system of frequency curves.
- Positive skewness (right skew) means the distribution has a longer tail on the right side. Most values cluster below the mean, and the mean is pulled above the median by extreme high values. The tail extends toward larger numbers. Common examples: income distributions, house prices, survival times for diseases, stock returns, city population sizes, and word frequencies in language. Negative skewness (left skew) means the distribution has a longer tail on the left side. Most values cluster above the mean, and the mean is pulled below the median by extreme low values. The tail extends toward smaller numbers. Common examples: age at death in developed countries (most people die in their 70s-80s, with the tail extending left toward early deaths), scores on easy tests (most students score high, few score very low), and reaction times with a left floor imposed by minimum possible speed. A useful memory trick: the direction of skew is the direction of the long tail, not the peak. A right-skewed distribution has a peak on the left with a tail going right — this confuses many beginners who expect the peak to indicate the direction.
- The mean is sensitive to extreme values (outliers) in a way the median is not. In a skewed distribution, the mean is pulled toward the long tail and away from the typical value. In a highly right-skewed income distribution with most people earning 40,000-60,000 and a few billionaires, the mean might be 80,000 while the median is 52,000 — the mean is not a useful description of a typical person's income. When skewness is high (|skewness| > 1), the median plus interquartile range (IQR) gives a more informative and robust description of the data than mean plus standard deviation. Standard deviation is also inflated by outliers and tail values. For approximately symmetric data (|skewness| < 0.5), the mean and median are close, and mean + SD is appropriate. For moderately skewed data (0.5 to 1), report both. Skewness also affects statistical tests: many parametric tests (t-test, ANOVA) assume normally distributed data, and severe skewness violates this assumption. In such cases, use nonparametric alternatives (Mann-Whitney U, Kruskal-Wallis) or transform the data (log, square root) to reduce skewness before applying parametric tests.
- Standard interpretation thresholds for the Fisher-Pearson sample skewness (G1): |skewness| < 0.5 is approximately symmetric — the distribution is close enough to symmetric for most purposes, and mean and SD are appropriate summaries. |skewness| between 0.5 and 1 indicates moderate skewness — visible asymmetry, worth noting; consider reporting both mean and median; parametric tests still generally robust for large samples. |skewness| > 1 indicates high skewness — significant asymmetry; median and IQR are preferred summaries; parametric tests may be inappropriate without transformation; the distribution departs substantially from normality. |skewness| > 2 is extreme skewness — found in distributions like the lognormal, Pareto, or exponential distributions, common in financial returns, internet traffic, and natural phenomena following power laws. Note that these thresholds are guidelines, not firm rules. The sample size matters: in small samples (n < 30), the standard error of skewness is large (roughly sqrt(6/n)), making the estimate unreliable. A formal test of skewness uses the Z-score = G1 / SE_skewness; values beyond ±2 suggest statistically significant skewness at the 5% level.
- Yes, data transformations are a standard tool for reducing skewness and making data more suitable for parametric statistical methods. For positive (right) skew, common transformations include: logarithm (log(x) or log(x+1) for zero values) — the most common transformation for exponential or multiplicative processes, effective when values span several orders of magnitude; square root (sqrt(x)) — milder transformation, useful for count data and moderate right skew; inverse (1/x) — strong transformation for highly skewed data. For negative (left) skew, reflect the data first (max(x) − x or c − x for some constant c), apply a right-skew transformation, then re-reflect. Box-Cox transformation is a systematic family that includes log and power transformations as special cases, with a parameter lambda chosen to optimize normality. The Johnson transformation system can handle virtually any skewness pattern. Important cautions: transformations change interpretation — results must be back-transformed and interpreted on the original scale; log(mean) does not equal mean(log), so the geometric mean corresponds to the median of the log-transformed data; always report which transformation was applied and justify the choice. Transformations should be chosen for scientific reasons, not just to pass normality tests.
Related Calculators
Sources & References (5) ▾
- Pearson K — Contributions to the Mathematical Theory of Evolution II: Skew Variation (1895) — Royal Society
- OpenStax Statistics — Chapter 2: Descriptive Statistics — OpenStax
- NIST/SEMATECH e-Handbook — Measures of Shape: Skewness and Kurtosis — NIST
- Walpole R, Myers R, Myers S — Probability and Statistics for Engineers and Scientists, 9th Ed. — Pearson
- Joanes D N & Gill C A — Comparing Measures of Sample Skewness and Kurtosis (1998) — Journal of the Royal Statistical Society Series D