Skip to main content

Descriptive Statistics With Charts and Outlier Checks

Compute mean, median, mode, variance, standard deviation, skewness, kurtosis, and outliers. Compare datasets or calculate weighted statistics with visual charts.

Last Updated: February 13, 2026

Descriptive statistics condense raw numbers into a handful of measures that reveal a dataset's center, spread, and shape—the foundation for any data analysis, from quarterly sales reviews to clinical trial reports. A marketing analyst exported 500 customer purchase amounts and wanted a quick summary. She computed the mean ($127), median ($89), and standard deviation ($156). The mean seemed high, the median much lower—red flag. The common mistake is reporting only the mean for skewed data: one whale customer spending $4,200 pulled the average up while most buyers spent under $100. When reading results, compare mean vs median: if they diverge, the distribution is skewed, and the median better represents the typical value.

Dataset Snapshot: Mean, Median, Mode

The mean (average) sums all values and divides by count. It uses every data point, making it sensitive to extremes. One outlier can drag the mean far from where most values cluster.

The median is the middle value after sorting. Half the data falls below, half above. It ignores how far outliers lie from the center, so it stays stable even when extreme values appear. For income, house prices, or anything with a long tail, median tells the truer story.

The mode is the most frequent value. A dataset can have one mode (unimodal), two (bimodal), or many (multimodal). When all values are unique, no mode exists. Mode shines for categorical data—like the most common shoe size sold—where arithmetic averages don't apply.

Quick formulas:

• Mean: μ = (Σxᵢ) / n

• Median: middle value after sorting (average of two middles if n is even)

• Mode: value(s) with highest frequency

Spread Breakdown: Variance, SD, IQR

Variance measures how far values scatter from the mean by averaging squared deviations. Squaring ensures positive and negative deviations don't cancel out, but it leaves units squared (dollars², cm²), making interpretation awkward.

Standard deviation (SD) is the square root of variance, restoring original units. A small SD means data clusters tightly around the mean; a large SD means wide scatter. For normally distributed data, roughly 68% falls within ±1 SD, 95% within ±2 SD, and 99.7% within ±3 SD.

The interquartile range (IQR) measures the middle 50% of data: Q3 minus Q1. Unlike SD, IQR ignores extremes entirely, making it robust to outliers. Box plots visualize IQR directly, with whiskers extending to 1.5×IQR beyond the quartiles.

Spread formulas:

• Variance: σ² = Σ(xᵢ − μ)² / n

• Std Dev: σ = √(variance)

• IQR = Q3 − Q1

Shape Signals: Skewness and Kurtosis

Skewness quantifies asymmetry. A positive (right) skew means a longer right tail—think income distributions where a few high earners stretch the mean above the median. Negative (left) skew means a longer left tail. Zero skewness indicates symmetry, like a bell curve.

Kurtosis measures tail heaviness relative to a normal distribution. High kurtosis (leptokurtic) means fatter tails and more extreme outliers than normal. Low kurtosis (platykurtic) means thinner tails and fewer outliers. A normal distribution has kurtosis of 3, or excess kurtosis of 0.

Together, skewness and kurtosis reveal whether standard parametric tests apply. Many statistical methods assume normality—if skewness exceeds ±1 or excess kurtosis exceeds ±2, consider transformations or non-parametric alternatives.

Rule of thumb: If |skewness| > 1 or |excess kurtosis| > 2, the distribution deviates substantially from normal and may require special handling.

Outlier Flags and Robust Alternatives

Outliers are data points far from the bulk of observations. The z-score method flags values beyond ±3 standard deviations from the mean. The IQR method flags anything below Q1 − 1.5×IQR or above Q3 + 1.5×IQR. Neither method is perfect—each catches different types of extremes.

An outlier isn't automatically an error. It might be a data entry mistake, a measurement glitch, or a genuine extreme value worth investigating. Before deleting outliers, understand their source. Legitimate extremes carry real information; erroneous ones distort analysis.

Robust statistics resist outlier influence. The median and IQR are robust; the mean and SD are not. Trimmed means (dropping extreme percentiles) or Winsorized statistics (capping extremes) offer middle-ground approaches when you suspect contamination but don't want to discard data entirely.

Caution: Automatic outlier removal can bias results. Always examine flagged values in context before deciding to exclude them.

Chart Readout: Histogram and Box Plot

A histogram bins continuous data into intervals and shows frequency per bin as bar height. The shape reveals distribution type: bell-shaped, skewed, bimodal, or uniform. Bin width matters—too few bins hide detail, too many create noise. A common rule: use √n bins, where n is sample size.

A box plot (box-and-whisker) compresses distribution into five numbers: minimum, Q1, median, Q3, and maximum. The box spans IQR; the line inside marks the median. Whiskers extend to 1.5×IQR, with individual points beyond marked as potential outliers. Box plots excel at comparing multiple groups side by side.

Always visualize before computing summary statistics. Two datasets with identical mean, median, and SD can have completely different shapes—Anscombe's quartet famously demonstrates this. Charts reveal patterns that numbers alone miss: clusters, gaps, outliers, multimodality.

Visualization rule: Histograms for detailed shape, box plots for quick comparison across groups. Use both when space allows.

Stats Summary Q&A

When should I use median instead of mean?

Use the median when data is skewed or contains outliers. Income, house prices, response times, and customer lifetime values often have long tails where the median better represents the typical case. If mean and median diverge significantly, report both and note the skew.

What's the difference between population and sample standard deviation?

Population SD divides by n; sample SD divides by n−1 (Bessel's correction). The n−1 adjustment makes sample SD an unbiased estimator of population SD. If your data is the entire population, use n. If it's a sample from a larger population, use n−1.

How do I compare two datasets with different units?

Use the coefficient of variation (CV = SD / mean × 100%). CV expresses variability relative to the mean, making it unit-free. A dataset with SD = 10 and mean = 100 (CV = 10%) is relatively less variable than one with SD = 5 and mean = 20 (CV = 25%).

Can a dataset have no mode?

Yes. If every value appears exactly once, no mode exists. If all values appear equally often, some define every value as a mode, others say none exists. Mode is most useful when some values repeat significantly more than others.

Why does my histogram look different with different bin widths?

Bin width is a tuning parameter, not a property of the data. Wide bins smooth out detail; narrow bins show noise. There's no single correct choice. Try multiple widths and use the one that reveals the underlying pattern without excessive chatter.

Limitations & Assumptions

• Data Quality: Descriptive statistics summarize what you provide. They cannot detect data entry errors, measurement bias, or sampling issues. Validate your data before computing summaries.

• Sample vs Population: Sample statistics estimate population parameters with inherent uncertainty. Descriptive measures don't quantify that uncertainty—inferential statistics (confidence intervals, hypothesis tests) do.

• Outlier Sensitivity: Mean and SD can be heavily influenced by outliers. Always check for extremes and consider robust alternatives (median, IQR) when outliers are present.

• Shape Matters: Summary statistics can mask important features. Two datasets with identical mean, median, and SD can have completely different distributions. Visualize your data, don't just compute numbers.

Disclaimer: This calculator demonstrates descriptive statistics for educational purposes. For research publications, business decisions, or regulatory submissions, use validated statistical software (R, Python, SPSS, SAS) and consult qualified statisticians.

Sources & References

Methods and formulas follow standard statistical references:

Frequently Asked Questions

Common questions about variance, standard deviation, skewness, kurtosis, weighted statistics, outliers, and effect sizes.

What's the difference between variance and standard deviation?

Variance measures the average squared deviation from the mean, expressed in squared units (e.g., dollars², cm²), which can be less intuitive to interpret. Standard deviation is simply the square root of variance and expresses variability in the same units as your original data, making it easier to understand. For example, if you're measuring heights in centimeters with a variance of 100 cm², the standard deviation is 10 cm—meaning heights typically vary by about 10 cm from the average. Both measure spread, but standard deviation is preferred for interpretation because it's in meaningful units. In statistical formulas and calculations, variance is often used because it has nice mathematical properties (e.g., variances add when combining independent random variables).

What does skewness indicate?

Skewness measures the asymmetry of a data distribution, indicating whether values are concentrated on one side with a tail extending in the opposite direction. Positive skewness (right-skewed) means the distribution has a longer tail on the right side, with most values clustered on the left and the mean greater than the median—common in income distributions, house prices, and response times. Negative skewness (left-skewed) has a longer left tail with values clustered on the right and mean less than median—seen in test scores with a ceiling effect or age at retirement. Zero or near-zero skewness indicates a symmetric distribution like the normal distribution, where mean ≈ median ≈ mode. Skewness values |skew| < 0.5 are considered fairly symmetric, 0.5-1.0 are moderately skewed, and > 1.0 are highly skewed. Understanding skewness helps choose appropriate statistical methods—for example, the median is often preferred over the mean for skewed data.

What does kurtosis mean?

Kurtosis measures how heavy or light the tails of a distribution are compared to a normal distribution, indicating the likelihood of extreme values (outliers). A normal distribution has kurtosis = 3 (or excess kurtosis = 0). High kurtosis (> 3, or excess > 0) indicates 'heavy tails' with more extreme values and a sharper peak than normal—common in financial returns, where rare but extreme events occur. Low kurtosis (< 3, or excess < 0) indicates 'light tails' with fewer outliers and a flatter distribution than normal. Platykurtic (low kurtosis) distributions have values clustered near the mean with few extremes, like uniform distributions. Leptokurtic (high kurtosis) distributions have long tails and many outliers, requiring robust statistical methods. In practice, kurtosis helps assess risk—high kurtosis in stock returns means more frequent crashes or booms than a normal model would predict.

When should I use weighted statistics?

Use weighted statistics when different observations contribute unequally to the analysis—i.e., some data points have more importance, frequency, or reliability than others. Common scenarios include: (1) Grade Point Average (GPA): weight grades by credit hours (a 4-credit A counts more than a 1-credit A). (2) Survey data: weight responses by the number of people each respondent represents (e.g., demographic surveys where one respondent represents 1000 people in that group). (3) Investment portfolios: weight returns by the dollar amount invested in each asset. (4) Quality scores: weight ratings by reliability or confidence (e.g., expert opinions weighted higher than novice). (5) Grouped data: when you have frequency counts rather than raw values (e.g., 10 students scored 85, 15 scored 90). The weighted mean formula is Σ(w_i · x_i) / Σw_i, where w_i are the weights. Using regular (unweighted) statistics when weights matter can produce misleading averages that don't reflect the true center of your data.

What's considered an outlier in this tool?

This tool uses two complementary methods to detect outliers: (1) Z-score method: A value is flagged as an outlier if it's more than 3 standard deviations away from the mean (|z| > 3). This works well for approximately normal distributions and identifies extreme values in terms of standard deviations. For example, if mean = 100, σ = 10, then values below 70 or above 130 are outliers. (2) IQR (Interquartile Range) method: A value is an outlier if it's below Q1 - 1.5×IQR or above Q3 + 1.5×IQR, where IQR = Q3 - Q1. This method is robust to non-normal distributions and doesn't assume any particular shape. The tool reports outliers detected by either method. Not all outliers are errors—some represent genuine extreme cases (e.g., a billionaire in income data, a genius in IQ scores). Investigate outliers to determine if they're measurement errors, data entry mistakes, or valid extreme observations that deserve special attention or separate analysis.

How do I interpret Cohen's d?

Cohen's d is a standardized effect size that measures the difference between two group means in units of standard deviation: d = (mean₁ - mean₂) / pooled_σ. It's 'standardized' because it's unit-free, making it comparable across different measurements and studies. Interpretation guidelines: |d| < 0.2 = negligible difference (groups are nearly identical), 0.2-0.5 = small effect (noticeable but subtle difference), 0.5-0.8 = medium effect (moderate practical significance), > 0.8 = large effect (substantial difference that's obvious in practice). For example, d = 0.5 means the groups differ by half a standard deviation—a medium effect where about 69% of one group scores above the mean of the other group. Cohen's d is used to assess practical significance beyond statistical significance—a statistically significant p-value with d = 0.1 may not be meaningful in practice, while d = 0.9 indicates a large real-world difference even if the sample is small. In clinical trials, d > 0.5 often indicates clinically significant improvement.

Related Statistics Calculators

Descriptive Stats Calculator: Mean to Outliers