Skip to main content

Convert Z-Scores to P-Values and Critical Z

Convert X to Z and find p-values, or get critical Z from p. Supports standard and custom normal distributions with shaded graph.

Last Updated: February 13, 2026

One-Tailed vs Two-Tailed: Choose the Right Tail

Z-to-p conversions require specifying tail direction. Selecting the wrong tail doubles or halves your p-value and can flip your hypothesis test conclusion. Before computing, match your alternative hypothesis to a tail type.

Left-Tailed: P(Z ≤ z)

Use left-tailed when your alternative hypothesis claims the parameter is less than the null value. "The new process reduces defect rate" or "Treatment lowers blood pressure" requires a left-tail test. The p-value equals the cumulative probability from negative infinity up to your z-score.

Right-Tailed: P(Z ≥ z)

Use right-tailed when your alternative claims the parameter is greater than the null value. "The drug increases survival time" or "Marketing campaign raised conversion rate" needs a right-tail test. The p-value equals 1 − CDF(z), the probability above your z-score.

Two-Tailed: P(|Z| ≥ |z|)

Use two-tailed when your alternative states "not equal to" without specifying direction. "The mean differs from the claimed value" or "There is a difference" requires a two-tail test. The p-value sums both extreme tails: 2 × (1 − Φ(|z|)). Two-tailed tests are more conservative—the same |z| produces a larger p-value than either one-tailed test.

Common mistake: Switching tail direction after seeing results. Your hypothesis determines the tail before data collection. Changing it post-hoc inflates false positive rates and undermines statistical validity.

Critical Z for Any Alpha (Replace the Z-Table)

Critical z-values define rejection regions for hypothesis tests. Instead of hunting through printed z-tables, enter your significance level α and let the calculator return the exact cutoff.

Common Critical Values

Confidence / αTwo-Tailed zOne-Tailed z
90% / α = 0.10±1.6451.282
95% / α = 0.05±1.9601.645
99% / α = 0.01±2.5762.326
99.9% / α = 0.001±3.2913.090

Using p → Z Mode

Enter your target p-value (e.g., 0.05) and select tail type. The calculator inverts the CDF to return the corresponding z. For two-tailed at α = 0.05, you get ±1.96. For one-tailed right at α = 0.05, you get +1.645. These values define rejection thresholds: if your test statistic exceeds the critical z, reject the null.

Non-Standard Alpha Levels

Not all tests use 0.05. Medical trials often use α = 0.01 for safety margins. Physics discovery claims require 5σ (p < 2.87 × 10⁻⁷). Exploratory research might use α = 0.10. Enter any α to get the corresponding critical z without interpolating tables.

Convert z ↔ p With Correct Rounding

Z-scores and p-values are two sides of the same coin. Converting between them requires the standard normal CDF (Φ) and its inverse. The calculator handles both directions with high precision.

Z → p: Finding Probability from Z-Score

Enter a z-score and select tail type. The calculator computes Φ(z) for left-tail, 1 − Φ(z) for right-tail, or 2 × (1 − Φ(|z|)) for two-tail. Results display to several decimal places—more precision than any printed table provides.

p → Z: Finding Z-Score from Probability

Enter a p-value and select tail type. The calculator inverts the CDF to find the z-score where the cumulative (or tail) probability equals your input. This is essential for constructing confidence intervals or finding rejection boundaries.

X → z → p: Full Conversion Chain

If you have a raw score x with known μ and σ, use X → z → p mode. The calculator first standardizes using z = (x − μ) / σ, then computes the tail probability. This connects raw measurements to statistical significance in one step.

Precision note: Results are accurate to many decimal places, matching R, Python SciPy, and professional statistical software. For extremely small p-values (below 10⁻¹⁰), minor rounding may appear.

Tail Shading Preview: What Your p-Value Represents

The interactive bell curve chart visualizes exactly which area corresponds to your p-value. Shading confirms that your tail selection matches your hypothesis and helps catch input errors before you act on results.

Left-Tail Shading

The region from negative infinity up to your z-score fills with color. The shaded area equals the cumulative probability P(Z ≤ z). Smaller (more negative) z-scores produce larger shaded areas.

Right-Tail Shading

The region from your z-score out to positive infinity fills. The shaded area equals P(Z ≥ z) = 1 − Φ(z). Larger z-scores produce smaller shaded areas—consistent with smaller p-values for more extreme observations.

Two-Tail Shading

Both extreme tails shade simultaneously: below −|z| and above +|z|. The total shaded area equals the two-tailed p-value. The center of the curve (the "acceptance region") remains unshaded, representing outcomes consistent with the null hypothesis.

Between-Bounds Shading

When computing interval probability P(a ≤ Z ≤ b), only the region between your lower and upper bounds shades. The rest of the curve stays clear. This mode answers "what fraction falls within this range" questions directly.

Decision Rule in Plain English (p vs Alpha)

Hypothesis testing compares your computed p-value against a pre-chosen significance level α. The decision rule is straightforward: if p ≤ α, reject the null; if p > α, fail to reject.

Rejection Interpretation

When p ≤ α, the observed data are unlikely under the null hypothesis—unlikely enough that you conclude the null is probably false. You "reject H₀" and accept the alternative. This does not prove the alternative true, but indicates sufficient evidence against the null at your chosen confidence level.

Failure to Reject

When p > α, the data are consistent with the null hypothesis—not proof that the null is true, but insufficient evidence to reject it. "Fail to reject H₀" is the correct phrasing; "accept H₀" overstates the conclusion.

Practical vs Statistical Significance

A small p-value indicates statistical significance—the effect is unlikely to be noise. It says nothing about effect size or practical importance. A tiny difference can be "significant" with large samples, yet irrelevant in practice. Always report effect sizes alongside p-values.

P-value misreading: A p-value of 0.03 does not mean "3% chance the null is true." It means "if the null were true, data this extreme would occur 3% of the time." The distinction matters for proper interpretation.

Z-to-P Quick Checks

What does z = 0 mean?

Z = 0 means the observation equals the mean exactly. Left-tail p = 0.5, right-tail p = 0.5, two-tail p = 1.0. No deviation from the null, no evidence against it.

Why is my two-tailed p-value double the one-tailed?

Two-tailed tests count extremity in both directions. If the right-tail p is 0.025, the two-tailed p is 0.05 because you also count equally extreme negative z. This makes two-tailed tests more conservative.

When should I use the t-distribution instead?

Use t when the population σ is unknown and estimated from sample data, especially with n < 30. The t-distribution has heavier tails, producing larger p-values and wider confidence intervals to account for estimation uncertainty. As sample size grows, t converges to z.

How do I interpret z = 3 or higher?

Z = 3 corresponds to about 3 standard deviations from the mean. Only 0.3% of values under a normal distribution fall beyond ±3σ. Values this extreme provide strong evidence against the null. In quality control, 3σ events are rare; in physics, 5σ (z ≈ 5) is the threshold for "discovery."

Can I use z-tests for proportions?

Yes—for large samples, the sampling distribution of a proportion is approximately normal. Compute z = (p̂ − p₀) / √(p₀(1−p₀)/n), then use this calculator to find the p-value. This is the basis of one-proportion z-tests.

What if my data are not normal?

Z-based inference assumes normality. For non-normal data, consider transformations (log, square root), non-parametric tests (Wilcoxon, Mann-Whitney), or bootstrapping. The Central Limit Theorem justifies z-tests for large samples even with non-normal populations, but check assumptions first.

Limitations

Normality assumed: Z-scores and z-based p-values require underlying normal distributions. Non-normal data may produce misleading results.

Known σ required: Z-tests assume you know the population standard deviation. Use t-tests when estimating σ from sample data.

P-value interpretation: A p-value is not the probability the null is true. It is the probability of observing data this extreme if the null were true.

Multiple comparisons: Running many tests inflates false positive rates. Apply corrections (Bonferroni, FDR) when testing multiple hypotheses.

Disclaimer: This calculator is for educational and informational purposes. Verify results with professional statistical software (R, Python SciPy, SAS, SPSS) for research, regulatory submissions, or critical decisions. Consult qualified statisticians for important analyses.

Sources

Frequently Asked Questions

Common questions about z-scores, p-values, tail types, hypothesis testing, and statistical accuracy.

What is a z-score and how do I interpret it?

A z-score is the number of standard deviations a value lies from the mean of a normal distribution. It standardizes any value using the formula z = (x - μ) / σ. A z-score of 0 means the value equals the mean, z = 1 means one standard deviation above the mean, z = -2 means two standard deviations below. Large absolute values (|z| > 2) indicate the value is far from the mean and relatively rare. By the empirical rule, about 95% of values fall within |z| < 2, and 99.7% within |z| < 3. Z-scores allow you to compare values from different distributions on a common scale and determine how unusual a value is.

What's the difference between left, right, and two-tailed p-values?

Left-tail p-value is P(Z ≤ z), the probability of getting a value less than or equal to z—used when testing if a parameter is lower than expected (H₁: μ < μ₀). Right-tail p-value is P(Z ≥ z), the probability of getting a value greater than or equal to z—used when testing if a parameter is higher than expected (H₁: μ > μ₀). Two-tailed p-value is 2 × min{P(Z ≤ z), P(Z ≥ z)} or equivalently 2 × (1 - Φ(|z|)), representing the probability of observing a value at least as extreme in either direction—used for non-directional tests (H₁: μ ≠ μ₀). The choice depends on your research hypothesis: one-sided hypotheses use left or right tail, while two-sided hypotheses use two-tailed.

How do I compute the area between two values?

To find the probability that a value falls between two bounds a and b, use the between-bounds feature by entering lower and upper values. The calculator computes P(a ≤ Z ≤ b) = Φ(b) - Φ(a), where Φ is the standard normal CDF. For example, to find what fraction of values fall between z = -1 and z = 1, enter lower = -1 and upper = 1 to get approximately 0.68 (68%). If working with raw scores instead of z-scores, enter your x-values along with μ and σ, and the calculator will standardize them first. The chart shades the exact interval, making it easy to visualize and communicate the result.

When should I use standard normal vs custom μ, σ?

Use the standard normal distribution (μ = 0, σ = 1) when you already have z-scores or standardized test statistics from hypothesis tests, or when you want to work with published z-tables and critical values. Use custom normal (specify μ and σ) when working with raw scores from a population with known mean and standard deviation—for example, test scores with μ = 75 and σ = 10, or manufacturing measurements with μ = 50.0 and σ = 0.3. The custom option automatically converts your raw x-values to z-scores using z = (x - μ) / σ, then computes probabilities. Both approaches yield the same probabilities for corresponding values; the difference is just the scale.

How do I interpret a p-value in hypothesis testing?

The p-value is the probability, assuming the null hypothesis is true, of observing a test statistic at least as extreme as the one you actually observed. It's NOT the probability that the null hypothesis is true. A small p-value (typically < 0.05 or < 0.01) suggests that your observed data would be unlikely under the null model, providing evidence against the null hypothesis. Compare the p-value to your chosen significance level α: if p < α, reject the null hypothesis; if p ≥ α, fail to reject. Remember that 'statistically significant' doesn't always mean 'practically important'—effect size and context matter. Also, failing to reject the null doesn't prove it's true; it just means you lack sufficient evidence against it.

How accurate are the numerical computations?

We use high-precision numerical approximations for the standard normal CDF (Φ) and its inverse, based on well-established statistical algorithms similar to those in R, Python SciPy, and Excel. The results are accurate to many decimal places and suitable for all typical statistical work, including research, quality control, finance, and engineering. For p-values down to 0.0001 or even lower, the calculator matches published statistical tables. Extremely far into the tails (probabilities below 10⁻¹⁰), small numerical errors may appear, but these are well beyond the range of practical statistical inference. The z-distribution assumes perfect normality; if your data are not normally distributed, consider data transformations, non-parametric methods, or robust statistics.

Related Math & Statistics Tools

Z-Score to P-Value Calculator: Critical Z & Plots