Z-Score Calculator: Percentile, CDF & Normal Distribution

Calculate Z-score, percentile rank, tail probability, shaded area under the curve, inverse Z for a given percentile, confidence intervals, and critical values. Visualize everything on interactive normal distribution graphs.

Z-Score & Percentile PDF / CDF Probability Inverse Z (Percentile→Z) Confidence Intervals Batch Z-Scores

Z-Score Calculator

Quick presets:
Original Distribution (μ, σ)
Standard Normal Z-Distribution

Batch Z-Score Calculator

Enter a list of values (one per line or comma-separated). Computes Z-scores using dataset mean & std, or custom μ/σ.

#Value (x)Z-ScorePercentileClassification

Inverse Z-Score: Percentile → Value

Common Critical Z Values

Confidence Interval Calculator

Theory, Formulas & Statistical Concepts

The Z-score (standard score) is a dimensionless measure indicating how many standard deviations a value is above or below the population mean. Z-scores are foundational to inferential statistics, quality control, standardised testing, and medical screening.

1. Z-Score Formula
Z = (x − μ) / σ
Z = standard score | x = observed value | μ = population mean | σ = population standard deviation

For a sample (when population σ is unknown), use the sample standard deviation s instead of σ. Positive Z means above mean; negative Z means below mean.

2. Probability Density Function (PDF)

The bell-shaped curve of the normal distribution is described by the PDF:

f(x) = (1 / (σ√(2π))) × exp(−(x−μ)² / (2σ²))
For the standard normal (μ=0, σ=1): φ(z) = (1/√(2π)) × exp(−z²/2)
3. Cumulative Distribution Function (CDF) - P(X ≤ x)

The CDF gives the probability that a normal random variable X is less than or equal to x. It is the area under the PDF curve from −∞ to x:

Φ(z) = P(Z ≤ z) = (1/2) × [1 + erf(z/√2)]
erf = error function (approximated numerically) | Φ(0) = 0.5 | Φ(1.645) = 0.95 | Φ(1.96) = 0.975
4. The Empirical Rule (68-95-99.7 Rule)
RangeZ-Score Range% Data WithinMeaning
μ ± 1σ−1 to +168.27%Most common values - within 1 SD
μ ± 2σ−2 to +295.45%Nearly all values - within 2 SD
μ ± 3σ−3 to +399.73%Extreme outliers - beyond 3 SD are very rare
μ ± 1.645σ−1.645 to +1.64590.0%90% confidence region
μ ± 1.96σ−1.96 to +1.9695.0%95% confidence region
μ ± 2.576σ−2.576 to +2.57699.0%99% confidence region
5. Confidence Interval for Population Mean
CI = x̄ ± z* × (σ / √n)
x̄ = sample mean | z* = critical z value | σ = standard deviation | n = sample size | σ/√n = standard error (SE)

The margin of error (ME) = z* × SE. For 95% CI: z* = 1.96. This means: if we repeated sampling 100 times, approximately 95 CIs would contain the true population mean.

6. Z-Scores in Practice
ApplicationWhat x representsμσUsage
Standardised Testing (SAT)Raw score500 per section100Percentile ranking of students
IQ TestsIQ score10015Z > 2 → top 2.3% (gifted)
Quality Control (SPC)MeasurementTarget specProcess σ|Z| > 3 triggers investigation
Finance (Z-score Model)Financial ratiosVariesVariesAltman Z-score predicts bankruptcy risk
Medical Reference RangesLab value (e.g. BMI)Population meanPopulation σIdentify outliers needing intervention
Growth Charts (CDC/WHO)Height/weightAge-sex specificAge-sex specificZ < −2 = stunting; Z < −3 = severe
7. Interpreting Z-Score Magnitude
|Z| ValueInterpretationPercentile (left tail)Rarity
0.00Exactly at the mean50thMost common
0.67Slightly above/below average75th/25thCommon (1 in 2)
1.001 SD from mean84th/16thModerately unusual
1.9695% CI boundary97.5th/2.5th1 in 20
2.5899% CI boundary99.5th/0.5th1 in 100
3.003 SD - extreme outlier99.87th/0.13th1 in 370
4.004 SD - very rare99.9968th1 in 15,787

Frequently Asked Questions

1. What is a Z-score and what does it measure?

A Z-score (standard score) measures how many standard deviations a particular value is above or below the mean of a distribution. Calculated as Z = (x − μ)/σ, it transforms any normally distributed variable into the standard normal distribution (μ=0, σ=1). A Z of +2 means the value is 2 standard deviations above the mean; Z = −1.5 means 1.5 SDs below. This standardisation allows comparison of values from different distributions with different means and units.

2. How do I convert a Z-score to a percentile?

A percentile tells you what percentage of the population falls below a given value. To convert Z to percentile, calculate the cumulative distribution function (CDF): P = Φ(Z) = (1/2)[1 + erf(Z/√2)]. Common conversions: Z = 0 → 50th percentile; Z = 1 → 84th; Z = 1.645 → 95th; Z = 1.96 → 97.5th; Z = 2.576 → 99.5th; Z = −1 → 16th. This is what a Z-table looks up.

3. What is the difference between left-tail, right-tail, and two-tail probabilities?

Left-tail P(X ≤ x) = Φ(Z): probability of getting a value ≤ x - area to the LEFT of z on the bell curve. Right-tail P(X ≥ x) = 1 − Φ(Z): probability of getting a value ≥ x - area to the RIGHT. Two-tail P(|Z| ≤ z) = 2Φ(Z) − 1: probability within ±z standard deviations of the mean. For hypothesis testing: two-tail tests use α/2 in each tail; one-tail tests use full α in one tail. Z = 1.96 gives a 5% two-tail significance level (2.5% in each tail).

4. What is the inverse Z-score (percentile to Z)?

The inverse Z (also called the quantile function or percent-point function) finds the Z-value corresponding to a given cumulative probability. For example, the 90th percentile corresponds to Z = 1.282; the 97.5th to Z = 1.960; the 99th to Z = 2.326. This is used in confidence interval construction (finding z* for a given confidence level) and in hypothesis testing (finding the critical z-value for a given α). Computed as Z = Φ⁻¹(p) using the inverse error function: Z = √2 × erf⁻¹(2p − 1).

5. What is a confidence interval and how does Z relate to it?

A confidence interval (CI) is a range of values that, with a specified probability (confidence level), contains the true population mean. Formula: x̄ ± z* × (σ/√n). The z* is the critical Z value for the chosen confidence level: 1.645 for 90%, 1.960 for 95%, 2.576 for 99%. The margin of error (ME) = z* × SE where SE = σ/√n is the standard error. A wider CI (higher z*) gives more confidence but less precision. Increasing n reduces ME without changing the confidence level.

6. What is a normal distribution and when does data follow it?

The normal (Gaussian) distribution is a symmetric, bell-shaped probability distribution characterised by its mean μ and standard deviation σ. It arises from the Central Limit Theorem: the sum (or mean) of a large number of independent random variables tends toward normality, regardless of the original distribution. Many natural phenomena approximate normality: heights, weights, test scores, measurement errors, blood pressure, and financial returns. Data does NOT need to be perfectly normal - the CLT ensures Z-score methods work well for large samples (n > 30) even with non-normal populations.

7. What Z-score is considered an outlier?

The common threshold for identifying outliers using Z-scores is |Z| > 2 (moderate outlier) or |Z| > 3 (extreme outlier). Using the empirical rule: values with |Z| > 3 occur only 0.27% of the time (3 in 1000) in a normal distribution. In quality control (Six Sigma), processes aim for ±6σ from the target, allowing only 3.4 defects per million opportunities. The choice of threshold depends on context - clinical studies might use |Z| > 2, financial risk models might use |Z| > 4.

8. What is the standard error and how is it different from standard deviation?

Standard deviation (σ or s) measures the spread of individual data values around the mean. Standard error (SE = σ/√n) measures the spread of the sample mean across different samples - it's the standard deviation of the sampling distribution of the mean. SE is always smaller than σ for n > 1. SE decreases as sample size increases (√n relationship), which is why larger samples give more precise estimates of the population mean.

9. Can Z-scores be used for non-normal distributions?

Z-scores can be calculated for any distribution, but the probability interpretations (percentiles, CDF values from the Z-table) only apply precisely to normal distributions. For non-normal data: (1) Chebyshev's theorem guarantees at least (1 − 1/z²) × 100% of data within ±z for any distribution - e.g. at least 75% within ±2σ (vs. 95.4% for normal). (2) If the sample is large (n > 30), the Central Limit Theorem means the sample mean will be approximately normal, enabling valid Z-based confidence intervals.

10. What is the difference between Z-score and T-score in statistics?

Both Z and t-scores measure distance from the mean in standard deviation units, but t-scores are used when the population standard deviation σ is unknown and must be estimated from sample data s. The t-distribution has heavier tails than the normal distribution (more probability in the extremes), especially for small n. As n increases, t → Z (the t-distribution approaches normal). Rule of thumb: use Z when σ is known or n > 30; use t when σ is unknown and n ≤ 30. The t-score is critical for t-tests comparing means.

11. How is the Z-score used in quality control (Six Sigma)?

In manufacturing quality control, Z-score (process Z or sigma level) measures process capability. A Six Sigma process has ±6σ between the mean and specification limits, corresponding to 3.4 DPMO (defects per million opportunities). The process capability index Cp = (USL − LSL)/(6σ) where USL/LSL are spec limits. A Cpk ≥ 1.33 (4σ) is typically required; Cpk ≥ 1.67 (5σ) for critical processes. Control charts use ±3σ limits (Z = ±3) as the standard control limits - points beyond these trigger investigation.

12. What is Altman's Z-score for financial analysis?

Altman's Z-score (1968) is a multi-factor financial model predicting corporate bankruptcy risk. It is not directly related to the statistical Z-score from normal distributions - instead, it is a weighted linear combination of financial ratios: Z = 1.2×T1 + 1.4×T2 + 3.3×T3 + 0.6×T4 + 1.0×T5, where T1–T5 are working capital/assets, retained earnings/assets, EBIT/assets, equity/debt, and sales/assets ratios. Z > 2.99 = safe zone; 1.81–2.99 = grey zone; Z < 1.81 = distress zone.

13. How do I standardise a dataset using Z-scores?

To standardise: (1) Calculate the mean μ and standard deviation σ of the dataset. (2) For each value x, compute Z = (x − μ)/σ. The resulting Z-scores have mean = 0 and std = 1 (standard normal form). This is called Z-score normalisation or standardisation, distinct from min-max normalisation. Standardisation is essential before: PCA (principal component analysis), k-means clustering, SVM classification, and any distance-based algorithm where features with different scales would otherwise dominate.

14. What does a Z-score of 0, +1, −1 mean practically?

Z = 0: The value equals exactly the mean - exactly average. Z = +1: The value is 1 standard deviation above the mean - in the 84th percentile (better than 84% of population). Z = −1: 1 SD below the mean - 16th percentile. Z = +2: 97.7th percentile - better than ~97.7% (top ~2.3%). For exam scores with μ=70, σ=10: Z=+1 means score=80, Z=−2 means score=50. For IQ with μ=100, σ=15: Z=+2 means IQ=130 (gifted range).

15. What is the central limit theorem and why does it matter for Z-scores?

The Central Limit Theorem (CLT) states that the distribution of the sample mean approaches a normal distribution as n increases, regardless of the shape of the original population distribution. Practically: for n ≥ 30, the sample mean x̄ is approximately normally distributed with mean μ and standard error SE = σ/√n, even if individual data points are skewed or bimodal. This is why Z-based confidence intervals and hypothesis tests work for large samples from any distribution - the CLT guarantees the sampling distribution of the mean is approximately normal.

Explore More Engineering Tools

Water budget, sieve analysis, simultaneous equations and many more free tools.

All Tools Blog