Normal distributions are the backbone of many statistical analyses. They're characterized by their symmetric bell shape and defined by two key parameters: the mean and standard deviation . These distributions are incredibly useful for modeling real-world phenomena and making predictions.
Understanding normal distributions is crucial for probability calculations. The empirical rule helps estimate probabilities within certain ranges, while z-scores standardize values for easier comparison. Probability density and cumulative distribution functions further aid in calculating specific probabilities for various scenarios.
Properties and Characteristics of Normal Distributions
Properties of normal distribution
Top images from around the web for Properties of normal distribution Normal distribution - Wikiversity View original
Is this image relevant?
The Empirical Rule – Math For Our World View original
Is this image relevant?
The Standard Normal Distribution | Introduction to Statistics – Gravina View original
Is this image relevant?
Normal distribution - Wikiversity View original
Is this image relevant?
The Empirical Rule – Math For Our World View original
Is this image relevant?
1 of 3
Top images from around the web for Properties of normal distribution Normal distribution - Wikiversity View original
Is this image relevant?
The Empirical Rule – Math For Our World View original
Is this image relevant?
The Standard Normal Distribution | Introduction to Statistics – Gravina View original
Is this image relevant?
Normal distribution - Wikiversity View original
Is this image relevant?
The Empirical Rule – Math For Our World View original
Is this image relevant?
1 of 3
Symmetric bell-shaped curve represents the shape of the distribution
Mean, median, and mode are equal located at the center (peak) of the distribution
Continuous probability distribution represents the data
Probability represented by the area under the curve (total area always equals 1)
Extends infinitely in both directions (left and right)
Asymptotically approaches the x-axis but never touches it (tails extend indefinitely)
Defined by two parameters: mean (μ \mu μ ) and standard deviation (σ \sigma σ )
68-95-99.7 rule (empirical rule) applies to normal distributions
Specifies the percentage of data within 1, 2, and 3 standard deviations of the mean
Parameters of normal distribution
Mean (μ \mu μ ) determines the location (center) of the distribution
Shifting the mean shifts the entire distribution left (negative shift) or right (positive shift)
Standard deviation (σ \sigma σ ) determines the spread (width) of the distribution
Larger standard deviation results in a wider, flatter distribution (more variability)
Smaller standard deviation results in a narrower, taller distribution (less variability)
Probability Calculations and Functions
Probabilities using empirical rule
Empirical rule (68-95-99.7 rule) specifies data percentages within standard deviations
Approximately 68% of data falls within one standard deviation of the mean (μ ± σ \mu \pm \sigma μ ± σ )
Approximately 95% of data falls within two standard deviations of the mean (μ ± 2 σ \mu \pm 2\sigma μ ± 2 σ )
Approximately 99.7% of data falls within three standard deviations of the mean (μ ± 3 σ \mu \pm 3\sigma μ ± 3 σ )
Z-scores standardize values to represent the number of standard deviations an observation is from the mean
Formula: z = x − μ σ z = \frac{x - \mu}{\sigma} z = σ x − μ (observation minus mean, divided by standard deviation)
Can be used to calculate probabilities using a standard normal distribution (μ = 0 , σ = 1 \mu = 0, \sigma = 1 μ = 0 , σ = 1 )
Density vs cumulative distribution functions
Probability density function (PDF) denoted as f ( x ) f(x) f ( x )
Describes the relative likelihood of a random variable taking on a specific value
Area under the curve between two points represents the probability of the random variable falling within that range (not the height of the curve)
Cumulative distribution function (CDF) denoted as F ( x ) F(x) F ( x )
Describes the probability that a random variable is less than or equal to a specific value
Monotonically increasing function, ranging from 0 to 1 (never decreases)
F ( x ) = P ( X ≤ x ) F(x) = P(X \leq x) F ( x ) = P ( X ≤ x ) , where X X X is the random variable
Can be used to calculate probabilities for intervals by subtracting CDF values (upper bound minus lower bound)