Covariance and correlation measure how two variables change together. Covariance shows if they move in the same or opposite directions, while correlation tells us how strong that relationship is on a scale from -1 to 1.
These concepts help us understand connections between things like height and weight or income and education. They're useful in finance, science, and social research to spot patterns and make predictions about related variables.
Covariance and Correlation
Definition of covariance and correlation
Top images from around the web for Definition of covariance and correlation Pearson correlation coefficient - Wikipedia View original
Is this image relevant?
CSUP Math 156 Correlation and Linear Regression View original
Is this image relevant?
Pearson correlation coefficient - Wikipedia View original
Is this image relevant?
CSUP Math 156 Correlation and Linear Regression View original
Is this image relevant?
1 of 3
Top images from around the web for Definition of covariance and correlation Pearson correlation coefficient - Wikipedia View original
Is this image relevant?
CSUP Math 156 Correlation and Linear Regression View original
Is this image relevant?
Pearson correlation coefficient - Wikipedia View original
Is this image relevant?
CSUP Math 156 Correlation and Linear Regression View original
Is this image relevant?
1 of 3
Covariance quantifies the joint variability of two random variables from their individual means
Positive covariance indicates variables tend to move in the same direction relative to their means (height and weight)
Negative covariance indicates variables tend to move in opposite directions relative to their means (price and demand)
Covariance formula: C o v ( X , Y ) = E [ ( X − μ X ) ( Y − μ Y ) ] Cov(X,Y) = E[(X - \mu_X)(Y - \mu_Y)] C o v ( X , Y ) = E [( X − μ X ) ( Y − μ Y )]
μ X \mu_X μ X and μ Y \mu_Y μ Y represent the means of random variables X X X and Y Y Y
Correlation measures the strength and direction of the linear relationship between two random variables
Ranges from -1 (perfect negative linear relationship) to 1 (perfect positive linear relationship)
Correlation of 0 implies no linear relationship (income and favorite color)
Correlation formula: ρ X Y = C o v ( X , Y ) σ X σ Y \rho_{XY} = \frac{Cov(X,Y)}{\sigma_X \sigma_Y} ρ X Y = σ X σ Y C o v ( X , Y )
σ X \sigma_X σ X and σ Y \sigma_Y σ Y represent the standard deviations of random variables X X X and Y Y Y
Calculation of joint distributions
Covariance calculation: C o v ( X , Y ) = E [ X Y ] − E [ X ] E [ Y ] Cov(X,Y) = E[XY] - E[X]E[Y] C o v ( X , Y ) = E [ X Y ] − E [ X ] E [ Y ]
E [ X Y ] E[XY] E [ X Y ] represents the expected value of the product of X X X and Y Y Y
E [ X ] E[X] E [ X ] and E [ Y ] E[Y] E [ Y ] represent the individual expected values (means) of X X X and Y Y Y
Correlation calculation: ρ X Y = C o v ( X , Y ) V a r ( X ) V a r ( Y ) \rho_{XY} = \frac{Cov(X,Y)}{\sqrt{Var(X)Var(Y)}} ρ X Y = Va r ( X ) Va r ( Y ) C o v ( X , Y )
V a r ( X ) Var(X) Va r ( X ) and V a r ( Y ) Var(Y) Va r ( Y ) represent the variances of random variables X X X and Y Y Y
For discrete random variables, calculate expected values using the probability mass function (PMF)
Example: Roll two fair dice, let X X X be the sum and Y Y Y be the product of the numbers rolled
For continuous random variables, calculate expected values using the joint probability density function (PDF)
Example: X X X and Y Y Y represent the heights of a randomly selected male and female student
Properties of statistical relationships
Covariance properties
C o v ( X , X ) = V a r ( X ) Cov(X,X) = Var(X) C o v ( X , X ) = Va r ( X ) , covariance of a variable with itself equals its variance
C o v ( X , Y ) = C o v ( Y , X ) Cov(X,Y) = Cov(Y,X) C o v ( X , Y ) = C o v ( Y , X ) , covariance is symmetric
C o v ( a X + b , c Y + d ) = a c ⋅ C o v ( X , Y ) Cov(aX + b, cY + d) = ac \cdot Cov(X,Y) C o v ( a X + b , c Y + d ) = a c ⋅ C o v ( X , Y ) for constants a a a , b b b , c c c , and d d d
Correlation properties
ρ X X = 1 \rho_{XX} = 1 ρ XX = 1 , a variable is perfectly correlated with itself
ρ X Y = ρ Y X \rho_{XY} = \rho_{YX} ρ X Y = ρ Y X , correlation is symmetric
∣ ρ X Y ∣ ≤ 1 |\rho_{XY}| \leq 1 ∣ ρ X Y ∣ ≤ 1 , correlation is bounded between -1 and 1
Relationship between independence and covariance/correlation
If X X X and Y Y Y are independent, then C o v ( X , Y ) = 0 Cov(X,Y) = 0 C o v ( X , Y ) = 0 and ρ X Y = 0 \rho_{XY} = 0 ρ X Y = 0
However, C o v ( X , Y ) = 0 Cov(X,Y) = 0 C o v ( X , Y ) = 0 or ρ X Y = 0 \rho_{XY} = 0 ρ X Y = 0 does not necessarily imply independence (non-linear relationships)
Applications in linear analysis
Interpret covariance and correlation values
Determine the direction of the linear relationship (positive or negative)
Assess the strength of the linear relationship (magnitude of correlation)
Covariance interpretation is scale-dependent and difficult to compare across different variable pairs
Correlation provides a standardized measure of linear relationship strength for easier comparison
Applications across various fields
Finance: Portfolio risk analysis and diversification (stocks and bonds)
Signal processing : Assessing similarity between signals (audio and video)
Machine learning: Feature selection and dimensionality reduction (customer preferences)
Social sciences: Studying relationships between variables (education and income)