Joint Probability Distributions
Joint probability mass/density functions, marginal and conditional distributions, covariance, and correlation.
In many engineering applications, we need to understand the relationship between two or more random variables simultaneously. For example, a structural engineer might study the joint distribution of wind speed () and atmospheric pressure () during a hurricane, or a transportation engineer might model the number of cars () and trucks () arriving at a toll booth.
Joint Probability Mass and Density Functions
Describing the simultaneous behavior of multiple random variables.
Joint Probability Mass Function (Discrete)
For two discrete random variables and , the joint probability mass function gives the probability that takes the specific value AND takes the specific value simultaneously.
It must satisfy two conditions:
- for all .
- .
Joint Probability Density Function (Continuous)
For two continuous random variables and , the joint probability density function represents the probability that falls within a specific two-dimensional region in the -plane. The probability is the volume under the surface over region .
It must satisfy two conditions:
- for all .
- .
Marginal Distributions
Isolating the behavior of one variable from the joint distribution.
Sometimes we have the joint distribution of and , but we only care about the distribution of alone, regardless of . This is called the marginal distribution.
Marginal Probability Distributions
To find the marginal distribution of one variable, we sum (or integrate) out the other variable over its entire range.
- Discrete case for :
- Continuous case for :
Similarly, is the marginal distribution for found by summing or integrating out .
Conditional Distributions and Independence
How knowledge of one variable affects the probability distribution of another.
Conditional Probability Distribution
The probability distribution of , given that has taken a specific value . This is analogous to basic conditional probability ().
Similarly, provided .
Independence of Random Variables
Two random variables and are independent if and only if their joint probability distribution is the product of their marginal distributions for all possible values of .
If this holds true, knowing the value of gives no information about the value of . (e.g., The compressive strength of concrete from Plant A vs. Plant B).
Covariance and Correlation
Measuring the linear relationship between two random variables.
Covariance ()
A measure of how much two random variables change together. A positive covariance indicates that when is above its mean, tends to be above its mean (e.g., traffic volume and noise levels). A negative covariance indicates an inverse relationship (e.g., age of asphalt and its flexibility).
- If and are statistically independent, their covariance is zero ().
- However, a covariance of zero does not necessarily mean they are independent (they could have a non-linear relationship).
Correlation Coefficient ()
A standardized measure of the linear relationship between two variables. Covariance depends on the units of and , making it hard to interpret the strength of the relationship. The correlation coefficient scales covariance by the standard deviations of both variables, producing a dimensionless value between -1 and 1.
- : Perfect positive linear relationship.
- : Perfect negative linear relationship.
- : No linear relationship.
The Bivariate Normal Distribution
The foundational model for two correlated continuous variables.
Bivariate Normal Distribution
When two continuous random variables are individually normally distributed and correlated, their joint behavior is described by the bivariate normal distribution. Its PDF forms a 3-dimensional bell surface (a mound) whose orientation depends on the correlation .
Key properties:
- The marginal distributions and are both normal.
- The conditional distributions and are both normal.
- If the correlation for a bivariate normal distribution, then and are strictly independent. (This is a special property; for other distributions, does not guarantee independence).
Key Takeaways
- Joint Distributions (): Describe the simultaneous behavior of two random variables.
- Marginal Distributions (): Isolate one variable by summing or integrating out the other.
- Conditional Distributions (): The behavior of given a specific value of .
- Independence: If and are independent, .
- Covariance and Correlation: Measure the linear relationship between variables. Correlation () is standardized, always falling between -1 and 1.
- Bivariate Normal: The standard 3D bell-shaped curve for two correlated continuous variables.