Specifies the probability of observing a combination of values for two or more random variables. Characterizes the relationship between multiple random variables, including their dependencies and correlations.
Definition
For two random variables X and Y, the joint probability distribution gives the probability that X and Y simultaneously take on specific values. This can be expressed as:
For discrete random variables: P(X=x,Y=y) or PX,Y(x,y)
For continuous random variables: fX,Y(x,y)
For a joint probability distribution P(A∩B), P(A) and P(B) are the marginal probabilities.
Covariance
Denoted by Cov(X,Y) or σXY. Measures the linear relationship between two random variables.
Cov(X,Y)=x∑y∑(x−μX)(y−μY)P(X=x,Y=y)
Positive covariance: Indicates that higher than mean values of one variable tend to be
paired with higher than mean values of other variable.
Negative covariance: Indicates that higher than mean values of one variable tend to be
paired with lower than mean values of other variable.
If the two random variables are independent then the covariance will
be zero.
Properties:
Cov(X,Y)=E[(X−E(X))(Y−E(Y))]
Cov(X,Y)=E(XY)−E(X)E(Y)
Cov(X,a)=0
Cov(X,X)=Var(X)
Cov(aX,bY)=abCov(X,Y)
Cov(X+Y,Z)=Cov(X,Z)+Cov(Y,Z)
Correlation
Corr(X,Y)=ρXY=Var(X)Var(Y)Cov(X,Y)
ρXY is the Pearson correlation coefficient.
Sample correlation coefficient
Denoted by r∈[−1,1].
r=(∑xi2−nxˉ2)(∑yi2−nyˉ2)∑(xiyi)−nxˉyˉ
Properties
Non-negativity
Discrete case: ∀x,yP(X=x,Y=y)≥0
Continuous case: ∀x,yfX,Y(x,y)≥0
Total probability equals 1
Discrete case: ∑x∑yP(X=x,Y=y)=1
Continuous case: ∫−∞∞∫−∞∞fX,Y(x,y)dydx=1
Marginal distributions
The distribution of an individual variable can be derived from the joint distribution:
Discrete case: P(X=x)=∑yP(X=x,Y=y)
Continuous case: fX(x)=∫−∞∞fX,Y(x,y)dy
Conditional distributions
The distribution of one variable given a specific value of the other:
Discrete case: P(X=x∣Y=y)=P(Y=y)P(X=x,Y=y)
Continuous case: fX∣Y(x∣y)=fY(y)fX,Y(x,y)
Independence
Random variables X and Y are independent iff:
Discrete case: ∀x,yP(X=x,Y=y)=P(X=x)⋅P(Y=y)
Continuous case: ∀x,yfX,Y(x,y)=fX(x)⋅fY(y)
Representation
Joint distributions can be represented in various ways:
For discrete variables: probability mass tables or matrices
For continuous variables: joint density functions or contour plots
Copulas: functions that describe the dependence structure between variables
Types
For Discrete Variables
For joint probability mass function, if x,y are independent, P(x,y)=P(x)P(y).
Cumulative probability:
P(X≤x,Y≤y)=x∑y∑P(x,y)
For marginal probability of X=a, ∑yP(a,y).
For Continuous Variables
Suppose f is the joint probability density function. The joint probability for any region A lying in x-y plane is: