Lesson 2.3: Performing Surgery on the MVN
This lesson dives into the mechanics of the Multivariate Normal distribution. We will learn how to perform 'statistical surgery' by deriving the precise formulas for Marginal distributions (when we ignore variables) and Conditional distributions (when we observe variables). The conditional mean formula derived here is the fundamental justification for Linear Regression.
Part 1: Partitioning the System
Imagine a system of variables, . To analyze it, we often need to split it into two groups:
- : A vector of variables we want to predict or analyze.
- : A vector of variables that we have observed.
We partition our mean vector and covariance matrix to match this structure:
Deconstructing the Partitioned Covariance Matrix
- : The covariance matrix *within* the variables.
- : The covariance matrix *within* the variables.
- : The matrix of covariances *between* every variable in and every variable in . ()
Part 2: The Marginal Distribution (Ignoring Variables)
The first question is the simplest: If we have a joint model of 10 stocks but we only care about the first two, what is their distribution? The answer is called the **marginal distribution**.
Theorem: Marginal Distribution of an MVN
The marginal distribution of any subset of an MVN vector is also MVN, defined only by the corresponding sub-vector of the mean and sub-matrix of the covariance matrix.
Intuition: This is the "zooming in" property. If a whole system is jointly Normal, any part you zoom in on is also Normal. This is a wonderfully simple and convenient result.
Part 3: The Conditional Distribution (Incorporating Information)
This is the main event and one of the most beautiful results in statistics. The **conditional distribution** answers the question: "Now that I have observed that has the specific value , how does this new information change my beliefs about the distribution of ?"
Theorem: Conditional Distribution of an MVN
The conditional distribution is also Multivariate Normal, but with an updated mean and a smaller variance.
The updated mean, , is our new best guess for .
Formula: Conditional Mean
Deconstructing the Formula's Intuition
Updated Belief = Original Belief + Adjustment
- : Our starting best guess for before any new information.
- : The "surprise" in the data we observed. How far was from its own average?
- : The **"transmission mechanism."** This matrix acts like a set of regression coefficients, translating the "surprise" in into a specific adjustment for .
The updated variance, , represents our remaining uncertainty about after observing . It is always smaller than the marginal variance .
Formula: Conditional Variance
Deconstructing the Formula's Intuition
Remaining Uncertainty = Original Uncertainty - Information Gained
- : Our initial variance (uncertainty) in .
- : The amount of variance that was *explained* by 's linear relationship with . This term is always non-negative.
This proves that observing a correlated variable **never increases our uncertainty**. At worst (if they are uncorrelated), the "information gained" is zero and the variance is unchanged. Usually, it reduces our uncertainty.
The Grand Connection: Why Linear Regression Works
Look again at the formula for the conditional mean:
If we group the terms, this is a simple linear equation:
This proves that if two sets of variables are jointly Multivariate Normal, the best possible prediction for one, given the other, is a linear function. This is the absolute theoretical justification for using Ordinary Least Squares (OLS) regression to model such systems.