Welcome to the final module of our comprehensive study of Ridge Regression!
In Module 1, we uncovered various facets of Ridge Regression, starting with SVD (Singular Value Decomposition) approach. We carefully dissected the formula for the Ridge estimator: , unraveling its intricacies through calculus.
Our previous discussions in Module 2 illuminated the mean and variance of , leading to an interesting revelation: there exists an optimal value of
that intricately balances the increase in model bias with a corresponding decrease in variance. This balance is crucial for building robust predictive models.
Now, we are set to cover another method of determining the Ridge estimator. We’ll pivot from the previously discussed methods and delve into the realm of Bayesian statistics, particularly focusing on MAP (Maximum A Posteriori) Estimation. This approach refines the MLE (Maximum Likelihood Estimation) methodology by integrating a prior distribution into our prediction strategy, offering a richer understanding of Ridge Regression.
To set the stage, let’s briefly revisit the MLE approach. The MLE method for deducing the weights is framed as:
This equation represents the cornerstone of likelihood–based estimation, seeking the weight vector that maximizes the likelihood of observing our responses
given our inputs
. However, MLE has its limitations, particularly in situations where the dataset is small or when multicollinearity is present among the predictions.
Using MAP Estimation
The MAP approach incorporates prior beliefs about the distribution of the weights through . Namely we are finding the weight vector
that maximizes the posterior probability
. Lets lay down some mathematical groundwork.
Our goal is to determine , defined as:
This equation might seem daunting, but we can rewrite it using Bayes’ Theorem:
where is the likelihood, representing how probable our observed response
is, given
and input data
. In this equation,
is the marginal probability of
, given
, and does not contribute to determining
. Instead we might consider it acts as a normalizing constant. Therefore, we simplify our equation to:
Let’s assume our response variable follows a normal distribution with mean
and variance
, denoted as
. Similarly, we assume our weights
also follow a normal distribution, centered around zero with variance
, expressed as
. With these assumptions in place, we can now expand our expression for MAP estimation.
Our expanded expression for the MAP estimator is given by:
In this expression, the terms and
are constants with respect to the optimization process and thus can be omitted for simplicity. This simplification leads us to:
applying the logarithm trick for simplification, our expression becomes a problem of minimization:
Expanding and rearranging the terms, we have:
As this function is convex in , we will obtain a unique solution by setting the gradient to zero:
After computing the gradient and simplifying, we arrive at:
therefore, our Ridge Estimator, by re-arranging and solving for , is finally given by:
This derivation not only reinforces our understanding of the Ridge Estimator but also highlights the seamless blend of statistical theory and mathematical elegance inherent in Ridge Regression.