Linear Discriminant Analysis for p greater than 1

Linear Discriminant Analysis for p>1

Assume that X=(X1,X2,…,Xp) is drawn from a multivariate Gaussian (or multivariate normal) distribution, with a class-specific mean vector and a common covariance matrix.

The multivariate Gaussian distribution assumes that each individual predictor follows a one-dimensional normal distribution with some correlation between each pair of predictors.

To indicate that a p-dimensional random variable X has a multivariate Gaussian distribution, we write X∼N(μ,Σ). Here E(X)=μ is the mean of X (a vector with p components), and Cov(X)=Σ is the p×p covariance matrix of X. Formally, the multivariate Gaussian density is defined as:

f(x)=1(2π)p/2|Σ|1/2exp(−12(x−μ)TΣ−1(x−μ))

In the case of p>1 predictors, the LDA classifier assumes that the observations in the kth class are drawn from a multivariate Gaussian distribution N(μk,Σ), where μk is a class-specific mean vector, and Σ is a covariance matrix that is common to all K classes.

The Bayes classifier assigns an observation X=x to the class for which

δk(x)=xTΣ−1μk−12μkTΣ−1μk+log(πk)

is largest.

The Bayes decision boundaries are the values x for which δk(x)=δl(x):

xTΣ−1μk−12μkTΣ−1μk=xTΣ−1μl−12μlTΣ−1μl

for k≠1 and πk is the same for each class. The Bayes classifier will classify an observation according to the region in which it is located.

LDA method uses estimates for the Bayes classifier the same as the p=1 case.

Check for errors using a Confusion Matrix and accuracy with a ROC Curve

Sources: 1

Connect With Me!