Fisher's Linear Discriminant
While linear regression aims to predict continuous values, Linear Discriminant Analysis (LDA) aims to project data into a lower-dimensional space while preserving the maximum class separability. Fisher's Linear Discriminant is a specific approach to finding this projection.
1. The Goal: Separation and Compactness
The idea is to find a projection vector such that when we project the data points onto it:
- The distance between the class means is as large as possible.
- The variance within each class is as small as possible.
2. Fisher's Criterion
We define the between-class scatter and the within-class scatter :
Fisher's goal is to find the vector that maximizes this ratio .
For two classes:
where are the class means.
3. Projection and Classification
Once we find the optimal , we project any input onto it: . We then set a threshold on to classify the point.
Fisher's Linear Discriminant
Data points are projected onto a line (w) that maximizes class separation while minimizing within-class spread.
LDA vs. PCA
| Feature | PCA (Principal Component Analysis) | LDA (Linear Discriminant Analysis) |
|---|---|---|
| Type | Unsupervised | Supervised |
| Goal | Maximize variance (Signal) | Maximize class separability |
| Labels | Ignores labels | Uses class labels |
| Use Case | General dimensionality reduction | Pre-processing for classification |
Multiple Classes: Fisher's discriminant can be generalized to classes by seeking a projection into a -dimensional space.