Calculation
✏️ Calculation

Σ^k=UkDkUkTcan make this calculation more faster.
In normal dsitrubiton, the quadratic form means mahalanobis distance. This is the distance that(https://darkpgmr.tistory.com/41) how much each data is away from the mean over the standard deviation.
Using this expression, we can interpret it as assigningX∗to the loc μ∗. This expression mahalanobis distance can be changed into euclidiean distance. In a transformed space, each point is assigned into the closest median point. It controls the effect of prior probability(πk). The transformed space X∗is equal to Whitening transformation that makes the variance of X I.

✏️ High Dimensional Data
p: The dimension of input matrix
K: The number of centroid.
LDA(QDA) reduces p dimension into K. The number of centroids in p dimension is K, so the dimension is reduced into K−1. Let HK−1be a subspace spanned by these centroids. The distance between this subspace and X∗can be neglected. ( These centroids are already in the subspace, so this distance has same impact on these points. Project transformed X∗onto HK−1, and compare distance between the projected points. If the variance of these projected centroids is big, this is an optimal situation. Finding the optimal subspace is same to finding PC space of centroids.
Class centroids M(K×p) , Common covariance matrix W=Σ^=Σk=1KΣgi=k(xi−μ^k)(xi−μ^k)T/(N−K)
M∗=MW−1/2 using the eigen-decomposition of W
B∗=Cov(M∗)=V∗DBV∗T, Columns of V∗Tare vl∗
Zl=vlTXwithvl=W−1/2vl∗
W: Within-variance matrix, B: Between-variance matrix
Find the linear combination Z=aTXsuch that the between-class variance is maximized relative to the within-class variance.
The problem can be changed into this one:
The solution a is the biggest eigenvector of W−1B. We can represent our data in a reduced form using the axis of Zl=vlTX=alTX. Zl is called as a canonical variate and this becomes a new axis.
Last updated