Dimensionality Reduction Methods
Last updated
Was this helpful?
Last updated
Was this helpful?
A series on DR for dummies on medium part
in tensorflow
Machine learning mastery:
What is missing is how the EigenDecomposition is calculated.
Randomized svd
Incremental svd
There are two things we are trying to accomplish with whitening:
Make the features less correlated with one another.
Give all of the features the same variance.
Whitening has two simple steps:
Project the dataset onto the eigenvectors. This rotates the dataset so that there is no correlation between the components.
Normalize the the dataset to have a variance of 1 for all components. This is done by simply dividing each component by the square root of its eigenvalue.
Finally, showing results how KPCA works well on noisy images, compared to PCA.
PCA vs LDA:
Both Linear Discriminant Analysis (LDA) and Principal Component Analysis (PCA) are linear transformation techniques used for dimensionality reduction.
PCA can be described as an “unsupervised” algorithm, since it “ignores” class labels and its goal is to find the directions (the so-called principal components) that maximize the variance in a dataset.
In contrast to PCA, LDA is “supervised” and computes the directions (“linear discriminants”) that will represent the axes that maximize the separation between multiple classes.
Although it might sound intuitive that LDA is superior to PCA for a multi-class classification task where the class labels are known, this might not always the case.
For example, comparisons between classification accuracies for image recognition after using PCA or LDA show that :
In practice, it is also not uncommon to use both LDA and PCA in combination:
Best Practice: PCA for dimensionality reduction can be followed by an LDA. But before we skip to the results of the respective linear transformations, let us quickly recapitulate the purposes of PCA and LDA: PCA finds the axes with maximum variance for the whole data set where LDA tries to find the axes for best class separability. In practice, often a LDA is done followed by a PCA for dimensionality reduction.
** To fully understand the details please follow the LDA link to the original and very informative article
*** TODO: need some benchmarking for PCA\LDA\LSA\ETC..
reduction of the dimensionality
noise reduction
incorporating relations between terms into the representation.
SVD and PCA and "total least-squares" (and several other names) are the same thing. It computes the orthogonal transform that decorrelates the variables and keeps the ones with the largest variance. There are two numerical approaches: one by SVD of the (centered) data matrix, and one by Eigen decomposition of this matrix "squared" (covariance).
While PCA is global, it finds global variables (with images we get eigen faces, good for reconstruction) that maximizes variance in orthogonal directions, and is not influenced by the TRANSPOSE of the data matrix.
On the other hand, ICA is local and finds local variables (with images we get eyes ears, mouth, basically edges!, etc), ICA will result differently on TRANSPOSED matrices, unlike PCA, its also “directional” - consider the “cocktail party” problem. On documents, ICA gives topics.
It helps, similarly to PCA, to help us analyze our data.
- the jist of it is that we assume a t- distribution on distances and remove those that are farther.normalized for density. T-dist used so that clusters are not clamped in the middle.
Iteratively moving from the left to the right
How why, all here.
(remove the mean from A, calculate cov(A), calculate eig(cov), A*eigK = PCA)
- what is an eigen vector - simply put its a vector that satisfies A*v = lambda*v, how to use eig() and how to confirm an eigenvector/eigenvalue and reconstruct the original A matrix.
(did not read)
(did not read)
(did not read)
, it is unclear what he is trying to show.
**
**
. (bottom line, do it on the train only.)
, (pca/zca whitening), ,
Autoencoder is PCA based on their equation, i.e. minimize the reconstruction error formula.
Then they say that PCA cant separate certain non-linear situations (circle within a circle), therefore they introduce kernel based PCA (using the kernel trick - like svm) which mapps the space to another linearly separable space, and performs PCA on it,
() - Linear Discriminant Analysis (LDA) is most commonly used as dimensionality reduction technique in the pre-processing step for pattern-classification and machine learning applications. The goal is to project a dataset onto a lower-dimensional space with good class-separability in order avoid overfitting (“curse of dimensionality”) and also reduce computational costs.
PCA tends to outperform LDA if the number of samples per class is relatively small (, A.M. Martinez et al., 2001).
- has KDA - This package provides the classic algorithms of supervised distance metric learning, together with some of the newest proposals.
is quite simple, you just use SVD to perform dimensionality reduction on the tf-idf vectors–that’s really all there is to it! And
Here is a very nice with code, explaining what are the three matrices, word clustering, sentence clustering and vector importance. They say that for sentence space we need to remove the first vector as it is correlated with sentence length.
*how to
PCA vs LSA: (, )
Sparse
(pca, sammon, isomap, tsne)
****
, this is a very important read.
In contrary to what it says on sklearn’s website, TSNE is not suited ONLY for visualization, you
“t-Distributed Stochastic Neighbor Embedding (t-SNE) is a () technique for dimensionality reduction that is particularly well suited for the visualization of high-dimensional datasets.”
,