6 Dimension Reduction Theory
6 Dimension Reduction Theory
Modeling?
An intuitive example of dimensionality reduction can be discussed through a simple e-mail
classification problem, where we need to classify whether the e-mail is spam or not. This can
involve a large number of features, such as whether or not the e-mail has a generic title, the
content of the e-mail, whether the e-mail uses a template, etc. However, some of these features
may overlap. In another condition, a classification problem that relies on both humidity and
rainfall can be collapsed into just one underlying feature, since both of the aforementioned are
correlated to a high degree. Hence, we can reduce the number of features in such problems.
Step 2: Determine the eigenvalues of ATA and sort these in descending order, in
the absolute sense. Square roots these to obtain the singular values of A.
A=
7. Choose principal components and form a feature vector using Eigen vectors.
8. Derive the new data set Z.
Z = X A Where X is a matrix containing columns xi-𝐱̅ and yi-𝐲̅ and A is a
matrix of Eigen vectors.
Singular Value Decomposition(SVD)
It is one of the most widely used unsupervised learning algorithms, that is at the center of many
recommendation and Dimensionality reduction.
In simple terms, SVD is the factorization of a matrix into 3 matrices. So if we have a matrix A,
Recorded time signals, by the two microphones, are denoted by x1(t), x2(t).
Use statistical “latent variables“ system. Random variable sk instead of time signal are latent
variables & are unknown AND Mixing matrix A is also unknown. Task here is to estimate A and
So after estimating A,
we can compute W=A-1 and hence s = Wx = A-1x
Independent component analysis (ICA) is a method for finding underlying factors or components
from multivariate (multi-dimensional) statistical data. What distinguishes ICA from other
methods is that it looks for components that are both statistically independent, and nonGaussian.
In probability theory, two events are independent, statistically independent, or
stochastically independent if the occurrence of one does not affect the probability of occurrence
of the other.
In physics, a non-Gaussianity is the correction that modifies the expected Gaussian function
Gaussian functions are widely used in statistics to describe the normal distributions.
In probability theory, the normal (or Gaussian or Gauss or Laplace–Gauss) distribution is a very
Invertible matrix: If this is the case, then the matrix B is uniquely determined by A and is called
the inverse of A, denoted by A−1. A square matrix that is not invertible is called singular or
Latent variables are variables that are not directly observed but are rather inferred (through a
mathematical model) from other variables that are observed (directly measured)
Scanned by CamScanner
Scanned by CamScanner
Scanned by CamScanner
Scanned by CamScanner
Scanned by CamScanner
Scanned by CamScanner
Scanned by CamScanner
Scanned by CamScanner
Scanned by CamScanner