Pattern Recognition Resources Compiled
Pattern Recognition Resources Compiled
K-Means, GMM, EM
https://www.youtube.com/watch?v=LmpkKwsyQj4&list=PLoROMvodv4rNH7qL6-
efu_q2_bPuy0adh&index=16
https://www.youtube.com/watch?v=LMZZPneTcP4
gmm, em, knn , andrew ng- https://www.youtube.com/watch?v=rVfZHWTwXSA
DBScan
https://www.youtube.com/watch?
v=MrLPzBxG95I&list=PLl8OlHZGYOQ7bkVbuRthEsaLr7bONzbXS - Cornell
https://www.youtube.com/watch?
v=ZMxfDWPXmjc&list=PLoROMvodv4rNyWOpJg_Yh4NSqI4Z4vOYy&index=8 - Stanford
videos
https://towardsdatascience.com/decision-trees-for-classification-id3-algorithm-explained-
89df76e72df1 - ID3
SVD, PCA
https://www.youtube.com/playlist?list=PLMrJAkhIeNNSVjnsviglFoY2nXildDCcv
https://www.youtube.com/watch?v=FgakZw6K1QQ&vl=en
SVD sum
https://www.d.umn.edu/~mhampton/m4326svd_example.pdf
https://medium.com/intuition/singular-value-decomposition-svd-working-example-
c2b6135673b5 (this is the best do this)
https://www.youtube.com/watch?v=cOUTpqlX-Xs - MIT open courseware
https://www.youtube.com/watch?v=uOzMM13iElw (check this)
PCA sum
https://www.geeksforgeeks.org/mathematical-approach-to-pca/
https://www.youtube.com/watch?v=ZtS6sQUAh0c&list=RDCMUCPi23Ql765_5smMj2-
r0X4g&start_radio=1&rv=ZtS6sQUAh0c&t=0 (check this especially no.6)
https://www.youtube.com/watch?v=TS_9u_NAc24 - this one
LDA
https://sebastianraschka.com/Articles/2014_python_lda.html
https://medium.com/machine-learning-researcher/dimensionality-reduction-pca-and-
lda-6be91734f567 (PCA and LDA)
https://www.youtube.com/watch?v=txgqfG4rfos (see this - mahesh huddar)
practice the decision tree (first) (watch again) - done Gini, left ID3, concept theory left
SVM - done
PCA - done ( https://www.youtube.com/watch?v=TS_9u_NAc24)
LDA - done (https://www.youtube.com/watch?v=txgqfG4rfos)
ANOVA - done
Chi-Square - done
Naive Bayes - done (https://www.kdnuggets.com/2020/06/naive-bayes-algorithm-
everything.html)
Clustering - k-means done, GMM and EM left, Agglomerative and hierarchical
clustering, Linkage - Single, Complete, Average (https://www.youtube.com/watch?
v=YH0r47m0kFM&list=PL4gu8xQu0_5JBO1FKRO5p20wc8DprlOgn&index=117)
Fuzzy C-Means
Regularization
DBScan - done
ICA - not done
w, b
w
An interesting edge case is kernel-SVM. Here it depends very much which kernel we are
using. E.g.
linear SVMs are parametric (for the same reason as the Perceptron or logistic regression).
So if the
kernel is linear the algorithm is clearly parametric. However, if we use an RBF kernel then
we cannot
represent the classifier of a hyper-plane of finite dimensions. Instead we have to store the
support
vectors and their corresponding dual variables -- the number of which is a function of the
data
set size (and complexity). Hence, the kernel-SVM with an RBF kernel is non-parametric. A
strange
in-between case is the polynomial kernel. It represents a hyper-plane in an extremely high
but still
finite-dimensional space. So technically one could represent any solution of an SVM with a
polynomial kernel as a hyperplane in an extremely high dimensional space with a fixed
number of
parameters, and the algorithm is therefore (technically) parametric. However, in practice this
is not
practical. Instead, it is almost always more economical to store the support vectors and their
corresponding dual variables (just like with the RBF kernel). It therefore is technically
parametric
but for all means and purposes behaves like a non-parametric algorithm.
Decision Trees are also an interesting case. If they are trained to full depth they are non-
parametric,
as the depth of a decision tree scales as a function of the training data (in practice ). If
we however limit the tree depth by a maximum value they become parametric (as an upper
bound
of the model size is now known prior to observing the training data).
Sums -
https://www.youtube.com/watch?
v=35VgJ84sqqI&list=PL4gu8xQu0_5JBO1FKRO5p20wc8DprlOgn&index=2 - Average
Linkage Hierarchical Agglomerative Clustering Algorithm draw dendrogram in ML
EM - Later (https://www.youtube.com/watch?v=3oefV-
AoP0E&list=PL4gu8xQu0_5JBO1FKRO5p20wc8DprlOgn&index=12)
SVD - https://www.youtube.com/watch?v=uOzMM13iElw
PCA - https://www.youtube.com/watch?v=TS_9u_NAc24
LDA - https://www.youtube.com/watch?
v=txgqfG4rfos&list=PL4gu8xQu0_5JBO1FKRO5p20wc8DprlOgn&index=21
https://www.youtube.com/watch?
v=KzJORp8bgqs&list=PL4gu8xQu0_5JBO1FKRO5p20wc8DprlOgn&index=112
DBScan - https://www.youtube.com/watch?
v=ZOLYaa9Jex0&list=PL4gu8xQu0_5JBO1FKRO5p20wc8DprlOgn&index=118 (Similarity
index and MinPts)
https://www.youtube.com/watch?v=-
p354tQsKrs&list=PL4gu8xQu0_5JBO1FKRO5p20wc8DprlOgn&index=119 (Eps and MinPts)
Normalization - https://www.youtube.com/watch?
v=jMvlyoegui4&list=PL4gu8xQu0_5JBO1FKRO5p20wc8DprlOgn&index=216