Introduction of Machine Learning
Introduction of Machine Learning
Cluster analysis and Dimensional reduction algorithm are examples of unsupervised algorithm.
Cluster objects based on its attributes.
Cluster Analysis:
Cluster analysis is the examples of unsupervised learning.it aims to group objects into disjoint clusters or
groups. cluster analysis cluster objects based on its attributes. All the data objects of the partitions are similar
in some aspect and vary from the data objects in the other partitions significantly.
Examples: detection of abnormal growth in medical image, determine clusters of signatures in a gene
database.in clustering algorithm takes a set of dogs and cats images and groups it as two cluster dog and cat.
Dimensionality Reduction:
It takes the higher dimension data as input and outputs the data in lower dimension by taking advantage of the
variance of the data.it is task of reducing the dataset with few features without losing the generality.
Semi-Supervised Machine Learning:
where the dataset has a huge collection of labelled data and
unlabelled data. Labelling is a costly process and difficult to perform
by the humans. Semi-supervised algorithms use unlabelled data by
assigning pseudo label.
Reinforcement learning:
Reinforcement learning uses trial and error to train algorithms and
create models. During the training process, algorithms operate in
specific environments and then are provided with feedback following
each outcome.
Reinforcement learning is often used to create algorithms that must
effectively make sequences of decisions or actions to achieve their
aims, such as playing a game or summarizing an entire text.
Support Vector Machine
Support Vector Machine (SVM) is a supervised learning algorithm that takes a labelled
data as input and creates learning functions that can be used for classification of
unknown test data.
Though we say regression problems as well it’s best suited for classification. The main
objective of the SVM algorithm is to find the optimal hyperplane in an N-dimensional
space that can separate the data points in different classes in the feature space.
The hyperplane tries that the margin between the closest points of different classes
should be as maximum as possible. The dimension of the hyperplane depends upon the
number of features. If the number of input features is two, then the hyperplane is just a
line. If the number of input features is three, then the hyperplane becomes a 2-D plane.
It becomes difficult to imagine when the number of features exceeds three.
Let’s consider two independent variables x1, x2, and one dependent variable
which is either a blue circle or a red circle.
From the figure above it’s very clear that there are multiple lines (our
hyperplane here is a line because we are considering only two input features
x1, x2) that segregate our data points or do a classification between red and
blue circles.
Q. Suppose you are running a factory some sort of widget that requires steel
as a raw material. your cost are predominately human labor which is $20 per
hour for workers and the steel itself which runs $170 per ton. suppose your
revenue R is loosely modeled by the following equations
R(h,s) = 200. h2/3.s1/3 . If your budget is $20000 what is the maximum possible
revenue?
Advantages of SVM