CAT2 Key
CAT2 Key
1.4.1,
2.3.1
Ans:
3. The probabilities of five events are: P(first) = 0.5, P(second)=
P(third)=P(fourth)=P(fifth)= 0.125. Calculate entropy using the given
probabilities.
Ans:
1.4.1,
13.3.1
8. Construct space splitting and K-d tree representation for the following data points:
(5,4), (2,2), (3,7), (4,6), (7,5), (6,1), (6,8)
9. The pattern number and its corresponding data points (X) with label (Y) for the training set are
given in the following table.
Pattern X Y
1 1 -1
2 3 -1
3 7 +1
The SVM objective function is “g(x) = wx + b”. Using the training set, find the values of x, w
and b in the objective function. Also, validate the function, for the test data x=9 to find the class
label.
Part – C (2×10 = 20 Marks)
10. Construct the decision tree for the dataset given in the following table.
Use the impurity functions: Entropy and Information gain to draw the
first level tree representation.
Height Hair Eyes Attractive?
Small Blonde Brown No
Tall Dark Brown No
Tall Blonde Blue Yes
Tall Dark Blue No
Small Dark Blue No
Tall Red Blue Yes
Tall Blonde Brown No
Small Blonde Blue Yes
11. Construct the decision tree for the dataset given in the following table.
Use the Gini Index impurity function to draw the final decision tree.
Write the corresponding rules in the Implication form (if condition –
then).
Past Trend Open Trading Return
Interest Volume
https://blog.quantinsti.com/gini-index/
12. Explain Linear Discriminant Analysis (LDA) as a sequence of steps
with necessary equations and state few applications of LDA.
Definition, Equations -8mark
Applications – 2 mark
13. Explain Principal Component Analysis (PCA) as a sequence of steps
with its importance in dimensionality reduction. Compare PCA with
LDA (7 + 3)
Definition, steps – 7mark
Comparison -3mark