Iv Vii: 20CSPC702
Iv Vii: 20CSPC702
UNIT-V
ADVANCED LEARNING
IV VII
20CSPC702
MACHINE LEARNING TECHNIQUES
CS8082
MACHINE LEARNING TECHNIQUES
● Sequential covering algorithms perform n*k primitive search steps to learn n rules
each containing k attribute- value tests. If the decision trees is a complete binary
tree, it makes (n-1) primitive search steps where n is the number of paths (i.e.,
rules).
● So Sequential Covering Algorithms must be supported by additional data, but have
the advantage of allowing rules with different tests.
CS8082
MACHINE LEARNING TECHNIQUES
● In specific to general there are many maximally specific hypothesis (the training
data).
● Golem chooses several randomly and picks the best learned hypothesis.
CS8082
MACHINE LEARNING TECHNIQUES
(iv) Post-pruning :
● Relative frequency :
AQ -
● m-estimate of accuracy :
CN2 -
p prior probability that a randomly drawn example from the entire data set
will have the classification assigned by the rule
m weight, or equivalent number of examples for weighting this prior p
CS8082
MACHINE LEARNING TECHNIQUES
● Entropy :
CN2 -
Example :
Target concept : Daughter (x, y), defined over pairs of people x and y
The value of Daughter(x, y) is True when x is the daughter of y, and False
otherwise.
Target attribute : Daughter
The subscript on each attribute name indicates which of the two persons is being
described.
CS8082
MACHINE LEARNING TECHNIQUES
To collect a number of such training examples for the target concept Daughter1,2 and
provide them to a propositional rule learner such as CN2 or C4.5,
The problem is that propositional representations offer no general way to describe the
essential relations among the values of the attributes.
Hence,
IF Parents(x, z) ^ Ancestor(z, y)
THEN Ancestor(x, y)
Terminology I :
● Every well-formed expression is composed of constants (e.g., Mary, 23, or Joe),
variables (e.g., x), predicates (e.g., Female, as in Female(Mary)), and functions
(e.g., age is in age(Mary)).
● A term is any constant, any variable, or any function applied to any term.
Examples include (Mary, x, age(Mary), age(x)).
● A literal is any predicate (or its negation) applied to any set of terms. Examples
include (Female(Mary),¬Female(x), Greater than(age(Mary),20)).
● A ground literal is a literal that does not contain any variables (e.g.,
¬Female(Joe)).
CS8082
MACHINE LEARNING TECHNIQUES
Terminology II :
● A negative literal is a literal containing a negated predicate (e.g., ¬Female(Joe)).
○
CS8082
MACHINE LEARNING TECHNIQUES
Terminology III :
● For any literals A and B, the expression (A¬B) is equivalent to (A v ¬B), and the
expression ¬(A ^ B) is equivalent to (¬A v ¬B). Therefore a Horn clause can
equivalently be written as the disjunction H v ¬L1 v … v ¬Ln.
● A substitution is any function that replaces variables by terms. For example, the
substitution {x/3, y/z} replaces the variable x by the term 3 and replaces the
variable y by the term z. Given a substitution q and a literal L we write Lq to denote
the result of applying substitution q to L.
● A unifying substitution for two literals L1 and L2 is any substitution q such that
L1q = L2q.