Different Paradigms of Pattern Recognition
Different Paradigms of Pattern Recognition
1
Different Paradigms for Pattern Recognition
• There are several paradigms in use to solve the pattern recognition
problem.
• Of the two, the statistical pattern recognition has been more popular
and received a major attention in the literature.
• The main reason for this is that most of the practical problems in this
area have to deal with noisy data and uncertainty and statistics and
probability are good tools to deal with such problems.
• On the other hand, formal language theory provides the background for
syntactic pattern recognition. Systems based on such linguistic tools,
more often than not, are not ideally suited to deal with noisy envi-
ronments. However, they are powerful in dealing with well-structured
domains. Also, recently there is a growing interest in statistical pattern
recognition because of the influence of statistical learning theory.
2
• There are several soft computing tools associated with this notion. Soft
computing techniques are tolerant of imprecision, uncertainty and ap-
proximation. These tools include neural networks, fuzzy systems and
evolutionary computation.
– neural networks,
– fuzzy set and rough set based pattern recognition schemes.
X4 X5
f2
X X1 P X6 X
8
3
X
X7 9
X2
f1
3
• The pattern P is a new sample (test sample) which has to be assigned
either to Class ‘X’ or Class ‘+’. There are different possibilities; some
of them are
4
neighbours. Suppose such a weighted classifier assigns a weight of
0.4 to the first neighbour (pattern X1 , labelled ‘X’), a weight of
0.35 to the second neighbour (pattern X6 from class ‘+’) and a
weight of 0.25 to the third neighbour (pattern X7 from class ‘+’).
We first add the weights of the neighbours of P coming from the
same class. So, the sum of the weights for class ‘X’, WX is 0.4 as
only the first neighbour is from ‘X’. The sum of the weights for
class ‘+’, W+ is 0.6 (0.35 ‘ +′ 0.25) corresponding the remaining
two neighbours (8 and 6) from class ‘+’. So, P is assigned class
label ‘+’.
– In a system that is built to classify humans into tall, medium and
short, the abstractions, learnt from examples, facilitate assigning
one of these class labels (tall, medium or short) to a newly en-
countered human. Here, the class labels are semantic; they convey
some meaning.
– In the case of clustering, we can group a collection of unlabelled
patterns also; in such a case, the labels assigned to each group of
patterns is syntactic, simply the cluster identity.
– Several times, it is possible that there is a large training data
which can be directly used for classification. In such a context,
clustering can be used to generate abstractions of the data and use
these abstractions for classification. For example, sets of patterns
corresponding to each of the classes can be clustered to form sub-
classes. Each such subclass (cluster) can berepresented by a single
prototypical pattern; these representative patterns can be used to
build the classifier instead of the entire data set.
Importance of Representation
5
• Such a similarity function is computed based on the representation of
patterns; the representation scheme plays a crucial role in classification.
Example
Consider the following data where humans are to be categorized into tall
and short. The classes are represented using the feature Weight. If a newly
encountered person weighs 46 KGs, then he/she may be assigned the class
label short because 46 is closer to 50. However, such an assignment does not
appeal to us because we know that weight and the class labels tall and short
do not correlate well; a feature such as Height is more appropriate. Module
2 deals with representation of patterns and classes.
6
Assignment
1. Consider a collection of data items bought in a supermarket. The
features include cost of the item, size of the item and the class label.
The data is shown in the following table. Consider a new item with
cost = 34 and volume = 8. How do you classify this item using the
NNC? How about KNNC with K = 3?
3. Consider a variant of the previous problem where the classes are small
circle and big circle. How do you classify such objects?
7
item no cost in Rs. volume in cm3 Class label
1 10 6 inexpensive
2 15 6 inexpensive
3 25 6 inexpensive
4 50 10 expensive
5 45 10 expensive
6 47 12 expensive
Further Reading
References
[1] V. Susheela Devi, M. Narasimha Murty. Pattern Recognition: An In-
troduction Universities Press, Hyderabad, 2011.
[2] R.O. Duda, P.E. Hart, D.G. Stork. Pattern Classification John Wiley
and Sons, 2000.