0% found this document useful (0 votes)
39 views

What Is Cluster Analysis?: Unsupervised Learning Stand-Alone Tool Preprocessing Step

The document discusses cluster analysis and its applications. Cluster analysis involves grouping similar data objects into clusters, where objects within a cluster are similar to each other and dissimilar to objects in other clusters. It is an unsupervised learning technique used to gain insight into data distribution or as a preprocessing step. Examples of applications include marketing, land use analysis, insurance, and city planning. The document also discusses evaluating the quality of clustering results.

Uploaded by

drazil
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPT, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
39 views

What Is Cluster Analysis?: Unsupervised Learning Stand-Alone Tool Preprocessing Step

The document discusses cluster analysis and its applications. Cluster analysis involves grouping similar data objects into clusters, where objects within a cluster are similar to each other and dissimilar to objects in other clusters. It is an unsupervised learning technique used to gain insight into data distribution or as a preprocessing step. Examples of applications include marketing, land use analysis, insurance, and city planning. The document also discusses evaluating the quality of clustering results.

Uploaded by

drazil
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPT, PDF, TXT or read online on Scribd
You are on page 1/ 21

What is Cluster Analysis?

 Cluster: a collection of data objects


 Similar to one another within the same cluster
 Dissimilar to the objects in other clusters
 Cluster analysis
 Finding similarities between data according to the
characteristics found in the data and grouping similar
data objects into clusters
 Unsupervised learning: no predefined classes
 Typical applications
 As a stand-alone tool to get insight into data distribution
 As a preprocessing step for other algorithms
March 28, 2019 Data Mining: Concepts and Techniques 1
Examples of Clustering Applications
 Marketing: Help marketers discover distinct groups in their customer
bases, and then use this knowledge to develop targeted marketing
programs
 Land use: Identification of areas of similar land use in an earth
observation database
 Insurance: Identifying groups of motor insurance policy holders with
a high average claim cost
 City-planning: Identifying groups of houses according to their house
type, value, and geographical location
 Earth-quake studies: Observed earth quake epicenters should be
clustered along continent faults

March 28, 2019 Data Mining: Concepts and Techniques 2


Quality: What Is Good Clustering?
 A good clustering method will produce high quality
clusters with
 high intra-class similarity
 low inter-class similarity
 The quality of a clustering result depends on both the
similarity measure used by the method and its
implementation
 The quality of a clustering method is also measured by its
ability to discover some or all of the hidden patterns

March 28, 2019 Data Mining: Concepts and Techniques 3


Measure the Quality of Clustering

 Dissimilarity/Similarity metric: Similarity is expressed in


terms of a distance function, typically metric: d(i, j)
 There is a separate “quality” function that measures the
“goodness” of a cluster.
 The definitions of distance functions are usually very
different for interval-scaled, boolean, categorical, ordinal
ratio, and vector variables.
 Weights should be associated with different variables
based on applications and data semantics.
 It is hard to define “similar enough” or “good enough”
 the answer is typically highly subjective.
March 28, 2019 Data Mining: Concepts and Techniques 4
Requirements of Clustering in Data Mining
 Scalability
 Ability to deal with different types of attributes
 Ability to handle dynamic data
 Discovery of clusters with arbitrary shape
 Able to deal with noise and outliers
 Insensitive to order of input records
 High dimensionality
 Incorporation of user-specified constraints
 Interpretability and usability

March 28, 2019 Data Mining: Concepts and Techniques 5


Data Structures
 Data matrix
 x11 ... x1f ... x1p 
 (two modes)  
 ... ... ... ... ... 
x ... xif ... xip 
 i1 
 ... ... ... ... ... 
x ... xnf ... xnp 
 n1 

 Dissimilarity matrix  0 
 (one mode)  d(2,1) 0 
 
 d(3,1) d ( 3,2) 0 
 
 : : : 
d ( n,1) d ( n,2) ... ... 0

March 28, 2019 Data Mining: Concepts and Techniques 6


Type of data in clustering analysis

 Interval-scaled variables
 Binary variables
 Nominal, ordinal, and ratio variables
 Variables of mixed types

March 28, 2019 Data Mining: Concepts and Techniques 7


Interval-valued variables

 Standardize data
 Calculate the mean absolute deviation:
s f  1n (| x1 f  m f |  | x2 f  m f | ... | xnf  m f |)

where m f  1n (x1 f  x2 f  ...  xnf )


.

 Calculate the standardized measurement (z-score)


xif  m f
zif  sf
 Using mean absolute deviation is more robust than using
standard deviation

March 28, 2019 Data Mining: Concepts and Techniques 8


Similarity and Dissimilarity Between
Objects
 Distances are normally used to measure the similarity or
dissimilarity between two data objects
 Some popular ones include: Minkowski distance:
d (i, j)  q (| x  x |q  | x  x | q ... | x  x |q )
i1 j1 i2 j2 ip jp
where i = (xi1, xi2, …, xip) and j = (xj1, xj2, …, xjp) are
two p-dimensional data objects, and q is a positive
integer
 If q = 1, d is Manhattan distance
d (i, j) | x  x |  | x  x | ... | x  x |
i1 j1 i2 j 2 ip jp

March 28, 2019 Data Mining: Concepts and Techniques 9


Similarity and Dissimilarity Between
Objects (Cont.)
 If q = 2, d is Euclidean distance:
d (i, j)  (| x  x |2  | x  x |2 ... | x  x |2 )
i1 j1 i2 j2 ip jp
 Properties
 d(i,j)  0
 d(i,i) = 0
 d(i,j) = d(j,i)
 d(i,j)  d(i,k) + d(k,j)
 Also, one can use weighted distance, parametric
Pearson product moment correlation, or other
disimilarity measures

March 28, 2019 Data Mining: Concepts and Techniques 10


Binary Variables
Object j
1 0 sum
 A contingency table for
1 a b a b
binary data Object i
0 c d cd
sum a  c b  d p
 Distance measure for
symmetric binary variables:
 Distance measure for bc
d (i, j) 
asymmetric binary variables: a bc
 Jaccard coefficient (similarity
measure for asymmetric simJaccard (i, j)  a
a bc
binary variables):
March 28, 2019 Data Mining: Concepts and Techniques 11
Dissimilarity between Binary Variables

 Example
Name Gender Fever Cough Test-1 Test-2 Test-3 Test-4
Jack M Y N P N N N
Mary F Y N P N P N
Jim M Y P N N N N

 gender is a symmetric attribute


 the remaining attributes are asymmetric binary
 let the values Y and P be set to 1, and the value N be set to 0
01
d ( jack , mary )   0.33
2 01
11
d ( jack , jim )   0.67
111
1 2
d ( jim , mary )   0.75
11 2
March 28, 2019 Data Mining: Concepts and Techniques 12
Nominal Variables

 A generalization of the binary variable in that it can take


more than 2 states, e.g., red, yellow, blue, green
 Method 1: Simple matching
 m: # of matches, p: total # of variables

d (i, j)  p 
p
m

 Method 2: use a large number of binary variables


 creating a new binary variable for each of the M
nominal states

March 28, 2019 Data Mining: Concepts and Techniques 13


Ordinal Variables

 An ordinal variable can be discrete or continuous


 Order is important, e.g., rank
 Can be treated like interval-scaled
 replace xif by their rank rif {1,..., M f }
 map the range of each variable onto [0, 1] by replacing
i-th object in the f-th variable by
rif 1
zif 
M f 1
 compute the dissimilarity using methods for interval-
scaled variables

March 28, 2019 Data Mining: Concepts and Techniques 14


Ratio-Scaled Variables

 Ratio-scaled variable: a positive measurement on a


nonlinear scale, approximately at exponential scale,
such as AeBt or Ae-Bt
 Methods:
 treat them like interval-scaled variables— not a good
choice! (why?—the scale can be distorted)
 apply logarithmic transformation
yif = log(xif)
 treat them as continuous ordinal data treat their rank as
interval-scaled

March 28, 2019 Data Mining: Concepts and Techniques 15


Variables of Mixed Types

 A database may contain all the six types of variables


 symmetric binary, asymmetric binary, nominal,

ordinal, interval and ratio


 One may use a weighted formula to combine their
effects  pf  1 ij( f ) d ij( f )
d (i, j ) 
 pf  1 ij( f )
 f is binary or nominal:

dij(f) = 0 if xif = xjf , or dij(f) = 1 otherwise


 f is interval-based: use the normalized distance

 f is ordinal or ratio-scaled

 compute ranks r and


if
z 
r 1
if

 and treat z as interval-scaled if M 1


if f

March 28, 2019 Data Mining: Concepts and Techniques 16


Vector Objects

 Vector objects: keywords in documents, gene


features in micro-arrays, etc.
 Broad applications: information retrieval, biologic
taxonomy, etc.
 Cosine measure

 A variant: Tanimoto coefficient

March 28, 2019 Data Mining: Concepts and Techniques 17


Major Clustering Approaches (I)

 Partitioning approach:
 Construct various partitions and then evaluate them by some criterion,
e.g., minimizing the sum of square errors
 Typical methods: k-means, k-medoids, CLARANS
 Hierarchical approach:
 Create a hierarchical decomposition of the set of data (or objects) using
some criterion
 Typical methods: Diana, Agnes, BIRCH, ROCK, CAMELEON
 Density-based approach:
 Based on connectivity and density functions
 Typical methods: DBSACN, OPTICS, DenClue

March 28, 2019 Data Mining: Concepts and Techniques 18


Major Clustering Approaches (II)
 Grid-based approach:
 based on a multiple-level granularity structure
 Typical methods: STING, WaveCluster, CLIQUE
 Model-based:
 A model is hypothesized for each of the clusters and tries to find the best
fit of that model to each other
 Typical methods: EM, SOM, COBWEB
 Frequent pattern-based:
 Based on the analysis of frequent patterns
 Typical methods: pCluster
 User-guided or constraint-based:
 Clustering by considering user-specified or application-specific constraints
 Typical methods: COD (obstacles), constrained clustering
March 28, 2019 Data Mining: Concepts and Techniques 19
Summary
 Cluster analysis groups objects based on their similarity
and has wide applications
 Measure of similarity can be computed for various types
of data
 Clustering algorithms can be categorized into partitioning
methods, hierarchical methods, density-based methods,
grid-based methods, and model-based methods
 Outlier detection and analysis are very useful for fraud
detection, etc. and can be performed by statistical,
distance-based or deviation-based approaches
 There are still lots of research issues on cluster analysis

March 28, 2019 Data Mining: Concepts and Techniques 20


Problems and Challenges

 Considerable progress has been made in scalable


clustering methods
 Partitioning: k-means, k-medoids, CLARANS
 Hierarchical: BIRCH, ROCK, CHAMELEON
 Density-based: DBSCAN, OPTICS, DenClue
 Grid-based: STING, WaveCluster, CLIQUE
 Model-based: EM, Cobweb, SOM
 Frequent pattern-based: pCluster
 Constraint-based: COD, constrained-clustering
 Current clustering techniques do not address all the
requirements adequately, still an active area of research
March 28, 2019 Data Mining: Concepts and Techniques 21

You might also like