Professional Documents
Culture Documents
Clustering
Clustering
Clustering
1
Preview
Introduction
Partitioning methods
Hierarchical methods
Model-based methods
Density-based methods
2
Examples of Clustering Applications
Marketing: Help marketers discover distinct groups in their
customer bases, and then use this knowledge to develop
targeted marketing programs
Land use: Identification of areas of similar land use in an
earth observation database
Insurance: Identifying groups of motor insurance policy
holders with a high average claim cost
Urban planning: Identifying groups of houses according to
their house type, value, and geographical location
Seismology: Observed earth quake epicenters should be
clustered along continent faults
4
What Is a Good Clustering?
5
Requirements for Clustering
in Data Mining
Scalability
Ability to deal with different types of attributes
Discovery of clusters with arbitrary shape
Minimal domain knowledge required to determine
input parameters
Ability to deal with noise and outliers
Insensitivity to order of input records
Robustness wrt high dimensionality
Incorporation of user-specified constraints
Interpretability and usability
6
Similarity and Dissimilarity
Between Objects
Same we used for IBL (e.g, Lp norm)
Euclidean distance (p = 2):
d(i,j)
2 2 2
d (i, j )
Properties (| x x | | x x | ... | x x | )
i1 of j1 a metrici2 j2 ip : jp
d(i,j) 0
d(i,i) = 0
d(i,j) = d(j,i)
7
Major Clustering Approaches
8
Partitioning Algorithms
nearest centroid.
Compute each centroid as the mean of the
10
K-Means Clustering (contd.)
Example
10 10
9 9
8 8
7 7
6 6
5 5
4 4
3 3
2 2
1 1
0 0
0 1 2 3 4 5 6 7 8 9 10 0 1 2 3 4 5 6 7 8 9 10
10 10
9 9
8 8
7 7
6 6
5 5
4 4
3 3
2 2
1 1
0 0
0 1 2 3 4 5 6 7 8 9 10 0 1 2 3 4 5 6 7 8 9 10
11
Comments on the K-Means Method
Strengths
Relatively efficient: O(tkn), where n is # objects, k is
# clusters, and t is # iterations. Normally, k, t << n.
Often terminates at a local optimum. The global optimum
may be found using techniques such as simulated
annealing and genetic algorithms
Weaknesses
Applicable only when mean is defined (what about
categorical data?)
Need to specify k, the number of clusters, in advance
12
Hierarchical Clustering
Use distance matrix as clustering criteria. This method
does not require the number of clusters k as an input,
but needs a termination condition
Step 0 Step 1 Step 2 Step 3 Step 4
agglomerative
(AGNES)
a ab
b abcde
c
cde
d
de
e
divisive
Step 4 Step 3 Step 2 Step 1 Step 0 (DIANA)
13
AGNES (Agglomerative Nesting)
Produces tree of clusters (nodes)
Initially: each object is a cluster (leaf)
Recursively merges nodes that have the least dissimilarity
Criteria: min distance, max distance, avg distance, center
distance
Eventually all nodes belong to the same cluster (root)
10 10 10
9 9 9
8 8 8
7 7 7
6 6 6
5 5 5
4 4 4
3 3 3
2 2 2
1 1 1
0 0 0
0 1 2 3 4 5 6 7 8 9 10 0 1 2 3 4 5 6 7 8 9 10 0 1 2 3 4 5 6 7 8 9 10
14
A Dendrogram Shows How the
Clusters are Merged Hierarchically
15
DIANA (Divisive Analysis)
9 9
9
8 8
8
7 7
7
6 6
6
5 5
5
4 4
4
3 3
3
2 2
2
1 1
1
0 0
0
0 1 2 3 4 5 6 7 8 9 10 0 1 2 3 4 5 6 7 8 9 10
0 1 2 3 4 5 6 7 8 9 10
16
Other Hierarchical Clustering Methods
Major weakness of agglomerative clustering methods
Do not scale well: time complexity of at least O(n2),
quality of sub-clusters
CURE: selects well-scattered points from the cluster and
17
BIRCH
BIRCH: Balanced Iterative Reducing and Clustering using
Hierarchies (Zhang, Ramakrishnan & Livny, 1996)
Incrementally construct a CF (Clustering Feature) tree
Parameters: max diameter, max children
Phase 1: scan DB to build an initial in-memory CF tree
(each node: #points, sum, sum of squares)
Phase 2: use an arbitrary clustering algorithm to cluster
the leaf nodes of the CF-tree
Scales linearly: finds a good clustering with a single scan
and improves the quality with a few additional scans
Weaknesses: handles only numeric data, sensitive to order
of data records.
18
Clustering Feature Vector
9
(3,4)
(2,6)
8
(4,5)
5
1
(4,7)
(3,8)
0
0 1 2 3 4 5 6 7 8 9 10
19
CF Tree Root
Non-leaf node
CF1 CF2 CF3 CF5
child1 child2 child3 child5
20
CURE (Clustering Using
REpresentatives)
23
Data Partitioning and Clustering
s = 50
p=2 s/pq = 5
s/p = 25
y
y y
x
y y
x x
x x
24
Cure: Shrinking Representative Points
y y
x x
27
EM Algorithm for Mixtures
Prob(mean|point) = Prob(mean)
Prob(point|mean) / Prob(point)
M step:
Each mean = Weighted avg. of points
Weight = Prob(mean|point)
28
EM Algorithm (contd.)
29
AutoClass
31
A COBWEB Tree
32
Neural Network Approaches
33
Competitive Learning
34
Self-Organizing Feature Maps
35
Density-Based Clustering
Clustering based on density (local cluster criterion),
such as density-connected points
Major features:
Discover clusters of arbitrary shape
Handle noise
One scan
Need density parameters as termination condition
Representative algorithms:
DBSCAN (Ester et al., 1996)
36
Definitions (I)
Two parameters:
Eps: Maximum radius of neighborhood
MinPts: Minimum number of points in an Eps-
neighborhood of a point
NEps(p) ={q Є D | dist(p,q) <= Eps}
Directly density-reachable: A point p is directly density-
reachable from a point q wrt. Eps, MinPts iff
1) p belongs to NEps(q)
p MinPts = 5
2) q is a core point: q
Eps = 1 cm
|NEps (q)| >= MinPts
37
Definitions (II)
Density-reachable:
p
A point p is density-reachable
from a point q wrt. Eps, MinPts if p1
q
there is a chain of points p1, …, pn,
p1 = q, pn = p such that pi+1 is
directly density-reachable from pi
Density-connected
p q
A point p is density-connected to a
point q wrt. Eps, MinPts if there is
a point o such that both, p and q o
are density-reachable from o wrt.
Eps and MinPts.
38
DBSCAN: Density Based Spatial
Clustering of Applications with Noise
Outlier
Border
Eps = 1cm
Core MinPts = 5
39
DBSCAN: The Algorithm
41
DENCLUE
Uses grid cells but only keeps information about grid cells
that do actually contain data points and manages these
cells in a tree-based access structure.
Influence function: describes the impact of a data point
within its neighborhood.
Overall density of the data space can be calculated as the
sum of the influence function of all data points.
Clusters can be determined mathematically by identifying
density attractors.
Density attractors are local maxima of the overall density
function.
42
Influence Functions
Example
d ( x , y )2
f Gaussian ( x , y ) e 2 2
d ( x , xi ) 2
( x ) i 1 e
D N
2 2
f Gaussian
d ( x , xi ) 2
( x, xi ) i 1 ( xi x) e
N
f D
Gaussian
2 2
43
Density Attractors
44
Center-Defined & Arbitrary Clusters
45
Clustering: Summary
Introduction
Partitioning methods
Hierarchical methods
Model-based methods
Density-based methods
46