Clustering

advertisement
Clustering
Quan Zou
P.H.D. Assistant Professor
http://datamining.xmu.edu.cn/~zq/
Outline
Introduction of Clustering
K-means Clustering
Hierarchical Clustering
What is Clustering
 Cluster: A collection of data objects
 similar (or related) to one another within the same group
 dissimilar (or unrelated) to the objects in other groups
 Cluster analysis (or clustering, data segmentation, …)
 Finding similarities between data according to the characteristics
found in the data and grouping similar data objects into clusters
 Unsupervised learning: no predefined classes (i.e., learning by
observations vs. learning by examples: supervised)
3
Clustering for Data Understanding
and Applications
 Biology: taxonomy of living things: kingdom, phylum, class, order,
family, genus and species
 Information retrieval: document clustering
 Marketing: Help marketers discover distinct groups in their customer
bases, and then use this knowledge to develop targeted marketing
programs
 Typical applications
 As a stand-alone tool to get insight into data distribution
 As a preprocessing step for other algorithms
5
Clustering as a Preprocessing Tool
(Utility)
 Summarization:
 Preprocessing for regression, PCA, classification, and association
analysis
 Compression:
 Image processing: vector quantization
 Finding K-nearest Neighbors
 Localizing search to one or a small number of clusters
 Outlier detection
 Outliers are often viewed as those “far away” from any cluster
6
聚类分析的几种方法
 划分方法
 将数据对象分为几组
 每组至少包含一个对象
 每个对象属于一组(扩展)
 层次方法
 凝聚:自底向上
 分裂:自顶向下
 基于网格的方法
 对象空间作为一个网格结构
向量化
聚类分析的一般准备
属性
颜色
长宽比
1
255106106
1.2
1
2
25511486
1.0
2
3
255239219
0.5
3
相异度矩阵
1
2
3
0
0
0
The K-Means Clustering Method
 Given k, the k-means algorithm is implemented in four steps:
 Partition objects into k nonempty subsets
 Compute seed points as the centroids of the clusters of the
current partitioning (the centroid is the center, i.e., mean
point, of the cluster)
 Assign each object to the cluster with the nearest seed point
 Go back to Step 2, stop when the assignment does not change
9
An Example of K-Means
Clustering
K=2
Arbitrarily
partition
objects into
k groups
The initial data set
Update the
cluster
centroids
Loop if
needed
Reassign objects
Do loop
Until no
change
Update the
cluster
centroids
10
Example
 Problem
Suppose we have 4 types of medicines and each has two attributes (pH
and weight index). Our goal is to group these objects into K=2 group of
medicine.
Medicin
D
Weight
e
pH-
C
Index
A
1
1
B
2
1
C
4
3
D
5
4
A
B
11
Example
 Step 1: Use initial seed points for
partitioning
c1  A , c 2  B
Euclidean distance
d( D , c1 )  ( 5  1)2  ( 4  1)2  5
d( D , c2 )  ( 5  2)2  ( 4  1)2  4.24
Assign each object to the cluster
with the nearest seed point
12
Example
 Step 2: Compute new centroids of the
current partition
Knowing the members of each
cluster, now we compute the new
centroid of each group based on
these new memberships.
c1  (1, 1)
2  4  5 1 3  4 

c2  
,

3
3


 (11/ 3 , 8 / 3)
 ( 3.67, 2.67)
13
Example
 Step 2: Renew membership based on new
centroids
Compute the distance of all
objects to the new centroids
Assign the membership to objects
14
Example
 Step 3: Repeat the first two steps until its
convergence
Knowing the members of each
cluster, now we compute the new
centroid of each group based on
these new memberships.
1
 1 2 11
c1  
,
  (1 , 1)
2 
2
 2
1
1
 45 34
c2  
,
  (4 , 3 )
2 
2
2
 2
15
Example
 Step 3: Repeat the first two steps until its
convergence
Compute the distance of all
objects to the new centroids
Stop due to no new assignment
16
K-means Demo
1. User set up the number of
clusters they’d like. (e.g. k=5)
17
K-means Demo
1. User set up the number of
clusters they’d like. (e.g. K=5)
2. Randomly guess K cluster
Center locations
18
K-means Demo
1. User set up the number of
clusters they’d like. (e.g. K=5)
2. Randomly guess K cluster
Center locations
3. Each data point finds out
which Center it’s closest to.
(Thus each Center “owns” a
set of data points)
19
K-means Demo
1. User set up the number of
clusters they’d like. (e.g. K=5)
2. Randomly guess K cluster
centre locations
3. Each data point finds out
which centre it’s closest to.
(Thus each Center “owns” a
set of data points)
4. Each centre finds the centroid
of the points it owns
20
K-means Demo
1. User set up the number of
clusters they’d like. (e.g. K=5)
2. Randomly guess K cluster
centre locations
3. Each data point finds out
which centre it’s closest to.
(Thus each centre “owns” a
set of data points)
4. Each centre finds the centroid
of the points it owns
5. …and jumps there
21
K-means Demo
1. User set up the number of
clusters they’d like. (e.g. K=5)
2. Randomly guess K cluster
centre locations
3. Each data point finds out
which centre it’s closest to.
(Thus each centre “owns” a
set of data points)
4. Each centre finds the centroid
of the points it owns
5. …and jumps there
6. …Repeat until terminated!
22
K-mean Algorithm
23
Variations of the K-Means Method
• Most of the variants of the k-means which differ in
– Selection of the initial k means
– Dissimilarity calculations
– Strategies to calculate cluster means
• Handling categorical data: k-modes
– Replacing means of clusters with modes
– Using new dissimilarity measures to deal with categorical objects
– Using a frequency-based method to update modes of clusters
24
讨论
 准则带来问题
局部最优
(选择恰当的初始点)
离群点
(每次循环的时候删除离群点)
K值的选取
(Xmeans)
球状簇
(层次聚类)
局部最优
较好的初始点
不好的初始点
讨论
 准则带来问题
局部最优
(选择恰当的初始点)
离群点
(每次循环的时候删除离群点)
K值的选取
(Xmeans)
球状簇
(层次聚类)
讨论
 准则带来问题
局部最优
(选择恰当的初始点)
离群点
(每次循环的时候删除离群点)
K值的选取
(Xmeans,平均直径或半径)
球状簇
(层次聚类)
讨论
 准则带来问题
局部最优
(选择恰当的初始点)
离群点
(每次循环的时候删除离群点)
K值的选取
(Xmeans)
球状簇
(层次聚类)
球状簇
Hierarchical Clustering
 Use distance matrix as clustering criteria. This method does not
require the number of clusters k as an input, but needs a termination
condition
 Termination condition: Number of clusters
Step 0
a
Step 1
Step 2 Step 3 Step 4
ab
b
abcde
c
cde
d
de
e
Step 4
agglomerative
(AGNES)
Step 3
Step 2 Step 1 Step 0
divisive
(DIANA)
32
层次聚类方法概述
 层次聚类方法对给定的数据集进行层次的分解,直到某种条
件满足为止。具体又可分为:
凝聚的层次聚类:一种自底向上的策略,首先将每个对象作为一个簇
,然后合并这些原子簇为越来越大的簇,直到某个终结条件被满足。
分裂的层次聚类:采用自顶向下的策略,它首先将所有对象置于一个
簇中,然后逐渐细分为越来越小的簇,直到达到了某个终结条件。
 层次凝聚的代表是AGNES算法。层次分裂的代表是DIANA
算法。
AGNES算法
 AGNES(Agglomerative NESting)算法最初将每个对象作为一
个簇,然后这些簇根据某些准则被一步步地合并。两个簇间的
相似度由这两个不同簇中距离最近的数据点对的相似度来确定
。聚类的合并过程反复进行直到所有的对象最终满足簇数目。
自底向上凝聚算法(AGNES):
输入:包含n个对象的数据库,终止条件簇的数目k。
输出:k个簇,达到终止条件规定簇数目。
(1)将每个对象当成一个初始簇;
(2)REPEAT
(3)根据两个簇中最近的数据点找到最近的两个簇;
(4)合并两个簇,生成新的簇的集合;
(5)UNTIL达到定义的簇的数目;
AGNES算法例题
序号
属性1
属性2
第1步:根据初始簇计算每个簇之间的距离,随机找出距离最小的两
个簇,进行合并,最小距离为1,合并后1,2两个点合并为一个簇。
1
1
1
2
1
2
3
2
1
第2步:对上一次合并后的簇计算簇间距离,找出距离最近的两个簇
进行合并,合并后3,4点成为一簇。
4
2
2
第3步:重复第2步的工作,5,6点成为一簇。
5
3
4
第4步:重复第2步的工作,7,8点成为一簇。
6
3
5
第5步:合并{1,2},{3,4}成为一个包含四个点的簇。
7
4
4
第6步:合并{5,6},{7,8},由于合并后的簇的数目已经达到了用户输
入的终止条件,程序终止。
8
4
5
步骤 最近的簇距离
最近的两个簇
合并后的新簇
1
1
{1},{2}
{1,2},{3},{4},{5},{6},{7},{8}
2
1
{3},{4}
{1,2},{3,4},{5},{6},{7},{8}
3
1
{5},{6}
{1,2},{3,4},{5,6},{7},{8}
4
1
{7},{8}
{1,2},{3,4},{5,6},{7,8}
5
1
{1,2},{3,4}
{1,2,3,4},{5,6},{7,8}
6
1
{5,6},{7,8}
{1,2,3,4},{5,6,7,8}结束
AGNES性能分析
 AGNES算法比较简单,但经常会遇到合并点选择的困难。假如一旦一组
对象被合并,下一步的处理将在新生成的簇上进行。已做处理不能撤销
,聚类之间也不能交换对象。如果在某一步没有很好的选择合并的决定
,可能会导致低质量的聚类结果。
 假定在开始的时候有n个簇,在结束的时候有1个簇,因此在主循环中有
n此迭代,在第i次迭代中,我们必须在n-i+1个簇中找到最靠近的两个聚
类。另外算法必须计算所有对象两两之间的距离,因此这个算法的复杂
度为O(n²),该算法对于n很大的情况是不适用的。
DIANA算法
 DIANA(DIvisive ANAlysis)算法是典型的分裂聚类方法。
 在聚类中,用户能定义希望得到的簇数目作为一个结束条件。
同时,它使用下面测度方法:
簇的直径:在一个簇中的任意两个数据点的距离中的最大值。
算法 DIANA(自顶向下分裂算法)
输入:包含n个对象的数据库,终止条件簇的数目k。
输出:k个簇,达到终止条件规定簇数目。
(1)将所有对象整个当成一个初始簇;
(2) FOR (i=1; i≠k; i++) DO BEGIN
(3)
在所有簇中挑出具有最大直径的簇C;
(4)
找出C中与其它点平均相异度最大的一个点p并把p放入splinter
group,剩余的放在old party中;
(5). REPEAT
(6)
在old party里找出到最近的splinter group中的点的距离不
大于到old party中最近点的距离的点,并将该点加入splinter
group。
(7)
UNTIL 没有新的old party的点被分配给splinter group;
(8) splinter group和old party为被选中的簇分裂成的两个簇,与其它
簇一起组成新的簇集合。
(9) END.
DIANA算法例题
序号
属性 1
属性 2
1
1
1
2
1
2
3
2
1
4
2
2
5
3
4
6
3
5
7
4
4
8
4
5
步骤
1
2
3
4
5
第1步,找到具有最大直径的簇,对簇中的每个点计算平均相异度(假定采用
是欧式距离)。
1的平均距离:(1+1+1.414+3.6+4.24+4.47+5)/7=2.96
类似地,2的平均距离为2.526;3的平均距离为2.68;4的平均距离为2.18;
5的平均距离为2.18;6的平均距离为2.68;7的平均距离为2.526;8的
平均距离为2.96。
挑出平均相异度最大的点1放到splinter group中,剩余点在old party中。
第2步,在old party里找出到最近的splinter group中的点的距离不大于到old
party中最近的点的距离的点,将该点放入splinter group中,该点是2。
第3步,重复第2步的工作,splinter group中放入点3。
第4步,重复第2步的工作,splinter group中放入点4。
第5步,没有在old party中的点放入了splinter group中且达到终止条件
(k=2),程序终止。如果没有到终止条件,因该从分裂好的簇中选一
个直径最大的簇继续分裂。
具有最大直径的簇
splinter group
{1,2,3,4,5,6,7,8} {1}
{1,2,3,4,5,6,7,8} {1,2}
{1,2,3,4,5,6,7,8} {1,2,3}
{1,2,3,4,5,6,7,8} {1,2,3,4}
{1,2,3,4,5,6,7,8} {1,2,3,4}
Old party
{2,3,4,5,6,7,8}
{3,4,5,6,7,8}
{4,5,6,7,8}
{5,6,7,8}
{5,6,7,8} 终止
关于层次聚类的讨论
高维灾难
维度非常高时,各个点之间的距离都近似
维度非常高时,各个向量几乎都正交
什么时候会高维?
• 文本、图像、基因表达数据
关于层次聚类的讨论
停止准则
簇的个数事先已知
自底向上
• 新合并的簇出现不合理的情况
– 要求簇内所有点到质心的距离都大于阈值
生成进化树,找出大于阈值的边
启发式课题
在本章介绍的聚类算法中,频繁用到了“
质心”的概念,对于欧式空间中的向量,
我们很容易计算质心。但对于非欧式空间
的样本,比如string、text,我们如何计算
一个cluster的质心?
Email: zouquan@xmu.edu.cn
作业
请描述K-means算法的原理,并分析其主要缺
点。
层次聚类的基本策略有哪两种?分别写出其算
法
有监督学习和无监督学习的区别?
Download