SlideShare a Scribd company logo
Prof. Pier Luca Lanzi
Clustering: Introduction
Data Mining andText Mining (UIC 583 @ Politecnico di Milano)
Prof. Pier Luca Lanzi
Readings
•  Mining of Massive Datasets (Chapter 7, Section 3.5)
2
Prof. Pier Luca Lanzi
3
Prof. Pier Luca Lanzi
Prof. Pier Luca Lanzi
Prof. Pier Luca Lanzi
Clustering algorithms group a collection of data points
into “clusters” according to some distance measure
Data points in the same cluster should have
a small distance from one another
Data points in different clusters should be at
a large distance from one another.
Prof. Pier Luca Lanzi
Clustering finds “natural” grouping/structure in un-labeled data
(Unsupervised Learning)
Prof. Pier Luca Lanzi
What is Cluster Analysis?
•  A cluster is a collection of data objects
§ Similar to one another within the same cluster
§ Dissimilar to the objects in other clusters
•  Cluster analysis
§ Given a set data points try to understand their structure
§ Finds similarities between data according to the characteristics
found in the data
§ Groups similar data objects into clusters
§ It is unsupervised learning since there is no predefined classes
•  Typical applications
§ Stand-alone tool to get insight into data
§ Preprocessing step for other algorithms
8
Prof. Pier Luca Lanzi
Clustering Methods
•  Hierarchical vs point assignment
•  Numeric and/or symbolic data
•  Deterministic vs. probabilistic
•  Exclusive vs. overlapping
•  Hierarchical vs. flat
•  Top-down vs. bottom-up
9
Prof. Pier Luca Lanzi
Clustering Applications
•  Marketing
§ Help marketers discover distinct groups in their customer bases,
and then use this knowledge to develop targeted marketing
programs
•  Land use
§ Identification of areas of similar land use in an earth observation
database
•  Insurance
§ Identifying groups of motor insurance policy holders with a high
average claim cost
•  City-planning
§ Identifying groups of houses according to their house type, value,
and geographical location
•  Earth-quake studies
§ Observed earth quake epicenters should be clustered along
continent faults
10
Prof. Pier Luca Lanzi
What Is Good Clustering?
•  A good clustering consists of high quality clusters with
§ High intra-class similarity
§ Low inter-class similarity
•  The quality of a clustering result depends on both the similarity
measure used by the method and its implementation
•  The quality of a clustering method is also measured by its ability
to discover some or all of the hidden patterns
•  Evaluation
§ Various measure of intra/inter cluster similarity
§ Manual inspection
§ Benchmarking on existing labels
11
Prof. Pier Luca Lanzi
Measure the Quality of Clustering
•  Dissimilarity/Similarity metric: Similarity is expressed in terms of a
distance function, typically metric d(i, j)
•  There is a separate “quality” function that measures the “goodness” of
a cluster
•  The definitions of distance functions are usually very different for
interval-scaled, boolean, categorical, ordinal ratio, and vector variables
•  Weights should be associated with different variables based on
applications and data semantics
•  It is hard to define “similar enough” or “good enough” as the answer is
typically highly subjective
12
Prof. Pier Luca Lanzi
Data Structures
0
d(2,1) 0
d(3,1) d(3,2) 0
: : :
d(n,1) d(n,2) ... ... 0
!

#
#
#
#
#
#
$
%






Outlook	
   Temp	
   Humidity	
   Windy	
   Play	
  
Sunny	
   Hot	
   High	
   False	
   No	
  
Sunny	
   Hot	
  	
   High	
  	
   True	
   No	
  
Overcast	
  	
   Hot	
  	
  	
   High	
   False	
   Yes	
  
…	
   …	
   …	
   …	
   …	
  
x
11
... x
1f
... x
1p
... ... ... ... ...
x
i1
... x
if
... x
ip
... ... ... ... ...
x
n1
... x
nf
... x
np
!

#
#
#
#
#
#
#
#
$
%








Data Matrix
13
Dis/Similarity Matrix
Prof. Pier Luca Lanzi
Type of Data in Clustering Analysis
•  Interval-scaled variables
•  Binary variables
•  Nominal, ordinal, and ratio variables
•  Variables of mixed types
14
Prof. Pier Luca Lanzi
Distance Measures
Prof. Pier Luca Lanzi
Distance Measures
•  Given a space and a set of points on this space, a distance
measure d(x,y) maps two points x and y to a real number, 
and satisfies three axioms
•  d(x,y) ≥	
 0
•  d(x,y) = 0 if and only x=y
•  d(x,y) = d(y,x)
•  d(x,y) ≤ d(x,z) + d(z,y)
16
Prof. Pier Luca Lanzi
Euclidean Distances 17
here are other distance measures that have been used for Euclidean
any constant r, we can define the Lr-norm to be the distance me
ed by:
d([x1, x2, . . . , xn], [y1, y2, . . . , yn]) = (
n
i=1
|xi − yi|r
)1/r
case r = 2 is the usual L2-norm just mentioned. Another common d
ure is the L1-norm, or Manhattan distance. There, the distance b
points is the sum of the magnitudes of the differences in each dim
called “Manhattan distance” because it is the distance one would
•  Lr-norm
•  Euclidean distance (r=2)
•  Manhattan distance (r=1)
•  L∞-norm
2 Euclidean Distances
most familiar distance measure is the one we normally think of as “dis-
e.” An n-dimensional Euclidean space is one where points are vectors of n
numbers. The conventional distance measure in this space, which we shall
to as the L2-norm, is defined:
d([x1, x2, . . . , xn], [y1, y2, . . . , yn]) =
n
i=1
(xi − yi)2
is, we square the distance in each dimension, sum the squares, and take
positive square root.
is easy to verify the first three requirements for a distance measure are
fied. The Euclidean distance between two points cannot be negative, be-
e the positive square root is intended. Since all squares of real numbers are
egative, any i such that xi ̸= yi forces the distance to be strictly positive.
he other hand, if xi = yi for all i, then the distance is clearly 0. Symmetry
ws because (xi − yi)2
= (yi − xi)2
. The triangle inequality requires a good
of algebra to verify. However, it is well understood to be a property of
Prof. Pier Luca Lanzi
Jaccard Distance
•  Jaccard distance is defined as d(x,y) = 1 – SIM(x,y) where SIM is
the Jaccard similarity,
•  Which can also be interpreted as the percentage of identical
attributes
18
Prof. Pier Luca Lanzi
Cosine Distance
•  The cosine distance between x, y is the angle that the vectors to
those points make
•  This angle will be in the range 0 to 180 degrees, regardless of
how many dimensions the space has.
•  Example: given x = (1,2,-1) and y = (2,1,1) the angle between the
two vectors is 60
19
Prof. Pier Luca Lanzi
Edit Distance
•  Used when the data points are strings
•  The distance between a string x=x1x2…xn and y=y1y2…ym is the smallest
number of insertions and deletions of single characters that will transform x
into y
•  Alternatively, the edit distance d(x, y) can be compute as the longest common
subsequence (LCS) of x and y and then,

d(x,y) = |x| + |y| - 2|LCS|
•  Example: the edit distance between x=abcde and y=acfdeg is 3 (delete b,
insert f, insert g), the LCS is acde which is coherent with the previous result
20
Prof. Pier Luca Lanzi
Hamming Distance
•  Hamming distance between two vectors is the number of
components in which they differ
•  Or equivalently, given the number of variables n, and the number
m of matching components, we define
•  Example: the Hamming distance between the vectors 10101 and
11110 is 3.
21
Prof. Pier Luca Lanzi
Ordinal Variables
•  An ordinal variable can be discrete or continuous
•  Order is important, e.g., rank
•  It can be treated as an interval-scaled
§ replace xif with their rank
§ map the range of each variable onto [0, 1] by replacing 
i-th object in the f-th variable by
§ compute the dissimilarity using methods for interval-scaled variables
22
Prof. Pier Luca Lanzi
Requirements of Clustering in Data Mining
•  Scalability
•  Ability to deal with different types of attributes
•  Ability to handle dynamic data
•  Discovery of clusters with arbitrary shape
•  Minimal requirements for domain knowledge to determine input
parameters
•  Able to deal with noise and outliers
•  Insensitive to order of input records
•  High dimensionality
•  Incorporation of user-specified constraints
•  Interpretability and usability
23
Prof. Pier Luca Lanzi
Curse of Dimensionality
in high dimensions, almost all pairs of points
are equally far away from one another
almost any two vectors are almost orthogonal

More Related Content

PDF
DMTM Lecture 11 Clustering
PDF
DMTM Lecture 20 Data preparation
PDF
DMTM 2015 - 04 Data Exploration
PDF
DMTM Lecture 04 Classification
PDF
DMTM Lecture 10 Classification ensembles
PDF
DMTM 2015 - 15 Classification Ensembles
PDF
DMTM Lecture 13 Representative based clustering
PDF
DMTM 2015 - 13 Naive bayes, Nearest Neighbours and Other Methods
DMTM Lecture 11 Clustering
DMTM Lecture 20 Data preparation
DMTM 2015 - 04 Data Exploration
DMTM Lecture 04 Classification
DMTM Lecture 10 Classification ensembles
DMTM 2015 - 15 Classification Ensembles
DMTM Lecture 13 Representative based clustering
DMTM 2015 - 13 Naive bayes, Nearest Neighbours and Other Methods

What's hot (20)

PDF
DMTM Lecture 09 Other classificationmethods
PDF
DMTM 2015 - 09 Density Based Clustering
PDF
DMTM Lecture 15 Clustering evaluation
PDF
DMTM 2015 - 10 Introduction to Classification
PDF
DMTM Lecture 03 Regression
PDF
DMTM 2015 - 08 Representative-Based Clustering
PDF
DMTM 2015 - 14 Evaluation of Classification Models
PDF
DMTM 2015 - 03 Data Representation
PDF
DMTM Lecture 06 Classification evaluation
PDF
DMTM Lecture 08 Classification rules
PDF
DMTM 2015 - 16 Data Preparation
PDF
DMTM Lecture 07 Decision trees
PDF
DMTM Lecture 12 Hierarchical clustering
PDF
DMTM Lecture 05 Data representation
PDF
DMTM 2015 - 12 Classification Rules
PDF
DMTM 2015 - 17 Text Mining Part 1
PDF
DMTM Lecture 18 Graph mining
PDF
DMTM Lecture 19 Data exploration
PDF
DMTM Lecture 16 Association rules
PDF
DMTM 2015 - 11 Decision Trees
DMTM Lecture 09 Other classificationmethods
DMTM 2015 - 09 Density Based Clustering
DMTM Lecture 15 Clustering evaluation
DMTM 2015 - 10 Introduction to Classification
DMTM Lecture 03 Regression
DMTM 2015 - 08 Representative-Based Clustering
DMTM 2015 - 14 Evaluation of Classification Models
DMTM 2015 - 03 Data Representation
DMTM Lecture 06 Classification evaluation
DMTM Lecture 08 Classification rules
DMTM 2015 - 16 Data Preparation
DMTM Lecture 07 Decision trees
DMTM Lecture 12 Hierarchical clustering
DMTM Lecture 05 Data representation
DMTM 2015 - 12 Classification Rules
DMTM 2015 - 17 Text Mining Part 1
DMTM Lecture 18 Graph mining
DMTM Lecture 19 Data exploration
DMTM Lecture 16 Association rules
DMTM 2015 - 11 Decision Trees
Ad

Viewers also liked (16)

PDF
DMTM 2015 - 02 Data Mining
PDF
DMTM 2015 - 01 Course Introduction
PDF
DMTM 2015 - 05 Association Rules
PDF
DMTM 2015 - 07 Hierarchical Clustering
PDF
Course Introduction
PDF
DMTM 2015 - 19 Graph Mining
PDF
Focus Junior - 14 Maggio 2016
PDF
Machine Learning and Data Mining: 12 Classification Rules
PDF
Lecture 9: Machine Learning in Practice (2)
PPT
Videogame Design and Programming: Conferenza d'Ateneo 18 Maggio 2011
PPT
Fitness Inheritance in Evolutionary and
PPTX
GECCO-2014 Learning Classifier Systems: A Gentle Introduction
PDF
Lecture 02 Machine Learning For Data Mining
PDF
Evolving Rules to Solve Problems: The Learning Classifier Systems Way
PDF
Lecture 04 Association Rules Basics
PDF
Machine Learning and Data Mining: 02 Machine Learning
DMTM 2015 - 02 Data Mining
DMTM 2015 - 01 Course Introduction
DMTM 2015 - 05 Association Rules
DMTM 2015 - 07 Hierarchical Clustering
Course Introduction
DMTM 2015 - 19 Graph Mining
Focus Junior - 14 Maggio 2016
Machine Learning and Data Mining: 12 Classification Rules
Lecture 9: Machine Learning in Practice (2)
Videogame Design and Programming: Conferenza d'Ateneo 18 Maggio 2011
Fitness Inheritance in Evolutionary and
GECCO-2014 Learning Classifier Systems: A Gentle Introduction
Lecture 02 Machine Learning For Data Mining
Evolving Rules to Solve Problems: The Learning Classifier Systems Way
Lecture 04 Association Rules Basics
Machine Learning and Data Mining: 02 Machine Learning
Ad

Similar to DMTM 2015 - 06 Introduction to Clustering (20)

PDF
Clustering Algorithms - Kmeans,Min ALgorithm
PPTX
Cluster analysis
PPTX
Cluster Analysis
PDF
Introduction to machine learning
PPTX
W5_CLASSIFICATION.pptxW5_CLASSIFICATION.pptx
PPT
UnSupervised Machincs4811-ch23a-clustering.ppt
PPT
cs4811-ch23a-K-means clustering algorithm .ppt
PPT
[PPT]
PDF
GIS in Public Health Research: Understanding Spatial Analysis and Interpretin...
PPTX
ODSC India 2018: Topological space creation & Clustering at BigData scale
PDF
09_dm1_knn_2022_23.pdf
PPT
cs4811-ch10c-clusuughv hgyf yfyf tering.ppt
PPT
Cs345 cl
PDF
PR07.pdf
PDF
CSA 3702 machine learning module 3
PPT
clustering and their types explanation of data mining
PPT
phylogenetics (1)...............................ppt
PPT
Plant Molecular Systematics Phylogenetics.ppt
PPT
distance based phylogenetics-methodology
PPT
Spatial data mining
Clustering Algorithms - Kmeans,Min ALgorithm
Cluster analysis
Cluster Analysis
Introduction to machine learning
W5_CLASSIFICATION.pptxW5_CLASSIFICATION.pptx
UnSupervised Machincs4811-ch23a-clustering.ppt
cs4811-ch23a-K-means clustering algorithm .ppt
[PPT]
GIS in Public Health Research: Understanding Spatial Analysis and Interpretin...
ODSC India 2018: Topological space creation & Clustering at BigData scale
09_dm1_knn_2022_23.pdf
cs4811-ch10c-clusuughv hgyf yfyf tering.ppt
Cs345 cl
PR07.pdf
CSA 3702 machine learning module 3
clustering and their types explanation of data mining
phylogenetics (1)...............................ppt
Plant Molecular Systematics Phylogenetics.ppt
distance based phylogenetics-methodology
Spatial data mining

More from Pier Luca Lanzi (14)

PDF
11 Settembre 2021 - Giocare con i Videogiochi
PDF
Breve Viaggio al Centro dei Videogiochi
PDF
Global Game Jam 19 @ POLIMI - Morning Welcome
PPTX
Data Driven Game Design @ Campus Party 2018
PDF
GGJ18 al Politecnico di Milano - Presentazione che precede la presentazione d...
PDF
GGJ18 al Politecnico di Milano - Presentazione di apertura
PDF
Presentation for UNITECH event - January 8, 2018
PDF
DMTM Lecture 17 Text mining
PDF
DMTM Lecture 14 Density based clustering
PDF
DMTM Lecture 01 Introduction
PDF
DMTM Lecture 02 Data mining
PDF
VDP2016 - Lecture 16 Rendering pipeline
PDF
VDP2016 - Lecture 15 PCG with Unity
PDF
VDP2016 - Lecture 14 Procedural content generation
11 Settembre 2021 - Giocare con i Videogiochi
Breve Viaggio al Centro dei Videogiochi
Global Game Jam 19 @ POLIMI - Morning Welcome
Data Driven Game Design @ Campus Party 2018
GGJ18 al Politecnico di Milano - Presentazione che precede la presentazione d...
GGJ18 al Politecnico di Milano - Presentazione di apertura
Presentation for UNITECH event - January 8, 2018
DMTM Lecture 17 Text mining
DMTM Lecture 14 Density based clustering
DMTM Lecture 01 Introduction
DMTM Lecture 02 Data mining
VDP2016 - Lecture 16 Rendering pipeline
VDP2016 - Lecture 15 PCG with Unity
VDP2016 - Lecture 14 Procedural content generation

Recently uploaded (20)

PDF
FourierSeries-QuestionsWithAnswers(Part-A).pdf
PPTX
Pharma ospi slides which help in ospi learning
PPTX
Week 4 Term 3 Study Techniques revisited.pptx
PPTX
Microbial diseases, their pathogenesis and prophylaxis
PPTX
Cell Structure & Organelles in detailed.
PDF
Basic Mud Logging Guide for educational purpose
PDF
Supply Chain Operations Speaking Notes -ICLT Program
PPTX
Cell Types and Its function , kingdom of life
PDF
Origin of periodic table-Mendeleev’s Periodic-Modern Periodic table
PDF
Classroom Observation Tools for Teachers
PPTX
BOWEL ELIMINATION FACTORS AFFECTING AND TYPES
PDF
Physiotherapy_for_Respiratory_and_Cardiac_Problems WEBBER.pdf
PPTX
The Healthy Child – Unit II | Child Health Nursing I | B.Sc Nursing 5th Semester
PDF
2.FourierTransform-ShortQuestionswithAnswers.pdf
PDF
Abdominal Access Techniques with Prof. Dr. R K Mishra
PDF
Business Ethics Teaching Materials for college
PDF
Complications of Minimal Access Surgery at WLH
PDF
Anesthesia in Laparoscopic Surgery in India
PDF
RMMM.pdf make it easy to upload and study
PDF
STATICS OF THE RIGID BODIES Hibbelers.pdf
FourierSeries-QuestionsWithAnswers(Part-A).pdf
Pharma ospi slides which help in ospi learning
Week 4 Term 3 Study Techniques revisited.pptx
Microbial diseases, their pathogenesis and prophylaxis
Cell Structure & Organelles in detailed.
Basic Mud Logging Guide for educational purpose
Supply Chain Operations Speaking Notes -ICLT Program
Cell Types and Its function , kingdom of life
Origin of periodic table-Mendeleev’s Periodic-Modern Periodic table
Classroom Observation Tools for Teachers
BOWEL ELIMINATION FACTORS AFFECTING AND TYPES
Physiotherapy_for_Respiratory_and_Cardiac_Problems WEBBER.pdf
The Healthy Child – Unit II | Child Health Nursing I | B.Sc Nursing 5th Semester
2.FourierTransform-ShortQuestionswithAnswers.pdf
Abdominal Access Techniques with Prof. Dr. R K Mishra
Business Ethics Teaching Materials for college
Complications of Minimal Access Surgery at WLH
Anesthesia in Laparoscopic Surgery in India
RMMM.pdf make it easy to upload and study
STATICS OF THE RIGID BODIES Hibbelers.pdf

DMTM 2015 - 06 Introduction to Clustering

  • 1. Prof. Pier Luca Lanzi Clustering: Introduction Data Mining andText Mining (UIC 583 @ Politecnico di Milano)
  • 2. Prof. Pier Luca Lanzi Readings •  Mining of Massive Datasets (Chapter 7, Section 3.5) 2
  • 3. Prof. Pier Luca Lanzi 3
  • 6. Prof. Pier Luca Lanzi Clustering algorithms group a collection of data points into “clusters” according to some distance measure Data points in the same cluster should have a small distance from one another Data points in different clusters should be at a large distance from one another.
  • 7. Prof. Pier Luca Lanzi Clustering finds “natural” grouping/structure in un-labeled data (Unsupervised Learning)
  • 8. Prof. Pier Luca Lanzi What is Cluster Analysis? •  A cluster is a collection of data objects § Similar to one another within the same cluster § Dissimilar to the objects in other clusters •  Cluster analysis § Given a set data points try to understand their structure § Finds similarities between data according to the characteristics found in the data § Groups similar data objects into clusters § It is unsupervised learning since there is no predefined classes •  Typical applications § Stand-alone tool to get insight into data § Preprocessing step for other algorithms 8
  • 9. Prof. Pier Luca Lanzi Clustering Methods •  Hierarchical vs point assignment •  Numeric and/or symbolic data •  Deterministic vs. probabilistic •  Exclusive vs. overlapping •  Hierarchical vs. flat •  Top-down vs. bottom-up 9
  • 10. Prof. Pier Luca Lanzi Clustering Applications •  Marketing § Help marketers discover distinct groups in their customer bases, and then use this knowledge to develop targeted marketing programs •  Land use § Identification of areas of similar land use in an earth observation database •  Insurance § Identifying groups of motor insurance policy holders with a high average claim cost •  City-planning § Identifying groups of houses according to their house type, value, and geographical location •  Earth-quake studies § Observed earth quake epicenters should be clustered along continent faults 10
  • 11. Prof. Pier Luca Lanzi What Is Good Clustering? •  A good clustering consists of high quality clusters with § High intra-class similarity § Low inter-class similarity •  The quality of a clustering result depends on both the similarity measure used by the method and its implementation •  The quality of a clustering method is also measured by its ability to discover some or all of the hidden patterns •  Evaluation § Various measure of intra/inter cluster similarity § Manual inspection § Benchmarking on existing labels 11
  • 12. Prof. Pier Luca Lanzi Measure the Quality of Clustering •  Dissimilarity/Similarity metric: Similarity is expressed in terms of a distance function, typically metric d(i, j) •  There is a separate “quality” function that measures the “goodness” of a cluster •  The definitions of distance functions are usually very different for interval-scaled, boolean, categorical, ordinal ratio, and vector variables •  Weights should be associated with different variables based on applications and data semantics •  It is hard to define “similar enough” or “good enough” as the answer is typically highly subjective 12
  • 13. Prof. Pier Luca Lanzi Data Structures 0 d(2,1) 0 d(3,1) d(3,2) 0 : : : d(n,1) d(n,2) ... ... 0 ! # # # # # # $ % Outlook   Temp   Humidity   Windy   Play   Sunny   Hot   High   False   No   Sunny   Hot     High     True   No   Overcast     Hot       High   False   Yes   …   …   …   …   …   x 11 ... x 1f ... x 1p ... ... ... ... ... x i1 ... x if ... x ip ... ... ... ... ... x n1 ... x nf ... x np ! # # # # # # # # $ % Data Matrix 13 Dis/Similarity Matrix
  • 14. Prof. Pier Luca Lanzi Type of Data in Clustering Analysis •  Interval-scaled variables •  Binary variables •  Nominal, ordinal, and ratio variables •  Variables of mixed types 14
  • 15. Prof. Pier Luca Lanzi Distance Measures
  • 16. Prof. Pier Luca Lanzi Distance Measures •  Given a space and a set of points on this space, a distance measure d(x,y) maps two points x and y to a real number, and satisfies three axioms •  d(x,y) ≥ 0 •  d(x,y) = 0 if and only x=y •  d(x,y) = d(y,x) •  d(x,y) ≤ d(x,z) + d(z,y) 16
  • 17. Prof. Pier Luca Lanzi Euclidean Distances 17 here are other distance measures that have been used for Euclidean any constant r, we can define the Lr-norm to be the distance me ed by: d([x1, x2, . . . , xn], [y1, y2, . . . , yn]) = ( n i=1 |xi − yi|r )1/r case r = 2 is the usual L2-norm just mentioned. Another common d ure is the L1-norm, or Manhattan distance. There, the distance b points is the sum of the magnitudes of the differences in each dim called “Manhattan distance” because it is the distance one would •  Lr-norm •  Euclidean distance (r=2) •  Manhattan distance (r=1) •  L∞-norm 2 Euclidean Distances most familiar distance measure is the one we normally think of as “dis- e.” An n-dimensional Euclidean space is one where points are vectors of n numbers. The conventional distance measure in this space, which we shall to as the L2-norm, is defined: d([x1, x2, . . . , xn], [y1, y2, . . . , yn]) = n i=1 (xi − yi)2 is, we square the distance in each dimension, sum the squares, and take positive square root. is easy to verify the first three requirements for a distance measure are fied. The Euclidean distance between two points cannot be negative, be- e the positive square root is intended. Since all squares of real numbers are egative, any i such that xi ̸= yi forces the distance to be strictly positive. he other hand, if xi = yi for all i, then the distance is clearly 0. Symmetry ws because (xi − yi)2 = (yi − xi)2 . The triangle inequality requires a good of algebra to verify. However, it is well understood to be a property of
  • 18. Prof. Pier Luca Lanzi Jaccard Distance •  Jaccard distance is defined as d(x,y) = 1 – SIM(x,y) where SIM is the Jaccard similarity, •  Which can also be interpreted as the percentage of identical attributes 18
  • 19. Prof. Pier Luca Lanzi Cosine Distance •  The cosine distance between x, y is the angle that the vectors to those points make •  This angle will be in the range 0 to 180 degrees, regardless of how many dimensions the space has. •  Example: given x = (1,2,-1) and y = (2,1,1) the angle between the two vectors is 60 19
  • 20. Prof. Pier Luca Lanzi Edit Distance •  Used when the data points are strings •  The distance between a string x=x1x2…xn and y=y1y2…ym is the smallest number of insertions and deletions of single characters that will transform x into y •  Alternatively, the edit distance d(x, y) can be compute as the longest common subsequence (LCS) of x and y and then, d(x,y) = |x| + |y| - 2|LCS| •  Example: the edit distance between x=abcde and y=acfdeg is 3 (delete b, insert f, insert g), the LCS is acde which is coherent with the previous result 20
  • 21. Prof. Pier Luca Lanzi Hamming Distance •  Hamming distance between two vectors is the number of components in which they differ •  Or equivalently, given the number of variables n, and the number m of matching components, we define •  Example: the Hamming distance between the vectors 10101 and 11110 is 3. 21
  • 22. Prof. Pier Luca Lanzi Ordinal Variables •  An ordinal variable can be discrete or continuous •  Order is important, e.g., rank •  It can be treated as an interval-scaled § replace xif with their rank § map the range of each variable onto [0, 1] by replacing i-th object in the f-th variable by § compute the dissimilarity using methods for interval-scaled variables 22
  • 23. Prof. Pier Luca Lanzi Requirements of Clustering in Data Mining •  Scalability •  Ability to deal with different types of attributes •  Ability to handle dynamic data •  Discovery of clusters with arbitrary shape •  Minimal requirements for domain knowledge to determine input parameters •  Able to deal with noise and outliers •  Insensitive to order of input records •  High dimensionality •  Incorporation of user-specified constraints •  Interpretability and usability 23
  • 24. Prof. Pier Luca Lanzi Curse of Dimensionality in high dimensions, almost all pairs of points are equally far away from one another almost any two vectors are almost orthogonal