SlideShare a Scribd company logo
2
Most read
3
Most read
10
Most read
Clustering
Model based techniques and Handling
high dimensional data
1
2
Model-Based Clustering Methods
 Attempt to optimize the fit between the data and some mathematical
model
 Assumption: Data are generated by a mixture of underlying probability
distributions
 Techniques
 Expectation-Maximization
 Conceptual Clustering
 Neural Networks Approach
Expectation Maximization
 Each cluster is represented mathematically by a
parametric probability distribution
 Component distribution
 Data is a mixture of these distributions
 Mixture density model
 Problem: To estimate parameters of probability distributions
3
Expectation Maximization
 Iterative Refinement Algorithm – used to find parameter
estimates
 Extension of k-means
 Assigns an object to a cluster according to a weight representing
probability of membership
 Initial estimate of parameters
 Iteratively reassigns scores
4
Expectation Maximization
 Initial guess for parameters; randomly select k objects to
represent cluster means or centers
 Iteratively refine parameters / clusters
 Expectation Step
 Assign each object xi to cluster Ck with probability
where
 Maximization Step
 Re-estimate model parameters
 Simple and easy to implement
 Complexity depends on features, objects and iterations
5
6
Conceptual Clustering
 Conceptual clustering
 A form of clustering in machine learning
 Produces a classification scheme for a set of unlabeled objects
 Finds characteristic description for each concept (class)
 COBWEB
 A popular and simple method of incremental conceptual learning
 Creates a hierarchical clustering in the form of a classification tree
 Each node refers to a concept and contains a probabilistic description
of that concept
7
COBWEB Clustering Method
A classification tree
COBWEB
 Classification tree
 Each node – Concept and its probabilistic distribution
(Summary of objects under that node)
 Description – Conditional probabilities P(Ai=vij / Ck)
 Sibling nodes at given level form a partition
 Category Utility
 Increase in the expected number of attribute values that can
be correctly guessed given a partition
8
COBWEB
 Category Utility rewards:
 Intra-class similarity P(Ai=vij|Ck)
 High value indicates many class members share this attribute-value
pair
 Inter-class dissimilarity P(Ck|Ai=vij)
 High values – fewer objects in different classes share this attribute-
value
 Placement of new objects
 Descend tree
 Identify best host
 Temporarily place object in each node and compute CU of resulting
partition
 Placement with highest CU is chosen
 COBWEB may also forms new nodes if object does not fit into the
existing tree
9
COBWEB
 COBWEB is sensitive to order of records
 Additional operations
 Merging and Splitting
 Two best hosts are considered for merging
 Best host is considered for splitting
 Limitations
 The assumption that the attributes are independent of each
other is often too strong because correlation may exist
 Not suitable for clustering large database data
 CLASSIT - an extension of COBWEB for incremental clustering of
continuous data
10
Neural Network Approach
 Represent each cluster as an exemplar, acting as a “prototype” of
the cluster
 New objects are distributed to the cluster whose exemplar is the
most similar according to some distance measure
 Self Organizing Map
 Competitive learning
 Involves a hierarchical architecture of several units (neurons)
 Neurons compete in a “winner-takes-all” fashion for the object currently
being presented
 Organization of units – forms a feature map
 Web Document Clustering
11
Kohenen SOM
12
Clustering High-Dimensional data
 As dimensionality increases
 number of irrelevant dimensions may produce noise and mask real clusters
 data becomes sparse
 Distance measures –meaningless
 Feature transformation methods
 PCA, SVD – Summarize data by creating linear combinations of attributes
 But do not remove any attributes; transformed attributes – complex to
interpret
 Feature Selection methods
 Most relevant set of attributes with respect to class labels
 Entropy Analysis
 Subspace Clustering – searches for groups of clusters within different
subspaces of the same data set
13
CLIQUE: CLustering In QUest
 Dimension growth subspace clustering
 Starts at 1-D and grows upwards to higher dimensions
 Partitions each dimension – grids – determines whether
cell is dense
 CLIQUE
 Determines sparse and crowded units
 Dense unit – fraction of data points > threshold
 Cluster – maximal set of connected dense units
14
CLIQUE
 First partitions d-dimensional space into non-overlapping units
 Performed in 1-D
 Based on Apriori property: If a k-dimensional unit is dense so are its
projections in (k-1) dimensional space
 Search space size is reduced
 Determines the maximal dense region and Generates a minimal
description
15
CLIQUE
 Finds subspace of highest dimension
 Insensitive to order of inputs
 Performance depends on grid size and density threshold
 Difficult to determine across all dimensions
 Several lower dimensional subspaces will have to be
processed
 Can use adaptive strategy
16
PROCLUS – PROjected CLUStering
 Dimension-reduction Subspace Clustering technique
 Finds initial approximation of clusters in high dimensional
space
 Avoids generation of large number of overlapped
clusters of lower dimensionality
 Finds best set of medoids by hill-climbing process
(Similar to CLARANS)
 Manhattan Segmental distance measure
17
PROCLUS
 Initialization phase
 Greedy algorithm to select a set of initial medoids that are far
apart
 Iteration Phase
 Selects a random set of k-medoids
 Replaces bad medoids
 For each medoid a set of dimensions is chosen whose average
distances are small
 Refinement Phase
 Computes new dimensions for each medoid based on clusters
found, reasigns points to medoids and removes outliers
18
Frequent Pattern based Clustering
 Frequent patterns may also form clusters
 Instead of growing clusters dimension by dimension sets
of frequent itemsets are determined
 Two common technqiues
 Frequent term-based text Clustering
 Clustering by Pattern similarity
19
Frequent-term based text clustering
 Text documents are clustered based on frequent terms
they contain
 Documents – terms
 Dimensionality is very high
 Frequent term based analysis
 Well selected subset of set of all frequent terms must be
discovered
 Fi – Set of frequent term sets
 Cov(Fi) – set of documents covered by Fi
 ∑i=1 k
cov(Fi) = D and overlap between Fi and Fj must be
minimized
 Description of clusters – their frequent term sets
20
Clustering by Pattern Similarity
 pCluster on micro-array data analysis
 DNA micro-array analysis – expression levels of two
genes may rise and fall synchronously in response to
stimuli
 Two objects are similar if they exhibit a coherent pattern
on a subset of dimensions
21
pCluster
 Shift Pattern discovery
 Euclidean distance – not suitable
 Derive new attributes
 Bi-Clustering based on mean squared residue score
 pCluster
 Objects –x, y; attributes – a, b
 A pair (O,T) forms a δ-pCluster if for any 2 x 2 matrix X in (O, T)
pScore(X) <= δ
 Each pair of objects and their features must satisfy threshold
22
pCluster
 Scaling patterns
 pCluster can be used in other applications
also
23

More Related Content

What's hot (20)

PPTX
Clustering in Data Mining
Archana Swaminathan
 
PPT
3.3 hierarchical methods
Krish_ver2
 
PPTX
Data Mining: clustering and analysis
DataminingTools Inc
 
PPT
K mean-clustering algorithm
parry prabhu
 
PPT
2.4 rule based classification
Krish_ver2
 
PPT
3. mining frequent patterns
Azad public school
 
PPTX
Data mining: Classification and prediction
DataminingTools Inc
 
PPT
2.3 bayesian classification
Krish_ver2
 
PPT
3.2 partitioning methods
Krish_ver2
 
PPTX
Clustering in data Mining (Data Mining)
Mustafa Sherazi
 
PPTX
05 Clustering in Data Mining
Valerii Klymchuk
 
PPT
3.6 constraint based cluster analysis
Krish_ver2
 
PPT
Ontology engineering
Aliabbas Petiwala
 
PPTX
Density based methods
SVijaylakshmi
 
PPTX
Data Integration and Transformation in Data mining
kavitha muneeshwaran
 
PPT
2.2 decision tree
Krish_ver2
 
PPTX
K-Means Clustering Algorithm.pptx
JebaRaj26
 
PPT
2.5 backpropagation
Krish_ver2
 
PPTX
Classification and prediction in data mining
Er. Nawaraj Bhandari
 
PPTX
Introduction to Data Mining
DataminingTools Inc
 
Clustering in Data Mining
Archana Swaminathan
 
3.3 hierarchical methods
Krish_ver2
 
Data Mining: clustering and analysis
DataminingTools Inc
 
K mean-clustering algorithm
parry prabhu
 
2.4 rule based classification
Krish_ver2
 
3. mining frequent patterns
Azad public school
 
Data mining: Classification and prediction
DataminingTools Inc
 
2.3 bayesian classification
Krish_ver2
 
3.2 partitioning methods
Krish_ver2
 
Clustering in data Mining (Data Mining)
Mustafa Sherazi
 
05 Clustering in Data Mining
Valerii Klymchuk
 
3.6 constraint based cluster analysis
Krish_ver2
 
Ontology engineering
Aliabbas Petiwala
 
Density based methods
SVijaylakshmi
 
Data Integration and Transformation in Data mining
kavitha muneeshwaran
 
2.2 decision tree
Krish_ver2
 
K-Means Clustering Algorithm.pptx
JebaRaj26
 
2.5 backpropagation
Krish_ver2
 
Classification and prediction in data mining
Er. Nawaraj Bhandari
 
Introduction to Data Mining
DataminingTools Inc
 

Similar to 3.5 model based clustering (20)

PPT
Basic Clustering Algorithms in Data Warehouisng and Data Miningppt
MrSKanthiKiran
 
PPT
Data Mining and Warehousing Concept and Techniques
AnilkumarBrahmane2
 
PPT
Data Mining Concepts and Techniques, Chapter 10. Cluster Analysis: Basic Conc...
Salah Amean
 
PPT
Chapter 10. Cluster Analysis Basic Concepts and Methods.ppt
Subrata Kumer Paul
 
PPT
10 clusbasic
engrasi
 
PPT
CLUSTERING
Aman Jatain
 
PPT
data mining cocepts and techniques chapter
NaveenKumar5162
 
PPT
data mining cocepts and techniques chapter
NaveenKumar5162
 
PDF
10 clusbasic
JoonyoungJayGwak
 
PPT
Data mining concepts and techniques Chapter 10
mqasimsheikh5
 
PPT
Capter10 cluster basic : Han & Kamber
Houw Liong The
 
PPT
Capter10 cluster basic
Houw Liong The
 
PPT
My8clst
ketan533
 
PPT
Dataa miining
SUBBIAH SURESH
 
PPTX
Grid based method & model based clustering method
rajshreemuthiah
 
PPT
multiarmed bandit.ppt
LPrashanthi
 
PPT
Chapter 10 ClusBasic ppt file for clear understaning
my123lapto
 
PPT
Chapter -10-Clus_Basic.ppt -DataMinning
nayabkainat470
 
PPT
ClusetrigBasic.ppt
ChaitanyaKulkarni451137
 
PPT
UniT_A_Clustering machine learning .ppt
HarshPanchal455289
 
Basic Clustering Algorithms in Data Warehouisng and Data Miningppt
MrSKanthiKiran
 
Data Mining and Warehousing Concept and Techniques
AnilkumarBrahmane2
 
Data Mining Concepts and Techniques, Chapter 10. Cluster Analysis: Basic Conc...
Salah Amean
 
Chapter 10. Cluster Analysis Basic Concepts and Methods.ppt
Subrata Kumer Paul
 
10 clusbasic
engrasi
 
CLUSTERING
Aman Jatain
 
data mining cocepts and techniques chapter
NaveenKumar5162
 
data mining cocepts and techniques chapter
NaveenKumar5162
 
10 clusbasic
JoonyoungJayGwak
 
Data mining concepts and techniques Chapter 10
mqasimsheikh5
 
Capter10 cluster basic : Han & Kamber
Houw Liong The
 
Capter10 cluster basic
Houw Liong The
 
My8clst
ketan533
 
Dataa miining
SUBBIAH SURESH
 
Grid based method & model based clustering method
rajshreemuthiah
 
multiarmed bandit.ppt
LPrashanthi
 
Chapter 10 ClusBasic ppt file for clear understaning
my123lapto
 
Chapter -10-Clus_Basic.ppt -DataMinning
nayabkainat470
 
ClusetrigBasic.ppt
ChaitanyaKulkarni451137
 
UniT_A_Clustering machine learning .ppt
HarshPanchal455289
 
Ad

More from Krish_ver2 (20)

PPT
5.5 back tracking
Krish_ver2
 
PPT
5.5 back track
Krish_ver2
 
PPT
5.5 back tracking 02
Krish_ver2
 
PPT
5.4 randomized datastructures
Krish_ver2
 
PPT
5.4 randomized datastructures
Krish_ver2
 
PPT
5.4 randamized algorithm
Krish_ver2
 
PPT
5.3 dynamic programming 03
Krish_ver2
 
PPT
5.3 dynamic programming
Krish_ver2
 
PPT
5.3 dyn algo-i
Krish_ver2
 
PPT
5.2 divede and conquer 03
Krish_ver2
 
PPT
5.2 divide and conquer
Krish_ver2
 
PPT
5.2 divede and conquer 03
Krish_ver2
 
PPT
5.1 greedyyy 02
Krish_ver2
 
PPT
5.1 greedy
Krish_ver2
 
PPT
5.1 greedy 03
Krish_ver2
 
PPT
4.4 hashing02
Krish_ver2
 
PPT
4.4 hashing
Krish_ver2
 
PPT
4.4 hashing ext
Krish_ver2
 
PPT
4.4 external hashing
Krish_ver2
 
PPT
4.2 bst
Krish_ver2
 
5.5 back tracking
Krish_ver2
 
5.5 back track
Krish_ver2
 
5.5 back tracking 02
Krish_ver2
 
5.4 randomized datastructures
Krish_ver2
 
5.4 randomized datastructures
Krish_ver2
 
5.4 randamized algorithm
Krish_ver2
 
5.3 dynamic programming 03
Krish_ver2
 
5.3 dynamic programming
Krish_ver2
 
5.3 dyn algo-i
Krish_ver2
 
5.2 divede and conquer 03
Krish_ver2
 
5.2 divide and conquer
Krish_ver2
 
5.2 divede and conquer 03
Krish_ver2
 
5.1 greedyyy 02
Krish_ver2
 
5.1 greedy
Krish_ver2
 
5.1 greedy 03
Krish_ver2
 
4.4 hashing02
Krish_ver2
 
4.4 hashing
Krish_ver2
 
4.4 hashing ext
Krish_ver2
 
4.4 external hashing
Krish_ver2
 
4.2 bst
Krish_ver2
 
Ad

Recently uploaded (20)

DOCX
Modul Ajar Deep Learning Bahasa Inggris Kelas 11 Terbaru 2025
wahyurestu63
 
PPTX
Translation_ Definition, Scope & Historical Development.pptx
DhatriParmar
 
PDF
Antianginal agents, Definition, Classification, MOA.pdf
Prerana Jadhav
 
PDF
BÀI TẬP TEST BỔ TRỢ THEO TỪNG CHỦ ĐỀ CỦA TỪNG UNIT KÈM BÀI TẬP NGHE - TIẾNG A...
Nguyen Thanh Tu Collection
 
PDF
TOP 10 AI TOOLS YOU MUST LEARN TO SURVIVE IN 2025 AND ABOVE
digilearnings.com
 
PPTX
Sonnet 130_ My Mistress’ Eyes Are Nothing Like the Sun By William Shakespear...
DhatriParmar
 
PDF
John Keats introduction and list of his important works
vatsalacpr
 
PPTX
Basics and rules of probability with real-life uses
ravatkaran694
 
PPTX
CONCEPT OF CHILD CARE. pptx
AneetaSharma15
 
PPTX
Electrophysiology_of_Heart. Electrophysiology studies in Cardiovascular syste...
Rajshri Ghogare
 
PPTX
How to Close Subscription in Odoo 18 - Odoo Slides
Celine George
 
PPTX
How to Track Skills & Contracts Using Odoo 18 Employee
Celine George
 
PPTX
ENGLISH 8 WEEK 3 Q1 - Analyzing the linguistic, historical, andor biographica...
OliverOllet
 
DOCX
pgdei-UNIT -V Neurological Disorders & developmental disabilities
JELLA VISHNU DURGA PRASAD
 
PDF
EXCRETION-STRUCTURE OF NEPHRON,URINE FORMATION
raviralanaresh2
 
PPTX
TOP 10 AI TOOLS YOU MUST LEARN TO SURVIVE IN 2025 AND ABOVE
digilearnings.com
 
PDF
My Thoughts On Q&A- A Novel By Vikas Swarup
Niharika
 
PPTX
Artificial Intelligence in Gastroentrology: Advancements and Future Presprec...
AyanHossain
 
PPTX
Command Palatte in Odoo 18.1 Spreadsheet - Odoo Slides
Celine George
 
PPTX
Introduction to pediatric nursing in 5th Sem..pptx
AneetaSharma15
 
Modul Ajar Deep Learning Bahasa Inggris Kelas 11 Terbaru 2025
wahyurestu63
 
Translation_ Definition, Scope & Historical Development.pptx
DhatriParmar
 
Antianginal agents, Definition, Classification, MOA.pdf
Prerana Jadhav
 
BÀI TẬP TEST BỔ TRỢ THEO TỪNG CHỦ ĐỀ CỦA TỪNG UNIT KÈM BÀI TẬP NGHE - TIẾNG A...
Nguyen Thanh Tu Collection
 
TOP 10 AI TOOLS YOU MUST LEARN TO SURVIVE IN 2025 AND ABOVE
digilearnings.com
 
Sonnet 130_ My Mistress’ Eyes Are Nothing Like the Sun By William Shakespear...
DhatriParmar
 
John Keats introduction and list of his important works
vatsalacpr
 
Basics and rules of probability with real-life uses
ravatkaran694
 
CONCEPT OF CHILD CARE. pptx
AneetaSharma15
 
Electrophysiology_of_Heart. Electrophysiology studies in Cardiovascular syste...
Rajshri Ghogare
 
How to Close Subscription in Odoo 18 - Odoo Slides
Celine George
 
How to Track Skills & Contracts Using Odoo 18 Employee
Celine George
 
ENGLISH 8 WEEK 3 Q1 - Analyzing the linguistic, historical, andor biographica...
OliverOllet
 
pgdei-UNIT -V Neurological Disorders & developmental disabilities
JELLA VISHNU DURGA PRASAD
 
EXCRETION-STRUCTURE OF NEPHRON,URINE FORMATION
raviralanaresh2
 
TOP 10 AI TOOLS YOU MUST LEARN TO SURVIVE IN 2025 AND ABOVE
digilearnings.com
 
My Thoughts On Q&A- A Novel By Vikas Swarup
Niharika
 
Artificial Intelligence in Gastroentrology: Advancements and Future Presprec...
AyanHossain
 
Command Palatte in Odoo 18.1 Spreadsheet - Odoo Slides
Celine George
 
Introduction to pediatric nursing in 5th Sem..pptx
AneetaSharma15
 

3.5 model based clustering

  • 1. Clustering Model based techniques and Handling high dimensional data 1
  • 2. 2 Model-Based Clustering Methods  Attempt to optimize the fit between the data and some mathematical model  Assumption: Data are generated by a mixture of underlying probability distributions  Techniques  Expectation-Maximization  Conceptual Clustering  Neural Networks Approach
  • 3. Expectation Maximization  Each cluster is represented mathematically by a parametric probability distribution  Component distribution  Data is a mixture of these distributions  Mixture density model  Problem: To estimate parameters of probability distributions 3
  • 4. Expectation Maximization  Iterative Refinement Algorithm – used to find parameter estimates  Extension of k-means  Assigns an object to a cluster according to a weight representing probability of membership  Initial estimate of parameters  Iteratively reassigns scores 4
  • 5. Expectation Maximization  Initial guess for parameters; randomly select k objects to represent cluster means or centers  Iteratively refine parameters / clusters  Expectation Step  Assign each object xi to cluster Ck with probability where  Maximization Step  Re-estimate model parameters  Simple and easy to implement  Complexity depends on features, objects and iterations 5
  • 6. 6 Conceptual Clustering  Conceptual clustering  A form of clustering in machine learning  Produces a classification scheme for a set of unlabeled objects  Finds characteristic description for each concept (class)  COBWEB  A popular and simple method of incremental conceptual learning  Creates a hierarchical clustering in the form of a classification tree  Each node refers to a concept and contains a probabilistic description of that concept
  • 7. 7 COBWEB Clustering Method A classification tree
  • 8. COBWEB  Classification tree  Each node – Concept and its probabilistic distribution (Summary of objects under that node)  Description – Conditional probabilities P(Ai=vij / Ck)  Sibling nodes at given level form a partition  Category Utility  Increase in the expected number of attribute values that can be correctly guessed given a partition 8
  • 9. COBWEB  Category Utility rewards:  Intra-class similarity P(Ai=vij|Ck)  High value indicates many class members share this attribute-value pair  Inter-class dissimilarity P(Ck|Ai=vij)  High values – fewer objects in different classes share this attribute- value  Placement of new objects  Descend tree  Identify best host  Temporarily place object in each node and compute CU of resulting partition  Placement with highest CU is chosen  COBWEB may also forms new nodes if object does not fit into the existing tree 9
  • 10. COBWEB  COBWEB is sensitive to order of records  Additional operations  Merging and Splitting  Two best hosts are considered for merging  Best host is considered for splitting  Limitations  The assumption that the attributes are independent of each other is often too strong because correlation may exist  Not suitable for clustering large database data  CLASSIT - an extension of COBWEB for incremental clustering of continuous data 10
  • 11. Neural Network Approach  Represent each cluster as an exemplar, acting as a “prototype” of the cluster  New objects are distributed to the cluster whose exemplar is the most similar according to some distance measure  Self Organizing Map  Competitive learning  Involves a hierarchical architecture of several units (neurons)  Neurons compete in a “winner-takes-all” fashion for the object currently being presented  Organization of units – forms a feature map  Web Document Clustering 11
  • 13. Clustering High-Dimensional data  As dimensionality increases  number of irrelevant dimensions may produce noise and mask real clusters  data becomes sparse  Distance measures –meaningless  Feature transformation methods  PCA, SVD – Summarize data by creating linear combinations of attributes  But do not remove any attributes; transformed attributes – complex to interpret  Feature Selection methods  Most relevant set of attributes with respect to class labels  Entropy Analysis  Subspace Clustering – searches for groups of clusters within different subspaces of the same data set 13
  • 14. CLIQUE: CLustering In QUest  Dimension growth subspace clustering  Starts at 1-D and grows upwards to higher dimensions  Partitions each dimension – grids – determines whether cell is dense  CLIQUE  Determines sparse and crowded units  Dense unit – fraction of data points > threshold  Cluster – maximal set of connected dense units 14
  • 15. CLIQUE  First partitions d-dimensional space into non-overlapping units  Performed in 1-D  Based on Apriori property: If a k-dimensional unit is dense so are its projections in (k-1) dimensional space  Search space size is reduced  Determines the maximal dense region and Generates a minimal description 15
  • 16. CLIQUE  Finds subspace of highest dimension  Insensitive to order of inputs  Performance depends on grid size and density threshold  Difficult to determine across all dimensions  Several lower dimensional subspaces will have to be processed  Can use adaptive strategy 16
  • 17. PROCLUS – PROjected CLUStering  Dimension-reduction Subspace Clustering technique  Finds initial approximation of clusters in high dimensional space  Avoids generation of large number of overlapped clusters of lower dimensionality  Finds best set of medoids by hill-climbing process (Similar to CLARANS)  Manhattan Segmental distance measure 17
  • 18. PROCLUS  Initialization phase  Greedy algorithm to select a set of initial medoids that are far apart  Iteration Phase  Selects a random set of k-medoids  Replaces bad medoids  For each medoid a set of dimensions is chosen whose average distances are small  Refinement Phase  Computes new dimensions for each medoid based on clusters found, reasigns points to medoids and removes outliers 18
  • 19. Frequent Pattern based Clustering  Frequent patterns may also form clusters  Instead of growing clusters dimension by dimension sets of frequent itemsets are determined  Two common technqiues  Frequent term-based text Clustering  Clustering by Pattern similarity 19
  • 20. Frequent-term based text clustering  Text documents are clustered based on frequent terms they contain  Documents – terms  Dimensionality is very high  Frequent term based analysis  Well selected subset of set of all frequent terms must be discovered  Fi – Set of frequent term sets  Cov(Fi) – set of documents covered by Fi  ∑i=1 k cov(Fi) = D and overlap between Fi and Fj must be minimized  Description of clusters – their frequent term sets 20
  • 21. Clustering by Pattern Similarity  pCluster on micro-array data analysis  DNA micro-array analysis – expression levels of two genes may rise and fall synchronously in response to stimuli  Two objects are similar if they exhibit a coherent pattern on a subset of dimensions 21
  • 22. pCluster  Shift Pattern discovery  Euclidean distance – not suitable  Derive new attributes  Bi-Clustering based on mean squared residue score  pCluster  Objects –x, y; attributes – a, b  A pair (O,T) forms a δ-pCluster if for any 2 x 2 matrix X in (O, T) pScore(X) <= δ  Each pair of objects and their features must satisfy threshold 22
  • 23. pCluster  Scaling patterns  pCluster can be used in other applications also 23