SlideShare a Scribd company logo
Data Mining Classification: Alternative Techniques
Rule-Based ClassifierClassify records by using a collection of “if…then…” rulesRule:    (Condition)  ywhere  Condition is a conjunctions of attributes  y is the class labelLHS: rule antecedent or conditionRHS: rule consequent
Characteristics of Rule-Based ClassifierMutually exclusive rulesClassifier contains mutually exclusive rules if the rules are independent of each otherEvery record is covered by at most one ruleExhaustive rulesClassifier has exhaustive coverage if it accounts for every possible combination of attribute valuesEach record is covered by at least one rule
Building Classification RulesDirect Method:  Extract rules directly from data e.g.: RIPPER, CN2, Holte’s 1RIndirect Method: Extract rules from other classification models (e.g.    decision trees, neural networks, etc).e.g: C4.5rules
Direct Method: Sequential CoveringStart from an empty ruleGrow a rule using the Learn-One-Rule functionRemove training records covered by the ruleRepeat Step (2) and (3) until stopping criterion is met
Aspects of Sequential CoveringRule GrowingInstance EliminationRule EvaluationStopping CriterionRule Pruning
Contd…Grow a single ruleRemove Instances from rulePrune the rule (if necessary)Add rule to Current Rule SetRepeat
Indirect Method: C4.5rulesExtract rules from an unpruned decision treeFor each rule, r: A  y, consider an alternative rule r’: A’  y where A’ is obtained by removing one of the conjuncts in ACompare the pessimistic error rate for r against all r’sPrune if one of the r’s has lower pessimistic error rateRepeat until we can no longer improve generalization error
Indirect Method: C4.5rulesInstead of ordering the rules, order subsets of rules (class ordering)Each subset is a collection of rules with the same rule consequent (class)Compute description length of each subset Description length = L(error) + g L(model) g is a parameter that takes into account the presence of redundant attributes in a rule set (default value = 0.5)
Advantages of Rule-Based ClassifiersAs highly expressive as decision treesEasy to interpretEasy to generateCan classify new instances rapidlyPerformance comparable to decision trees
Nearest Neighbor ClassifiersRequires three things
The set of stored records
Distance Metric to compute distance between records
The value of k, the number of nearest neighbors to retrieve
To classify an unknown record:
Compute distance to other training records
Identify k nearest neighbors
Use class labels of nearest neighbors to determine the class label of unknown record (e.g., by taking majority voteDefinition of Nearest NeighborK-nearest neighbors of a record x are data points that have the k smallest distance to x
Nearest Neighbor Classification…Choosing the value of k:If k is too small, sensitive to noise pointsIf k is too large, neighborhood may include points from other classesScaling issuesAttributes may have to be scaled to prevent distance measures from being dominated by one of the attributesExample: height of a person may vary from 1.5m to 1.8m weight of a person may vary from 90lb to 300lb
Nearest neighbor Classification…k-NN classifiers are lazy learners It does not build models explicitlyUnlike eager learners such as decision tree induction and rule-based systemsClassifying unknown records are relatively expensive
Bayes ClassifierA probabilistic framework for solving classification problemsConditional Probability:Bayes theorem:
Example of Bayes TheoremGiven: A doctor knows that meningitis causes stiff neck 50% of the timePrior probability of any patient having meningitis is 1/50,000Prior probability of any patient having stiff neck is 1/20 If a patient has stiff neck, what’s the probability he/she has meningitis?
Naïve Bayes ClassifierAssume independence among attributes Ai when class is given:    P(A1, A2, …, An |C) = P(A1| Cj) P(A2| Cj)… P(An| Cj)Can estimate P(Ai| Cj) for all Ai and Cj.New point is classified to Cj if  P(Cj)  P(Ai| Cj)  is maximal.
Naïve Bayes ClassifierIf one of the conditional probability is zero, then the entire expression becomes zeroProbability estimation:c: number of classesp: prior probabilitym: parameter
Naïve Bayes (Summary)Robust to isolated noise pointsHandle missing values by ignoring the instance during probability estimate calculationsRobust to irrelevant attributesIndependence assumption may not hold for some attributesUse other techniques such as Bayesian Belief Networks (BBN)
Artificial Neural Networks (ANN)Model is an assembly of inter-connected nodes and weighted linksOutput node sums up each of its input value according to the weights of its linksCompare output node against some threshold t
General Structure of ANNTraining ANN means learning the weights of the neurons
Algorithm for learning ANNInitialize the weights (w0, w1, …, wk)Adjust the weights in such a way that the output of ANN is consistent with class labels of training examplesObjective function:Find the weights wi’s that minimize the above objective function e.g., backpropagation algorithm
Ensemble MethodsConstruct a set of classifiers from the training dataPredict class label of previously unseen records by aggregating predictions made by multiple classifiers

More Related Content

What's hot (18)

PPT
Ap Power Point Chpt9
dplunkett
 
PPTX
Linear models and multiclass classification
NdSv94
 
PDF
List classes
Ravi_Kant_Sahu
 
PDF
Lecture1
Ritu Chaturvedi
 
PPTX
Vector space classification
Ujjawal
 
ODP
C++ arrays part1
Subhasis Nayak
 
PPT
Arrays
Faisal Aziz
 
PPT
Ap Power Point Chpt6
dplunkett
 
PPT
Finding everything about findings about (fa)
Ram Gali
 
PPTX
Array 2 hina
heena94
 
DOCX
Datastructures and algorithms prepared by M.V.Brehmanada Reddy
Malikireddy Bramhananda Reddy
 
PDF
M v bramhananda reddy dsa complete notes
Malikireddy Bramhananda Reddy
 
PPTX
Collections (1)
abdullah619
 
PPTX
Chap4java5th
Asfand Hassan
 
PPTX
Types of methods in python
Aravindreddy Mokireddy
 
PPTX
264finalppt (1)
Mahima Verma
 
Ap Power Point Chpt9
dplunkett
 
Linear models and multiclass classification
NdSv94
 
List classes
Ravi_Kant_Sahu
 
Lecture1
Ritu Chaturvedi
 
Vector space classification
Ujjawal
 
C++ arrays part1
Subhasis Nayak
 
Arrays
Faisal Aziz
 
Ap Power Point Chpt6
dplunkett
 
Finding everything about findings about (fa)
Ram Gali
 
Array 2 hina
heena94
 
Datastructures and algorithms prepared by M.V.Brehmanada Reddy
Malikireddy Bramhananda Reddy
 
M v bramhananda reddy dsa complete notes
Malikireddy Bramhananda Reddy
 
Collections (1)
abdullah619
 
Chap4java5th
Asfand Hassan
 
Types of methods in python
Aravindreddy Mokireddy
 
264finalppt (1)
Mahima Verma
 

Viewers also liked (20)

PPTX
Txomin Hartz Txikia
irantzugoitia86
 
PDF
Norihicodanch
Filip Yang
 
PPTX
MED dra Coding -MSSO
drabhishekpitti
 
PPTX
SPSS: Data Editor
DataminingTools Inc
 
PPTX
XL-Miner: Timeseries
DataminingTools Inc
 
PPTX
MS SQL SERVER: Microsoft sequence clustering and association rules
DataminingTools Inc
 
PPTX
LISP:Loops In Lisp
DataminingTools Inc
 
PPTX
Mysql:Operators
DataminingTools Inc
 
PPTX
Data Applied: Association
DataminingTools Inc
 
PPT
Asha & Beckis Nc Presentation
Asha Stremcha
 
PPTX
Quick Look At Classification
DataminingTools Inc
 
PPTX
R: Apply Functions
DataminingTools Inc
 
PPTX
RapidMiner: Setting Up A Process
DataminingTools Inc
 
PPT
LíRica Latina 2ºBac Lara Lozano
lara
 
PPTX
Control Statements in Matlab
DataminingTools Inc
 
PPTX
Data Applied:Decision Trees
DataminingTools Inc
 
PPTX
SPSS: Quick Look
DataminingTools Inc
 
PPT
Épica Latina Latín II
lara
 
Txomin Hartz Txikia
irantzugoitia86
 
Norihicodanch
Filip Yang
 
MED dra Coding -MSSO
drabhishekpitti
 
SPSS: Data Editor
DataminingTools Inc
 
XL-Miner: Timeseries
DataminingTools Inc
 
MS SQL SERVER: Microsoft sequence clustering and association rules
DataminingTools Inc
 
LISP:Loops In Lisp
DataminingTools Inc
 
Mysql:Operators
DataminingTools Inc
 
Data Applied: Association
DataminingTools Inc
 
Asha & Beckis Nc Presentation
Asha Stremcha
 
Quick Look At Classification
DataminingTools Inc
 
R: Apply Functions
DataminingTools Inc
 
RapidMiner: Setting Up A Process
DataminingTools Inc
 
LíRica Latina 2ºBac Lara Lozano
lara
 
Control Statements in Matlab
DataminingTools Inc
 
Data Applied:Decision Trees
DataminingTools Inc
 
SPSS: Quick Look
DataminingTools Inc
 
Épica Latina Latín II
lara
 
Ad

Similar to Classification Continued (20)

PPTX
�datamining-lect7.pptx literature of data mining and summary
mohammedalbohiry85
 
PPT
594503964-Introduction-to-Classification-PPT-Slides-1.ppt
snehajuly2004
 
PPT
[ppt]
butest
 
PPT
[ppt]
butest
 
PPT
Introduction to Machine Learning Aristotelis Tsirigos
butest
 
PPTX
UNIT 3: Data Warehousing and Data Mining
Nandakumar P
 
PDF
machine_learning.pptx
Panchami V U
 
PPTX
Deep learning from mashine learning AI..
premkumarlive
 
PDF
Introduction to conventional machine learning techniques
Xavier Rafael Palou
 
PPTX
BAS 250 Lecture 8
Wake Tech BAS
 
PDF
Machine Learning Algorithms Introduction.pdf
Vinodh58
 
PPT
ai4.ppt
akshatsharma823122
 
PPT
ai4.ppt
ssuser448ad3
 
PPT
ai4.ppt
atul404633
 
PPTX
machine leraning : main principles and techniques
johngeorgakis99
 
PDF
Classification Based Machine Learning Algorithms
Md. Main Uddin Rony
 
PPT
coppin chapter 10e.ppt
butest
 
PPTX
Machine Learning
GaytriDhingra1
 
PPTX
AI -learning and machine learning.pptx
GaytriDhingra1
 
PDF
1_5_AI_edx_ml_51intro_240204_104838machine learning lecture 1
MostafaHazemMostafaa
 
�datamining-lect7.pptx literature of data mining and summary
mohammedalbohiry85
 
594503964-Introduction-to-Classification-PPT-Slides-1.ppt
snehajuly2004
 
[ppt]
butest
 
[ppt]
butest
 
Introduction to Machine Learning Aristotelis Tsirigos
butest
 
UNIT 3: Data Warehousing and Data Mining
Nandakumar P
 
machine_learning.pptx
Panchami V U
 
Deep learning from mashine learning AI..
premkumarlive
 
Introduction to conventional machine learning techniques
Xavier Rafael Palou
 
BAS 250 Lecture 8
Wake Tech BAS
 
Machine Learning Algorithms Introduction.pdf
Vinodh58
 
ai4.ppt
ssuser448ad3
 
ai4.ppt
atul404633
 
machine leraning : main principles and techniques
johngeorgakis99
 
Classification Based Machine Learning Algorithms
Md. Main Uddin Rony
 
coppin chapter 10e.ppt
butest
 
Machine Learning
GaytriDhingra1
 
AI -learning and machine learning.pptx
GaytriDhingra1
 
1_5_AI_edx_ml_51intro_240204_104838machine learning lecture 1
MostafaHazemMostafaa
 
Ad

More from DataminingTools Inc (20)

PPTX
Terminology Machine Learning
DataminingTools Inc
 
PPTX
Techniques Machine Learning
DataminingTools Inc
 
PPTX
Machine learning Introduction
DataminingTools Inc
 
PPTX
Areas of machine leanring
DataminingTools Inc
 
PPTX
AI: Planning and AI
DataminingTools Inc
 
PPTX
AI: Logic in AI 2
DataminingTools Inc
 
PPTX
AI: Logic in AI
DataminingTools Inc
 
PPTX
AI: Learning in AI 2
DataminingTools Inc
 
PPTX
AI: Learning in AI
DataminingTools Inc
 
PPTX
AI: Introduction to artificial intelligence
DataminingTools Inc
 
PPTX
AI: Belief Networks
DataminingTools Inc
 
PPTX
AI: AI & Searching
DataminingTools Inc
 
PPTX
AI: AI & Problem Solving
DataminingTools Inc
 
PPTX
Data Mining: Text and web mining
DataminingTools Inc
 
PPTX
Data Mining: Outlier analysis
DataminingTools Inc
 
PPTX
Data Mining: Mining stream time series and sequence data
DataminingTools Inc
 
PPTX
Data Mining: Mining ,associations, and correlations
DataminingTools Inc
 
PPTX
Data Mining: Graph mining and social network analysis
DataminingTools Inc
 
PPTX
Data warehouse and olap technology
DataminingTools Inc
 
PPTX
Data Mining: Data processing
DataminingTools Inc
 
Terminology Machine Learning
DataminingTools Inc
 
Techniques Machine Learning
DataminingTools Inc
 
Machine learning Introduction
DataminingTools Inc
 
Areas of machine leanring
DataminingTools Inc
 
AI: Planning and AI
DataminingTools Inc
 
AI: Logic in AI 2
DataminingTools Inc
 
AI: Logic in AI
DataminingTools Inc
 
AI: Learning in AI 2
DataminingTools Inc
 
AI: Learning in AI
DataminingTools Inc
 
AI: Introduction to artificial intelligence
DataminingTools Inc
 
AI: Belief Networks
DataminingTools Inc
 
AI: AI & Searching
DataminingTools Inc
 
AI: AI & Problem Solving
DataminingTools Inc
 
Data Mining: Text and web mining
DataminingTools Inc
 
Data Mining: Outlier analysis
DataminingTools Inc
 
Data Mining: Mining stream time series and sequence data
DataminingTools Inc
 
Data Mining: Mining ,associations, and correlations
DataminingTools Inc
 
Data Mining: Graph mining and social network analysis
DataminingTools Inc
 
Data warehouse and olap technology
DataminingTools Inc
 
Data Mining: Data processing
DataminingTools Inc
 

Recently uploaded (20)

PDF
The Rise of AI and IoT in Mobile App Tech.pdf
IMG Global Infotech
 
PDF
"Beyond English: Navigating the Challenges of Building a Ukrainian-language R...
Fwdays
 
PPTX
AI Penetration Testing Essentials: A Cybersecurity Guide for 2025
defencerabbit Team
 
PDF
Agentic AI lifecycle for Enterprise Hyper-Automation
Debmalya Biswas
 
PPTX
Webinar: Introduction to LF Energy EVerest
DanBrown980551
 
PDF
CIFDAQ Market Wrap for the week of 4th July 2025
CIFDAQ
 
PDF
LLMs.txt: Easily Control How AI Crawls Your Site
Keploy
 
PPTX
Building Search Using OpenSearch: Limitations and Workarounds
Sease
 
PDF
NewMind AI - Journal 100 Insights After The 100th Issue
NewMind AI
 
PDF
Biography of Daniel Podor.pdf
Daniel Podor
 
PDF
Newgen Beyond Frankenstein_Build vs Buy_Digital_version.pdf
darshakparmar
 
PDF
Mastering Financial Management in Direct Selling
Epixel MLM Software
 
PPTX
"Autonomy of LLM Agents: Current State and Future Prospects", Oles` Petriv
Fwdays
 
PDF
Empower Inclusion Through Accessible Java Applications
Ana-Maria Mihalceanu
 
PDF
CIFDAQ Weekly Market Wrap for 11th July 2025
CIFDAQ
 
PDF
Bitcoin for Millennials podcast with Bram, Power Laws of Bitcoin
Stephen Perrenod
 
PDF
Reverse Engineering of Security Products: Developing an Advanced Microsoft De...
nwbxhhcyjv
 
PDF
How Startups Are Growing Faster with App Developers in Australia.pdf
India App Developer
 
PDF
CIFDAQ Market Insights for July 7th 2025
CIFDAQ
 
PDF
Fl Studio 24.2.2 Build 4597 Crack for Windows Free Download 2025
faizk77g
 
The Rise of AI and IoT in Mobile App Tech.pdf
IMG Global Infotech
 
"Beyond English: Navigating the Challenges of Building a Ukrainian-language R...
Fwdays
 
AI Penetration Testing Essentials: A Cybersecurity Guide for 2025
defencerabbit Team
 
Agentic AI lifecycle for Enterprise Hyper-Automation
Debmalya Biswas
 
Webinar: Introduction to LF Energy EVerest
DanBrown980551
 
CIFDAQ Market Wrap for the week of 4th July 2025
CIFDAQ
 
LLMs.txt: Easily Control How AI Crawls Your Site
Keploy
 
Building Search Using OpenSearch: Limitations and Workarounds
Sease
 
NewMind AI - Journal 100 Insights After The 100th Issue
NewMind AI
 
Biography of Daniel Podor.pdf
Daniel Podor
 
Newgen Beyond Frankenstein_Build vs Buy_Digital_version.pdf
darshakparmar
 
Mastering Financial Management in Direct Selling
Epixel MLM Software
 
"Autonomy of LLM Agents: Current State and Future Prospects", Oles` Petriv
Fwdays
 
Empower Inclusion Through Accessible Java Applications
Ana-Maria Mihalceanu
 
CIFDAQ Weekly Market Wrap for 11th July 2025
CIFDAQ
 
Bitcoin for Millennials podcast with Bram, Power Laws of Bitcoin
Stephen Perrenod
 
Reverse Engineering of Security Products: Developing an Advanced Microsoft De...
nwbxhhcyjv
 
How Startups Are Growing Faster with App Developers in Australia.pdf
India App Developer
 
CIFDAQ Market Insights for July 7th 2025
CIFDAQ
 
Fl Studio 24.2.2 Build 4597 Crack for Windows Free Download 2025
faizk77g
 

Classification Continued

  • 1. Data Mining Classification: Alternative Techniques
  • 2. Rule-Based ClassifierClassify records by using a collection of “if…then…” rulesRule: (Condition)  ywhere Condition is a conjunctions of attributes y is the class labelLHS: rule antecedent or conditionRHS: rule consequent
  • 3. Characteristics of Rule-Based ClassifierMutually exclusive rulesClassifier contains mutually exclusive rules if the rules are independent of each otherEvery record is covered by at most one ruleExhaustive rulesClassifier has exhaustive coverage if it accounts for every possible combination of attribute valuesEach record is covered by at least one rule
  • 4. Building Classification RulesDirect Method: Extract rules directly from data e.g.: RIPPER, CN2, Holte’s 1RIndirect Method: Extract rules from other classification models (e.g. decision trees, neural networks, etc).e.g: C4.5rules
  • 5. Direct Method: Sequential CoveringStart from an empty ruleGrow a rule using the Learn-One-Rule functionRemove training records covered by the ruleRepeat Step (2) and (3) until stopping criterion is met
  • 6. Aspects of Sequential CoveringRule GrowingInstance EliminationRule EvaluationStopping CriterionRule Pruning
  • 7. Contd…Grow a single ruleRemove Instances from rulePrune the rule (if necessary)Add rule to Current Rule SetRepeat
  • 8. Indirect Method: C4.5rulesExtract rules from an unpruned decision treeFor each rule, r: A  y, consider an alternative rule r’: A’  y where A’ is obtained by removing one of the conjuncts in ACompare the pessimistic error rate for r against all r’sPrune if one of the r’s has lower pessimistic error rateRepeat until we can no longer improve generalization error
  • 9. Indirect Method: C4.5rulesInstead of ordering the rules, order subsets of rules (class ordering)Each subset is a collection of rules with the same rule consequent (class)Compute description length of each subset Description length = L(error) + g L(model) g is a parameter that takes into account the presence of redundant attributes in a rule set (default value = 0.5)
  • 10. Advantages of Rule-Based ClassifiersAs highly expressive as decision treesEasy to interpretEasy to generateCan classify new instances rapidlyPerformance comparable to decision trees
  • 12. The set of stored records
  • 13. Distance Metric to compute distance between records
  • 14. The value of k, the number of nearest neighbors to retrieve
  • 15. To classify an unknown record:
  • 16. Compute distance to other training records
  • 17. Identify k nearest neighbors
  • 18. Use class labels of nearest neighbors to determine the class label of unknown record (e.g., by taking majority voteDefinition of Nearest NeighborK-nearest neighbors of a record x are data points that have the k smallest distance to x
  • 19. Nearest Neighbor Classification…Choosing the value of k:If k is too small, sensitive to noise pointsIf k is too large, neighborhood may include points from other classesScaling issuesAttributes may have to be scaled to prevent distance measures from being dominated by one of the attributesExample: height of a person may vary from 1.5m to 1.8m weight of a person may vary from 90lb to 300lb
  • 20. Nearest neighbor Classification…k-NN classifiers are lazy learners It does not build models explicitlyUnlike eager learners such as decision tree induction and rule-based systemsClassifying unknown records are relatively expensive
  • 21. Bayes ClassifierA probabilistic framework for solving classification problemsConditional Probability:Bayes theorem:
  • 22. Example of Bayes TheoremGiven: A doctor knows that meningitis causes stiff neck 50% of the timePrior probability of any patient having meningitis is 1/50,000Prior probability of any patient having stiff neck is 1/20 If a patient has stiff neck, what’s the probability he/she has meningitis?
  • 23. Naïve Bayes ClassifierAssume independence among attributes Ai when class is given: P(A1, A2, …, An |C) = P(A1| Cj) P(A2| Cj)… P(An| Cj)Can estimate P(Ai| Cj) for all Ai and Cj.New point is classified to Cj if P(Cj)  P(Ai| Cj) is maximal.
  • 24. Naïve Bayes ClassifierIf one of the conditional probability is zero, then the entire expression becomes zeroProbability estimation:c: number of classesp: prior probabilitym: parameter
  • 25. Naïve Bayes (Summary)Robust to isolated noise pointsHandle missing values by ignoring the instance during probability estimate calculationsRobust to irrelevant attributesIndependence assumption may not hold for some attributesUse other techniques such as Bayesian Belief Networks (BBN)
  • 26. Artificial Neural Networks (ANN)Model is an assembly of inter-connected nodes and weighted linksOutput node sums up each of its input value according to the weights of its linksCompare output node against some threshold t
  • 27. General Structure of ANNTraining ANN means learning the weights of the neurons
  • 28. Algorithm for learning ANNInitialize the weights (w0, w1, …, wk)Adjust the weights in such a way that the output of ANN is consistent with class labels of training examplesObjective function:Find the weights wi’s that minimize the above objective function e.g., backpropagation algorithm
  • 29. Ensemble MethodsConstruct a set of classifiers from the training dataPredict class label of previously unseen records by aggregating predictions made by multiple classifiers
  • 31. Why does it work?Suppose there are 25 base classifiersEach classifier has error rate,  = 0.35Assume classifiers are independentProbability that the ensemble classifier makes a wrong prediction:
  • 32. Examples of Ensemble MethodsHow to generate an ensemble of classifiers?BaggingBoosting
  • 33. BaggingSampling with replacementBuild classifier on each bootstrap sampleEach sample has probability (1 – 1/n)n of being selected
  • 34. BoostingAn iterative procedure to adaptively change distribution of training data by focusing more on previously misclassified recordsInitially, all N records are assigned equal weightsUnlike bagging, weights may change at the end of boosting round
  • 35. Visit more self help tutorialsPick a tutorial of your choice and browse through it at your own pace.The tutorials section is free, self-guiding and will not involve any additional support.Visit us at www.dataminingtools.net