SlideShare a Scribd company logo
Applications of Classification Algorithm
•Classification algorithms are widely used in many real-world
applications across various domains, including:
Email spam filtering
Credit risk assessment
Medical diagnosis
Image classification
Sentiment analysis.
Fraud detection
Quality control
Recommendation systems
Classification Error Rate
In machine learning, misclassification rate is a metric that tells us the percentage of
observations that were incorrectly predicted by some classification model.
• It is calculated as:
• Misclassification Rate = # incorrect predictions / # total predictions The value for
misclassification rate can range from 0 to 1 where:
 0 represents a model that had zero incorrect predictions.
1 represents a model that had completely incorrect predictions.
• The lower the value for the misclassification rate, the better a classification model
is able to predict the outcomes of the response variable.
• The following example show how to calculate misclassification rate for a logistic
regression model in practice.
Example: Calculating Misclassification Rate for a Logistic Regression Model
Suppose we use a logistic regression model to predict whether or not 400 different college
basketball players get drafted into the NBA.
The following confusion matrix summarizes the predictions made by the model:
Here is how to calculate the misclassification rate for the model:
•Misclassification Rate = # incorrect predictions / # total predictions
•Misclassification Rate = (false positive + false negative) / (total predictions)
•Misclassification Rate = (70 + 40) / (400)
•Misclassification Rate = 0.275
• The misclassification rate for this model is 0.275 or 27.5%.
• This means the model incorrectly predicted the outcome for 27.5% of the players.
• The opposite of misclassification rate would be accuracy,
which is calculated as:
• Accuracy = 1 – Misclassification rate
 Accuracy = 1 – 0.275
 Accuracy = 0.725
• This means the model correctly predicted the outcome for 72.5% of the players.
Bayes Classification Rule
What is Naive Bayes classifiers?
• Naive Bayes classifiers are a collection of classification algorithms based on Bayes’ Theorem. It is
not a single algorithm but a family of algorithms where all of them share a common principle, i.e.
every pair of features being classified is independent of each other. To start with, let us consider a
dataset.
•Why it is called Naive Bayes?
•The “Naive” part of the name indicates the simplifying assumption
made by the Naïve Bayes classifier. The classifier assumes that the
features used to describe an observation are conditionally
independent, given the class label. The “Bayes” part of the name
refers to Reverend Thomas Bayes, an 18th-century statistician and
theologian who formulated Bayes’ theorem.
•Assumption of Naive Bayes
•The fundamental Naive Bayes assumption is that each feature
makes an:
1. Feature independence: The features of the data are conditionally independent of each other,
given the class label.
2. Continuous features are normally distributed: If a feature is continuous, then it is assumed to
be normally distributed within each class.
3. Discrete features have multinomial distributions: If a feature is discrete, then it is assumed to
have a multinomial distribution within each class.
4. Features are equally important: All features are assumed to contribute equally
.
•No missing data: The data should not contain any missing values.
• Bayes’ Theorem
• Bayes’ Theorem finds the probability of an event occurring given the probability of another event that
has already occurred. Bayes’ theorem is stated mathematically as the following
• equation:
where A and B are events and P(B) ≠ 0
Basically, we are trying to find probability of event A, given the event B is true. to the prediction of the class
label
Event B is also termed as evidence.
P(A) is the priori of A (the prior probability, i.e. Probability of event before
evidence is seen). The evidence is an attribute value of an unknown
instance (here, it is event B).
P(B) is Marginal Probability: Probability of Evidence.
P(A|B) is a posteriori probability of B, i.e. probability of event after
evidence is seen.
P(B|A) is Likelihood probability i.e the likelihood that a hypothesis will
come true based on the evidence.
Types of Naive Bayes Model
• There are three types of Naive Bayes Model:
• Gaussian Naive Bayes classifier
• In Gaussian Naive Bayes, continuous values associated with each feature are assumed to be distributed
according to a Gaussian distribution. A Gaussian distribution is also called Normal distribution
When plotted, it gives a bell shaped curve which is symmetric about the mean of the feature values as
shown below:
Multinomial Naive Bayes
Feature vectors represent the frequencies with which certain events have been generated by a
multinomial distribution. This is the event model typically used for document classification.
Bernoulli Naive Bayes
In the multivariate Bernoulli event model, features are independent booleans (binary variables)
describing inputs. Like the multinomial model, this model is popular for document classification
tasks, where binary term occurrence(i.e. a word occurs in a document or not) features are used
rather than term frequencies(i.e. frequency of a word in the document).
Advantages of Naive Bayes Classifier
Easy to implement and computationally efficient.
Effective in cases with a large number of features.
Performs well even with limited training data.
Disadvantages of Naive Bayes Classifier
 Assumes that features are independent, which may not always hold in real-world data.
 Can be influenced by irrelevant attributes.
 May assign zero probability to unseen events, leading to poor generalization.
Applications of Naive Bayes Classifier
 Spam Email Filtering: Classifies emails as spam or non-spam based on features.
 Text Classification: Used in sentiment analysis, document categorization, and topic
classification.
 Medical Diagnosis: Helps in predicting the likelihood of a disease based on
symptoms.
 Credit Scoring: Evaluates creditworthiness of individuals for loan approval.
 Weather Prediction: Classifies weather conditions based on various factors
Linear Methods for Classification
• Linear methods for classification are techniques that use linear functions to separate different classes of data. They are
based on the idea of finding a decision boundary that minimizes the classification error or maximizes the likelihood of the
data. Some examples of linear methods for classification are:
 Logistic regression: This method models the probability of a binary response variable as a logistic function of a linear
combination of predictor variables. It estimates the parameters of the linear function by maximizing the likelihood of
the observed data1
.
 Linear discriminant analysis (LDA): This method assumes that the data from each class follows a multivariate
normal distribution with a common covariance matrix, and derives a linear function that best discriminates between
the classes. It estimates the parameters of the normal distributions by using the sample means and covariance matrix of
the data2
.
 Support vector machines (SVMs): This method finds a linear function that maximizes the margin between the classes,
where the margin is defined as the distance from the decision boundary to the closest data points. It uses a technique
called kernel trick to transform the data into a higher-dimensional space where a linear separation is possible3
.

More Related Content

Similar to Applications of Classification Algorithm.pptx (20)

PPTX
Ml part2
Leon Gladston
 
PPTX
supervised machine learning algorithms support vector machine
pranalisonawane8600
 
PPTX
Naive Bayes | Statistics
Transweb Global Inc
 
PDF
NAIVE BAYES ALGORITHM
Rang Technologies
 
PPTX
UNIT 3: Data Warehousing and Data Mining
Nandakumar P
 
DOCX
introduction to machine learning unit iv
GayathriRHICETCSESTA
 
PPTX
regression.pptx
aneeshs28
 
PDF
classification_clean.pdf
Afaq Ahmad
 
PPTX
Navies bayes
HassanRaza323
 
PPT
Supervised and unsupervised learning
AmAn Singh
 
PPTX
Naïve Bayes Classifier Algorithm.pptx
Shubham Jaybhaye
 
PPTX
Big Data Analytics.pptx
Kaviya452563
 
PPTX
Unit 3 – AIML.pptx
hiblooms
 
PPTX
Supervised Learning.pptx
Abdul Wali Khan University Mardan(AWKUM)
 
PPTX
classification.pptx
VaishnaviGaikwad67
 
PPTX
CHAPTER 11 LOGISTIC REGRESSION.pptx
UmaDeviAnanth
 
PPTX
EDAB Module 5 Singular Value Decomposition (SVD).pptx
rajalakshmi5921
 
PPTX
Singular Value Decomposition (SVD).pptx
rajalakshmi5921
 
PDF
Prediciting restaurant and popularity based on Yelp Dataset project report
ALIN BABU
 
PDF
76a15ed521b7679e372aab35412ab78ab583436a-1602816156135.pdf
triwicak1
 
Ml part2
Leon Gladston
 
supervised machine learning algorithms support vector machine
pranalisonawane8600
 
Naive Bayes | Statistics
Transweb Global Inc
 
NAIVE BAYES ALGORITHM
Rang Technologies
 
UNIT 3: Data Warehousing and Data Mining
Nandakumar P
 
introduction to machine learning unit iv
GayathriRHICETCSESTA
 
regression.pptx
aneeshs28
 
classification_clean.pdf
Afaq Ahmad
 
Navies bayes
HassanRaza323
 
Supervised and unsupervised learning
AmAn Singh
 
Naïve Bayes Classifier Algorithm.pptx
Shubham Jaybhaye
 
Big Data Analytics.pptx
Kaviya452563
 
Unit 3 – AIML.pptx
hiblooms
 
classification.pptx
VaishnaviGaikwad67
 
CHAPTER 11 LOGISTIC REGRESSION.pptx
UmaDeviAnanth
 
EDAB Module 5 Singular Value Decomposition (SVD).pptx
rajalakshmi5921
 
Singular Value Decomposition (SVD).pptx
rajalakshmi5921
 
Prediciting restaurant and popularity based on Yelp Dataset project report
ALIN BABU
 
76a15ed521b7679e372aab35412ab78ab583436a-1602816156135.pdf
triwicak1
 

More from nagarajan740445 (20)

PPTX
Visual impairment is a significant global.pptx
nagarajan740445
 
PPTX
development and effectiveness of assistive technologies for visually.pptx
nagarajan740445
 
PPTX
Corporate governance and sustainablility G.pptx
nagarajan740445
 
PPTX
FDP from 17th Feb 2025 LCD PPT Banner.pptx
nagarajan740445
 
PPTX
FDP from 18th Feb 2025 LCD PPT Banner.pptx
nagarajan740445
 
PPTX
Nurturing-Cooperative-Excellence-Through-Strategic-Leadership.pptx
nagarajan740445
 
PPTX
20.02.2025 artificial intelligence Banner.pptx
nagarajan740445
 
PPTX
social media marketing S B kalatage.pptx
nagarajan740445
 
PPT
Individual Differences Individual Differences & Traits.ppt
nagarajan740445
 
PPTX
Role Role of financial of financial Institutions.pptx
nagarajan740445
 
PPTX
promotion of venture promotion of venture.pptx
nagarajan740445
 
PPTX
Successful Project Successful Project Delivery.pptx
nagarajan740445
 
PPTX
Computer Security National Security Day.pptx
nagarajan740445
 
PPTX
business plan process business plan process .pptx
nagarajan740445
 
PPTX
theories of entrepreneurship theories of entrepreneurship new.pptx
nagarajan740445
 
PPTX
business decision making Review of probability theory.pptx
nagarajan740445
 
PPTX
elements of theorories in managment.pptx
nagarajan740445
 
PPTX
Remuneration of Employees Remuneration of Employees.pptx
nagarajan740445
 
PPTX
Design thinking in indian business. pptx
nagarajan740445
 
PPTX
CASE STUDY ON CASE STUDY ON HR782024.pptx
nagarajan740445
 
Visual impairment is a significant global.pptx
nagarajan740445
 
development and effectiveness of assistive technologies for visually.pptx
nagarajan740445
 
Corporate governance and sustainablility G.pptx
nagarajan740445
 
FDP from 17th Feb 2025 LCD PPT Banner.pptx
nagarajan740445
 
FDP from 18th Feb 2025 LCD PPT Banner.pptx
nagarajan740445
 
Nurturing-Cooperative-Excellence-Through-Strategic-Leadership.pptx
nagarajan740445
 
20.02.2025 artificial intelligence Banner.pptx
nagarajan740445
 
social media marketing S B kalatage.pptx
nagarajan740445
 
Individual Differences Individual Differences & Traits.ppt
nagarajan740445
 
Role Role of financial of financial Institutions.pptx
nagarajan740445
 
promotion of venture promotion of venture.pptx
nagarajan740445
 
Successful Project Successful Project Delivery.pptx
nagarajan740445
 
Computer Security National Security Day.pptx
nagarajan740445
 
business plan process business plan process .pptx
nagarajan740445
 
theories of entrepreneurship theories of entrepreneurship new.pptx
nagarajan740445
 
business decision making Review of probability theory.pptx
nagarajan740445
 
elements of theorories in managment.pptx
nagarajan740445
 
Remuneration of Employees Remuneration of Employees.pptx
nagarajan740445
 
Design thinking in indian business. pptx
nagarajan740445
 
CASE STUDY ON CASE STUDY ON HR782024.pptx
nagarajan740445
 
Ad

Recently uploaded (20)

PDF
Generative AI: it's STILL not a robot (CIJ Summer 2025)
Paul Bradshaw
 
PDF
ARAL-Orientation_Morning-Session_Day-11.pdf
JoelVilloso1
 
PDF
The Constitution Review Committee (CRC) has released an updated schedule for ...
nservice241
 
PDF
Women's Health: Essential Tips for Every Stage.pdf
Iftikhar Ahmed
 
PPTX
GRADE-3-PPT-EVE-2025-ENG-Q1-LESSON-1.pptx
EveOdrapngimapNarido
 
PPTX
QUARTER 1 WEEK 2 PLOT, POV AND CONFLICTS
KynaParas
 
PPTX
PPT-Q1-WK-3-ENGLISH Revised Matatag Grade 3.pptx
reijhongidayawan02
 
PDF
Stokey: A Jewish Village by Rachel Kolsky
History of Stoke Newington
 
PPTX
I AM MALALA The Girl Who Stood Up for Education and was Shot by the Taliban...
Beena E S
 
PDF
ARAL_Orientation_Day-2-Sessions_ARAL-Readung ARAL-Mathematics ARAL-Sciencev2.pdf
JoelVilloso1
 
PPTX
How to Handle Salesperson Commision in Odoo 18 Sales
Celine George
 
PPTX
How to Set Up Tags in Odoo 18 - Odoo Slides
Celine George
 
PDF
Aprendendo Arquitetura Framework Salesforce - Dia 03
Mauricio Alexandre Silva
 
PDF
Chapter-V-DED-Entrepreneurship: Institutions Facilitating Entrepreneurship
Dayanand Huded
 
PPTX
Universal immunization Programme (UIP).pptx
Vishal Chanalia
 
PPTX
HUMAN RESOURCE MANAGEMENT: RECRUITMENT, SELECTION, PLACEMENT, DEPLOYMENT, TRA...
PRADEEP ABOTHU
 
PPTX
Cultivation practice of Litchi in Nepal.pptx
UmeshTimilsina1
 
PPTX
How to Create Odoo JS Dialog_Popup in Odoo 18
Celine George
 
PDF
Geographical Diversity of India 100 Mcq.pdf/ 7th class new ncert /Social/Samy...
Sandeep Swamy
 
PPTX
grade 5 lesson matatag ENGLISH 5_Q1_PPT_WEEK4.pptx
SireQuinn
 
Generative AI: it's STILL not a robot (CIJ Summer 2025)
Paul Bradshaw
 
ARAL-Orientation_Morning-Session_Day-11.pdf
JoelVilloso1
 
The Constitution Review Committee (CRC) has released an updated schedule for ...
nservice241
 
Women's Health: Essential Tips for Every Stage.pdf
Iftikhar Ahmed
 
GRADE-3-PPT-EVE-2025-ENG-Q1-LESSON-1.pptx
EveOdrapngimapNarido
 
QUARTER 1 WEEK 2 PLOT, POV AND CONFLICTS
KynaParas
 
PPT-Q1-WK-3-ENGLISH Revised Matatag Grade 3.pptx
reijhongidayawan02
 
Stokey: A Jewish Village by Rachel Kolsky
History of Stoke Newington
 
I AM MALALA The Girl Who Stood Up for Education and was Shot by the Taliban...
Beena E S
 
ARAL_Orientation_Day-2-Sessions_ARAL-Readung ARAL-Mathematics ARAL-Sciencev2.pdf
JoelVilloso1
 
How to Handle Salesperson Commision in Odoo 18 Sales
Celine George
 
How to Set Up Tags in Odoo 18 - Odoo Slides
Celine George
 
Aprendendo Arquitetura Framework Salesforce - Dia 03
Mauricio Alexandre Silva
 
Chapter-V-DED-Entrepreneurship: Institutions Facilitating Entrepreneurship
Dayanand Huded
 
Universal immunization Programme (UIP).pptx
Vishal Chanalia
 
HUMAN RESOURCE MANAGEMENT: RECRUITMENT, SELECTION, PLACEMENT, DEPLOYMENT, TRA...
PRADEEP ABOTHU
 
Cultivation practice of Litchi in Nepal.pptx
UmeshTimilsina1
 
How to Create Odoo JS Dialog_Popup in Odoo 18
Celine George
 
Geographical Diversity of India 100 Mcq.pdf/ 7th class new ncert /Social/Samy...
Sandeep Swamy
 
grade 5 lesson matatag ENGLISH 5_Q1_PPT_WEEK4.pptx
SireQuinn
 
Ad

Applications of Classification Algorithm.pptx

  • 1. Applications of Classification Algorithm •Classification algorithms are widely used in many real-world applications across various domains, including: Email spam filtering Credit risk assessment Medical diagnosis Image classification Sentiment analysis. Fraud detection Quality control Recommendation systems
  • 2. Classification Error Rate In machine learning, misclassification rate is a metric that tells us the percentage of observations that were incorrectly predicted by some classification model. • It is calculated as: • Misclassification Rate = # incorrect predictions / # total predictions The value for misclassification rate can range from 0 to 1 where:  0 represents a model that had zero incorrect predictions. 1 represents a model that had completely incorrect predictions. • The lower the value for the misclassification rate, the better a classification model is able to predict the outcomes of the response variable. • The following example show how to calculate misclassification rate for a logistic regression model in practice.
  • 3. Example: Calculating Misclassification Rate for a Logistic Regression Model Suppose we use a logistic regression model to predict whether or not 400 different college basketball players get drafted into the NBA. The following confusion matrix summarizes the predictions made by the model: Here is how to calculate the misclassification rate for the model: •Misclassification Rate = # incorrect predictions / # total predictions •Misclassification Rate = (false positive + false negative) / (total predictions) •Misclassification Rate = (70 + 40) / (400) •Misclassification Rate = 0.275
  • 4. • The misclassification rate for this model is 0.275 or 27.5%. • This means the model incorrectly predicted the outcome for 27.5% of the players. • The opposite of misclassification rate would be accuracy, which is calculated as: • Accuracy = 1 – Misclassification rate  Accuracy = 1 – 0.275  Accuracy = 0.725 • This means the model correctly predicted the outcome for 72.5% of the players. Bayes Classification Rule What is Naive Bayes classifiers? • Naive Bayes classifiers are a collection of classification algorithms based on Bayes’ Theorem. It is not a single algorithm but a family of algorithms where all of them share a common principle, i.e. every pair of features being classified is independent of each other. To start with, let us consider a dataset.
  • 5. •Why it is called Naive Bayes? •The “Naive” part of the name indicates the simplifying assumption made by the Naïve Bayes classifier. The classifier assumes that the features used to describe an observation are conditionally independent, given the class label. The “Bayes” part of the name refers to Reverend Thomas Bayes, an 18th-century statistician and theologian who formulated Bayes’ theorem. •Assumption of Naive Bayes •The fundamental Naive Bayes assumption is that each feature makes an:
  • 6. 1. Feature independence: The features of the data are conditionally independent of each other, given the class label. 2. Continuous features are normally distributed: If a feature is continuous, then it is assumed to be normally distributed within each class. 3. Discrete features have multinomial distributions: If a feature is discrete, then it is assumed to have a multinomial distribution within each class. 4. Features are equally important: All features are assumed to contribute equally . •No missing data: The data should not contain any missing values. • Bayes’ Theorem • Bayes’ Theorem finds the probability of an event occurring given the probability of another event that has already occurred. Bayes’ theorem is stated mathematically as the following • equation: where A and B are events and P(B) ≠ 0 Basically, we are trying to find probability of event A, given the event B is true. to the prediction of the class label
  • 7. Event B is also termed as evidence. P(A) is the priori of A (the prior probability, i.e. Probability of event before evidence is seen). The evidence is an attribute value of an unknown instance (here, it is event B). P(B) is Marginal Probability: Probability of Evidence. P(A|B) is a posteriori probability of B, i.e. probability of event after evidence is seen. P(B|A) is Likelihood probability i.e the likelihood that a hypothesis will come true based on the evidence.
  • 8. Types of Naive Bayes Model • There are three types of Naive Bayes Model: • Gaussian Naive Bayes classifier • In Gaussian Naive Bayes, continuous values associated with each feature are assumed to be distributed according to a Gaussian distribution. A Gaussian distribution is also called Normal distribution When plotted, it gives a bell shaped curve which is symmetric about the mean of the feature values as shown below:
  • 9. Multinomial Naive Bayes Feature vectors represent the frequencies with which certain events have been generated by a multinomial distribution. This is the event model typically used for document classification. Bernoulli Naive Bayes In the multivariate Bernoulli event model, features are independent booleans (binary variables) describing inputs. Like the multinomial model, this model is popular for document classification tasks, where binary term occurrence(i.e. a word occurs in a document or not) features are used rather than term frequencies(i.e. frequency of a word in the document). Advantages of Naive Bayes Classifier Easy to implement and computationally efficient. Effective in cases with a large number of features. Performs well even with limited training data.
  • 10. Disadvantages of Naive Bayes Classifier  Assumes that features are independent, which may not always hold in real-world data.  Can be influenced by irrelevant attributes.  May assign zero probability to unseen events, leading to poor generalization. Applications of Naive Bayes Classifier  Spam Email Filtering: Classifies emails as spam or non-spam based on features.  Text Classification: Used in sentiment analysis, document categorization, and topic classification.  Medical Diagnosis: Helps in predicting the likelihood of a disease based on symptoms.  Credit Scoring: Evaluates creditworthiness of individuals for loan approval.  Weather Prediction: Classifies weather conditions based on various factors
  • 11. Linear Methods for Classification • Linear methods for classification are techniques that use linear functions to separate different classes of data. They are based on the idea of finding a decision boundary that minimizes the classification error or maximizes the likelihood of the data. Some examples of linear methods for classification are:  Logistic regression: This method models the probability of a binary response variable as a logistic function of a linear combination of predictor variables. It estimates the parameters of the linear function by maximizing the likelihood of the observed data1 .  Linear discriminant analysis (LDA): This method assumes that the data from each class follows a multivariate normal distribution with a common covariance matrix, and derives a linear function that best discriminates between the classes. It estimates the parameters of the normal distributions by using the sample means and covariance matrix of the data2 .  Support vector machines (SVMs): This method finds a linear function that maximizes the margin between the classes, where the margin is defined as the distance from the decision boundary to the closest data points. It uses a technique called kernel trick to transform the data into a higher-dimensional space where a linear separation is possible3 .