Machine Learning
By: Dhananjay Birmole
INTRODUCTION
• Machine learning is a branch of artificial intelligence
• Machine learning name is derived from the concept that it deals with
“construction and study of systems that can learn from data”
• Machine learning can be seen as building blocks to make computers learn to
behave more intelligently
• it is a theoretical concept
Learning types
Machine Learning
Regression
• A regression problem is when the output variable is a real or
continuous value.
• It is a measure of the relation between the mean value of one
variable (e.g. output) and corresponding values of other variables.
• Regression analysis is a statistical process for estimating the
relationships among variables.
• Regression means to predict the output value using training data.
• Common function used are MSE, Gradient Descent
Classification
• A classification problem is when the output variable is a category,
such as “red” or “blue” or “disease” and “no disease”.
• Classification is technique to categorize our data into a desired and
distinct number of classes where we can assign label to each class.
• Classification means to divide the output using training data.
• Works on discrete values
• Common function used are Bayer classifier, Support Vector Machine
Classification vs
Regression• Classification means to group
the output into a class.
• Classification to predict the
type of tumor i.e. harmful or not
harmful using training data
• If it is discrete/categorical
variable, then it is classification
problem
• Regression means to predict the
output value using training
data.
• Regression to predict the house
price from training data
• If it is a real
number/continuous, then it is
regression problem.
Clustering
• An unsupervised learning method is a method in which we
draw references from datasets consisting of input data
without labeled responses.
• It is used as a process to find meaningful structure,
explanatory underlying processes, generative features, and
groupings inherent in a set of examples.
• It is basically a collection of objects on the basis of similarity and
dissimilarity between them.
• Common function used is K-mean
Decision Tree Learning
•Decision tree learning uses a
decision tree as a predictive model
which maps observations about an
item to conclusions about the item's
target value.
•To classify a new instance, we start
at the root and traverse the tree to
reach a leaf; at an internal node we
evaluate the predicate(or function)
on the data instance, to find which
child to go. The process continues
till we reach a leaf node 
• The decision tree in Figure3 classifies whether a
person will buy a sports car or a minivan depending on
their age and marital status. If the person is over 30
years and is not married, we walk the tree as follows :
‘over 30 years?’ -> yes -> ’married?’ -> no. Hence, the
model outputs a sportscar.
Random Forest
•Random Forest is a trademark term for an
ensemble of decision trees. In Random
Forest, we’ve collection of decision trees
(so known as “Forest”). To classify a new
object based on attributes, each tree
gives a classification and we say the tree
“votes” for that class. The forest chooses
the classification having the most votes
(over all the trees in the forest).
•Each tree is planted & grown as follows:
•If the number of cases in the training set
is N, then sample of N cases is taken at
random but with replacement. This
sample will be the training set for
growing the tree.
•If there are M input variables, a number
m<<M is specified such that at each node,
m variables are selected at random out of
the M and the best split on these m is
used to split the node. The value of m is
held constant during the forest growing.
•Each tree is grown to the largest extent
possible. There is no pruning.
True vs. False and Positive
vs. Negative
•The Boy Who Cried Wolf example
•"Wolf" is a positive class.
•"No wolf" is a negative class.
•We can summarize our "wolf-prediction" model
using a 2x2 confusion matrix that depicts all four
possible outcomes:
Machine Learning
•A true positive is an outcome where the
model correctly predicts
the positive class. Similarly, a true
negative is an outcome where the
model correctly predicts
the negative class.
•A false positive is an outcome where the
model incorrectly predicts
the positive class. And a false negative is
an outcome where the
model incorrectly predicts
the negative class.
•In the following sections, we'll look at
how to evaluate classification models
using metrics derived from these four
outcomes.
Linear Regression
•In ML, we have a set of input variables (x) that
are used to determine the output variable (y). A
relationship exists between the input variables
and the output variable. The goal of ML is to
quantify this relationship.
Fig 1
•In Linear Regression, the relationship
between the input variables (x) and
output variable (y) is expressed as an
equation of the form y = a + bx. Thus, the
goal of linear regression is to find out the
values of coefficients a and b. Here, a is
the intercept and b is the slope of the
line.
•Figure 1 shows the plotted x and y values
for a dataset. The goal is to fit a line that
is nearest to most of the points. This
would reduce the distance (‘error’)
between the y value of a data point and
the line.
Logistic Regression
• Linear regression predictions are continuous values
(rainfall in cm),logistic regression predictions are
discrete values (whether a student passed/failed)
• Logistic regression is best suited for binary
classification (datasets where y = 0 or 1, where 1
denotes the default class. Example: In predicting
whether an event will occur or not, the event that it
occurs is classified as 1. In predicting whether a
person will be sick or not, the sick instances are
denoted as 1). It is named after the transformation
function used in it, called the logistic function h(x)= 1/
(1 + ex), which is an S-shaped curve.after applying a
transformation function.
• In logistic regression, the output is in the form
of probabilities of the default class (unlike
linear regression, where the output is directly
produced). As it is a probability, the output
lies in the range of 0-1. The output (y-value) is
generated by log transforming the x-value,
using the logistic function h(x)= 1/ (1 + e^ -x) .
A threshold is then applied to force this
probability into a binary classification.
Machine
Learning Applications
Future of Machine
Learning
•Machine Learning can be a
competitive advantage to any
company be it a top MNC or a startup
as things that are currently being
done manually will be done tomorrow
by machines. Machine Learning
revolution will stay with us for long
and so will be the future of Machine
Learning.

More Related Content

PPTX
Machine Learning
PPT
ARTIFICIAL INTELLIGENCE.PPT
PDF
Chicago AWS Solutions Architect Mehdy Haghy recaps the new AI/ML releases and...
PPTX
Machine Learning
PDF
Artificial Intelligence - Hill climbing.
PPTX
Artificial neural network
PPTX
Google BARD v/s ChatGPT _ A review
PPTX
Data structure - Graph
Machine Learning
ARTIFICIAL INTELLIGENCE.PPT
Chicago AWS Solutions Architect Mehdy Haghy recaps the new AI/ML releases and...
Machine Learning
Artificial Intelligence - Hill climbing.
Artificial neural network
Google BARD v/s ChatGPT _ A review
Data structure - Graph

What's hot (20)

PPTX
Machine learning overview
PPT
Machine Learning
PPTX
Supervised and Unsupervised Learning In Machine Learning | Machine Learning T...
PPTX
Supervised learning and Unsupervised learning
PPTX
Machine Learning Tutorial | Machine Learning Basics | Machine Learning Algori...
PDF
Feature selection
PDF
Decision trees in Machine Learning
PPT
Decision tree
PDF
Classification Based Machine Learning Algorithms
PPTX
Random forest
PPTX
Unsupervised learning clustering
PPTX
Machine learning clustering
PPTX
Machine learning with ADA Boost
PDF
Dimensionality Reduction
PDF
Introduction to Machine Learning Classifiers
PPTX
Machine learning ppt.
PPT
Machine learning
PPTX
Machine Learning Tutorial Part - 1 | Machine Learning Tutorial For Beginners ...
PPTX
Decision Tree Algorithm With Example | Decision Tree In Machine Learning | Da...
PPTX
Random forest algorithm
Machine learning overview
Machine Learning
Supervised and Unsupervised Learning In Machine Learning | Machine Learning T...
Supervised learning and Unsupervised learning
Machine Learning Tutorial | Machine Learning Basics | Machine Learning Algori...
Feature selection
Decision trees in Machine Learning
Decision tree
Classification Based Machine Learning Algorithms
Random forest
Unsupervised learning clustering
Machine learning clustering
Machine learning with ADA Boost
Dimensionality Reduction
Introduction to Machine Learning Classifiers
Machine learning ppt.
Machine learning
Machine Learning Tutorial Part - 1 | Machine Learning Tutorial For Beginners ...
Decision Tree Algorithm With Example | Decision Tree In Machine Learning | Da...
Random forest algorithm
Ad

Similar to Machine Learning (20)

PPTX
Unit 3 – AIML.pptx
PPTX
Lecture 3 ml
PDF
Module 5.pdf Machine Learning Types and examples
PPTX
ML Unjkfmvjmnb ,mit-2 - Rejhjmfnvhjmnv gression.pptx
PPTX
DECISION TREE AND PROBABILISTIC MODELS.pptx
PPT
Supervised and unsupervised learning
PPTX
ML SFCSE.pptx
PPTX
Supervised Learning-Unit 3.pptx
PDF
Introduction to machine learning
PPT
RM7.ppt
PPTX
Supervised and Unsupervised Learning.pptx
PPTX
Unit 4 Classification of data and more info on it
PPTX
Data mining approaches and methods
PPTX
Supervised learning
PPTX
UNIT 3: Data Warehousing and Data Mining
PPT
3 DM Classification HFCS kilometres .ppt
PPTX
UNIT 3.pptx.......................................
PPTX
03 Data Mining Techniques
PPTX
Discriminant analysis.pptx
PDF
Data Science Interview Preparation(#DAY 02).pdf
Unit 3 – AIML.pptx
Lecture 3 ml
Module 5.pdf Machine Learning Types and examples
ML Unjkfmvjmnb ,mit-2 - Rejhjmfnvhjmnv gression.pptx
DECISION TREE AND PROBABILISTIC MODELS.pptx
Supervised and unsupervised learning
ML SFCSE.pptx
Supervised Learning-Unit 3.pptx
Introduction to machine learning
RM7.ppt
Supervised and Unsupervised Learning.pptx
Unit 4 Classification of data and more info on it
Data mining approaches and methods
Supervised learning
UNIT 3: Data Warehousing and Data Mining
3 DM Classification HFCS kilometres .ppt
UNIT 3.pptx.......................................
03 Data Mining Techniques
Discriminant analysis.pptx
Data Science Interview Preparation(#DAY 02).pdf
Ad

Recently uploaded (20)

PDF
Unit1 - AIML Chapter 1 concept and ethics
PDF
VSL-Strand-Post-tensioning-Systems-Technical-Catalogue_2019-01.pdf
PPTX
chapter 1.pptx dotnet technology introduction
PPTX
SE unit 1.pptx aaahshdhajdviwhsiehebeiwheiebeiev
PDF
20250617 - IR - Global Guide for HR - 51 pages.pdf
PDF
Micro 4 New.ppt.pdf a servay of cells and microorganism
PPTX
Agentic Artificial Intelligence (Agentic AI).pptx
PPTX
Solar energy pdf of gitam songa hemant k
PPTX
Cisco Network Behaviour dibuywvdsvdtdstydsdsa
PPTX
Environmental studies, Moudle 3-Environmental Pollution.pptx
PPTX
Wireless sensor networks (WSN) SRM unit 2
DOCX
ENVIRONMENTAL PROTECTION AND MANAGEMENT (18CVL756)
PPT
UNIT-I Machine Learning Essentials for 2nd years
PPT
Programmable Logic Controller PLC and Industrial Automation
PPTX
CNS - Unit 1 (Introduction To Computer Networks) - PPT (2).pptx
PDF
UEFA_Carbon_Footprint_Calculator_Methology_2.0.pdf
PPTX
Module1.pptxrjkeieuekwkwoowkemehehehrjrjrj
PDF
Computer System Architecture 3rd Edition-M Morris Mano.pdf
PDF
Present and Future of Systems Engineering: Air Combat Systems
PPTX
Principal presentation for NAAC (1).pptx
Unit1 - AIML Chapter 1 concept and ethics
VSL-Strand-Post-tensioning-Systems-Technical-Catalogue_2019-01.pdf
chapter 1.pptx dotnet technology introduction
SE unit 1.pptx aaahshdhajdviwhsiehebeiwheiebeiev
20250617 - IR - Global Guide for HR - 51 pages.pdf
Micro 4 New.ppt.pdf a servay of cells and microorganism
Agentic Artificial Intelligence (Agentic AI).pptx
Solar energy pdf of gitam songa hemant k
Cisco Network Behaviour dibuywvdsvdtdstydsdsa
Environmental studies, Moudle 3-Environmental Pollution.pptx
Wireless sensor networks (WSN) SRM unit 2
ENVIRONMENTAL PROTECTION AND MANAGEMENT (18CVL756)
UNIT-I Machine Learning Essentials for 2nd years
Programmable Logic Controller PLC and Industrial Automation
CNS - Unit 1 (Introduction To Computer Networks) - PPT (2).pptx
UEFA_Carbon_Footprint_Calculator_Methology_2.0.pdf
Module1.pptxrjkeieuekwkwoowkemehehehrjrjrj
Computer System Architecture 3rd Edition-M Morris Mano.pdf
Present and Future of Systems Engineering: Air Combat Systems
Principal presentation for NAAC (1).pptx

Machine Learning

  • 2. INTRODUCTION • Machine learning is a branch of artificial intelligence • Machine learning name is derived from the concept that it deals with “construction and study of systems that can learn from data” • Machine learning can be seen as building blocks to make computers learn to behave more intelligently • it is a theoretical concept
  • 5. Regression • A regression problem is when the output variable is a real or continuous value. • It is a measure of the relation between the mean value of one variable (e.g. output) and corresponding values of other variables. • Regression analysis is a statistical process for estimating the relationships among variables. • Regression means to predict the output value using training data. • Common function used are MSE, Gradient Descent
  • 6. Classification • A classification problem is when the output variable is a category, such as “red” or “blue” or “disease” and “no disease”. • Classification is technique to categorize our data into a desired and distinct number of classes where we can assign label to each class. • Classification means to divide the output using training data. • Works on discrete values • Common function used are Bayer classifier, Support Vector Machine
  • 7. Classification vs Regression• Classification means to group the output into a class. • Classification to predict the type of tumor i.e. harmful or not harmful using training data • If it is discrete/categorical variable, then it is classification problem • Regression means to predict the output value using training data. • Regression to predict the house price from training data • If it is a real number/continuous, then it is regression problem.
  • 8. Clustering • An unsupervised learning method is a method in which we draw references from datasets consisting of input data without labeled responses. • It is used as a process to find meaningful structure, explanatory underlying processes, generative features, and groupings inherent in a set of examples. • It is basically a collection of objects on the basis of similarity and dissimilarity between them. • Common function used is K-mean
  • 9. Decision Tree Learning •Decision tree learning uses a decision tree as a predictive model which maps observations about an item to conclusions about the item's target value. •To classify a new instance, we start at the root and traverse the tree to reach a leaf; at an internal node we evaluate the predicate(or function) on the data instance, to find which child to go. The process continues till we reach a leaf node 
  • 10. • The decision tree in Figure3 classifies whether a person will buy a sports car or a minivan depending on their age and marital status. If the person is over 30 years and is not married, we walk the tree as follows : ‘over 30 years?’ -> yes -> ’married?’ -> no. Hence, the model outputs a sportscar.
  • 11. Random Forest •Random Forest is a trademark term for an ensemble of decision trees. In Random Forest, we’ve collection of decision trees (so known as “Forest”). To classify a new object based on attributes, each tree gives a classification and we say the tree “votes” for that class. The forest chooses the classification having the most votes (over all the trees in the forest).
  • 12. •Each tree is planted & grown as follows: •If the number of cases in the training set is N, then sample of N cases is taken at random but with replacement. This sample will be the training set for growing the tree. •If there are M input variables, a number m<<M is specified such that at each node, m variables are selected at random out of the M and the best split on these m is used to split the node. The value of m is held constant during the forest growing. •Each tree is grown to the largest extent possible. There is no pruning.
  • 13. True vs. False and Positive vs. Negative •The Boy Who Cried Wolf example •"Wolf" is a positive class. •"No wolf" is a negative class. •We can summarize our "wolf-prediction" model using a 2x2 confusion matrix that depicts all four possible outcomes:
  • 15. •A true positive is an outcome where the model correctly predicts the positive class. Similarly, a true negative is an outcome where the model correctly predicts the negative class. •A false positive is an outcome where the model incorrectly predicts the positive class. And a false negative is an outcome where the model incorrectly predicts the negative class. •In the following sections, we'll look at how to evaluate classification models using metrics derived from these four outcomes.
  • 16. Linear Regression •In ML, we have a set of input variables (x) that are used to determine the output variable (y). A relationship exists between the input variables and the output variable. The goal of ML is to quantify this relationship. Fig 1
  • 17. •In Linear Regression, the relationship between the input variables (x) and output variable (y) is expressed as an equation of the form y = a + bx. Thus, the goal of linear regression is to find out the values of coefficients a and b. Here, a is the intercept and b is the slope of the line. •Figure 1 shows the plotted x and y values for a dataset. The goal is to fit a line that is nearest to most of the points. This would reduce the distance (‘error’) between the y value of a data point and the line.
  • 18. Logistic Regression • Linear regression predictions are continuous values (rainfall in cm),logistic regression predictions are discrete values (whether a student passed/failed) • Logistic regression is best suited for binary classification (datasets where y = 0 or 1, where 1 denotes the default class. Example: In predicting whether an event will occur or not, the event that it occurs is classified as 1. In predicting whether a person will be sick or not, the sick instances are denoted as 1). It is named after the transformation function used in it, called the logistic function h(x)= 1/ (1 + ex), which is an S-shaped curve.after applying a transformation function.
  • 19. • In logistic regression, the output is in the form of probabilities of the default class (unlike linear regression, where the output is directly produced). As it is a probability, the output lies in the range of 0-1. The output (y-value) is generated by log transforming the x-value, using the logistic function h(x)= 1/ (1 + e^ -x) . A threshold is then applied to force this probability into a binary classification.
  • 21. Future of Machine Learning •Machine Learning can be a competitive advantage to any company be it a top MNC or a startup as things that are currently being done manually will be done tomorrow by machines. Machine Learning revolution will stay with us for long and so will be the future of Machine Learning.