SlideShare a Scribd company logo
Computational
decision making
Dr. Boris Adryan
@BorisAdryan
What I aim to
provide
✓basic vocabulary for
✓fundamental concepts of
computational decision making
✓phenomenological introduction
to machine learning methods
✓a rough idea when and how to
use these methods
What this
presentation isn’t
x hands-on tutorial
x thorough summary
x comprehensive guide
x technical deep dive
x statistics course
Is this artificial
intelligence?
word = input(‘Enter a word:’)
for key in British_dictionary.iteritems():
if key.startswith(word):
print(‘This is a British word.’)
Is this machine
learning?
temperature = float(input('What is the temperature?'))
if temperature >= 1.0:
print('Wear shorts.')
else:
print('Wear long underwear.’)
Definition
Rule-based decision making on
the basis of numeric thresholds
or string patterns etc is not
machine learning.
And most definitely it is
not artificial intelligence.
But what if…
…the threshold is inferred at
run time?
“Write a software that says how
close to Euston you can move if
you can afford to spend £650k.”
450
550
650
750
850
950
1050
0 3 6 9 12
Average property price
Northern Line, number of stops from Euston
table = input_table(‘cost at station’)
print(where_x_yields_a_low_enough_y)
Linear regression
is probably the most simple
“machine learning” method.
Average property price
Northern Line, number of stops from Euston
It is an example of supervised
learning, because we teach the
computer the relation between
an input variable (“feature”) and
an output variable (“label”).
450
550
650
750
850
950
1050
0 3 6 9 12
y = m . x + b
Linear regression
can become arbitrarily
complicated.
The difference between curve fitting
in statistics and machine learning is
mostly semantics.
f(number of stops to Euston,
square footage, bedrooms,
bathrooms, …) price
many features
Euston
High
Barnet
small
large
price
Classification tasks I. setosa
I. virginica
I. versicolor
All images from https://en.wikipedia.org/wiki/Iris_flower_data_set
Rather than projecting the feature
vector onto a continuous variable,
many supervised learning
methods identify “class labels”.
Classification tasks
Supervised learning requires
complete input matrices. Missing
or nonsensical values have to be
replaced or removed.
Non-numerical features (think, e.g.
“name of colour”, “smell”) have to
be encoded.
class
label
feature 1 feature 2 feature 3 feature 4
1 5.1 3.5 1.4 0.2
1 4.9 3.0 1.4 0.2
2 7.0 3.2 4.7 1.4
3 6.3 3.3 6.0 2.5
Classification tasks
sepal length
sepalwidth
I. setosa
I. virginica
In a first approximation,
classification (by regression) aims
to find a function that best
separates the different class labels.
f(sepal width, sepal length) (1,2)
“1”
“2”
sepal length
sepalwidth
I. setosa
I. virginica
I. versicolor
Decision trees
sepal width
petalwidth
A decision tree can be understood
as a series of linear separations of
the data.
ratio of sepal width : sepal length
I. virginica
ratio of petal width : sepal width
I. setosa I. versicolor
Random forests
A collection of decision trees, each
trained on a random subset of the
data, can minimise the risk of
over-fitting.
A single big decision tree trained
on all data can effectively describe
a single data point.
sepal width
petalwidth
Over-/Underfitting
Sloppy separation is called
underfitting, and greedy
separation overfitting.
Counteracting an overfit is called
regularisation, and works by
penalising too many features (L1)
or too strong feature weights (L2).
underfit (high bias)
okay
overfit (high variance)
Dimensionality
reduction
Dimensionality reduction aims to
reduce the complexity of a dataset
(in respect to number of features).
The first principal components are
dimensions that explain most of a
dataset’s variance.
Support vector
machine
The SVM aims to provide an ideal
separation plane by supporting it
with a training data vector.
A classification margin protects the
SVM against overfitting.
?
margin support vectors
decision boundary
wTx = 0
negative hyperplane
wTx = -1
positive hyperplane
wTx = 1
Kernel trick
Input data can be projected into a
higher-dimensional space that
allows linear separation of
otherwise inseparable data.
There are different kernels, such
as the radial basis function
(Gaussian) or polynomial kernel.
“2D” input space
Φ
“3D” feature space
http://scikit-learn.org/0.18/auto_examples/svm/plot_iris.html
features
weather forecast
airport location
# of gates
# of runways
# of
snowploughs
airline
aircraft
BLACK
BOX
training
flights
cancelled in
the past classifier
ranked list of
relevant features
weight of
features
thresholds for
features
performance
metric
prediction
new data
General
approach
An intuitive example from real life.
training
classifier
performance
assessment
good enough?
success!
moredatafortraining
data
no
yes
sensitivity
“truepositives”
1-specificity
“false positives”
0 0.2 0.4 0.6 0.8 1.0
1.0
0.8
0.6
0.4
0.2
worse than
random
guess
Classifier
performance
Not all machine learning behaves
ideal, and performance metrics
are important for quality checks
and parameter tuning.
https://en.wikipedia.org/wiki/Precision_and_recall
There is a wide range of performance
metrics, comprising combinations of
true & false positives as well as
true & false negatives.
Metrics zoo
positive class (P) negative class (N)
predicted positive true positive (TP) false positive (FP)
predicted negative false negative (FN) true negative (TN)
data acquisition model building test use in production
data recording
(production system)
evaluation
raw data clean-up feature engineering model learning model selection
labour intense compute intensebrain intense
development
production
ML pipeline
Choosing a method
from: Olson et al., 2017, https://arxiv.org/abs/1708.05070
There is no ‘one-size-fits-all’
machine learning method. Most
methods need to be carefully
tuned to perform ideal.
Often, there a ‘non-functional’
constraints on choosing a method.
Runtime, interpretability, etc.
What about
neural networks?
feature 1 feature 2 feature 3
weight 1 weight 2 weight 3
input function
activation function
class output
error for weight updates
a simple perceptron
Neural networks attempt to mimic
the integrative properties of neurons.
The perceptron is a single-layer
network.
inputs
outputs
Deep neural networks
http://www.asimovinstitute.org/neural-network-zoo
While many artificial neural networks
show great performance, the basis on
which features exactly the classification
works remains largely unknown.
In RL, the methods iteratively
learn to optimise an output from
an abstract representation of a
system
Reinforcement
learning
move (l/r)
or shoot
unknown
systemmap (210 x 160 pixels, 8-bit RGB)
actual score
choose action on basis
of map to optimise score
Mnih et al., Nature (2015)
Machine learning can help to
structure and explore an
unknown dataset.
These methods aid the
identification of classes where
their existence isn’t known yet.
Unsupervised
learning
• Hierarchical clustering
• K-means clustering
• Expectation maximisation
• Density-based clustering
plus
clever visualisation
Hierarchical clustering
Derives hierarchical dependencies of
individual rows and columns in the
dataset on the basis of similarity
(correlation) between their properties.
Combined with a heatmap, gives a good
first impression of a dataset.
k-means clustering
Defines k different centroids to which
data points are assigned by proximity. If
the distance doesn’t get much smaller, k
is the number of clusters in the set.
Density-based clustering and expectation
maximisation are conceptually related,
the latter giving a probability for
membership in any group.
k = 2
k vs centroid distance
Conclusions
https://badryan.github.io/2015/10/20/is-it-all-machine-learning.html
•People on the Internet steal
infographics.
•ML methods have been around in
the stats world for ages, but big
data sets and compute power
make them more widely known.
•Understanding key principles
behind ML should be part of the
school curriculum.

More Related Content

What's hot (20)

PDF
Deep Learning Primer: A First-Principles Approach
Maurizio Calo Caligaris
 
PDF
Deep Recurrent Neural Networks for Sequence Learning in Spark by Yves Mabiala
Spark Summit
 
PDF
IRJET - Object Detection using Deep Learning with OpenCV and Python
IRJET Journal
 
PPT
Eckovation machine learning project
Vinod Jatav
 
PDF
II-SDV 2017: The Next Era: Deep Learning for Biomedical Research
Dr. Haxel Consult
 
PDF
Recent Trends in Deep Learning
Sungjoon Choi
 
PPTX
Cat and dog classification
omaraldabash
 
PPTX
Daniel Shank, Data Scientist, Talla at MLconf SF 2016
MLconf
 
PDF
Notes from Coursera Deep Learning courses by Andrew Ng
dataHacker. rs
 
PDF
Bhadale group of companies ai neural networks and algorithms catalogue
Vijayananda Mohire
 
PDF
Machine Learning for Everyone
Dhiana Deva
 
PPTX
Deep learning with Tensorflow in R
mikaelhuss
 
PDF
Deep learning - A Visual Introduction
Lukas Masuch
 
PDF
Super Resolution with OCR Optimization
niveditJain
 
PDF
“An Introduction to Data Augmentation Techniques in ML Frameworks,” a Present...
Edge AI and Vision Alliance
 
PPTX
Automating Machine Learning - Is it feasible?
Manuel Martín
 
PDF
Deep Learning: Application Landscape - March 2018
Grigory Sapunov
 
PDF
Data Parallel Deep Learning
inside-BigData.com
 
PPTX
PPT - Deep and Confident Prediction For Time Series at Uber
Jisang Yoon
 
PDF
Image Classification Done Simply using Keras and TensorFlow
Rajiv Shah
 
Deep Learning Primer: A First-Principles Approach
Maurizio Calo Caligaris
 
Deep Recurrent Neural Networks for Sequence Learning in Spark by Yves Mabiala
Spark Summit
 
IRJET - Object Detection using Deep Learning with OpenCV and Python
IRJET Journal
 
Eckovation machine learning project
Vinod Jatav
 
II-SDV 2017: The Next Era: Deep Learning for Biomedical Research
Dr. Haxel Consult
 
Recent Trends in Deep Learning
Sungjoon Choi
 
Cat and dog classification
omaraldabash
 
Daniel Shank, Data Scientist, Talla at MLconf SF 2016
MLconf
 
Notes from Coursera Deep Learning courses by Andrew Ng
dataHacker. rs
 
Bhadale group of companies ai neural networks and algorithms catalogue
Vijayananda Mohire
 
Machine Learning for Everyone
Dhiana Deva
 
Deep learning with Tensorflow in R
mikaelhuss
 
Deep learning - A Visual Introduction
Lukas Masuch
 
Super Resolution with OCR Optimization
niveditJain
 
“An Introduction to Data Augmentation Techniques in ML Frameworks,” a Present...
Edge AI and Vision Alliance
 
Automating Machine Learning - Is it feasible?
Manuel Martín
 
Deep Learning: Application Landscape - March 2018
Grigory Sapunov
 
Data Parallel Deep Learning
inside-BigData.com
 
PPT - Deep and Confident Prediction For Time Series at Uber
Jisang Yoon
 
Image Classification Done Simply using Keras and TensorFlow
Rajiv Shah
 

Similar to Computational decision making (20)

PPTX
Machine Learning Algorithms (Part 1)
Zihui Li
 
PPTX
OBJECTRECOGNITION1.pptxjjjkkkkjjjjkkkkkkk
shesnasuneer
 
PPTX
OBJECTRECOGNITION1.pptxjjjkkkkjjjjkkkkkkk
shesnasuneer
 
PPTX
Introduction to Deep Learning and Tensorflow
Oswald Campesato
 
PPT
Computational Biology, Part 4 Protein Coding Regions
butest
 
DOC
Presentation on Machine Learning and Data Mining
butest
 
PPT
Supervised and unsupervised learning
AmAn Singh
 
PDF
X-TREPAN: A MULTI CLASS REGRESSION AND ADAPTED EXTRACTION OF COMPREHENSIBLE D...
cscpconf
 
PDF
X-TREPAN : A Multi Class Regression and Adapted Extraction of Comprehensible ...
csandit
 
PDF
Classifiers
Ayurdata
 
PDF
A Few Useful Things to Know about Machine Learning
nep_test_account
 
PPTX
Deep learning from a novice perspective
Anirban Santara
 
PDF
AI and Deep Learning
Subrat Panda, PhD
 
PDF
Radial Basis Function Neural Network (RBFNN), Induction Motor, Vector control...
cscpconf
 
PDF
ANALYTICAL STUDY OF FEATURE EXTRACTION TECHNIQUES IN OPINION MINING
csandit
 
PDF
Analytical study of feature extraction techniques in opinion mining
csandit
 
PPTX
Deep learning
Aman Kamboj
 
PPT
notes as .ppt
butest
 
PPTX
Artificial Neural Networks for NIU
Prof. Neeta Awasthy
 
PDF
Machine Learning Comparative Analysis - Part 1
Kaniska Mandal
 
Machine Learning Algorithms (Part 1)
Zihui Li
 
OBJECTRECOGNITION1.pptxjjjkkkkjjjjkkkkkkk
shesnasuneer
 
OBJECTRECOGNITION1.pptxjjjkkkkjjjjkkkkkkk
shesnasuneer
 
Introduction to Deep Learning and Tensorflow
Oswald Campesato
 
Computational Biology, Part 4 Protein Coding Regions
butest
 
Presentation on Machine Learning and Data Mining
butest
 
Supervised and unsupervised learning
AmAn Singh
 
X-TREPAN: A MULTI CLASS REGRESSION AND ADAPTED EXTRACTION OF COMPREHENSIBLE D...
cscpconf
 
X-TREPAN : A Multi Class Regression and Adapted Extraction of Comprehensible ...
csandit
 
Classifiers
Ayurdata
 
A Few Useful Things to Know about Machine Learning
nep_test_account
 
Deep learning from a novice perspective
Anirban Santara
 
AI and Deep Learning
Subrat Panda, PhD
 
Radial Basis Function Neural Network (RBFNN), Induction Motor, Vector control...
cscpconf
 
ANALYTICAL STUDY OF FEATURE EXTRACTION TECHNIQUES IN OPINION MINING
csandit
 
Analytical study of feature extraction techniques in opinion mining
csandit
 
Deep learning
Aman Kamboj
 
notes as .ppt
butest
 
Artificial Neural Networks for NIU
Prof. Neeta Awasthy
 
Machine Learning Comparative Analysis - Part 1
Kaniska Mandal
 
Ad

More from Boris Adryan (20)

PDF
Development and Deployment: The Human Factor
Boris Adryan
 
PDF
Zühlke Meetup - Mai 2017
Boris Adryan
 
PDF
Mehr und schneller ist nicht automatisch besser - data2day, 06.10.16
Boris Adryan
 
PDF
Industry of Things World - Berlin 19-09-16
Boris Adryan
 
PDF
Just because you can doesn't mean that you should - thingmonk 2016
Boris Adryan
 
PDF
Plattformen für das Internet der Dinge, solutions.hamburg, 05.09.16
Boris Adryan
 
PDF
Eclipse IoT - Day 0 of thingmonk 2016
Boris Adryan
 
PDF
Geo-IoT World, 25/05/16
Boris Adryan
 
PDF
Smart IoT London, 13th April 2016
Boris Adryan
 
PDF
Eclipse IoT - ecosystem
Boris Adryan
 
PDF
TopConf Linz, 02/02/2016
Boris Adryan
 
PDF
IoT - Be Open or Miss Out
Boris Adryan
 
PDF
Thingmonk 2015
Boris Adryan
 
PDF
Node-RED and Minecraft - CamJam September 2015
Boris Adryan
 
PDF
EclipseCon France 2015 - Science Track
Boris Adryan
 
PDF
Node-RED workshop at IoT Toulouse
Boris Adryan
 
PDF
Data Science London - Meetup, 28/05/15
Boris Adryan
 
PDF
O'Reilly Webcast: Organizing the Internet of Things - Actionable Insight Thro...
Boris Adryan
 
PDF
An introduction to workflow-based programming with Node-RED
Boris Adryan
 
PDF
What the IoT should learn from the life sciences
Boris Adryan
 
Development and Deployment: The Human Factor
Boris Adryan
 
Zühlke Meetup - Mai 2017
Boris Adryan
 
Mehr und schneller ist nicht automatisch besser - data2day, 06.10.16
Boris Adryan
 
Industry of Things World - Berlin 19-09-16
Boris Adryan
 
Just because you can doesn't mean that you should - thingmonk 2016
Boris Adryan
 
Plattformen für das Internet der Dinge, solutions.hamburg, 05.09.16
Boris Adryan
 
Eclipse IoT - Day 0 of thingmonk 2016
Boris Adryan
 
Geo-IoT World, 25/05/16
Boris Adryan
 
Smart IoT London, 13th April 2016
Boris Adryan
 
Eclipse IoT - ecosystem
Boris Adryan
 
TopConf Linz, 02/02/2016
Boris Adryan
 
IoT - Be Open or Miss Out
Boris Adryan
 
Thingmonk 2015
Boris Adryan
 
Node-RED and Minecraft - CamJam September 2015
Boris Adryan
 
EclipseCon France 2015 - Science Track
Boris Adryan
 
Node-RED workshop at IoT Toulouse
Boris Adryan
 
Data Science London - Meetup, 28/05/15
Boris Adryan
 
O'Reilly Webcast: Organizing the Internet of Things - Actionable Insight Thro...
Boris Adryan
 
An introduction to workflow-based programming with Node-RED
Boris Adryan
 
What the IoT should learn from the life sciences
Boris Adryan
 
Ad

Recently uploaded (20)

PPTX
✨Unleashing Collaboration: Salesforce Channels & Community Power in Patna!✨
SanjeetMishra29
 
PDF
Human-centred design in online workplace learning and relationship to engagem...
Tracy Tang
 
PPTX
Darren Mills The Migration Modernization Balancing Act: Navigating Risks and...
AWS Chicago
 
PDF
DevBcn - Building 10x Organizations Using Modern Productivity Metrics
Justin Reock
 
PDF
SWEBOK Guide and Software Services Engineering Education
Hironori Washizaki
 
PPTX
Webinar: Introduction to LF Energy EVerest
DanBrown980551
 
PDF
Impact of IEEE Computer Society in Advancing Emerging Technologies including ...
Hironori Washizaki
 
PDF
Wojciech Ciemski for Top Cyber News MAGAZINE. June 2025
Dr. Ludmila Morozova-Buss
 
PDF
Windsurf Meetup Ottawa 2025-07-12 - Planning Mode at Reliza.pdf
Pavel Shukhman
 
PDF
Apache CloudStack 201: Let's Design & Build an IaaS Cloud
ShapeBlue
 
PDF
Chris Elwell Woburn, MA - Passionate About IT Innovation
Chris Elwell Woburn, MA
 
PPTX
MSP360 Backup Scheduling and Retention Best Practices.pptx
MSP360
 
PDF
CIFDAQ Token Spotlight for 9th July 2025
CIFDAQ
 
PDF
NewMind AI - Journal 100 Insights After The 100th Issue
NewMind AI
 
PPTX
Building Search Using OpenSearch: Limitations and Workarounds
Sease
 
PDF
Why Orbit Edge Tech is a Top Next JS Development Company in 2025
mahendraalaska08
 
PPTX
WooCommerce Workshop: Bring Your Laptop
Laura Hartwig
 
PDF
Log-Based Anomaly Detection: Enhancing System Reliability with Machine Learning
Mohammed BEKKOUCHE
 
PDF
Blockchain Transactions Explained For Everyone
CIFDAQ
 
PDF
Complete JavaScript Notes: From Basics to Advanced Concepts.pdf
haydendavispro
 
✨Unleashing Collaboration: Salesforce Channels & Community Power in Patna!✨
SanjeetMishra29
 
Human-centred design in online workplace learning and relationship to engagem...
Tracy Tang
 
Darren Mills The Migration Modernization Balancing Act: Navigating Risks and...
AWS Chicago
 
DevBcn - Building 10x Organizations Using Modern Productivity Metrics
Justin Reock
 
SWEBOK Guide and Software Services Engineering Education
Hironori Washizaki
 
Webinar: Introduction to LF Energy EVerest
DanBrown980551
 
Impact of IEEE Computer Society in Advancing Emerging Technologies including ...
Hironori Washizaki
 
Wojciech Ciemski for Top Cyber News MAGAZINE. June 2025
Dr. Ludmila Morozova-Buss
 
Windsurf Meetup Ottawa 2025-07-12 - Planning Mode at Reliza.pdf
Pavel Shukhman
 
Apache CloudStack 201: Let's Design & Build an IaaS Cloud
ShapeBlue
 
Chris Elwell Woburn, MA - Passionate About IT Innovation
Chris Elwell Woburn, MA
 
MSP360 Backup Scheduling and Retention Best Practices.pptx
MSP360
 
CIFDAQ Token Spotlight for 9th July 2025
CIFDAQ
 
NewMind AI - Journal 100 Insights After The 100th Issue
NewMind AI
 
Building Search Using OpenSearch: Limitations and Workarounds
Sease
 
Why Orbit Edge Tech is a Top Next JS Development Company in 2025
mahendraalaska08
 
WooCommerce Workshop: Bring Your Laptop
Laura Hartwig
 
Log-Based Anomaly Detection: Enhancing System Reliability with Machine Learning
Mohammed BEKKOUCHE
 
Blockchain Transactions Explained For Everyone
CIFDAQ
 
Complete JavaScript Notes: From Basics to Advanced Concepts.pdf
haydendavispro
 

Computational decision making

  • 2. What I aim to provide ✓basic vocabulary for ✓fundamental concepts of computational decision making ✓phenomenological introduction to machine learning methods ✓a rough idea when and how to use these methods
  • 3. What this presentation isn’t x hands-on tutorial x thorough summary x comprehensive guide x technical deep dive x statistics course
  • 4. Is this artificial intelligence? word = input(‘Enter a word:’) for key in British_dictionary.iteritems(): if key.startswith(word): print(‘This is a British word.’)
  • 5. Is this machine learning? temperature = float(input('What is the temperature?')) if temperature >= 1.0: print('Wear shorts.') else: print('Wear long underwear.’)
  • 6. Definition Rule-based decision making on the basis of numeric thresholds or string patterns etc is not machine learning. And most definitely it is not artificial intelligence.
  • 7. But what if… …the threshold is inferred at run time? “Write a software that says how close to Euston you can move if you can afford to spend £650k.” 450 550 650 750 850 950 1050 0 3 6 9 12 Average property price Northern Line, number of stops from Euston table = input_table(‘cost at station’) print(where_x_yields_a_low_enough_y)
  • 8. Linear regression is probably the most simple “machine learning” method. Average property price Northern Line, number of stops from Euston It is an example of supervised learning, because we teach the computer the relation between an input variable (“feature”) and an output variable (“label”). 450 550 650 750 850 950 1050 0 3 6 9 12 y = m . x + b
  • 9. Linear regression can become arbitrarily complicated. The difference between curve fitting in statistics and machine learning is mostly semantics. f(number of stops to Euston, square footage, bedrooms, bathrooms, …) price many features Euston High Barnet small large price
  • 10. Classification tasks I. setosa I. virginica I. versicolor All images from https://en.wikipedia.org/wiki/Iris_flower_data_set Rather than projecting the feature vector onto a continuous variable, many supervised learning methods identify “class labels”.
  • 11. Classification tasks Supervised learning requires complete input matrices. Missing or nonsensical values have to be replaced or removed. Non-numerical features (think, e.g. “name of colour”, “smell”) have to be encoded. class label feature 1 feature 2 feature 3 feature 4 1 5.1 3.5 1.4 0.2 1 4.9 3.0 1.4 0.2 2 7.0 3.2 4.7 1.4 3 6.3 3.3 6.0 2.5
  • 12. Classification tasks sepal length sepalwidth I. setosa I. virginica In a first approximation, classification (by regression) aims to find a function that best separates the different class labels. f(sepal width, sepal length) (1,2) “1” “2”
  • 13. sepal length sepalwidth I. setosa I. virginica I. versicolor Decision trees sepal width petalwidth A decision tree can be understood as a series of linear separations of the data. ratio of sepal width : sepal length I. virginica ratio of petal width : sepal width I. setosa I. versicolor
  • 14. Random forests A collection of decision trees, each trained on a random subset of the data, can minimise the risk of over-fitting. A single big decision tree trained on all data can effectively describe a single data point. sepal width petalwidth
  • 15. Over-/Underfitting Sloppy separation is called underfitting, and greedy separation overfitting. Counteracting an overfit is called regularisation, and works by penalising too many features (L1) or too strong feature weights (L2). underfit (high bias) okay overfit (high variance)
  • 16. Dimensionality reduction Dimensionality reduction aims to reduce the complexity of a dataset (in respect to number of features). The first principal components are dimensions that explain most of a dataset’s variance.
  • 17. Support vector machine The SVM aims to provide an ideal separation plane by supporting it with a training data vector. A classification margin protects the SVM against overfitting. ? margin support vectors decision boundary wTx = 0 negative hyperplane wTx = -1 positive hyperplane wTx = 1
  • 18. Kernel trick Input data can be projected into a higher-dimensional space that allows linear separation of otherwise inseparable data. There are different kernels, such as the radial basis function (Gaussian) or polynomial kernel. “2D” input space Φ “3D” feature space http://scikit-learn.org/0.18/auto_examples/svm/plot_iris.html
  • 19. features weather forecast airport location # of gates # of runways # of snowploughs airline aircraft BLACK BOX training flights cancelled in the past classifier ranked list of relevant features weight of features thresholds for features performance metric prediction new data General approach An intuitive example from real life.
  • 20. training classifier performance assessment good enough? success! moredatafortraining data no yes sensitivity “truepositives” 1-specificity “false positives” 0 0.2 0.4 0.6 0.8 1.0 1.0 0.8 0.6 0.4 0.2 worse than random guess Classifier performance Not all machine learning behaves ideal, and performance metrics are important for quality checks and parameter tuning.
  • 21. https://en.wikipedia.org/wiki/Precision_and_recall There is a wide range of performance metrics, comprising combinations of true & false positives as well as true & false negatives. Metrics zoo positive class (P) negative class (N) predicted positive true positive (TP) false positive (FP) predicted negative false negative (FN) true negative (TN)
  • 22. data acquisition model building test use in production data recording (production system) evaluation raw data clean-up feature engineering model learning model selection labour intense compute intensebrain intense development production ML pipeline
  • 23. Choosing a method from: Olson et al., 2017, https://arxiv.org/abs/1708.05070 There is no ‘one-size-fits-all’ machine learning method. Most methods need to be carefully tuned to perform ideal. Often, there a ‘non-functional’ constraints on choosing a method. Runtime, interpretability, etc.
  • 24. What about neural networks? feature 1 feature 2 feature 3 weight 1 weight 2 weight 3 input function activation function class output error for weight updates a simple perceptron Neural networks attempt to mimic the integrative properties of neurons. The perceptron is a single-layer network. inputs outputs
  • 25. Deep neural networks http://www.asimovinstitute.org/neural-network-zoo While many artificial neural networks show great performance, the basis on which features exactly the classification works remains largely unknown.
  • 26. In RL, the methods iteratively learn to optimise an output from an abstract representation of a system Reinforcement learning move (l/r) or shoot unknown systemmap (210 x 160 pixels, 8-bit RGB) actual score choose action on basis of map to optimise score Mnih et al., Nature (2015)
  • 27. Machine learning can help to structure and explore an unknown dataset. These methods aid the identification of classes where their existence isn’t known yet. Unsupervised learning • Hierarchical clustering • K-means clustering • Expectation maximisation • Density-based clustering plus clever visualisation
  • 28. Hierarchical clustering Derives hierarchical dependencies of individual rows and columns in the dataset on the basis of similarity (correlation) between their properties. Combined with a heatmap, gives a good first impression of a dataset.
  • 29. k-means clustering Defines k different centroids to which data points are assigned by proximity. If the distance doesn’t get much smaller, k is the number of clusters in the set. Density-based clustering and expectation maximisation are conceptually related, the latter giving a probability for membership in any group. k = 2 k vs centroid distance
  • 30. Conclusions https://badryan.github.io/2015/10/20/is-it-all-machine-learning.html •People on the Internet steal infographics. •ML methods have been around in the stats world for ages, but big data sets and compute power make them more widely known. •Understanding key principles behind ML should be part of the school curriculum.