SlideShare a Scribd company logo
Object Classification/
Recognition
BASIC IDEA
2
Classifier
model
CAR
INPUT IMAGE
Two ways
1. Using
tensorflow API
2. Basic Classifier
from scratch
3
PLAN
01 Dataset
02 Import
Libraries
03
Build the CNN
Model
04
Train and
validate
4
Dataset
1
5
About Dataset
Dataset: Airplane and car
images
Source : Kaggle.com
Number of images: 1000
6
Import Libraries
2
7
Libraries, Tools and Technologies used.
8
● Cv2
● Spyder
● Anaconda
● NumPy
● Sklearn
● Matplotlib
● Keras
● Split the data into two parts using sklearn
- Training data and validation data
● Stored the resized images in array X and the label for it in array y.
- Resize to uniform dimension 150 by 150 for height and width and 3
channels(i.e., for RGB)
- Array Y will store label 0 for CAR and 1 for Airplane.
- X is now an array of image pixel values
- Y is a list of corresponding labels.
9
Data preprocessing
Most awaited
Model
3
10
CNN
▪ Convolutional neural network to train the model.
11
But, What is
convolutional
neural network?
● It’s deep, feed-forward artificial neural
network.(FFNN are also c/as Multilayer perceptrons).
● made up of neurons that have learnable weights and
biases.
● Each neuron receives some inputs, performs a dot
product.
● unlike a regular Neural Network, the layers of a
ConvNet have neurons arranged in 3 dimensions:
width, height, depth.
12
A visualization is worth a thousand words
Activation map or feature map
CNN
▪ Convolutional layer acts as a feature extractor.
▪ To extract features like edges, corners etc.,
13
Network Architecture
14
Activation Function
ReLU - Rectified linear units(non-linear activation
function)
A(x) = max(0,x)
Sigmoid Function suffers from vanishing Gradient problem.
❏ Sparsity - efficient in computations,
❏ more intuitive
❏ independent
❏ No complicated math.
PS : Find notes and important links with click to add notes section of the ppt.15
BUT WHY ONLY ReLU?
Pooling layer
▪ Using all features in classification is computationally hard when the image
size is large.
▪ For example 224x224, 480x480 size images.
▪ Prone to overfitting.
▪ Pooling (also called subsampling or downsampling) reduces the
dimensionality of each feature map.
▪ But retains the most important information
▪ Max pooling
16
Flatten layer and
Dropout Layer
▪ Flatten Layer - Convert 2D array to single linear vector.
▪ Dropout layer -
▫ Dropout randomly drops some layers in a neural
networks and then learns with the reduced network.
▫ This way, the network learns to be independent and not
reliable on a single layer. Bottom-line is that it helps in
overfitting.
▫ 0.5 means to randomly drop half of the layers.
17
Model.summary()
Formula to get number
of parameters
=
(((kernel size *
stride)*channels)+1)
*filters)
= (((3*3*1)*3))+1)*32
=((9*3)+1)*32
=28*32
= 896
Visualization of layers
Use big image
19
Model Fitting using
Back Propagation
▪ Back propagate errors(error calculation with loss function)
▪ Loss optimization (optimizer-update weights)
▪ Calculate gradient.
▪ With Learning rate 0.001 update the weight.
20
Positive Gradient Decrease weight
Negative Gradient Increase weight
Optimizer and
entropy
Entropy - binary_crossentropy since there are two classes (Loss
function)
Optimizer - RMSprop with a learning rate 0.0001.
▪ Minimize the loss incurred using RMSprop optimizer.
21
22
23
Snapshot of epochs
Number of Epochs=10
Training set has 768
images with batch
size of 32.
Hence Number of
iterations to complete
one epoch = 768/32=
24
24
When Softmax was used for final fully connected layer
Using Tensorflow
API
▪ Install object detection API
▪ Download pre-trained object detection model(COCO)
▪ Load images to be classified.
▪ Apply object detection on the load images.
▪ Based on the pre-trained model COCO dataset it tries
to predict the object in the image
25
Future Scope
Transfer Learning
http://cs231n.stanford.edu/repor
ts/2017/pdfs/411.pdf
By - Yangyang Yu, Olivier Jin,
Daniel Hsu , Stanford University
TL (Transfer learning) is a
popular training technique
used in deep learning;
where models that have
been trained for a task are
reused as base/starting
point for another model.
26
References
▪ Presentation template by SlidesCarnival
▪ https://towardsdatascience.com/a-guide-to-an-efficient-way-to-build-neural-network-
architectures-part-ii-hyper-parameter-42efca01e5d7
▪ https://towardsdatascience.com/activation-functions-and-its-types-which-is-better-
a9a5310cc8f
▪ https://www.tensorflow.org/tutorials/keras/basic_classification
▪ https://www.google.com/search?q=convolved+gif&rlz=1C1RLNS_enIN787IN787&tbm=isch&s
ource=iu&ictx=1&fir=1VHtR2R2SFEVMM%253A%252CXjlYoIPax9NDkM%252C_&vet=1&us
g=AI4_-kQpnIDbwPbGc-
UEQVXxHxuXIZ_CRA&sa=X&ved=2ahUKEwilyfryn73hAhXkmuYKHT22A38Q9QEwB3oECA
YQEg#imgrc=1VHtR2R2SFEVMM:
27
THANK YOU.
28
KADAMBINI INDURKAR
(BT15CSE035)

More Related Content

PPTX
Real Time Object Dectection using machine learning
pratik pratyay
 
PDF
GTC2019 NavInfo Europe Session
Hong Wang (Suzy)
 
PPTX
Practical Reinforcement Learning with TensorFlow
Illia Polosukhin
 
PDF
GTC Japan 2016 Chainer feature introduction
Kenta Oono
 
PDF
Introduction to Chainer
Seiya Tokui
 
PPTX
Deep Learning with Apache Spark: an Introduction
Emanuele Bezzi
 
PDF
Kaggle Lyft Motion Prediction for Autonomous Vehicles 4th Place Solution
Preferred Networks
 
PDF
NVIDIA 深度學習教育機構 (DLI): Image segmentation with tensorflow
NVIDIA Taiwan
 
Real Time Object Dectection using machine learning
pratik pratyay
 
GTC2019 NavInfo Europe Session
Hong Wang (Suzy)
 
Practical Reinforcement Learning with TensorFlow
Illia Polosukhin
 
GTC Japan 2016 Chainer feature introduction
Kenta Oono
 
Introduction to Chainer
Seiya Tokui
 
Deep Learning with Apache Spark: an Introduction
Emanuele Bezzi
 
Kaggle Lyft Motion Prediction for Autonomous Vehicles 4th Place Solution
Preferred Networks
 
NVIDIA 深度學習教育機構 (DLI): Image segmentation with tensorflow
NVIDIA Taiwan
 

What's hot (20)

PPT
Monte Carlo on GPUs
fcassier
 
PDF
Introduction to Chainer 11 may,2018
Preferred Networks
 
PDF
CUDA and Caffe for deep learning
Amgad Muhammad
 
PDF
IIBMP2019 講演資料「オープンソースで始める深層学習」
Preferred Networks
 
PPTX
Deep parking
Shintaro Shiba
 
PDF
K-Fashion 경진대회 3등 수상자 솔루션
DACON AI 데이콘
 
PDF
Common Design of Deep Learning Frameworks
Kenta Oono
 
PDF
Chainer ui v0.3 and imagereport
Preferred Networks
 
PPTX
cuTau Leaping
Amritesh Srivastava
 
PDF
Parallel Application Performance Prediction of Using Analysis Based Modeling
Jason Liu
 
PPTX
Ultrasound Nerve Segmentation
Sneha Ravikumar
 
PPTX
TPU paper slide
Dong-Hyun Hwang
 
PPTX
Chainer v3
Seiya Tokui
 
PDF
Distributed Multi-GPU Computing with Dask, CuPy and RAPIDS
PeterAndreasEntschev
 
PDF
Asynchronous Methods for Deep Reinforcement Learning
Ssu-Rui Lee
 
PDF
Learning stochastic neural networks with Chainer
Seiya Tokui
 
PDF
Efficient execution of quantized deep learning models a compiler approach
jemin lee
 
PPTX
Tianqi Chen, PhD Student, University of Washington, at MLconf Seattle 2017
MLconf
 
PDF
Profiling PyTorch for Efficiency & Sustainability
geetachauhan
 
PDF
MLPerf an industry standard benchmark suite for machine learning performance
jemin lee
 
Monte Carlo on GPUs
fcassier
 
Introduction to Chainer 11 may,2018
Preferred Networks
 
CUDA and Caffe for deep learning
Amgad Muhammad
 
IIBMP2019 講演資料「オープンソースで始める深層学習」
Preferred Networks
 
Deep parking
Shintaro Shiba
 
K-Fashion 경진대회 3등 수상자 솔루션
DACON AI 데이콘
 
Common Design of Deep Learning Frameworks
Kenta Oono
 
Chainer ui v0.3 and imagereport
Preferred Networks
 
cuTau Leaping
Amritesh Srivastava
 
Parallel Application Performance Prediction of Using Analysis Based Modeling
Jason Liu
 
Ultrasound Nerve Segmentation
Sneha Ravikumar
 
TPU paper slide
Dong-Hyun Hwang
 
Chainer v3
Seiya Tokui
 
Distributed Multi-GPU Computing with Dask, CuPy and RAPIDS
PeterAndreasEntschev
 
Asynchronous Methods for Deep Reinforcement Learning
Ssu-Rui Lee
 
Learning stochastic neural networks with Chainer
Seiya Tokui
 
Efficient execution of quantized deep learning models a compiler approach
jemin lee
 
Tianqi Chen, PhD Student, University of Washington, at MLconf Seattle 2017
MLconf
 
Profiling PyTorch for Efficiency & Sustainability
geetachauhan
 
MLPerf an industry standard benchmark suite for machine learning performance
jemin lee
 
Ad

Similar to Cv mini project (1) (20)

PDF
Eye deep
sveitser
 
PPTX
Deep learning requirement and notes for novoice
AmmarAhmedSiddiqui2
 
PPTX
High Performance Pedestrian Detection On TEGRA X1
NVIDIA
 
PDF
Lecture 5: Convolutional Neural Network Models
Mohamed Loey
 
PDF
dfdshofdifhdifhdfhgfoighfgofgfgfgfgdfdfdfdf
nguyenhoangy207
 
PPTX
Reducing the dimensionality of data with neural networks
Hakky St
 
PPTX
Nuts and Bolts of Transfer Learning.pptx
vmanjusundertamil21
 
PDF
AI optimizing HPC simulations (presentation from 6th EULAG Workshop)
byteLAKE
 
PDF
Convolutional Neural Network Models - Deep Learning
Mohamed Loey
 
PPTX
B.tech_project_ppt.pptx
supratikmondal6
 
PDF
Introduction to Applied Machine Learning
SheilaJimenezMorejon
 
PPTX
presentation of IntroductionDeepLearning.pptx
andani26
 
PDF
Power ai tensorflowworkloadtutorial-20171117
Ganesan Narayanasamy
 
PPTX
Image Classification using deep learning
Asma-AH
 
PDF
AIML4 CNN lab256 1hr (111-1).pdf
ssuserb4d806
 
PDF
深度學習在AOI的應用
CHENHuiMei
 
PPTX
Computer Vision for Beginners
Sanghamitra Deb
 
PPTX
IMAGE CLASSIFICATION USING CONVOLUTIONAL NEURAL NETWORK.P.pptx
usmanch7829
 
PPTX
Waste Classification System using Convolutional Neural Networks.pptx
JohnPrasad14
 
PDF
Point cloud mesh-investigation_report-lihang
Lihang Li
 
Eye deep
sveitser
 
Deep learning requirement and notes for novoice
AmmarAhmedSiddiqui2
 
High Performance Pedestrian Detection On TEGRA X1
NVIDIA
 
Lecture 5: Convolutional Neural Network Models
Mohamed Loey
 
dfdshofdifhdifhdfhgfoighfgofgfgfgfgdfdfdfdf
nguyenhoangy207
 
Reducing the dimensionality of data with neural networks
Hakky St
 
Nuts and Bolts of Transfer Learning.pptx
vmanjusundertamil21
 
AI optimizing HPC simulations (presentation from 6th EULAG Workshop)
byteLAKE
 
Convolutional Neural Network Models - Deep Learning
Mohamed Loey
 
B.tech_project_ppt.pptx
supratikmondal6
 
Introduction to Applied Machine Learning
SheilaJimenezMorejon
 
presentation of IntroductionDeepLearning.pptx
andani26
 
Power ai tensorflowworkloadtutorial-20171117
Ganesan Narayanasamy
 
Image Classification using deep learning
Asma-AH
 
AIML4 CNN lab256 1hr (111-1).pdf
ssuserb4d806
 
深度學習在AOI的應用
CHENHuiMei
 
Computer Vision for Beginners
Sanghamitra Deb
 
IMAGE CLASSIFICATION USING CONVOLUTIONAL NEURAL NETWORK.P.pptx
usmanch7829
 
Waste Classification System using Convolutional Neural Networks.pptx
JohnPrasad14
 
Point cloud mesh-investigation_report-lihang
Lihang Li
 
Ad

Recently uploaded (20)

PPTX
MULTI LEVEL DATA TRACKING USING COOJA.pptx
dollysharma12ab
 
PPTX
Information Retrieval and Extraction - Module 7
premSankar19
 
PDF
The Effect of Artifact Removal from EEG Signals on the Detection of Epileptic...
Partho Prosad
 
PDF
Packaging Tips for Stainless Steel Tubes and Pipes
heavymetalsandtubes
 
PDF
67243-Cooling and Heating & Calculation.pdf
DHAKA POLYTECHNIC
 
PPTX
business incubation centre aaaaaaaaaaaaaa
hodeeesite4
 
PDF
top-5-use-cases-for-splunk-security-analytics.pdf
yaghutialireza
 
PPTX
sunil mishra pptmmmmmmmmmmmmmmmmmmmmmmmmm
singhamit111
 
PDF
CAD-CAM U-1 Combined Notes_57761226_2025_04_22_14_40.pdf
shailendrapratap2002
 
PDF
AI-Driven IoT-Enabled UAV Inspection Framework for Predictive Maintenance and...
ijcncjournal019
 
PPTX
MT Chapter 1.pptx- Magnetic particle testing
ABCAnyBodyCanRelax
 
PDF
2025 Laurence Sigler - Advancing Decision Support. Content Management Ecommer...
Francisco Javier Mora Serrano
 
PDF
Construction of a Thermal Vacuum Chamber for Environment Test of Triple CubeS...
2208441
 
PDF
2010_Book_EnvironmentalBioengineering (1).pdf
EmilianoRodriguezTll
 
PPTX
quantum computing transition from classical mechanics.pptx
gvlbcy
 
PDF
STUDY OF NOVEL CHANNEL MATERIALS USING III-V COMPOUNDS WITH VARIOUS GATE DIEL...
ijoejnl
 
PDF
Zero Carbon Building Performance standard
BassemOsman1
 
PDF
20ME702-Mechatronics-UNIT-1,UNIT-2,UNIT-3,UNIT-4,UNIT-5, 2025-2026
Mohanumar S
 
PPTX
Civil Engineering Practices_BY Sh.JP Mishra 23.09.pptx
bineetmishra1990
 
PDF
EVS+PRESENTATIONS EVS+PRESENTATIONS like
saiyedaqib429
 
MULTI LEVEL DATA TRACKING USING COOJA.pptx
dollysharma12ab
 
Information Retrieval and Extraction - Module 7
premSankar19
 
The Effect of Artifact Removal from EEG Signals on the Detection of Epileptic...
Partho Prosad
 
Packaging Tips for Stainless Steel Tubes and Pipes
heavymetalsandtubes
 
67243-Cooling and Heating & Calculation.pdf
DHAKA POLYTECHNIC
 
business incubation centre aaaaaaaaaaaaaa
hodeeesite4
 
top-5-use-cases-for-splunk-security-analytics.pdf
yaghutialireza
 
sunil mishra pptmmmmmmmmmmmmmmmmmmmmmmmmm
singhamit111
 
CAD-CAM U-1 Combined Notes_57761226_2025_04_22_14_40.pdf
shailendrapratap2002
 
AI-Driven IoT-Enabled UAV Inspection Framework for Predictive Maintenance and...
ijcncjournal019
 
MT Chapter 1.pptx- Magnetic particle testing
ABCAnyBodyCanRelax
 
2025 Laurence Sigler - Advancing Decision Support. Content Management Ecommer...
Francisco Javier Mora Serrano
 
Construction of a Thermal Vacuum Chamber for Environment Test of Triple CubeS...
2208441
 
2010_Book_EnvironmentalBioengineering (1).pdf
EmilianoRodriguezTll
 
quantum computing transition from classical mechanics.pptx
gvlbcy
 
STUDY OF NOVEL CHANNEL MATERIALS USING III-V COMPOUNDS WITH VARIOUS GATE DIEL...
ijoejnl
 
Zero Carbon Building Performance standard
BassemOsman1
 
20ME702-Mechatronics-UNIT-1,UNIT-2,UNIT-3,UNIT-4,UNIT-5, 2025-2026
Mohanumar S
 
Civil Engineering Practices_BY Sh.JP Mishra 23.09.pptx
bineetmishra1990
 
EVS+PRESENTATIONS EVS+PRESENTATIONS like
saiyedaqib429
 

Cv mini project (1)

  • 3. Two ways 1. Using tensorflow API 2. Basic Classifier from scratch 3
  • 4. PLAN 01 Dataset 02 Import Libraries 03 Build the CNN Model 04 Train and validate 4
  • 6. About Dataset Dataset: Airplane and car images Source : Kaggle.com Number of images: 1000 6
  • 8. Libraries, Tools and Technologies used. 8 ● Cv2 ● Spyder ● Anaconda ● NumPy ● Sklearn ● Matplotlib ● Keras
  • 9. ● Split the data into two parts using sklearn - Training data and validation data ● Stored the resized images in array X and the label for it in array y. - Resize to uniform dimension 150 by 150 for height and width and 3 channels(i.e., for RGB) - Array Y will store label 0 for CAR and 1 for Airplane. - X is now an array of image pixel values - Y is a list of corresponding labels. 9 Data preprocessing
  • 11. CNN ▪ Convolutional neural network to train the model. 11 But, What is convolutional neural network? ● It’s deep, feed-forward artificial neural network.(FFNN are also c/as Multilayer perceptrons). ● made up of neurons that have learnable weights and biases. ● Each neuron receives some inputs, performs a dot product. ● unlike a regular Neural Network, the layers of a ConvNet have neurons arranged in 3 dimensions: width, height, depth.
  • 12. 12 A visualization is worth a thousand words Activation map or feature map
  • 13. CNN ▪ Convolutional layer acts as a feature extractor. ▪ To extract features like edges, corners etc., 13
  • 15. Activation Function ReLU - Rectified linear units(non-linear activation function) A(x) = max(0,x) Sigmoid Function suffers from vanishing Gradient problem. ❏ Sparsity - efficient in computations, ❏ more intuitive ❏ independent ❏ No complicated math. PS : Find notes and important links with click to add notes section of the ppt.15 BUT WHY ONLY ReLU?
  • 16. Pooling layer ▪ Using all features in classification is computationally hard when the image size is large. ▪ For example 224x224, 480x480 size images. ▪ Prone to overfitting. ▪ Pooling (also called subsampling or downsampling) reduces the dimensionality of each feature map. ▪ But retains the most important information ▪ Max pooling 16
  • 17. Flatten layer and Dropout Layer ▪ Flatten Layer - Convert 2D array to single linear vector. ▪ Dropout layer - ▫ Dropout randomly drops some layers in a neural networks and then learns with the reduced network. ▫ This way, the network learns to be independent and not reliable on a single layer. Bottom-line is that it helps in overfitting. ▫ 0.5 means to randomly drop half of the layers. 17
  • 18. Model.summary() Formula to get number of parameters = (((kernel size * stride)*channels)+1) *filters) = (((3*3*1)*3))+1)*32 =((9*3)+1)*32 =28*32 = 896
  • 20. Model Fitting using Back Propagation ▪ Back propagate errors(error calculation with loss function) ▪ Loss optimization (optimizer-update weights) ▪ Calculate gradient. ▪ With Learning rate 0.001 update the weight. 20 Positive Gradient Decrease weight Negative Gradient Increase weight
  • 21. Optimizer and entropy Entropy - binary_crossentropy since there are two classes (Loss function) Optimizer - RMSprop with a learning rate 0.0001. ▪ Minimize the loss incurred using RMSprop optimizer. 21
  • 22. 22
  • 23. 23 Snapshot of epochs Number of Epochs=10 Training set has 768 images with batch size of 32. Hence Number of iterations to complete one epoch = 768/32= 24
  • 24. 24 When Softmax was used for final fully connected layer
  • 25. Using Tensorflow API ▪ Install object detection API ▪ Download pre-trained object detection model(COCO) ▪ Load images to be classified. ▪ Apply object detection on the load images. ▪ Based on the pre-trained model COCO dataset it tries to predict the object in the image 25
  • 26. Future Scope Transfer Learning http://cs231n.stanford.edu/repor ts/2017/pdfs/411.pdf By - Yangyang Yu, Olivier Jin, Daniel Hsu , Stanford University TL (Transfer learning) is a popular training technique used in deep learning; where models that have been trained for a task are reused as base/starting point for another model. 26
  • 27. References ▪ Presentation template by SlidesCarnival ▪ https://towardsdatascience.com/a-guide-to-an-efficient-way-to-build-neural-network- architectures-part-ii-hyper-parameter-42efca01e5d7 ▪ https://towardsdatascience.com/activation-functions-and-its-types-which-is-better- a9a5310cc8f ▪ https://www.tensorflow.org/tutorials/keras/basic_classification ▪ https://www.google.com/search?q=convolved+gif&rlz=1C1RLNS_enIN787IN787&tbm=isch&s ource=iu&ictx=1&fir=1VHtR2R2SFEVMM%253A%252CXjlYoIPax9NDkM%252C_&vet=1&us g=AI4_-kQpnIDbwPbGc- UEQVXxHxuXIZ_CRA&sa=X&ved=2ahUKEwilyfryn73hAhXkmuYKHT22A38Q9QEwB3oECA YQEg#imgrc=1VHtR2R2SFEVMM: 27

Editor's Notes

  • #13: Where pooling layer reduces the resolution and fc layer acts as classifier which o/p image size = (imagesize + 2*padding - kernel size)/stride)+1 Stride,padding etc.. Smaller the filter size greater local information captured Receptive field technical jargons
  • #14: CNNs, like neural networks, are made up of neurons with learnable weights and biases. Each neuron receives several inputs, takes a weighted sum over them, pass it through an activation function and responds with an output. The whole network has a loss function and all the tips and tricks that we developed for neural networks still apply on CNNs. https://medium.com/technologymadeeasy/the-best-explanation-of-convolutional-neural-networks-on-the-internet-fbb8b1ad5df8
  • #15: https://towardsdatascience.com/a-guide-to-an-efficient-way-to-build-neural-network-architectures-part-ii-hyper-parameter-42efca01e5d7 Activation function to decide which activations to keep and which to be discarded….in simple terms layers will understand their language so we apply some activation on feature map When we do not have the activation function the weights and bias would simply do a linear transformation. A linear equation is simple to solve but is limited in its capacity to solve complex problems. A neural network without an activation function is essentially just a linear regression model. The activation function does the non-linear transformation to the input making it capable to learn and perform more complex tasks. Wide and shallow and narrow and deep https://towardsdatascience.com/activation-functions-and-its-types-which-is-better-a9a5310cc8f Non-linearity adapt with variety of data and to differentiate between the output. https://towardsdatascience.com/activation-functions-neural-networks-1cbd9f8d91d6
  • #16: The problem is that in some cases, the gradient will be vanishingly small, effectively preventing the weight from changing its value. In the worst case, this may completely stop the neural network from further training. https://towardsdatascience.com/activation-functions-and-its-types-which-is-better-a9a5310cc8f https://www.quora.com/What-is-special-about-rectifier-neural-units-used-in-NN-learning
  • #17: https://medium.com/nybles/create-your-first-image-recognition-classifier-using-cnn-keras-and-tensorflow-backend-6eaab98d14dd https://medium.com/technologymadeeasy/the-best-explanation-of-convolutional-neural-networks-on-the-internet-fbb8b1ad5df8 Overfitting - when our model memorizes the training data. The model will perform excellently at training time but fail at test time. allowing for assumptions to be made about features contained
  • #18: In such methods, each of the neural network's weights receives an update proportional to the partial derivative of the error function with respect to the current weight in each iteration of training.
  • #19: Formula to get number of parameters = kernel size * stride)*channels)+1)*filters
  • #21: https://medium.com/datathings/neural-networks-and-backpropagation-explained-in-a-simple-way-f540a3611f5eThis is very similar to genetic algorithms where after each generation we apply a small mutation rate and the fittest survives.
  • #24: https://stats.stackexchange.com/questions/233658/softmax-vs-sigmoid-function-in-logistic-classifier