SlideShare a Scribd company logo
2
Most read
11
Most read
24
Most read
Autoencoders
By
Harish.R
Introduction
Autoencoders are a specific type of feed
forward neural networks where the input is the
same as the output. They compress the input into
a lower-dimensional code and then reconstruct
the output from this representation. The code is
a compact “summary” or “compression” of the
input, also called the latent-space representation.
Autoencoder
Components of Autoencoder
An autoencoder consists of 3 components:
Encoder
Code
Decoder
Encoder compresses the input and produces
the code.
Decoder then reconstructs the input only
using this code.
Properties of Autoencoder
Data-specific
Lossy
Unsupervised
Properties of Autoencoder
Data-specific: Autoencoders are only able to
meaningfully compress data similar to what they
have been trained on.
Lossy: The output of the autoencoder will not be
exactly the same as the input, it will be a close
but degraded representation.
Unsupervised: Autoencoders are considered an
unsupervised learning technique since they don’t
need explicit labels to train on. But to be more
precise they are self-supervised because they
generate their own labels from the training data.
Architecture
Architecture
Both the encoder and decoder are fully-
connected feed forward neural networks.
The number of nodes in the code layer (code
size) is a hyper parameter that we set before
training the autoencoder.
Hyperparameters
There are 4 hyperparameters that we need to
set before training an autoencoder:
Code size: number of nodes in the middle
layer. Smaller size results in more
compression.
Number of layers: the autoencoder can be as
deep as we like. In the figure above we have 2
layers in both the encoder and decoder,
without considering the input and output.
Hyperparameters
Number of nodes per layer : The layers are
stacked one after another. The number of nodes
per layer decreases with each subsequent layer of
the encoder, and increases back in the decoder.
Loss function: We either use mean squared error
(mse) or binary crossentropy. If the input values
are in the range [0, 1] then we typically use
crossentropy, otherwise we use the mean
squared error.
Types of Autoencoder :
The Four Types of Autoencoders are :-
Vanilla autoencoder
Multilayer autoencoder
Convolutional autoencoder
Regularized autoencoder
Vanilla autoencoder
The autoencoder is a three layers of a neural
net with one hidden layer. The input and
output are the same, and we learn how to
reconstruct the input.
Here, we see that we have an undercomplete
autoencoder as the hidden layer dimension
(64) is smaller than the input (784). This
constraint will impose our neural net to learn
a compressed representation of data.
Multilayer autoencoder
If one hidden layer is not enough, we can
obviously extend the autoencoder to more
hidden layers.
Now our implementation uses 3 hidden layers
instead of just one. Any of the hidden layers
can be picked as the feature representation
but we will make the network symmetrical
and use the middle-most layer.
Convolution autoencoder
We may also ask ourselves: can autoencoders be
used with Convolutions instead of Fully-
connected layers
The answer is yes and the principle is the same,
but using images (3D vectors) instead of flattened
1D vectors. The input image is downsampled to
give a latent representation of smaller
dimensions and force the autoencoder to learn a
compressed version of the images.
Regularized autoencoder
There are other ways we can constraint the
reconstruction of an autoencoder than to
impose a hidden layer of smaller dimension
than the input. Rather than limiting the model
capacity by keeping the encoder and decoder
shallow and the code size
small, regularized autoencoders use a loss
function that encourages the model to have
other properties besides the ability to copy its
input to its output.
Visualization
Challenges
They are indeed pretty similar, but not exactly the
same. We can notice it more clearly in the last
digit “4”. Since this was a simple task our
autoencoder performed pretty well.
We have total control over the architecture of the
autoencoder. We can make it very powerful by
increasing the number of layers, nodes per layer
and most importantly the code size. Increasing
these hyperparameters will let the autoencoder
to learn more complex codings.
Challenges
But we should be careful to not make it too
powerful.
Over-fitting
The autoencoder will reconstruct the training
data perfectly, but it will be over fitting
without being able to generalize to new
instances, which is not what we want.
 Undercomplete
The autoencoder is said to be undercomplete.
It won’t be able to directly copy its inputs to
the output, and will be forced to learn
intelligent features. then an undercomplete
autoencoder won’t be able to recover it
perfectly.
Regularization
We would like to learn meaningful features
without altering the code’s dimensions
(Overcomplete or Undercomplete).
We usually find two types of regularized
autoencoder:
 Sparse Autoencoder
 Denoising Autoencoder
Sparse autoencoder
Sparse autoencoders are typically used to
learn features for another task such as
classification. An autoencoder that has been
regularized to be sparse must respond to
unique statistical features of the dataset it has
been trained on, rather than simply acting as
an identity function. In this way, training to
perform the copying task with a sparsity
penalty can yield a model that has learned
useful features as a by product.
Denoising Autoencoders
This way the autoencoder can’t simply
copy the input to its output because the input
also contains random noise. We are asking it
to subtract the noise and produce the
underlying meaningful data.
Visualization
Conclusion
Autoencoders are a very useful
dimensionality reduction technique. They are
very popular as a teaching material in
introductory deep learning courses, most
likely due to their simplicity.

More Related Content

What's hot (20)

ODP
Simple Introduction to AutoEncoder
Jun Lang
 
PPTX
Deep Learning With Neural Networks
Aniket Maurya
 
PPTX
An overview of gradient descent optimization algorithms
Hakky St
 
PDF
Deep Feed Forward Neural Networks and Regularization
Yan Xu
 
PDF
Naive Bayes
CloudxLab
 
PPT
Neural Networks
NikitaRuhela
 
PDF
Convolutional Neural Networks (CNN)
Gaurav Mittal
 
PDF
Deep Learning - Overview of my work II
Mohamed Loey
 
PPTX
Object detection with deep learning
Sushant Shrivastava
 
PPT
Back propagation
Nagarajan
 
PPTX
Feedforward neural network
Sopheaktra YONG
 
PPT
Artificial Neural Networks - ANN
Mohamed Talaat
 
PPTX
Feed forward ,back propagation,gradient descent
Muhammad Rasel
 
PDF
Transfer Learning
Hichem Felouat
 
PPT
Perceptron
Nagarajan
 
PPTX
Hough Transform By Md.Nazmul Islam
Nazmul Islam
 
PPTX
Activation functions
PRATEEK SAHU
 
PDF
Resnet
ashwinjoseph95
 
PPTX
Hyperparameter Tuning
Jon Lederman
 
Simple Introduction to AutoEncoder
Jun Lang
 
Deep Learning With Neural Networks
Aniket Maurya
 
An overview of gradient descent optimization algorithms
Hakky St
 
Deep Feed Forward Neural Networks and Regularization
Yan Xu
 
Naive Bayes
CloudxLab
 
Neural Networks
NikitaRuhela
 
Convolutional Neural Networks (CNN)
Gaurav Mittal
 
Deep Learning - Overview of my work II
Mohamed Loey
 
Object detection with deep learning
Sushant Shrivastava
 
Back propagation
Nagarajan
 
Feedforward neural network
Sopheaktra YONG
 
Artificial Neural Networks - ANN
Mohamed Talaat
 
Feed forward ,back propagation,gradient descent
Muhammad Rasel
 
Transfer Learning
Hichem Felouat
 
Perceptron
Nagarajan
 
Hough Transform By Md.Nazmul Islam
Nazmul Islam
 
Activation functions
PRATEEK SAHU
 
Hyperparameter Tuning
Jon Lederman
 

Similar to Autoencoder (20)

PDF
UNIT-4.pdf
NiharikaThakur32
 
PDF
UNIT-4.pdf
NiharikaThakur32
 
PDF
Autoencoder in Deep Learning and its types
DivyaMeenaS
 
PPTX
Introduction to Autoencoders: Types and Applications
Amr Rashed
 
PDF
Explanation of Autoencoder to Variontal Auto Encoder
seshathirid
 
PDF
Lecture 7-8 From Autoencoder to VAE.pdf
EmadAbdelkader5
 
PPTX
Lec16 - Autoencoders.pptx
Sameer Gulshan
 
PPTX
DL-unite4-Autoencoders.pptx..............
kirankamblecoin03
 
PPTX
Lecture 7-8 From Autoencoder to VAE.pptx
yosrghozzi2023
 
PPTX
Autoencoders in Computer Vision: A Deep Learning Approach for Image Denoising...
ShubhamMittal569818
 
PPTX
A Comprehensive Overview of Encoder and Decoder Architectures in Deep Learnin...
ShubhamMittal569818
 
PPTX
AUTO ENCODERS (Deep Learning fundamentals)
aayanshsingh0401
 
PPTX
Autoencoders for image_classification
Cenk Bircanoğlu
 
PPTX
Piotr Mirowski - Review Autoencoders (Deep Learning) - CIUUK14
Daniel Lewis
 
PDF
Autoencoder
Wataru Hirota
 
PPTX
Autoecoders.pptx
MirzaJahanzeb5
 
PDF
Autoencoders
Ashok Govindarajan
 
PPTX
Understanding Autoencoder (Deep Learning Book, Chapter 14)
Entrepreneur / Startup
 
PDF
Structure learning with Deep Neural Networks
Patrick Michl
 
PDF
Structure learning with Deep Neural Networks
Patrick Michl
 
UNIT-4.pdf
NiharikaThakur32
 
UNIT-4.pdf
NiharikaThakur32
 
Autoencoder in Deep Learning and its types
DivyaMeenaS
 
Introduction to Autoencoders: Types and Applications
Amr Rashed
 
Explanation of Autoencoder to Variontal Auto Encoder
seshathirid
 
Lecture 7-8 From Autoencoder to VAE.pdf
EmadAbdelkader5
 
Lec16 - Autoencoders.pptx
Sameer Gulshan
 
DL-unite4-Autoencoders.pptx..............
kirankamblecoin03
 
Lecture 7-8 From Autoencoder to VAE.pptx
yosrghozzi2023
 
Autoencoders in Computer Vision: A Deep Learning Approach for Image Denoising...
ShubhamMittal569818
 
A Comprehensive Overview of Encoder and Decoder Architectures in Deep Learnin...
ShubhamMittal569818
 
AUTO ENCODERS (Deep Learning fundamentals)
aayanshsingh0401
 
Autoencoders for image_classification
Cenk Bircanoğlu
 
Piotr Mirowski - Review Autoencoders (Deep Learning) - CIUUK14
Daniel Lewis
 
Autoencoder
Wataru Hirota
 
Autoecoders.pptx
MirzaJahanzeb5
 
Autoencoders
Ashok Govindarajan
 
Understanding Autoencoder (Deep Learning Book, Chapter 14)
Entrepreneur / Startup
 
Structure learning with Deep Neural Networks
Patrick Michl
 
Structure learning with Deep Neural Networks
Patrick Michl
 
Ad

Recently uploaded (20)

PPTX
VITEEE 2026 Exam Details , Important Dates
SonaliSingh127098
 
PPTX
Introduction to Basic Renewable Energy.pptx
examcoordinatormesu
 
PPTX
美国电子版毕业证南卡罗莱纳大学上州分校水印成绩单USC学费发票定做学位证书编号怎么查
Taqyea
 
PPTX
Evaluation and thermal analysis of shell and tube heat exchanger as per requi...
shahveer210504
 
PPTX
Big Data and Data Science hype .pptx
SUNEEL37
 
PDF
Biomechanics of Gait: Engineering Solutions for Rehabilitation (www.kiu.ac.ug)
publication11
 
PDF
AI TECHNIQUES FOR IDENTIFYING ALTERATIONS IN THE HUMAN GUT MICROBIOME IN MULT...
vidyalalltv1
 
PPTX
Heart Bleed Bug - A case study (Course: Cryptography and Network Security)
Adri Jovin
 
PPTX
fatigue in aircraft structures-221113192308-0ad6dc8c.pptx
aviatecofficial
 
PDF
MAD Unit - 2 Activity and Fragment Management in Android (Diploma IT)
JappanMavani
 
PPTX
Thermal runway and thermal stability.pptx
godow93766
 
PPTX
Solar Thermal Energy System Seminar.pptx
Gpc Purapuza
 
PPTX
265587293-NFPA 101 Life safety code-PPT-1.pptx
chandermwason
 
PDF
PORTFOLIO Golam Kibria Khan — architect with a passion for thoughtful design...
MasumKhan59
 
PPTX
Presentation 2.pptx AI-powered home security systems Secure-by-design IoT fr...
SoundaryaBC2
 
DOCX
CS-802 (A) BDH Lab manual IPS Academy Indore
thegodhimself05
 
PDF
MAD Unit - 1 Introduction of Android IT Department
JappanMavani
 
PDF
smart lot access control system with eye
rasabzahra
 
PPTX
Introduction to Design of Machine Elements
PradeepKumarS27
 
PDF
Viol_Alessandro_Presentazione_prelaurea.pdf
dsecqyvhbowrzxshhf
 
VITEEE 2026 Exam Details , Important Dates
SonaliSingh127098
 
Introduction to Basic Renewable Energy.pptx
examcoordinatormesu
 
美国电子版毕业证南卡罗莱纳大学上州分校水印成绩单USC学费发票定做学位证书编号怎么查
Taqyea
 
Evaluation and thermal analysis of shell and tube heat exchanger as per requi...
shahveer210504
 
Big Data and Data Science hype .pptx
SUNEEL37
 
Biomechanics of Gait: Engineering Solutions for Rehabilitation (www.kiu.ac.ug)
publication11
 
AI TECHNIQUES FOR IDENTIFYING ALTERATIONS IN THE HUMAN GUT MICROBIOME IN MULT...
vidyalalltv1
 
Heart Bleed Bug - A case study (Course: Cryptography and Network Security)
Adri Jovin
 
fatigue in aircraft structures-221113192308-0ad6dc8c.pptx
aviatecofficial
 
MAD Unit - 2 Activity and Fragment Management in Android (Diploma IT)
JappanMavani
 
Thermal runway and thermal stability.pptx
godow93766
 
Solar Thermal Energy System Seminar.pptx
Gpc Purapuza
 
265587293-NFPA 101 Life safety code-PPT-1.pptx
chandermwason
 
PORTFOLIO Golam Kibria Khan — architect with a passion for thoughtful design...
MasumKhan59
 
Presentation 2.pptx AI-powered home security systems Secure-by-design IoT fr...
SoundaryaBC2
 
CS-802 (A) BDH Lab manual IPS Academy Indore
thegodhimself05
 
MAD Unit - 1 Introduction of Android IT Department
JappanMavani
 
smart lot access control system with eye
rasabzahra
 
Introduction to Design of Machine Elements
PradeepKumarS27
 
Viol_Alessandro_Presentazione_prelaurea.pdf
dsecqyvhbowrzxshhf
 
Ad

Autoencoder

  • 2. Introduction Autoencoders are a specific type of feed forward neural networks where the input is the same as the output. They compress the input into a lower-dimensional code and then reconstruct the output from this representation. The code is a compact “summary” or “compression” of the input, also called the latent-space representation.
  • 4. Components of Autoencoder An autoencoder consists of 3 components: Encoder Code Decoder Encoder compresses the input and produces the code. Decoder then reconstructs the input only using this code.
  • 6. Properties of Autoencoder Data-specific: Autoencoders are only able to meaningfully compress data similar to what they have been trained on. Lossy: The output of the autoencoder will not be exactly the same as the input, it will be a close but degraded representation. Unsupervised: Autoencoders are considered an unsupervised learning technique since they don’t need explicit labels to train on. But to be more precise they are self-supervised because they generate their own labels from the training data.
  • 8. Architecture Both the encoder and decoder are fully- connected feed forward neural networks. The number of nodes in the code layer (code size) is a hyper parameter that we set before training the autoencoder.
  • 9. Hyperparameters There are 4 hyperparameters that we need to set before training an autoencoder: Code size: number of nodes in the middle layer. Smaller size results in more compression. Number of layers: the autoencoder can be as deep as we like. In the figure above we have 2 layers in both the encoder and decoder, without considering the input and output.
  • 10. Hyperparameters Number of nodes per layer : The layers are stacked one after another. The number of nodes per layer decreases with each subsequent layer of the encoder, and increases back in the decoder. Loss function: We either use mean squared error (mse) or binary crossentropy. If the input values are in the range [0, 1] then we typically use crossentropy, otherwise we use the mean squared error.
  • 11. Types of Autoencoder : The Four Types of Autoencoders are :- Vanilla autoencoder Multilayer autoencoder Convolutional autoencoder Regularized autoencoder
  • 12. Vanilla autoencoder The autoencoder is a three layers of a neural net with one hidden layer. The input and output are the same, and we learn how to reconstruct the input. Here, we see that we have an undercomplete autoencoder as the hidden layer dimension (64) is smaller than the input (784). This constraint will impose our neural net to learn a compressed representation of data.
  • 13. Multilayer autoencoder If one hidden layer is not enough, we can obviously extend the autoencoder to more hidden layers. Now our implementation uses 3 hidden layers instead of just one. Any of the hidden layers can be picked as the feature representation but we will make the network symmetrical and use the middle-most layer.
  • 14. Convolution autoencoder We may also ask ourselves: can autoencoders be used with Convolutions instead of Fully- connected layers The answer is yes and the principle is the same, but using images (3D vectors) instead of flattened 1D vectors. The input image is downsampled to give a latent representation of smaller dimensions and force the autoencoder to learn a compressed version of the images.
  • 15. Regularized autoencoder There are other ways we can constraint the reconstruction of an autoencoder than to impose a hidden layer of smaller dimension than the input. Rather than limiting the model capacity by keeping the encoder and decoder shallow and the code size small, regularized autoencoders use a loss function that encourages the model to have other properties besides the ability to copy its input to its output.
  • 17. Challenges They are indeed pretty similar, but not exactly the same. We can notice it more clearly in the last digit “4”. Since this was a simple task our autoencoder performed pretty well. We have total control over the architecture of the autoencoder. We can make it very powerful by increasing the number of layers, nodes per layer and most importantly the code size. Increasing these hyperparameters will let the autoencoder to learn more complex codings.
  • 18. Challenges But we should be careful to not make it too powerful. Over-fitting The autoencoder will reconstruct the training data perfectly, but it will be over fitting without being able to generalize to new instances, which is not what we want.
  • 19.  Undercomplete The autoencoder is said to be undercomplete. It won’t be able to directly copy its inputs to the output, and will be forced to learn intelligent features. then an undercomplete autoencoder won’t be able to recover it perfectly.
  • 20. Regularization We would like to learn meaningful features without altering the code’s dimensions (Overcomplete or Undercomplete). We usually find two types of regularized autoencoder:  Sparse Autoencoder  Denoising Autoencoder
  • 21. Sparse autoencoder Sparse autoencoders are typically used to learn features for another task such as classification. An autoencoder that has been regularized to be sparse must respond to unique statistical features of the dataset it has been trained on, rather than simply acting as an identity function. In this way, training to perform the copying task with a sparsity penalty can yield a model that has learned useful features as a by product.
  • 22. Denoising Autoencoders This way the autoencoder can’t simply copy the input to its output because the input also contains random noise. We are asking it to subtract the noise and produce the underlying meaningful data.
  • 24. Conclusion Autoencoders are a very useful dimensionality reduction technique. They are very popular as a teaching material in introductory deep learning courses, most likely due to their simplicity.