1. INDIAN INSTITUTE OF TECHNOLOGY ROORKEE
Introduction to Neural networks and Recent
Advances
ASHA RANI
2. Perceptron learning algorithm
repeat until convergence (or for some # of iterations):
for each training example (f1, f2, …, fn, label):
if prediction * label ≤ 0: // they don’t agree
for each wi:
wi = wi + fi*label
b = b + label
Why is it called the “perceptron” learning algorithm if
what it learns is a line? Why not “line learning”
algorithm?
4. Our nervous system:
the computer science view
the human brain is a large collection
of interconnected neurons
a NEURON is a brain cell
– they collect, process, and disseminate
electrical signals
– they are connected via synapses
– they FIRE depending on the
conditions of the neighboring neurons
5. Output y
Input x1
Input x2
Input x3
Input x4
Weight w1
Weight w2
Weight w3
Weight w4
A neuron/perceptron
activation function
How is this a linear
classifier (i.e. perceptron)?
6. Hard threshold = linear classifier
hard threshold:
…
output
x1
x2
xm
w1
w2
wm
7. Neural Networks
Neural Networks try to mimic the structure and function of our
nervous system
People like biologically motivated approaches
9. 9
W is the strength of signal sent between A and B.
If A fires and w is positive, then A stimulates B.
If A fires and w is negative, then A inhibits B.
Weight w
Node A Node B
(perceptron) (perceptron)
44. What does the decision boundary look like?
Input x1
Input x2
Output = x1 xor x2
linear splits of the
feature space
combination of
these linear
spaces
50. What does the decision boundary look like?
Input x1
Input x2
Output = x1 xor x2
linear splits of the
feature space
combination of
these linear
spaces
53. NN decision boundaries
For DT, as the tree gets larger, the model gets more complex
The same is true for neural networks: more hidden nodes =
more complexity
Adding more layers adds even more complexity (and much
more quickly)
Good rule of thumb:
number of 2-layer hidden nodes ≤
number of examples
number of dimensions
Editor's Notes
#10:Allow us to model more than just 0/1 outputs
Differentiable! (this will come into play in a minute)