1. Deep Learning
1. Deep Learning
Deep Learning
Agenda
Ex Machina (2015)
Visualized here are 3% of the neurons and 0.0001% of the synapses in the brain.
Thalamocortical system visualization via DigiCortex Engine.
https://deeplearning.mit.edu 2019
For the full list of references visit:
https://hcai.mit.edu/references
[286]
History of Deep Learning Ideas and Milestones*
• 1943: Neural networks
We are here
• 1957: Perceptron
• 1974-86: Backpropagation, RBM, RNN
• 1989-98: CNN, MNIST, LSTM, Bidirectional RNN
• 2006: “Deep Learning”, DBN
• 2009: ImageNet
Perspective:
• 2012: AlexNet, Dropout
• Universe created
13.8 billion years ago • 2014: GANs
• Earth created • 2014: DeepFace
4.54 billion years ago
• Modern humans • 2016: AlphaGo
300,000 years ago
• 2017: AlphaZero, Capsule Networks
• Civilization
12,000 years ago • 2018: BERT
• Written record * Dates are for perspective and not as definitive historical
5,000 years ago record of invention or credit
Human Brain
• Thalamocortical system:
3 million neurons
476 million synapses
• Full brain:
100 billion neurons
1,000 trillion synapses
Step 2 Step 5
Step 3 Animated
“output wires”
Neurons in the brain
Deep
Learning
Representation
Learning
Machine
Learning
Artificial
Intelligence
Task: Draw a line to separate the green triangles and blue circles.
Task: Draw a line to separate the blue curve and red curve
Deep Learning
Self-Driving Cars
𝑦
ො
© deeplearning.ai
Why Deep Learning and Why
Now?
https://www.prowesscorp.com/whats-the-difference-between-artificial-intelligence-ai-machine-learning-and-deep-learning/
Machine Learning definition
Herbert Simon.
Learning is any process by which a system improves performance
from experience.
Machine Learning is concerned with computer
programs that automatically improve their
performance through experience.
Machine Learning definition
Labels
A label is the thing we're predicting.
Features
A feature is an input variable
Examples or samples
An example is a particular instance of data, x.
labeled examples. A labeled example includes both feature(s) and the label.
unlabeled examples. An unlabeled example contains features but not the
label.
Models
A model defines the relationship between features and label.
Training means creating or learning the model.
Inference means applying the trained model to unlabeled examples.
Key types of Machine Learning problems
Supervised machine learning: Learn to predict target values from labelled data/
•Examine the statistics of two football teams, and predicting which team will
win tomorrow's match (given historical data of teams' wins/losses to learn
from).
300
Price ($)
200
in 1000’s
100
0
0 500 1000 1500 2000 2500
Size in feet2
Image Segmentation
https://phenaki.github.io/
Image to Text Image Generation
VisionEye
With object detection and voice delivery systems,
the project aims to help individuals with visual
impairments navigate easily and safely. Our
students promote an inclusive society where
everyone has access to the resources that they
need.
Computer Vision and Its role in Digital Inclusion
Computer Vision and Its role in Digital Inclusion
Autonomous Car Driving
Path
Planning
Laser Terrain
Mapping
Sebastian
Stanley
Face Recognition
object models
object parts
(combination
of edges)
edges
pixels
Image Segmentation
Deep Learning in One Slide
Input Image:
TensorFlow Neural 5
Model: Network
6
Output: 5
(with 87% confidence)
“Encoded in the large, highly evolve sensory and motor portions of the human brain is a billion
years of experience about the nature of the world and how to survive in it.… Abstract thought,
though, is a new trick, perhaps less than 100 thousand years old. We have not yet mastered it. It
is not all that intrinsically difficult; it just seems so when we do it.”
- Hans Moravec, Mind Children (1988)
For the full list of references visit:
https://hcai.mit.edu/references
[6, 7, 11, 68] https://deeplearning.mit.edu
Measuring Progress: Einstein vs Savant
Human Augmented
Supervised
Machine Learning
Human Semi-
Supervised
Machine Learning
Human Reinforcement
Machine Learning
Machine Unsupervised
Learning
https://deeplearning.mit.edu 2019
Data Augmentation
Crop: Flip:
Scale: Rotate:
Noise:
Translation:
Training Stage:
Testing Stage:
https://deeplearning.mit.edu
How Neural Networks Learn: Backpropagation
Forward Pass:
Input Neural
Prediction
Data Network
Neural Measure
Network of Error
Adjust to Reduce Error
https://deeplearning.mit.edu
What can we do with Deep Learning?
• Number • Number
• Vector of numbers • Vector of numbers
• Sequence of numbers • Sequence of numbers
• Sequence of vectors of numbers • Sequence of vectors of numbers
Tanh
• Vanishing gradients
ReLU
• Not zero centered
Subtasks:
1. Forward pass to compute network output and “error”
2. Backward pass to compute gradients
3. A fraction of the weight’s gradient is subtracted from the weight.
Hard to break symmetry Vanilla SGD gets your there, but can be slow
“Training with large minibatches is bad for your health. More importantly, it's
bad for your test error. Friends don’t let friends use minibatches larger than 32.”
- Yann LeCun
Revisiting Small Batch Training for Deep Neural Networks (2018)
• Other options:
• Layer normalization (LN) – conceived for RNNs
• Instance normalization (IN) – conceived for Style Transfer
• Group normalization (GN) – conceived for CNNs
• Convolutional filters:
take advantage of
spatial invariance
• Unsupervised learning
• Gives embedding
• Typically better embeddings
come from discriminative task
http://projector.tensorflow.org/
For the full updated list of references visit:
https://selfdrivingcars.mit.edu/references
[298] https://deeplearning.mit.edu 2019
Generative Adversarial Network (GANs)
Word Vector
• Applications
• Sequence Data
• Text
• Speech
• Audio
• Video
• Generation
• Short-term dependence:
Bob is eating an apple.
Context • Long-term dependence:
Bob likes apples. He is hungry and decided to
have a snack. So now he is eating an apple.
In theory, vanilla RNNs
can handle arbitrarily
long-term dependence.
Chapter 1 - Zhang, Aston & Lipton, Zachary & Li, Mu & Smola,
Alexander. (2023). Dive into Deep Learning, Cambridge University
Press.
https://d2l.ai/chapter_introduction/index.html