Deep Learning Introduction Class (1)
Deep Learning Introduction Class (1)
http://playground.tensorflow.org/
cat
dog
Loss function
• The quantity to be minimized (optimized) during training
• the only thing the network cares about
• there might also be other metrics you care about
• Common tasks have “standard” loss functions:
• mean squared error for regression
• binary cross-entropy for two-class classification
• categorical cross-entropy for multi-class classification
• etc.
• https://lossfunctions.tumblr.com/
Optimizer
• How to update the weights
based on the loss function
• Learning rate (+scheduling)
• Stochastic gradient descent,
momentum, and their variants
• RMSProp is usually a good
first choice
• more info:
http://ruder.io/optimizing-gradient-descent/
graphs
• Automatic differentiation ✕ ✕
• Python/numpy or R interfaces
• instead of C, C++, CUDA or HIP
• Open source
Deep learning Lasagn
e
Keras
TF
Estimator
torch.nn Gluon
Algorithms
• Activation function (ReLU)
• Algorithms that help convergence
• Adam
• RMSProp
Extraction of features in
layers of CNN
Lower layers of a multilayered convolutional neural network will learn
basic features such as corners and edges.
The higher layers of the neural network will learn to detect more
complex features such as faces, objects, full scenes, etc.
Ideally, a neural network’s latent space would represent concepts that
are relevant to the classes of images it is meant to detect.
Further resources
• Recommended textbook: “Deep learning”
by Goodfellow, Bengio, Courville
•“Deep Learning with Python” by François
Chollet