https://github.com/telecombcn-dl/dlmm-2017-dcu
Deep learning technologies are at the core of the current revolution in artificial intelligence for multimedia data analysis. The convergence of big annotated data and affordable GPU hardware has allowed the training of neural networks for data analysis tasks which had been addressed until now with hand-crafted features. Architectures such as convolutional neural networks, recurrent neural networks and Q-nets for reinforcement learning have shaped a brand new scenario in signal processing. This course will cover the basic principles and applications of deep learning to computer vision problems, such as image classification, object detection or text captioning.
6. 6
Recurrent Neural Network (RNN)
Alex Graves, “Supervised Sequence Labelling with Recurrent Neural Networks”
The hidden layers and the
output depend from previous
states of the hidden layers
7. 7
Alex Graves, “Supervised Sequence Labelling with Recurrent Neural Networks”
The hidden layers and the
output depend from previous
states of the hidden layers
Recurrent Neural Network (RNN)
8. 8
Recurrent Neural Network (RNN)
F. Van Veen, “The Neural Network Zoo” (2016)
Further details:
D2L8, “Recurrent”
9. 9
Recurrent Neural Network (RNN)
F. Van Veen, “The Neural Network Zoo” (2016)
Further details:
D2L8, “Recurrent”
14. 14
Convolutional Neural Network (CNN)
Orange
A Krizhevsky, I Sutskever, GE Hinton “Imagenet classification with deep convolutional neural networks” Part of: Advances in
Neural Information Processing Systems 25 (NIPS 2012)
AlexNet
19. 19Long, Jonathan, Evan Shelhamer, and Trevor Darrell. "Fully convolutional networks for semantic segmentation." CVPR 2015 & PAMI 2016.
Deep Residual Learning / Skip connections
20. 20
Ronneberger, Olaf, Philipp Fischer, and Thomas Brox. "U-net: Convolutional networks for biomedical image segmentation." In
International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 234-241. Springer.
U-Net
Deep Residual Learning / Skip connections
21. 21Manel Baradad, Amaia Salvador, Xavier Giró-i-Nieto, Ferran Marqués (work under progress)
Deep Residual Learning / Skip connections
22. 22
Milletari, Fausto, Nassir Navab, and Seyed-Ahmad Ahmadi. "V-net: Fully convolutional neural networks for volumetric medical
image segmentation." In 3D Vision (3DV), 2016 Fourth International Conference on, pp. 565-571. IEEE, 2016.
V-Net
Deep Residual Learning / Skip connections
23. 23
Dense connections
Dense Block of 5-layers
with a growth rate of k=4
Connect every layer to every other layer of the same filter size.
Huang, Gao, Zhuang Liu, Kilian Q. Weinberger, and Laurens van der Maaten. "Densely connected
convolutional networks." arXiv preprint arXiv:1608.06993 (2016). [code]
DenseNet
24. 24
Dense connections
Huang, Gao, Zhuang Liu, Kilian Q. Weinberger, and Laurens van der Maaten. "Densely connected
convolutional networks." arXiv preprint arXiv:1608.06993 (2016). [code]
DenseNet
25. 25
Dense connections
Huang, Gao, Zhuang Liu, Kilian Q. Weinberger, and Laurens van der Maaten. "Densely connected
convolutional networks." arXiv preprint arXiv:1608.06993 (2016). [code]
26. 26
Dense connections
Jégou, Simon, Michal Drozdzal, David Vazquez, Adriana Romero, and Yoshua Bengio. "The One Hundred Layers
Tiramisu: Fully Convolutional DenseNets for Semantic Segmentation." arXiv preprint arXiv:1611.09326 (2016). [code]
[slides]
28. 28
Autoencoder (AE)
“Deep Learning Tutorial”, Dept. Computer Science, Stanford
Autoencoders:
● Predict at the output the
same input data.
● Do not need labels:
Unsupervised
learning
29. 29
Autoencoder (AE)
“Deep Learning Tutorial”, Dept. Computer Science, Stanford
Dimensionality reduction:
● Use hidden layer as
a feature extractor of
the desired size.
Unsupervised
learning
32. 32
Variational Autoencoder (VAE)
Kevin Frans, “Variational Autoencoders explained” (2016)
The latent vector learned in the hidden layer of the basic autoencoder (in green)...
36. 36F. Van Veen, “The Neural Network Zoo” (2016)
Restricted Boltzmann Machine (RBM)
37. 37
Restricted Boltzmann Machine (RBM)
Figure: Geoffrey Hinton (2013)
Salakhutdinov, Ruslan, Andriy Mnih, and Geoffrey Hinton. "Restricted Boltzmann machines for
collaborative filtering." Proceedings of the 24th international conference on Machine learning. ACM, 2007.
● Shallow two-layer net.
● Restricted=No two nodes in a layer share a
connection
● Bipartite graph.
● Bidirectional graph
○ Shared weights.
○ Different biases.
38. 38
Restricted Boltzmann Machine (RBM)
Figure: Geoffrey Hinton (2013)
Salakhutdinov, Ruslan, Andriy Mnih, and Geoffrey Hinton. "Restricted Boltzmann machines for
collaborative filtering." Proceedings of the 24th international conference on Machine learning. ACM, 2007.
39. 39
Restricted Boltzmann Machine (RBM)
DeepLearning4j, “A Beginner’s Tutorial for Restricted Boltzmann Machines”.
RBMs are a specific type of
autoencoder.
Unsupervised
learning
42. 42
Deep Belief Networks (DBN)
Hinton, Geoffrey E., Simon Osindero, and Yee-Whye Teh. "A fast learning algorithm for deep belief
nets." Neural computation 18, no. 7 (2006): 1527-1554.
● Architecture like an MLP.
● Training as a stack of
RBMs…
● ...so they do not need
labels:
Unsupervised
learning
43. 43
Deep Belief Networks (DBN)
Hinton, Geoffrey E., Simon Osindero, and Yee-Whye Teh. "A fast learning algorithm for deep belief
nets." Neural computation 18, no. 7 (2006): 1527-1554.
After the DBN is trained, it can
be fine-tuned with a reduced
amount of labels to solve a
supervised task with superior
performance.
Supervised
learning
Softmax
45. 45
Deep Belief Networks (DBN)
Geoffrey Hinton, "Introduction to Deep Learning & Deep Belief Nets” (2012)
Geoorey Hinton, “Tutorial on Deep Belief Networks”. NIPS 2007.
50. 50
Differentiable Neural Computers (DNC)
Add a trainable external memory to a neural network.
Graves, Alex, Greg Wayne, Malcolm Reynolds, Tim Harley, Ivo Danihelka, Agnieszka Grabska-Barwińska,
Sergio Gómez Colmenarejo et al. "Hybrid computing using a neural network with dynamic external memory."
Nature 538, no. 7626 (2016): 471-476. [Post by DeepMind]
51. 51
Differentiable Neural Computers (DNC)
DNC can solve tasks reading information from a trained memory.
Graves, Alex, Greg Wayne, Malcolm Reynolds, Tim Harley, Ivo Danihelka, Agnieszka Grabska-Barwińska,
Sergio Gómez Colmenarejo et al. "Hybrid computing using a neural network with dynamic external memory."
Nature 538, no. 7626 (2016): 471-476. [Post by DeepMind]
52. 52
Reinforcement Learning (RL)
Mnih, Volodymyr, Koray Kavukcuoglu, David Silver, Alex Graves, Ioannis Antonoglou, Daan Wierstra, and Martin
Riedmiller. "Playing atari with deep reinforcement learning." arXiv preprint arXiv:1312.5602 (2013).
55. 55
Reinforcement Learning (RL)
An agent that is a decision-maker interacts with the environment and learns
through trial-and-error
Slide credit: UCL Course on RL by David Silver
56. 56
Reinforcement Learning (RL)
Deep Q-Network (DQN)
Mnih, Volodymyr, Koray Kavukcuoglu, David Silver, Andrei A. Rusu, Joel Veness, Marc G. Bellemare, Alex Graves et al.
"Human-level control through deep reinforcement learning." Nature 518, no. 7540 (2015): 529-533.
57. 57
Reinforcement Learning (RL)
Deep Q-Network (DQN)
Source: Tambet Matiisen, Demystifying Deep Reinforcement Learning (Nervana)
Naive DQN Refined DQN
59. 59
Reinforcement Learning (RL)
Slide credit: Míriam Bellver
actor
state
critic
‘q-value’action (5)
state
action (5)
actor performs an
action
critic assesses how good the action was, and the
gradients are used to train the actor and the critic
Actor-Critic algorithm
Grondman, Ivo, Lucian Busoniu, Gabriel AD Lopes, and Robert Babuska. "A survey of actor-critic reinforcement learning: Standard and natural
policy gradients." IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews) 42, no. 6 (2012): 1291-1307.
60. 60
Reinforcement Learning (RL)
Silver, D., Huang, A., Maddison, C.J., Guez, A., Sifre, L., Van Den Driessche, G., Schrittwieser, J., Antonoglou, I.,
Panneershelvam, V., Lanctot, M. and Dieleman, S., 2016. Mastering the game of Go with deep neural networks and tree
search. Nature, 529(7587), pp.484-489