Presentation on 'Deep Learning: Evolution of ML from Statistical to Brain-like Computing'
Speaker- Dr. Vijay Srinivas Agneeswaran,Director, Big Data Labs, Impetus
The main objective of the presentation is to give an overview of our cutting edge work on realizing distributed deep learning networks over GraphLab. The objectives can be summarized as below:
- First-hand experience and insights into implementation of distributed deep learning networks.
- Thorough view of GraphLab (including descriptions of code) and the extensions required to implement these networks.
- Details of how the extensions were realized/implemented in GraphLab source – they have been submitted to the community for evaluation.
- Arrhythmia detection use case as an application of the large scale distributed deep learning network.
Reference : http://neuralnetworksanddeeplearning.com/chap1.html
Consider the problem to identify the individual digits from the input image
Each image 28 by 28 pixel image. Then network is designed as follows
Input layer (image) -> 28*28 = 784 neurons. Each neuron corresponds to a pixel
The output layer can be identified by the number of digits to be identified i.e. 10 (0 to 9)
The intermediate hidden layer can be experimented with varied number of neurons. Let us fix at 10 nodes in hidden layer
Reference: http://neuralnetworksanddeeplearning.com/chap1.html
How about recognizing a human face from given set of random images?
Attack this problem in the similar fashion explained earlier. Input -> Image pixels, output -> Is it a face or not? (a single node)
A face can be recognized by answering some questions like “Is there an eye in the top left?”, “Is there a nose in the middle?” etc..
Each question corresponds to a hidden layer
ANN for face recognition?
Why SVMs or any kernel based approach cannot be used here.
Implicit assumption of a locally smooth function around each training example.
Problem decomposition into sub-problems
Breakdown into sub-problems, solvable by sub-networks. Complex problem requires more sub-networks, more hidden layers, hence need for deep neural networks.
http://deeplearning4j.org/convolutionalnets.html
Refined by Lecun in 1989 – mainly to apply CNNs to identify variability in 2D image data.
Introduced in 1980 by Fukushima
A type of RBMs where the communication is absent across the nodes in the same layer
Nodes are not connected to every other node of next layer. Symmetry is not there
Convolution networks learn images by pieces rather than learning as a whole (RBM does this)
Designed to use minimal amounts of pre processing