SlideShare a Scribd company logo
1 of 66
Download to read offline
Terry Taewoong Um (terry.t.um@gmail.com)
University of Waterloo
Department of Electrical & Computer Engineering
Terry Taewoong Um
MACHINE LEARNING,
DEEP LEARNING, AND
MOTION ANALYSIS
1
Terry Taewoong Um (terry.t.um@gmail.com)
CAUTION
• I cannot explain everything
• You cannot get every details
2
• Try to get a big picture
• Get some useful keywords
• Connect with your research
Terry Taewoong Um (terry.t.um@gmail.com)
CONTENTS
1. What is Machine Learning?
(Part 1 Q & A)
2. What is Deep Learning?
(Part 2 Q & A)
3. Machine Learning in Motion Analysis
(Part 3 Q & A)
3
Terry Taewoong Um (terry.t.um@gmail.com)
CONTENTS
4
1. What is Machine Learning?
Terry Taewoong Um (terry.t.um@gmail.com)
WHAT IS MACHINE LEARNING?
"A computer program is said to learn from experience E
with respect to some class of tasks T and performance
measure P, if its performance at tasks in T, as measured
by P, improves with experience E“ – T. Michell (1997)
Example: A program for soccer tactics
5
T : Win the game
P : Goals
E : (x) Players’ movements
(y) Evaluation
Terry Taewoong Um (terry.t.um@gmail.com)
WHAT IS MACHINE LEARNING?
6
“Toward learning robot table tennis”, J. Peters et al. (2012)
https://youtu.be/SH3bADiB7uQ
"A computer program is said to learn from experience E
with respect to some class of tasks T and performance
measure P, if its performance at tasks in T, as measured
by P, improves with experience E“ – T. Michell (1997)
Terry Taewoong Um (terry.t.um@gmail.com)
TASKS
7
classification
discrete target values
x : pixels (28*28)
y : 0,1, 2,3,…,9
regression
real target values
x ∈ (0,100)
y : 0,1, 2,3,…,9
clustering
no target values
x ∈ (-3,3)×(-3,3)
"A computer program is said to learn from experience E
with respect to some class of tasks T and performance
measure P, if its performance at tasks in T, as measured
by P, improves with experience E“ – T. Michell (1997)
Terry Taewoong Um (terry.t.um@gmail.com)
PERFORMANCE
8
"A computer program is said to learn from experience E
with respect to some class of tasks T and performance
measure P, if its performance at tasks in T, as measured
by P, improves with experience E“ – T. Michell (1997)
classification
0-1 loss function
regression
L2 loss function
clustering
Terry Taewoong Um (terry.t.um@gmail.com)
EXPERIENCE
9
"A computer program is said to learn from experience E
with respect to some class of tasks T and performance
measure P, if its performance at tasks in T, as measured
by P, improves with experience E“ – T. Michell (1997)
classification
labeled data
(pixels)→(number)
regression
labeled data
(x) → (y)
clustering
unlabeled data
(x1,x2)
Terry Taewoong Um (terry.t.um@gmail.com)
A TOY EXAMPLE
10
? Height(cm)
Weight
(kg)
[Input X]
[Output Y]
Terry Taewoong Um (terry.t.um@gmail.com)
11
180 Height(cm)
Weight
(kg)
80
Y = aX+b
Model : Y = aX+b Parameter : (a, b)
[Goal] Find (a,b) which best fits the given data
A TOY EXAMPLE
Terry Taewoong Um (terry.t.um@gmail.com)
12
[Analytic Solution]
Least square problem
(from AX = b, X=A#b where
A# is A’s pseudo inverse)
Not always available
[Numerical Solution]
1. Set a cost function
2. Apply an optimization method
(e.g. Gradient Descent (GD) Method)
L
(a,b)
http://www.yaldex.com/game-
development/1592730043_ch18lev1sec4.html
Local minima problem
http://mnemstudio.org/neural-networks-
multilayer-perceptron-design.htm
A TOY EXAMPLE
Terry Taewoong Um (terry.t.um@gmail.com)
13
32 Age(year)
Running
Record
(min)
140
WHAT WOULD BE THE CORRECT MODEL?
Select a model → Set a cost function → Optimization
Terry Taewoong Um (terry.t.um@gmail.com)
14
? X
Y
WHAT WOULD BE THE CORRECT MODEL?
1. Regularization 2. Nonparametric model
“overfitting”
Terry Taewoong Um (terry.t.um@gmail.com)
15
L2 REGULARIZATION
(e.g. w=(a,b) where Y=aX+b)
Avoid a complicated model!
• Another interpretation :
: Maximum a Posteriori (MAP)
http://goo.gl/6GE2ix
http://goo.gl/6GE2ix
Terry Taewoong Um (terry.t.um@gmail.com)
16
L2 REGULARIZATION
• Another interpretation :
: Maximum a Posteriori (MAP)
http://goo.gl/6GE2ix
http://goo.gl/6GE2ix
• Bayesian inference
𝑃 𝐵𝑒𝑙𝑖𝑒𝑓 𝐷𝑎𝑡𝑎 =
𝑃 𝐵𝑒𝑙𝑖𝑒𝑓 𝑃(𝐷𝑎𝑡𝑎|𝐵𝑒𝑙𝑖𝑒𝑓)
𝑃(𝐷𝑎𝑡𝑎)
posterior
prior likelihood
ex) fair coin : 50% H, 50% T
falsified coin : 80% H, 20% T
Let’s say we observed ten heads consecutively.
What’s the probability for being a fair coin?
𝑃 𝐵𝑒𝑙𝑖𝑒𝑓 = 0.2
𝑃 𝐷𝑎𝑡𝑎|𝐵𝑒𝑙𝑖𝑒𝑓 = 0.510
≈ 0.001
𝑃 𝐵𝑒𝑙𝑖𝑒𝑓|𝐷𝑎𝑡𝑎 ∝ 0.2 ∗ 0.001 = 0.0002
normalization
(you don’t believe this coin is fair)
Fair
coin?
Falsified
coin?
𝑃 𝐵𝑒𝑙𝑖𝑒𝑓 = 0.8
𝑃 𝐷𝑎𝑡𝑎|𝐵𝑒𝑙𝑖𝑒𝑓 = 0.810
≈ 0.107
𝑃 𝐵𝑒𝑙𝑖𝑒𝑓|𝐷𝑎𝑡𝑎 ∝ 0.8 ∗ 0.107 = 0.0856
Fair =
0.0002
0.0002+0.0856
= 0.23% , Unfair = 99.77%
Terry Taewoong Um (terry.t.um@gmail.com)
17
WHAT WOULD BE THE CORRECT MODEL?
1. Regularization 2. Nonparametric model
training time
error
training error
test error
we should
stop here
training
set
validation
set
test
set
for training
(parameter
optimization)
for early
stopping
(avoid
overfitting)
for evaluation
(measure the
performance)
keep watching the validation error
Terry Taewoong Um (terry.t.um@gmail.com)
18
NONPARAMETRIC MODEL
• It does not assume any parametric models (e.g. Y = aX+b, Y=aX2+bX+c, etc.)
• It often requires much more samples
• Kernel methods are frequently applied for modeling the data
• Gaussian Process Regression (GPR), a sort of kernel method, is a widely-used
nonparametric regression method
• Support Vector Machine (SVM), also a sort of kernel method, is a widely-used
nonparametric classification method
kernel function
[Input space] [Feature space]
Terry Taewoong Um (terry.t.um@gmail.com)
19
SUPPORT VECTOR MACHINE (SVM)
“Myo”, Thalmic Labs (2013)
https://youtu.be/oWu9TFJjHaM
[Linear classifiers] [Maximum margin]
Support vector Machine Tutorial, J. Weston, http://goo.gl/19ywcj
[Dual formulation] ( )
kernel function
kernel function
Terry Taewoong Um (terry.t.um@gmail.com)
20
GAUSSIAN PROCESS REGRESSION (GPR)
https://youtu.be/YqhLnCm0KXY
https://youtu.be/kvPmArtVoFE
• Gaussian Distribution
• Multivariate regression likelihood
posterior
prior
likelihood
prediction conditioning the joint distribution of the observed & predicted values
https://goo.gl/EO54WN
http://goo.gl/XvOOmf
Terry Taewoong Um (terry.t.um@gmail.com)
21
DIMENSION REDUCTION
[Original space] [Feature space]
low dim. high dim.
high dim. low dim.
𝑋 → ∅(𝑋)
• Principal Component Analysis
: Find the best orthogonal axes
(=principal components) which
maximize the variance of the data
Y = P X
* The rows in P are m largest eigenvectors
of
1
𝑁
𝑋𝑋 𝑇
(covariance matrix)
Terry Taewoong Um (terry.t.um@gmail.com)
22
DIMENSION REDUCTION
http://jbhuang0604.blogspot.kr/2013/04/miss-korea-2013-contestants-face.html
Terry Taewoong Um (terry.t.um@gmail.com)
23
SUMMARY - PART 1
• Machine Learning
- Tasks : Classification, Regression, Clustering, etc.
- Performance : 0-1 loss, L2 loss, etc.
- Experience : labeled data, unlabelled data
• Machine Learning Process
(1) Select a parametric / nonparametric model
(2) Set a performance measurement including regularization term
(3) Training data (optimizing parameters) until validation error increases
(4) Evaluate the final performance using test set
• Nonparametric model : Support Vector Machine, Gaussian Process Regression
• Dimension reduction : used as pre-processing data
Terry Taewoong Um (terry.t.um@gmail.com)
CONTENTS
24
Questions about Part 1?
Terry Taewoong Um (terry.t.um@gmail.com)
CONTENTS
25
2. What is Deep Learning?
Terry Taewoong Um (terry.t.um@gmail.com)
26
PARADIGM CHANGE
PAST
Knowledge
ML
Method
(e.g.
GPR, SVM)
PRESENT
What is the best
ML method for
the target task?
Knowledge
Representation
How can we find a
good representation?
Terry Taewoong Um (terry.t.um@gmail.com)
27
PARADIGM CHANGE
Knowledge
PRESENT
Representation
How can we find a
good representation?
kernel function
Terry Taewoong Um (terry.t.um@gmail.com)
28
PARADIGM CHANGE
Knowledge
PRESENT
Representation
(Features)
How can we find a
good representation?
IMAGE
SPEECH
Hand-Crafted Features
Terry Taewoong Um (terry.t.um@gmail.com)
29
PARADIGM CHANGE
IMAGE
SPEECH
Hand-Crafted Features
Knowledge
PRESENT
Representation
(Features)
Can we learn a good representation
(feature) for the target task as well?
Terry Taewoong Um (terry.t.um@gmail.com)
30
DEEP LEARNING
• What is Deep Learning (DL) ?
- Learning methods which have deep (not shallow) architecture
- It often allows end-to-end learning
- It automatically finds intermediate representation. Thus,
it can be regarded as a representation learning
- It often contains stacked “neural network”. Thus,
Deep learning usually indicates “deep neural network”
“Deep Gaussian Process” (2013)
https://youtu.be/NwoGqYsQifg
http://goo.gl/fxmmPE
http://goo.gl/5Ry08S
Terry Taewoong Um (terry.t.um@gmail.com)
31
OUTSTANDING PERFORMANCE OF DL
error rate : 28% → 15% → 8%
(2010) (2014)(2012)
- Object recognition (Simonyan et al., 2015)
- Natural machine translation (Bahdanau et al., 2014)
- Speech recognition (Chorowski et al., 2014)
- Face recognition (Taigman et al., 2014)
- Emotion recognition (Ebrahimi-Kahou et al., 2014)
- Human pose estimation (Jain et al., 2014)
- Deep reinforcement learning(mnih et al., 2013)
- Image/Video caption (Xu et al., 2015)
- Particle physics (Baldi et al., 2014)
- Bioinformatics (Leung et al., 2014)
- And so on….
• State-of-art results achieved by DL
DL has won most of ML challenges!
K. Cho, https://goo.gl/vdfGpu
Terry Taewoong Um (terry.t.um@gmail.com)
32
BIOLOGICAL EVIDENCE
• Somatosensory cortex learns to see
• Why do we need different ML methods
for different task?
Yann LeCun, https://goo.gl/VVQXJG
• The vental pathway in the visual cortex has multiple stages
• There exist a lot of intermediate representations
Andrew Ng, https://youtu.be/ZmNOAtZIgIk
Terry Taewoong Um (terry.t.um@gmail.com)
33
BIG MOVEMENT
http://goo.gl/zNbBE2 http://goo.gl/Lk64Q4
Going deeper and deeper….
Terry Taewoong Um (terry.t.um@gmail.com)
34
NEURAL NETWORK (NN)
Hugo Larochelle, http://www.dmi.usherb.ca/~larocheh/index_en.html
• Universal approximation theorem (Hornik, 1991)
- A single hidden layer NN w/ linear output can approximate any cont. func. arbitrarily well,
given enough hidden units
- This does not imply we have learning method to train them
Terry Taewoong Um (terry.t.um@gmail.com)
35
TRAINING NN
Hugo Larochelle, http://www.dmi.usherb.ca/~larocheh/index_en.html
• First, calculate the output using data & initial parameters (W ,b)
• Activation functions
http://goo.gl/qMQk5H
1
Terry Taewoong Um (terry.t.um@gmail.com)
36
TRAINING NN
Hugo Larochelle, http://www.dmi.usherb.ca/~larocheh/index_en.html
• Then, calculate the error and update the weights from top to bottom
• Parameter gradients
http://goo.gl/qMQk5H
: Backpropagation algorithm
2
known
Terry Taewoong Um (terry.t.um@gmail.com)
37
TRAINING NN
Hugo Larochelle, http://www.dmi.usherb.ca/~larocheh/index_en.html
• Then, calculate the error and update the weights from top to bottom
• Parameter gradients
http://goo.gl/qMQk5H
: Backpropagation algorithm
2
known
Terry Taewoong Um (terry.t.um@gmail.com)
38
TRAINING NN
Hugo Larochelle, http://www.dmi.usherb.ca/~larocheh/index_en.html
• Then, calculate the error and update the weights from top to bottom
• Parameter gradients
http://goo.gl/qMQk5H
: Backpropagation algorithm
2
known
Terry Taewoong Um (terry.t.um@gmail.com)
39
TRAINING NN
Hugo Larochelle, http://www.dmi.usherb.ca/~larocheh/index_en.html
• Then, calculate the error and update the weights from top to bottom
• Parameter gradients
http://goo.gl/qMQk5H
: Backpropagation algorithm
2
known
Terry Taewoong Um (terry.t.um@gmail.com)
40
TRAINING NN
• Repeat this process with different dataset(mini-batches)
http://goo.gl/qMQk5H
- Forward propagation (calculate the output values)
- Evaluate the error
- Backward propagation (update the weights)
- Repeat this process until the error converges
3
• As you can see here, NN is not a fancy algorithm,
but just a iterative gradient descent method with
huge number of parameters
• NN is often likely to be
stuck in local minima pitfall
Terry Taewoong Um (terry.t.um@gmail.com)
41
FROM NN TO DEEP NN
• From NN to deep NN (since 2006)
- NN requires expert’s skill to tune the hyperparameters
- It sometimes gives a good result, but sometimes gives a bad result.
The result is highly depend on the quality of initialization, regularization,
hyperparameters, data, etc.
- Local minima is always problematic
• A long winter of NN
Yann LeCun
(NYU, Facebook)
Yoshua Bengio
(U. Montreal)
Geoffrey Hinton
(U. Toronto, Google)
Terry Taewoong Um (terry.t.um@gmail.com)
42
WHY IS DL SO SUCCESSFUL?
http://t-robotics.blogspot.kr/2015/05/deep-learning.html
• Pre-training with unsupervised learning
• Convolutional Neural Network
• Recurrent Neural Net
• GPGPU (parallel processing) & big data
• Advanced algorithms for optimization,
activation, regularization
• Huge research society
(Vision, Speech, NLP, Biology, etc.)
Terry Taewoong Um (terry.t.um@gmail.com)
43
UNSUPERVISED LEARNING
• How can we avoid pathologic local minima cases?
(1) First, pre-train the data with unsupervised learning method
and get a new representation
(2) Stack up this block structures
(3) Training each layer in end-to-end manner
(4) Fine tune the final structure with (ordinary) fully-connected NN
• Unsupervised learning method
- Restricted Boltzmann Machine (RBM)
→ Deep RBM, Deep Belief Network (DBN)
- Autoencoder
→ Deep Auto-encoder
http://goo.gl/QGJm5k
Autoencoder http://goo.gl/s6kmqY
Terry Taewoong Um (terry.t.um@gmail.com)
44
UNSUPERVISED LEARNING
“Convolutional deep belief networks for scalable unsupervised learning of hierarchical representation”, Lee et al., 2012
Terry Taewoong Um (terry.t.um@gmail.com)
45
CONVOLUTIONAL NN
• How can we deal with real images which is
much bigger than MNIST digit images?
- Use not fully-connected, but locally-connected NN
- Use convolutions to get various feature maps
- Abstract the results into higher layer by using pooling
- Fine tune with fully-connected NN
https://goo.gl/G7kBjI
https://goo.gl/Xswsbd
http://goo.gl/5OR5oH
Terry Taewoong Um (terry.t.um@gmail.com)
46
CONVOLUTIONAL NN
“Visualization and Understanding Convolutional Network”, Zeiler et al., 2012
Terry Taewoong Um (terry.t.um@gmail.com)
47
CONVNET + RNN
“Large-scale Video Classification with Convolutional Neural Network”,
A. Karpathy 2014, https://youtu.be/qrzQ_AB1DZk
Terry Taewoong Um (terry.t.um@gmail.com)
48
RECURRENT NEURAL NETWORK (RNN)
t-1 t t+1
[Neural Network] [Recurrent Neural Network]
http://www.dmi.usherb.ca/~larocheh/index_en.html
Terry Taewoong Um (terry.t.um@gmail.com)
49
RECURRENT NEURAL NETWORK (RNN)
[Neural Network] [Recurrent Neural Network]
back propagation
back propagation
through time
(BPTT)
• Vanishing gradient problem : Can’t have long memory!
“Training Recurrent Neural Networks, I. Sutskever, 2013
Terry Taewoong Um (terry.t.um@gmail.com)
50
RNN + LSTM
• Long Short-Term Memory (LSTM) (Hochreiter and Schmidhuber, 1997)
“Training Recurrent Neural Networks, I. Sutskever, 2013
Terry Taewoong Um (terry.t.um@gmail.com)
51
INTERESTING RESULTS FROM RNN
http://pail.unist.ac.kr/carpedm20/poet/
http://cs.stanford.edu/people/karpathy/deepimagesent/
“generating sequences with RNN”,
A.Graves, 2013
Terry Taewoong Um (terry.t.um@gmail.com)
52
WHY IS DL SO SUCCESSFUL?
http://t-robotics.blogspot.kr/2015/05/deep-learning.html
• Pre-training with unsupervised learning
• Convolutional Neural Network
• Recurrent Neural Net
• GPGPU (parallel processing) & big data
• Advanced algorithms for optimization,
activation, regularization
• Huge research society
(Vision, Speech, NLP, Biology, etc.)
Terry Taewoong Um (terry.t.um@gmail.com)
CONTENTS
53
Questions about Part 2?
Terry Taewoong Um (terry.t.um@gmail.com)
CONTENTS
54
3. Machine Learning in
Motion Analysis
Terry Taewoong Um (terry.t.um@gmail.com)
55
MOTION DATA
“츄리닝”, 이상신 국중록
Terry Taewoong Um (terry.t.um@gmail.com)
56
MOTION DATA
We need to know the state not only at time t
but also at time t-1, t-2, t-3, etc.
𝑓 = 𝑓(𝑥, 𝑡)
“츄리닝”, 이상신 국중록
Terry Taewoong Um (terry.t.um@gmail.com)
57
MOTION DATA
• Why do motion data need special treatment?
- In general, most machine learning techniques assume i.i.d. (independent
& identically distributed) sampling condition.
e.g.) coins tossing
- However, motion data is temporally & spatially correlated
http://goo.gl/LQulvcswing motion manipulability ellipsoid https://goo.gl/dHjFO9
Terry Taewoong Um (terry.t.um@gmail.com)
58
MOTION DATA
http://goo.gl/ll3sq6
We can infer the next state
based on the temporal &
spatial information
But, how can we exploit
those benefits in ML method?
Terry Taewoong Um (terry.t.um@gmail.com)
59
WHAT CAN WE DO WITH MOTION DATA?
• Learning the kinematic/dynamic model
• Motion segmentation
• Motion generation / synthesis
• Motion imitation (Imitation learning)
• Activity / Gesture recognition
TASKS
Data
• Motion capture data
• Vision Data
• Dynamic-level data
Applications
• Biomechanics
• Humanoid
• Animation
http://goo.gl/gFOVWL
Terry Taewoong Um (terry.t.um@gmail.com)
60
HIDDEN MARKOV MODEL (HMM)
Prob. of (n+1) state only depends on state at (n+1)
Terry Taewoong Um (terry.t.um@gmail.com)
61
LIMITATIONS OF HMM
1. Extract features (e.g. PCA)
2. Define the HMM structure (e.g. using GMM)
3. Train a separate HMM per class (Baum-Welch algorithm)
4. Evaluate probability under each HMM (Fwd/Bwd algorithm)
or 3. Choose most probable sequence (Viterbi algorithm)
- HMM handle discrete states only!
- HMM has short memory! (using just the previous state)
- HMM has limited expressive power!
- [Trend1] features-GMM → unsupervised learning methods
- [Trend2] features-GMM-HMM → recurrent neural network
• A common procedure of HMM for motion analysis
• Limitations & trend change in speech recognition area
Terry Taewoong Um (terry.t.um@gmail.com)
62
CAPTURE TEMPORAL INFORMATION
• 3D ConvNet
- “3D Convolutional Neural Network for
Human Action Recognition” (Ji et al., 2010)
- 3D convolution
- Activity recognition / Pose estimation from video
“Joint Training of a Convolutional Network
and a Graphical Model for Human Pose
Estimation”, Tompson et al., 2014
Terry Taewoong Um (terry.t.um@gmail.com)
63
CAPTURE TEMPORAL INFORMATION
• Recurrent Neural Network (RNN)
“Hierarchical Recurrent Neural Network for Skeleton Based Action Recognition”, Y. Du et al., 2015
• However, how can we capture the
spatial information about motions?
Terry Taewoong Um (terry.t.um@gmail.com)
64
CHALLENGES
We should connect the geometric information with deep neural network!
• The link transformation from the i-1 th link to the i th link
• Forward Kinematics
constant, Mvariable, 𝜃
c.f.)
𝑋𝑖−1,𝑖 = 𝑅𝑜𝑡 𝑧, 𝜃𝑖 𝑇𝑟𝑎𝑛𝑠 𝑧, 𝑑𝑖 𝑇𝑟𝑎𝑛𝑠 𝑥, 𝑎𝑖 𝑅𝑜𝑡 𝑧, 𝛼𝑖 = 𝑒[𝐴 𝑖]𝜃 𝑖 𝑀𝑖−1,𝑖
𝑋0,𝑛 = 𝑒[𝐴1]𝜃1 𝑀0,1 𝑒[𝐴2]𝜃2 𝑀1,2 ⋯ 𝑒 𝐴 𝑛 𝜃 𝑛 𝑀 𝑛−1,𝑛
= 𝑒[𝑆1]𝜃1 𝑒[𝑆2]𝜃2 ⋯ 𝑒[𝑆 𝑛]𝜃 𝑛 𝑀0,𝑛
𝑆𝑖 = 𝐴𝑑 𝑀01⋯𝑀 𝑖−2,𝑖−1
𝐴𝑖 , 𝑖 = 1, ⋯ , 𝑛
propagated forces
external force acting
on the ith body where
• Newton-Euler formulation for inverse dynamics
Lie group & Lie algebra,
http://goo.gl/uqilDV
Terry Taewoong Um (terry.t.um@gmail.com)
65
CHALLENGES
https://www.youtube.com/watch?v=oxA2O-tHftI
Terry Taewoong Um (terry.t.um@gmail.com)
66
Thank you

More Related Content

What's hot

CNN Machine learning DeepLearning
CNN Machine learning DeepLearningCNN Machine learning DeepLearning
CNN Machine learning DeepLearningAbhishek Sharma
 
Machine Learning project presentation
Machine Learning project presentationMachine Learning project presentation
Machine Learning project presentationRamandeep Kaur Bagri
 
Introduction to Artificial Neural Networks
Introduction to Artificial Neural NetworksIntroduction to Artificial Neural Networks
Introduction to Artificial Neural NetworksAdri Jovin
 
Search problems in Artificial Intelligence
Search problems in Artificial IntelligenceSearch problems in Artificial Intelligence
Search problems in Artificial Intelligenceananth
 
Deep Learning - CNN and RNN
Deep Learning - CNN and RNNDeep Learning - CNN and RNN
Deep Learning - CNN and RNNAshray Bhandare
 
Convolutional Neural Network (CNN)
Convolutional Neural Network (CNN)Convolutional Neural Network (CNN)
Convolutional Neural Network (CNN)Muhammad Haroon
 
Artificial intelligence NEURAL NETWORKS
Artificial intelligence NEURAL NETWORKSArtificial intelligence NEURAL NETWORKS
Artificial intelligence NEURAL NETWORKSREHMAT ULLAH
 
Convolution Neural Network (CNN)
Convolution Neural Network (CNN)Convolution Neural Network (CNN)
Convolution Neural Network (CNN)Suraj Aavula
 
What Is Deep Learning? | Introduction to Deep Learning | Deep Learning Tutori...
What Is Deep Learning? | Introduction to Deep Learning | Deep Learning Tutori...What Is Deep Learning? | Introduction to Deep Learning | Deep Learning Tutori...
What Is Deep Learning? | Introduction to Deep Learning | Deep Learning Tutori...Simplilearn
 
Deep Learning Explained
Deep Learning ExplainedDeep Learning Explained
Deep Learning ExplainedMelanie Swan
 
Artificial Intelligence with Python | Edureka
Artificial Intelligence with Python | EdurekaArtificial Intelligence with Python | Edureka
Artificial Intelligence with Python | EdurekaEdureka!
 
Deep learning - what is it and why now?
Deep learning - what is it and why now?Deep learning - what is it and why now?
Deep learning - what is it and why now?Natalia Konstantinova
 
Hill Climbing Algorithm in Artificial Intelligence
Hill Climbing Algorithm in Artificial IntelligenceHill Climbing Algorithm in Artificial Intelligence
Hill Climbing Algorithm in Artificial IntelligenceBharat Bhushan
 
AI_Session 7 Greedy Best first search algorithm.pptx
AI_Session 7 Greedy Best first search algorithm.pptxAI_Session 7 Greedy Best first search algorithm.pptx
AI_Session 7 Greedy Best first search algorithm.pptxAsst.prof M.Gokilavani
 
An introduction to Deep Learning
An introduction to Deep LearningAn introduction to Deep Learning
An introduction to Deep LearningJulien SIMON
 
Vanishing & Exploding Gradients
Vanishing & Exploding GradientsVanishing & Exploding Gradients
Vanishing & Exploding GradientsSiddharth Vij
 
Artificial intelligence and knowledge representation
Artificial intelligence and knowledge representationArtificial intelligence and knowledge representation
Artificial intelligence and knowledge representationSajan Sahu
 

What's hot (20)

CNN Machine learning DeepLearning
CNN Machine learning DeepLearningCNN Machine learning DeepLearning
CNN Machine learning DeepLearning
 
Machine Learning project presentation
Machine Learning project presentationMachine Learning project presentation
Machine Learning project presentation
 
Introduction to Artificial Neural Networks
Introduction to Artificial Neural NetworksIntroduction to Artificial Neural Networks
Introduction to Artificial Neural Networks
 
Introduction to Deep Learning
Introduction to Deep Learning Introduction to Deep Learning
Introduction to Deep Learning
 
Search problems in Artificial Intelligence
Search problems in Artificial IntelligenceSearch problems in Artificial Intelligence
Search problems in Artificial Intelligence
 
Deep Learning - CNN and RNN
Deep Learning - CNN and RNNDeep Learning - CNN and RNN
Deep Learning - CNN and RNN
 
Convolutional Neural Network (CNN)
Convolutional Neural Network (CNN)Convolutional Neural Network (CNN)
Convolutional Neural Network (CNN)
 
Artificial intelligence NEURAL NETWORKS
Artificial intelligence NEURAL NETWORKSArtificial intelligence NEURAL NETWORKS
Artificial intelligence NEURAL NETWORKS
 
Convolution Neural Network (CNN)
Convolution Neural Network (CNN)Convolution Neural Network (CNN)
Convolution Neural Network (CNN)
 
What Is Deep Learning? | Introduction to Deep Learning | Deep Learning Tutori...
What Is Deep Learning? | Introduction to Deep Learning | Deep Learning Tutori...What Is Deep Learning? | Introduction to Deep Learning | Deep Learning Tutori...
What Is Deep Learning? | Introduction to Deep Learning | Deep Learning Tutori...
 
Deep Learning Explained
Deep Learning ExplainedDeep Learning Explained
Deep Learning Explained
 
Artificial Intelligence with Python | Edureka
Artificial Intelligence with Python | EdurekaArtificial Intelligence with Python | Edureka
Artificial Intelligence with Python | Edureka
 
Deep learning - what is it and why now?
Deep learning - what is it and why now?Deep learning - what is it and why now?
Deep learning - what is it and why now?
 
Hill Climbing Algorithm in Artificial Intelligence
Hill Climbing Algorithm in Artificial IntelligenceHill Climbing Algorithm in Artificial Intelligence
Hill Climbing Algorithm in Artificial Intelligence
 
AI_Session 7 Greedy Best first search algorithm.pptx
AI_Session 7 Greedy Best first search algorithm.pptxAI_Session 7 Greedy Best first search algorithm.pptx
AI_Session 7 Greedy Best first search algorithm.pptx
 
An introduction to Deep Learning
An introduction to Deep LearningAn introduction to Deep Learning
An introduction to Deep Learning
 
machine learning
machine learningmachine learning
machine learning
 
Vanishing & Exploding Gradients
Vanishing & Exploding GradientsVanishing & Exploding Gradients
Vanishing & Exploding Gradients
 
Artificial intelligence and knowledge representation
Artificial intelligence and knowledge representationArtificial intelligence and knowledge representation
Artificial intelligence and knowledge representation
 
One shot learning
One shot learningOne shot learning
One shot learning
 

Viewers also liked

기계학습(Machine learning) 입문하기
기계학습(Machine learning) 입문하기기계학습(Machine learning) 입문하기
기계학습(Machine learning) 입문하기Terry Taewoong Um
 
Human Motion Forecasting (Generation) with RNNs
Human Motion Forecasting (Generation) with RNNsHuman Motion Forecasting (Generation) with RNNs
Human Motion Forecasting (Generation) with RNNsTerry Taewoong Um
 
Understanding Black-box Predictions via Influence Functions (2017)
Understanding Black-box Predictions via Influence Functions (2017)Understanding Black-box Predictions via Influence Functions (2017)
Understanding Black-box Predictions via Influence Functions (2017)Terry Taewoong Um
 
Learning with side information through modality hallucination (2016)
Learning with side information through modality hallucination (2016)Learning with side information through modality hallucination (2016)
Learning with side information through modality hallucination (2016)Terry Taewoong Um
 
Deformable Convolutional Network (2017)
Deformable Convolutional Network (2017)Deformable Convolutional Network (2017)
Deformable Convolutional Network (2017)Terry Taewoong Um
 
기계학습 / 딥러닝이란 무엇인가
기계학습 / 딥러닝이란 무엇인가기계학습 / 딥러닝이란 무엇인가
기계학습 / 딥러닝이란 무엇인가Yongha Kim
 
Lie Group Formulation for Robot Mechanics
Lie Group Formulation for Robot MechanicsLie Group Formulation for Robot Mechanics
Lie Group Formulation for Robot MechanicsTerry Taewoong Um
 
Lie Group Formulation for Robot Mechanics
Lie Group Formulation for Robot MechanicsLie Group Formulation for Robot Mechanics
Lie Group Formulation for Robot MechanicsTerry Taewoong Um
 
쫄지말자딥러닝2 - CNN RNN 포함버전
쫄지말자딥러닝2 - CNN RNN 포함버전쫄지말자딥러닝2 - CNN RNN 포함버전
쫄지말자딥러닝2 - CNN RNN 포함버전Modulabs
 
알파고 (바둑 인공지능)의 작동 원리
알파고 (바둑 인공지능)의 작동 원리알파고 (바둑 인공지능)의 작동 원리
알파고 (바둑 인공지능)의 작동 원리Shane (Seungwhan) Moon
 
[2A4]DeepLearningAtNAVER
[2A4]DeepLearningAtNAVER[2A4]DeepLearningAtNAVER
[2A4]DeepLearningAtNAVERNAVER D2
 
알파고 해부하기 1부
알파고 해부하기 1부알파고 해부하기 1부
알파고 해부하기 1부Donghun Lee
 
[모두의연구소] 쫄지말자딥러닝
[모두의연구소] 쫄지말자딥러닝[모두의연구소] 쫄지말자딥러닝
[모두의연구소] 쫄지말자딥러닝Modulabs
 
인공 신경망 구현에 관한 간단한 설명
인공 신경망 구현에 관한 간단한 설명인공 신경망 구현에 관한 간단한 설명
인공 신경망 구현에 관한 간단한 설명Woonghee Lee
 
R 프로그래밍 기본 문법
R 프로그래밍 기본 문법R 프로그래밍 기본 문법
R 프로그래밍 기본 문법Terry Cho
 
머신 러닝 입문 #1-머신러닝 소개와 kNN 소개
머신 러닝 입문 #1-머신러닝 소개와 kNN 소개머신 러닝 입문 #1-머신러닝 소개와 kNN 소개
머신 러닝 입문 #1-머신러닝 소개와 kNN 소개Terry Cho
 
Introduction to Deep Learning with TensorFlow
Introduction to Deep Learning with TensorFlowIntroduction to Deep Learning with TensorFlow
Introduction to Deep Learning with TensorFlowTerry Taewoong Um
 
인공지능, 기계학습 그리고 딥러닝
인공지능, 기계학습 그리고 딥러닝인공지능, 기계학습 그리고 딥러닝
인공지능, 기계학습 그리고 딥러닝Jinwon Lee
 
Large Scale Deep Learning with TensorFlow
Large Scale Deep Learning with TensorFlow Large Scale Deep Learning with TensorFlow
Large Scale Deep Learning with TensorFlow Jen Aman
 
Deview deep learning-김정희
Deview deep learning-김정희Deview deep learning-김정희
Deview deep learning-김정희NAVER D2
 

Viewers also liked (20)

기계학습(Machine learning) 입문하기
기계학습(Machine learning) 입문하기기계학습(Machine learning) 입문하기
기계학습(Machine learning) 입문하기
 
Human Motion Forecasting (Generation) with RNNs
Human Motion Forecasting (Generation) with RNNsHuman Motion Forecasting (Generation) with RNNs
Human Motion Forecasting (Generation) with RNNs
 
Understanding Black-box Predictions via Influence Functions (2017)
Understanding Black-box Predictions via Influence Functions (2017)Understanding Black-box Predictions via Influence Functions (2017)
Understanding Black-box Predictions via Influence Functions (2017)
 
Learning with side information through modality hallucination (2016)
Learning with side information through modality hallucination (2016)Learning with side information through modality hallucination (2016)
Learning with side information through modality hallucination (2016)
 
Deformable Convolutional Network (2017)
Deformable Convolutional Network (2017)Deformable Convolutional Network (2017)
Deformable Convolutional Network (2017)
 
기계학습 / 딥러닝이란 무엇인가
기계학습 / 딥러닝이란 무엇인가기계학습 / 딥러닝이란 무엇인가
기계학습 / 딥러닝이란 무엇인가
 
Lie Group Formulation for Robot Mechanics
Lie Group Formulation for Robot MechanicsLie Group Formulation for Robot Mechanics
Lie Group Formulation for Robot Mechanics
 
Lie Group Formulation for Robot Mechanics
Lie Group Formulation for Robot MechanicsLie Group Formulation for Robot Mechanics
Lie Group Formulation for Robot Mechanics
 
쫄지말자딥러닝2 - CNN RNN 포함버전
쫄지말자딥러닝2 - CNN RNN 포함버전쫄지말자딥러닝2 - CNN RNN 포함버전
쫄지말자딥러닝2 - CNN RNN 포함버전
 
알파고 (바둑 인공지능)의 작동 원리
알파고 (바둑 인공지능)의 작동 원리알파고 (바둑 인공지능)의 작동 원리
알파고 (바둑 인공지능)의 작동 원리
 
[2A4]DeepLearningAtNAVER
[2A4]DeepLearningAtNAVER[2A4]DeepLearningAtNAVER
[2A4]DeepLearningAtNAVER
 
알파고 해부하기 1부
알파고 해부하기 1부알파고 해부하기 1부
알파고 해부하기 1부
 
[모두의연구소] 쫄지말자딥러닝
[모두의연구소] 쫄지말자딥러닝[모두의연구소] 쫄지말자딥러닝
[모두의연구소] 쫄지말자딥러닝
 
인공 신경망 구현에 관한 간단한 설명
인공 신경망 구현에 관한 간단한 설명인공 신경망 구현에 관한 간단한 설명
인공 신경망 구현에 관한 간단한 설명
 
R 프로그래밍 기본 문법
R 프로그래밍 기본 문법R 프로그래밍 기본 문법
R 프로그래밍 기본 문법
 
머신 러닝 입문 #1-머신러닝 소개와 kNN 소개
머신 러닝 입문 #1-머신러닝 소개와 kNN 소개머신 러닝 입문 #1-머신러닝 소개와 kNN 소개
머신 러닝 입문 #1-머신러닝 소개와 kNN 소개
 
Introduction to Deep Learning with TensorFlow
Introduction to Deep Learning with TensorFlowIntroduction to Deep Learning with TensorFlow
Introduction to Deep Learning with TensorFlow
 
인공지능, 기계학습 그리고 딥러닝
인공지능, 기계학습 그리고 딥러닝인공지능, 기계학습 그리고 딥러닝
인공지능, 기계학습 그리고 딥러닝
 
Large Scale Deep Learning with TensorFlow
Large Scale Deep Learning with TensorFlow Large Scale Deep Learning with TensorFlow
Large Scale Deep Learning with TensorFlow
 
Deview deep learning-김정희
Deview deep learning-김정희Deview deep learning-김정희
Deview deep learning-김정희
 

Similar to Introduction to Machine Learning and Deep Learning

Deep learning (Machine learning) tutorial for beginners
Deep learning (Machine learning) tutorial for beginnersDeep learning (Machine learning) tutorial for beginners
Deep learning (Machine learning) tutorial for beginnersTerry Taewoong Um
 
Machine Learning, Financial Engineering and Quantitative Investing
Machine Learning, Financial Engineering and Quantitative InvestingMachine Learning, Financial Engineering and Quantitative Investing
Machine Learning, Financial Engineering and Quantitative InvestingShengyuan Wang Steven
 
Introduction to Algorithms
Introduction to AlgorithmsIntroduction to Algorithms
Introduction to AlgorithmsVenkatesh Iyer
 
Data Structure: Algorithm and analysis
Data Structure: Algorithm and analysisData Structure: Algorithm and analysis
Data Structure: Algorithm and analysisDr. Rajdeep Chatterjee
 
Week 2 - ML models and Linear Regression.pptx
Week 2 - ML models and Linear Regression.pptxWeek 2 - ML models and Linear Regression.pptx
Week 2 - ML models and Linear Regression.pptxHafizAliHummad
 
Data Structures and Algorithm - Week 5 - AVL Trees
Data Structures and Algorithm - Week 5 - AVL TreesData Structures and Algorithm - Week 5 - AVL Trees
Data Structures and Algorithm - Week 5 - AVL TreesFerdin Joe John Joseph PhD
 
Learning to Reconstruct
Learning to ReconstructLearning to Reconstruct
Learning to ReconstructJonas Adler
 
Lec 1 Ds
Lec 1 DsLec 1 Ds
Lec 1 DsQundeel
 
Data Structure
Data StructureData Structure
Data Structuresheraz1
 
Lec 1 Ds
Lec 1 DsLec 1 Ds
Lec 1 DsQundeel
 
Deep Reinforcement Learning Through Policy Optimization, John Schulman, OpenAI
Deep Reinforcement Learning Through Policy Optimization, John Schulman, OpenAIDeep Reinforcement Learning Through Policy Optimization, John Schulman, OpenAI
Deep Reinforcement Learning Through Policy Optimization, John Schulman, OpenAIJack Clark
 
Machine learning (1)
Machine learning (1)Machine learning (1)
Machine learning (1)NYversity
 
Scala 3 Is Coming: Martin Odersky Shares What To Know
Scala 3 Is Coming: Martin Odersky Shares What To KnowScala 3 Is Coming: Martin Odersky Shares What To Know
Scala 3 Is Coming: Martin Odersky Shares What To KnowLightbend
 
Intro to Approximate Bayesian Computation (ABC)
Intro to Approximate Bayesian Computation (ABC)Intro to Approximate Bayesian Computation (ABC)
Intro to Approximate Bayesian Computation (ABC)Umberto Picchini
 
Introduction to computing Processing and performance.pdf
Introduction to computing Processing and performance.pdfIntroduction to computing Processing and performance.pdf
Introduction to computing Processing and performance.pdfTulasiramKandula1
 
Chap 8. Optimization for training deep models
Chap 8. Optimization for training deep modelsChap 8. Optimization for training deep models
Chap 8. Optimization for training deep modelsYoung-Geun Choi
 
Abductive commonsense reasoning
Abductive commonsense reasoningAbductive commonsense reasoning
Abductive commonsense reasoningSan Kim
 

Similar to Introduction to Machine Learning and Deep Learning (20)

Deep learning (Machine learning) tutorial for beginners
Deep learning (Machine learning) tutorial for beginnersDeep learning (Machine learning) tutorial for beginners
Deep learning (Machine learning) tutorial for beginners
 
Machine Learning, Financial Engineering and Quantitative Investing
Machine Learning, Financial Engineering and Quantitative InvestingMachine Learning, Financial Engineering and Quantitative Investing
Machine Learning, Financial Engineering and Quantitative Investing
 
Introduction to Algorithms
Introduction to AlgorithmsIntroduction to Algorithms
Introduction to Algorithms
 
Data Structure: Algorithm and analysis
Data Structure: Algorithm and analysisData Structure: Algorithm and analysis
Data Structure: Algorithm and analysis
 
Chapter two
Chapter twoChapter two
Chapter two
 
Week 2 - ML models and Linear Regression.pptx
Week 2 - ML models and Linear Regression.pptxWeek 2 - ML models and Linear Regression.pptx
Week 2 - ML models and Linear Regression.pptx
 
Pythonic Math
Pythonic MathPythonic Math
Pythonic Math
 
Data Structures and Algorithm - Week 5 - AVL Trees
Data Structures and Algorithm - Week 5 - AVL TreesData Structures and Algorithm - Week 5 - AVL Trees
Data Structures and Algorithm - Week 5 - AVL Trees
 
Learning to Reconstruct
Learning to ReconstructLearning to Reconstruct
Learning to Reconstruct
 
Lec 1 Ds
Lec 1 DsLec 1 Ds
Lec 1 Ds
 
Data Structure
Data StructureData Structure
Data Structure
 
Lec 1 Ds
Lec 1 DsLec 1 Ds
Lec 1 Ds
 
Deep Reinforcement Learning Through Policy Optimization, John Schulman, OpenAI
Deep Reinforcement Learning Through Policy Optimization, John Schulman, OpenAIDeep Reinforcement Learning Through Policy Optimization, John Schulman, OpenAI
Deep Reinforcement Learning Through Policy Optimization, John Schulman, OpenAI
 
Machine learning (1)
Machine learning (1)Machine learning (1)
Machine learning (1)
 
Scala 3 Is Coming: Martin Odersky Shares What To Know
Scala 3 Is Coming: Martin Odersky Shares What To KnowScala 3 Is Coming: Martin Odersky Shares What To Know
Scala 3 Is Coming: Martin Odersky Shares What To Know
 
Intro to Approximate Bayesian Computation (ABC)
Intro to Approximate Bayesian Computation (ABC)Intro to Approximate Bayesian Computation (ABC)
Intro to Approximate Bayesian Computation (ABC)
 
Computer Network Assignment Help
Computer Network Assignment HelpComputer Network Assignment Help
Computer Network Assignment Help
 
Introduction to computing Processing and performance.pdf
Introduction to computing Processing and performance.pdfIntroduction to computing Processing and performance.pdf
Introduction to computing Processing and performance.pdf
 
Chap 8. Optimization for training deep models
Chap 8. Optimization for training deep modelsChap 8. Optimization for training deep models
Chap 8. Optimization for training deep models
 
Abductive commonsense reasoning
Abductive commonsense reasoningAbductive commonsense reasoning
Abductive commonsense reasoning
 

More from Terry Taewoong Um

#44. KAIST에서 "대학 유죄"를 외치다: ART Lab의 도전
#44. KAIST에서 "대학 유죄"를 외치다: ART Lab의 도전#44. KAIST에서 "대학 유죄"를 외치다: ART Lab의 도전
#44. KAIST에서 "대학 유죄"를 외치다: ART Lab의 도전Terry Taewoong Um
 
A brief introduction to OCR (Optical character recognition)
A brief introduction to OCR (Optical character recognition)A brief introduction to OCR (Optical character recognition)
A brief introduction to OCR (Optical character recognition)Terry Taewoong Um
 
Deep Reinforcement Learning in a Handful of Trials using Probabilistic Dynami...
Deep Reinforcement Learning in a Handful of Trials using Probabilistic Dynami...Deep Reinforcement Learning in a Handful of Trials using Probabilistic Dynami...
Deep Reinforcement Learning in a Handful of Trials using Probabilistic Dynami...Terry Taewoong Um
 
인공지능의 사회정의의 편이 될 수 있을까? (인공지능과 법)
인공지능의 사회정의의 편이 될 수 있을까? (인공지능과 법)인공지능의 사회정의의 편이 될 수 있을까? (인공지능과 법)
인공지능의 사회정의의 편이 될 수 있을까? (인공지능과 법)Terry Taewoong Um
 
Deep Variational Bayes Filters (2017)
Deep Variational Bayes Filters (2017)Deep Variational Bayes Filters (2017)
Deep Variational Bayes Filters (2017)Terry Taewoong Um
 
On Calibration of Modern Neural Networks (2017)
On Calibration of Modern Neural Networks (2017)On Calibration of Modern Neural Networks (2017)
On Calibration of Modern Neural Networks (2017)Terry Taewoong Um
 
Deep Learning: A Critical Appraisal (2018)
Deep Learning: A Critical Appraisal (2018)Deep Learning: A Critical Appraisal (2018)
Deep Learning: A Critical Appraisal (2018)Terry Taewoong Um
 
About Two Motion Planning Papers
About Two Motion Planning PapersAbout Two Motion Planning Papers
About Two Motion Planning PapersTerry Taewoong Um
 
로봇과 인공지능, 그리고 미래의 노동
로봇과 인공지능, 그리고 미래의 노동로봇과 인공지능, 그리고 미래의 노동
로봇과 인공지능, 그리고 미래의 노동Terry Taewoong Um
 

More from Terry Taewoong Um (9)

#44. KAIST에서 "대학 유죄"를 외치다: ART Lab의 도전
#44. KAIST에서 "대학 유죄"를 외치다: ART Lab의 도전#44. KAIST에서 "대학 유죄"를 외치다: ART Lab의 도전
#44. KAIST에서 "대학 유죄"를 외치다: ART Lab의 도전
 
A brief introduction to OCR (Optical character recognition)
A brief introduction to OCR (Optical character recognition)A brief introduction to OCR (Optical character recognition)
A brief introduction to OCR (Optical character recognition)
 
Deep Reinforcement Learning in a Handful of Trials using Probabilistic Dynami...
Deep Reinforcement Learning in a Handful of Trials using Probabilistic Dynami...Deep Reinforcement Learning in a Handful of Trials using Probabilistic Dynami...
Deep Reinforcement Learning in a Handful of Trials using Probabilistic Dynami...
 
인공지능의 사회정의의 편이 될 수 있을까? (인공지능과 법)
인공지능의 사회정의의 편이 될 수 있을까? (인공지능과 법)인공지능의 사회정의의 편이 될 수 있을까? (인공지능과 법)
인공지능의 사회정의의 편이 될 수 있을까? (인공지능과 법)
 
Deep Variational Bayes Filters (2017)
Deep Variational Bayes Filters (2017)Deep Variational Bayes Filters (2017)
Deep Variational Bayes Filters (2017)
 
On Calibration of Modern Neural Networks (2017)
On Calibration of Modern Neural Networks (2017)On Calibration of Modern Neural Networks (2017)
On Calibration of Modern Neural Networks (2017)
 
Deep Learning: A Critical Appraisal (2018)
Deep Learning: A Critical Appraisal (2018)Deep Learning: A Critical Appraisal (2018)
Deep Learning: A Critical Appraisal (2018)
 
About Two Motion Planning Papers
About Two Motion Planning PapersAbout Two Motion Planning Papers
About Two Motion Planning Papers
 
로봇과 인공지능, 그리고 미래의 노동
로봇과 인공지능, 그리고 미래의 노동로봇과 인공지능, 그리고 미래의 노동
로봇과 인공지능, 그리고 미래의 노동
 

Recently uploaded

Call Us ≽ 8377877756 ≼ Call Girls In Shastri Nagar (Delhi)
Call Us ≽ 8377877756 ≼ Call Girls In Shastri Nagar (Delhi)Call Us ≽ 8377877756 ≼ Call Girls In Shastri Nagar (Delhi)
Call Us ≽ 8377877756 ≼ Call Girls In Shastri Nagar (Delhi)dollysharma2066
 
Introduction to Machine Learning Unit-3 for II MECH
Introduction to Machine Learning Unit-3 for II MECHIntroduction to Machine Learning Unit-3 for II MECH
Introduction to Machine Learning Unit-3 for II MECHC Sai Kiran
 
Class 1 | NFPA 72 | Overview Fire Alarm System
Class 1 | NFPA 72 | Overview Fire Alarm SystemClass 1 | NFPA 72 | Overview Fire Alarm System
Class 1 | NFPA 72 | Overview Fire Alarm Systemirfanmechengr
 
Why does (not) Kafka need fsync: Eliminating tail latency spikes caused by fsync
Why does (not) Kafka need fsync: Eliminating tail latency spikes caused by fsyncWhy does (not) Kafka need fsync: Eliminating tail latency spikes caused by fsync
Why does (not) Kafka need fsync: Eliminating tail latency spikes caused by fsyncssuser2ae721
 
Piping Basic stress analysis by engineering
Piping Basic stress analysis by engineeringPiping Basic stress analysis by engineering
Piping Basic stress analysis by engineeringJuanCarlosMorales19600
 
complete construction, environmental and economics information of biomass com...
complete construction, environmental and economics information of biomass com...complete construction, environmental and economics information of biomass com...
complete construction, environmental and economics information of biomass com...asadnawaz62
 
Concrete Mix Design - IS 10262-2019 - .pptx
Concrete Mix Design - IS 10262-2019 - .pptxConcrete Mix Design - IS 10262-2019 - .pptx
Concrete Mix Design - IS 10262-2019 - .pptxKartikeyaDwivedi3
 
Call Girls Narol 7397865700 Independent Call Girls
Call Girls Narol 7397865700 Independent Call GirlsCall Girls Narol 7397865700 Independent Call Girls
Call Girls Narol 7397865700 Independent Call Girlsssuser7cb4ff
 
Architect Hassan Khalil Portfolio for 2024
Architect Hassan Khalil Portfolio for 2024Architect Hassan Khalil Portfolio for 2024
Architect Hassan Khalil Portfolio for 2024hassan khalil
 
Correctly Loading Incremental Data at Scale
Correctly Loading Incremental Data at ScaleCorrectly Loading Incremental Data at Scale
Correctly Loading Incremental Data at ScaleAlluxio, Inc.
 
welding defects observed during the welding
welding defects observed during the weldingwelding defects observed during the welding
welding defects observed during the weldingMuhammadUzairLiaqat
 
Work Experience-Dalton Park.pptxfvvvvvvv
Work Experience-Dalton Park.pptxfvvvvvvvWork Experience-Dalton Park.pptxfvvvvvvv
Work Experience-Dalton Park.pptxfvvvvvvvLewisJB
 
Past, Present and Future of Generative AI
Past, Present and Future of Generative AIPast, Present and Future of Generative AI
Past, Present and Future of Generative AIabhishek36461
 
UNIT III ANALOG ELECTRONICS (BASIC ELECTRONICS)
UNIT III ANALOG ELECTRONICS (BASIC ELECTRONICS)UNIT III ANALOG ELECTRONICS (BASIC ELECTRONICS)
UNIT III ANALOG ELECTRONICS (BASIC ELECTRONICS)Dr SOUNDIRARAJ N
 
Gurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort service
Gurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort serviceGurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort service
Gurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort servicejennyeacort
 
US Department of Education FAFSA Week of Action
US Department of Education FAFSA Week of ActionUS Department of Education FAFSA Week of Action
US Department of Education FAFSA Week of ActionMebane Rash
 
lifi-technology with integration of IOT.pptx
lifi-technology with integration of IOT.pptxlifi-technology with integration of IOT.pptx
lifi-technology with integration of IOT.pptxsomshekarkn64
 

Recently uploaded (20)

Call Us ≽ 8377877756 ≼ Call Girls In Shastri Nagar (Delhi)
Call Us ≽ 8377877756 ≼ Call Girls In Shastri Nagar (Delhi)Call Us ≽ 8377877756 ≼ Call Girls In Shastri Nagar (Delhi)
Call Us ≽ 8377877756 ≼ Call Girls In Shastri Nagar (Delhi)
 
Introduction to Machine Learning Unit-3 for II MECH
Introduction to Machine Learning Unit-3 for II MECHIntroduction to Machine Learning Unit-3 for II MECH
Introduction to Machine Learning Unit-3 for II MECH
 
Class 1 | NFPA 72 | Overview Fire Alarm System
Class 1 | NFPA 72 | Overview Fire Alarm SystemClass 1 | NFPA 72 | Overview Fire Alarm System
Class 1 | NFPA 72 | Overview Fire Alarm System
 
Why does (not) Kafka need fsync: Eliminating tail latency spikes caused by fsync
Why does (not) Kafka need fsync: Eliminating tail latency spikes caused by fsyncWhy does (not) Kafka need fsync: Eliminating tail latency spikes caused by fsync
Why does (not) Kafka need fsync: Eliminating tail latency spikes caused by fsync
 
Piping Basic stress analysis by engineering
Piping Basic stress analysis by engineeringPiping Basic stress analysis by engineering
Piping Basic stress analysis by engineering
 
9953056974 Call Girls In South Ex, Escorts (Delhi) NCR.pdf
9953056974 Call Girls In South Ex, Escorts (Delhi) NCR.pdf9953056974 Call Girls In South Ex, Escorts (Delhi) NCR.pdf
9953056974 Call Girls In South Ex, Escorts (Delhi) NCR.pdf
 
complete construction, environmental and economics information of biomass com...
complete construction, environmental and economics information of biomass com...complete construction, environmental and economics information of biomass com...
complete construction, environmental and economics information of biomass com...
 
Concrete Mix Design - IS 10262-2019 - .pptx
Concrete Mix Design - IS 10262-2019 - .pptxConcrete Mix Design - IS 10262-2019 - .pptx
Concrete Mix Design - IS 10262-2019 - .pptx
 
Design and analysis of solar grass cutter.pdf
Design and analysis of solar grass cutter.pdfDesign and analysis of solar grass cutter.pdf
Design and analysis of solar grass cutter.pdf
 
Call Girls Narol 7397865700 Independent Call Girls
Call Girls Narol 7397865700 Independent Call GirlsCall Girls Narol 7397865700 Independent Call Girls
Call Girls Narol 7397865700 Independent Call Girls
 
Architect Hassan Khalil Portfolio for 2024
Architect Hassan Khalil Portfolio for 2024Architect Hassan Khalil Portfolio for 2024
Architect Hassan Khalil Portfolio for 2024
 
Correctly Loading Incremental Data at Scale
Correctly Loading Incremental Data at ScaleCorrectly Loading Incremental Data at Scale
Correctly Loading Incremental Data at Scale
 
welding defects observed during the welding
welding defects observed during the weldingwelding defects observed during the welding
welding defects observed during the welding
 
Exploring_Network_Security_with_JA3_by_Rakesh Seal.pptx
Exploring_Network_Security_with_JA3_by_Rakesh Seal.pptxExploring_Network_Security_with_JA3_by_Rakesh Seal.pptx
Exploring_Network_Security_with_JA3_by_Rakesh Seal.pptx
 
Work Experience-Dalton Park.pptxfvvvvvvv
Work Experience-Dalton Park.pptxfvvvvvvvWork Experience-Dalton Park.pptxfvvvvvvv
Work Experience-Dalton Park.pptxfvvvvvvv
 
Past, Present and Future of Generative AI
Past, Present and Future of Generative AIPast, Present and Future of Generative AI
Past, Present and Future of Generative AI
 
UNIT III ANALOG ELECTRONICS (BASIC ELECTRONICS)
UNIT III ANALOG ELECTRONICS (BASIC ELECTRONICS)UNIT III ANALOG ELECTRONICS (BASIC ELECTRONICS)
UNIT III ANALOG ELECTRONICS (BASIC ELECTRONICS)
 
Gurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort service
Gurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort serviceGurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort service
Gurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort service
 
US Department of Education FAFSA Week of Action
US Department of Education FAFSA Week of ActionUS Department of Education FAFSA Week of Action
US Department of Education FAFSA Week of Action
 
lifi-technology with integration of IOT.pptx
lifi-technology with integration of IOT.pptxlifi-technology with integration of IOT.pptx
lifi-technology with integration of IOT.pptx
 

Introduction to Machine Learning and Deep Learning

  • 1. Terry Taewoong Um (terry.t.um@gmail.com) University of Waterloo Department of Electrical & Computer Engineering Terry Taewoong Um MACHINE LEARNING, DEEP LEARNING, AND MOTION ANALYSIS 1
  • 2. Terry Taewoong Um (terry.t.um@gmail.com) CAUTION • I cannot explain everything • You cannot get every details 2 • Try to get a big picture • Get some useful keywords • Connect with your research
  • 3. Terry Taewoong Um (terry.t.um@gmail.com) CONTENTS 1. What is Machine Learning? (Part 1 Q & A) 2. What is Deep Learning? (Part 2 Q & A) 3. Machine Learning in Motion Analysis (Part 3 Q & A) 3
  • 4. Terry Taewoong Um (terry.t.um@gmail.com) CONTENTS 4 1. What is Machine Learning?
  • 5. Terry Taewoong Um (terry.t.um@gmail.com) WHAT IS MACHINE LEARNING? "A computer program is said to learn from experience E with respect to some class of tasks T and performance measure P, if its performance at tasks in T, as measured by P, improves with experience E“ – T. Michell (1997) Example: A program for soccer tactics 5 T : Win the game P : Goals E : (x) Players’ movements (y) Evaluation
  • 6. Terry Taewoong Um (terry.t.um@gmail.com) WHAT IS MACHINE LEARNING? 6 “Toward learning robot table tennis”, J. Peters et al. (2012) https://youtu.be/SH3bADiB7uQ "A computer program is said to learn from experience E with respect to some class of tasks T and performance measure P, if its performance at tasks in T, as measured by P, improves with experience E“ – T. Michell (1997)
  • 7. Terry Taewoong Um (terry.t.um@gmail.com) TASKS 7 classification discrete target values x : pixels (28*28) y : 0,1, 2,3,…,9 regression real target values x ∈ (0,100) y : 0,1, 2,3,…,9 clustering no target values x ∈ (-3,3)×(-3,3) "A computer program is said to learn from experience E with respect to some class of tasks T and performance measure P, if its performance at tasks in T, as measured by P, improves with experience E“ – T. Michell (1997)
  • 8. Terry Taewoong Um (terry.t.um@gmail.com) PERFORMANCE 8 "A computer program is said to learn from experience E with respect to some class of tasks T and performance measure P, if its performance at tasks in T, as measured by P, improves with experience E“ – T. Michell (1997) classification 0-1 loss function regression L2 loss function clustering
  • 9. Terry Taewoong Um (terry.t.um@gmail.com) EXPERIENCE 9 "A computer program is said to learn from experience E with respect to some class of tasks T and performance measure P, if its performance at tasks in T, as measured by P, improves with experience E“ – T. Michell (1997) classification labeled data (pixels)→(number) regression labeled data (x) → (y) clustering unlabeled data (x1,x2)
  • 10. Terry Taewoong Um (terry.t.um@gmail.com) A TOY EXAMPLE 10 ? Height(cm) Weight (kg) [Input X] [Output Y]
  • 11. Terry Taewoong Um (terry.t.um@gmail.com) 11 180 Height(cm) Weight (kg) 80 Y = aX+b Model : Y = aX+b Parameter : (a, b) [Goal] Find (a,b) which best fits the given data A TOY EXAMPLE
  • 12. Terry Taewoong Um (terry.t.um@gmail.com) 12 [Analytic Solution] Least square problem (from AX = b, X=A#b where A# is A’s pseudo inverse) Not always available [Numerical Solution] 1. Set a cost function 2. Apply an optimization method (e.g. Gradient Descent (GD) Method) L (a,b) http://www.yaldex.com/game- development/1592730043_ch18lev1sec4.html Local minima problem http://mnemstudio.org/neural-networks- multilayer-perceptron-design.htm A TOY EXAMPLE
  • 13. Terry Taewoong Um (terry.t.um@gmail.com) 13 32 Age(year) Running Record (min) 140 WHAT WOULD BE THE CORRECT MODEL? Select a model → Set a cost function → Optimization
  • 14. Terry Taewoong Um (terry.t.um@gmail.com) 14 ? X Y WHAT WOULD BE THE CORRECT MODEL? 1. Regularization 2. Nonparametric model “overfitting”
  • 15. Terry Taewoong Um (terry.t.um@gmail.com) 15 L2 REGULARIZATION (e.g. w=(a,b) where Y=aX+b) Avoid a complicated model! • Another interpretation : : Maximum a Posteriori (MAP) http://goo.gl/6GE2ix http://goo.gl/6GE2ix
  • 16. Terry Taewoong Um (terry.t.um@gmail.com) 16 L2 REGULARIZATION • Another interpretation : : Maximum a Posteriori (MAP) http://goo.gl/6GE2ix http://goo.gl/6GE2ix • Bayesian inference 𝑃 𝐵𝑒𝑙𝑖𝑒𝑓 𝐷𝑎𝑡𝑎 = 𝑃 𝐵𝑒𝑙𝑖𝑒𝑓 𝑃(𝐷𝑎𝑡𝑎|𝐵𝑒𝑙𝑖𝑒𝑓) 𝑃(𝐷𝑎𝑡𝑎) posterior prior likelihood ex) fair coin : 50% H, 50% T falsified coin : 80% H, 20% T Let’s say we observed ten heads consecutively. What’s the probability for being a fair coin? 𝑃 𝐵𝑒𝑙𝑖𝑒𝑓 = 0.2 𝑃 𝐷𝑎𝑡𝑎|𝐵𝑒𝑙𝑖𝑒𝑓 = 0.510 ≈ 0.001 𝑃 𝐵𝑒𝑙𝑖𝑒𝑓|𝐷𝑎𝑡𝑎 ∝ 0.2 ∗ 0.001 = 0.0002 normalization (you don’t believe this coin is fair) Fair coin? Falsified coin? 𝑃 𝐵𝑒𝑙𝑖𝑒𝑓 = 0.8 𝑃 𝐷𝑎𝑡𝑎|𝐵𝑒𝑙𝑖𝑒𝑓 = 0.810 ≈ 0.107 𝑃 𝐵𝑒𝑙𝑖𝑒𝑓|𝐷𝑎𝑡𝑎 ∝ 0.8 ∗ 0.107 = 0.0856 Fair = 0.0002 0.0002+0.0856 = 0.23% , Unfair = 99.77%
  • 17. Terry Taewoong Um (terry.t.um@gmail.com) 17 WHAT WOULD BE THE CORRECT MODEL? 1. Regularization 2. Nonparametric model training time error training error test error we should stop here training set validation set test set for training (parameter optimization) for early stopping (avoid overfitting) for evaluation (measure the performance) keep watching the validation error
  • 18. Terry Taewoong Um (terry.t.um@gmail.com) 18 NONPARAMETRIC MODEL • It does not assume any parametric models (e.g. Y = aX+b, Y=aX2+bX+c, etc.) • It often requires much more samples • Kernel methods are frequently applied for modeling the data • Gaussian Process Regression (GPR), a sort of kernel method, is a widely-used nonparametric regression method • Support Vector Machine (SVM), also a sort of kernel method, is a widely-used nonparametric classification method kernel function [Input space] [Feature space]
  • 19. Terry Taewoong Um (terry.t.um@gmail.com) 19 SUPPORT VECTOR MACHINE (SVM) “Myo”, Thalmic Labs (2013) https://youtu.be/oWu9TFJjHaM [Linear classifiers] [Maximum margin] Support vector Machine Tutorial, J. Weston, http://goo.gl/19ywcj [Dual formulation] ( ) kernel function kernel function
  • 20. Terry Taewoong Um (terry.t.um@gmail.com) 20 GAUSSIAN PROCESS REGRESSION (GPR) https://youtu.be/YqhLnCm0KXY https://youtu.be/kvPmArtVoFE • Gaussian Distribution • Multivariate regression likelihood posterior prior likelihood prediction conditioning the joint distribution of the observed & predicted values https://goo.gl/EO54WN http://goo.gl/XvOOmf
  • 21. Terry Taewoong Um (terry.t.um@gmail.com) 21 DIMENSION REDUCTION [Original space] [Feature space] low dim. high dim. high dim. low dim. 𝑋 → ∅(𝑋) • Principal Component Analysis : Find the best orthogonal axes (=principal components) which maximize the variance of the data Y = P X * The rows in P are m largest eigenvectors of 1 𝑁 𝑋𝑋 𝑇 (covariance matrix)
  • 22. Terry Taewoong Um (terry.t.um@gmail.com) 22 DIMENSION REDUCTION http://jbhuang0604.blogspot.kr/2013/04/miss-korea-2013-contestants-face.html
  • 23. Terry Taewoong Um (terry.t.um@gmail.com) 23 SUMMARY - PART 1 • Machine Learning - Tasks : Classification, Regression, Clustering, etc. - Performance : 0-1 loss, L2 loss, etc. - Experience : labeled data, unlabelled data • Machine Learning Process (1) Select a parametric / nonparametric model (2) Set a performance measurement including regularization term (3) Training data (optimizing parameters) until validation error increases (4) Evaluate the final performance using test set • Nonparametric model : Support Vector Machine, Gaussian Process Regression • Dimension reduction : used as pre-processing data
  • 24. Terry Taewoong Um (terry.t.um@gmail.com) CONTENTS 24 Questions about Part 1?
  • 25. Terry Taewoong Um (terry.t.um@gmail.com) CONTENTS 25 2. What is Deep Learning?
  • 26. Terry Taewoong Um (terry.t.um@gmail.com) 26 PARADIGM CHANGE PAST Knowledge ML Method (e.g. GPR, SVM) PRESENT What is the best ML method for the target task? Knowledge Representation How can we find a good representation?
  • 27. Terry Taewoong Um (terry.t.um@gmail.com) 27 PARADIGM CHANGE Knowledge PRESENT Representation How can we find a good representation? kernel function
  • 28. Terry Taewoong Um (terry.t.um@gmail.com) 28 PARADIGM CHANGE Knowledge PRESENT Representation (Features) How can we find a good representation? IMAGE SPEECH Hand-Crafted Features
  • 29. Terry Taewoong Um (terry.t.um@gmail.com) 29 PARADIGM CHANGE IMAGE SPEECH Hand-Crafted Features Knowledge PRESENT Representation (Features) Can we learn a good representation (feature) for the target task as well?
  • 30. Terry Taewoong Um (terry.t.um@gmail.com) 30 DEEP LEARNING • What is Deep Learning (DL) ? - Learning methods which have deep (not shallow) architecture - It often allows end-to-end learning - It automatically finds intermediate representation. Thus, it can be regarded as a representation learning - It often contains stacked “neural network”. Thus, Deep learning usually indicates “deep neural network” “Deep Gaussian Process” (2013) https://youtu.be/NwoGqYsQifg http://goo.gl/fxmmPE http://goo.gl/5Ry08S
  • 31. Terry Taewoong Um (terry.t.um@gmail.com) 31 OUTSTANDING PERFORMANCE OF DL error rate : 28% → 15% → 8% (2010) (2014)(2012) - Object recognition (Simonyan et al., 2015) - Natural machine translation (Bahdanau et al., 2014) - Speech recognition (Chorowski et al., 2014) - Face recognition (Taigman et al., 2014) - Emotion recognition (Ebrahimi-Kahou et al., 2014) - Human pose estimation (Jain et al., 2014) - Deep reinforcement learning(mnih et al., 2013) - Image/Video caption (Xu et al., 2015) - Particle physics (Baldi et al., 2014) - Bioinformatics (Leung et al., 2014) - And so on…. • State-of-art results achieved by DL DL has won most of ML challenges! K. Cho, https://goo.gl/vdfGpu
  • 32. Terry Taewoong Um (terry.t.um@gmail.com) 32 BIOLOGICAL EVIDENCE • Somatosensory cortex learns to see • Why do we need different ML methods for different task? Yann LeCun, https://goo.gl/VVQXJG • The vental pathway in the visual cortex has multiple stages • There exist a lot of intermediate representations Andrew Ng, https://youtu.be/ZmNOAtZIgIk
  • 33. Terry Taewoong Um (terry.t.um@gmail.com) 33 BIG MOVEMENT http://goo.gl/zNbBE2 http://goo.gl/Lk64Q4 Going deeper and deeper….
  • 34. Terry Taewoong Um (terry.t.um@gmail.com) 34 NEURAL NETWORK (NN) Hugo Larochelle, http://www.dmi.usherb.ca/~larocheh/index_en.html • Universal approximation theorem (Hornik, 1991) - A single hidden layer NN w/ linear output can approximate any cont. func. arbitrarily well, given enough hidden units - This does not imply we have learning method to train them
  • 35. Terry Taewoong Um (terry.t.um@gmail.com) 35 TRAINING NN Hugo Larochelle, http://www.dmi.usherb.ca/~larocheh/index_en.html • First, calculate the output using data & initial parameters (W ,b) • Activation functions http://goo.gl/qMQk5H 1
  • 36. Terry Taewoong Um (terry.t.um@gmail.com) 36 TRAINING NN Hugo Larochelle, http://www.dmi.usherb.ca/~larocheh/index_en.html • Then, calculate the error and update the weights from top to bottom • Parameter gradients http://goo.gl/qMQk5H : Backpropagation algorithm 2 known
  • 37. Terry Taewoong Um (terry.t.um@gmail.com) 37 TRAINING NN Hugo Larochelle, http://www.dmi.usherb.ca/~larocheh/index_en.html • Then, calculate the error and update the weights from top to bottom • Parameter gradients http://goo.gl/qMQk5H : Backpropagation algorithm 2 known
  • 38. Terry Taewoong Um (terry.t.um@gmail.com) 38 TRAINING NN Hugo Larochelle, http://www.dmi.usherb.ca/~larocheh/index_en.html • Then, calculate the error and update the weights from top to bottom • Parameter gradients http://goo.gl/qMQk5H : Backpropagation algorithm 2 known
  • 39. Terry Taewoong Um (terry.t.um@gmail.com) 39 TRAINING NN Hugo Larochelle, http://www.dmi.usherb.ca/~larocheh/index_en.html • Then, calculate the error and update the weights from top to bottom • Parameter gradients http://goo.gl/qMQk5H : Backpropagation algorithm 2 known
  • 40. Terry Taewoong Um (terry.t.um@gmail.com) 40 TRAINING NN • Repeat this process with different dataset(mini-batches) http://goo.gl/qMQk5H - Forward propagation (calculate the output values) - Evaluate the error - Backward propagation (update the weights) - Repeat this process until the error converges 3 • As you can see here, NN is not a fancy algorithm, but just a iterative gradient descent method with huge number of parameters • NN is often likely to be stuck in local minima pitfall
  • 41. Terry Taewoong Um (terry.t.um@gmail.com) 41 FROM NN TO DEEP NN • From NN to deep NN (since 2006) - NN requires expert’s skill to tune the hyperparameters - It sometimes gives a good result, but sometimes gives a bad result. The result is highly depend on the quality of initialization, regularization, hyperparameters, data, etc. - Local minima is always problematic • A long winter of NN Yann LeCun (NYU, Facebook) Yoshua Bengio (U. Montreal) Geoffrey Hinton (U. Toronto, Google)
  • 42. Terry Taewoong Um (terry.t.um@gmail.com) 42 WHY IS DL SO SUCCESSFUL? http://t-robotics.blogspot.kr/2015/05/deep-learning.html • Pre-training with unsupervised learning • Convolutional Neural Network • Recurrent Neural Net • GPGPU (parallel processing) & big data • Advanced algorithms for optimization, activation, regularization • Huge research society (Vision, Speech, NLP, Biology, etc.)
  • 43. Terry Taewoong Um (terry.t.um@gmail.com) 43 UNSUPERVISED LEARNING • How can we avoid pathologic local minima cases? (1) First, pre-train the data with unsupervised learning method and get a new representation (2) Stack up this block structures (3) Training each layer in end-to-end manner (4) Fine tune the final structure with (ordinary) fully-connected NN • Unsupervised learning method - Restricted Boltzmann Machine (RBM) → Deep RBM, Deep Belief Network (DBN) - Autoencoder → Deep Auto-encoder http://goo.gl/QGJm5k Autoencoder http://goo.gl/s6kmqY
  • 44. Terry Taewoong Um (terry.t.um@gmail.com) 44 UNSUPERVISED LEARNING “Convolutional deep belief networks for scalable unsupervised learning of hierarchical representation”, Lee et al., 2012
  • 45. Terry Taewoong Um (terry.t.um@gmail.com) 45 CONVOLUTIONAL NN • How can we deal with real images which is much bigger than MNIST digit images? - Use not fully-connected, but locally-connected NN - Use convolutions to get various feature maps - Abstract the results into higher layer by using pooling - Fine tune with fully-connected NN https://goo.gl/G7kBjI https://goo.gl/Xswsbd http://goo.gl/5OR5oH
  • 46. Terry Taewoong Um (terry.t.um@gmail.com) 46 CONVOLUTIONAL NN “Visualization and Understanding Convolutional Network”, Zeiler et al., 2012
  • 47. Terry Taewoong Um (terry.t.um@gmail.com) 47 CONVNET + RNN “Large-scale Video Classification with Convolutional Neural Network”, A. Karpathy 2014, https://youtu.be/qrzQ_AB1DZk
  • 48. Terry Taewoong Um (terry.t.um@gmail.com) 48 RECURRENT NEURAL NETWORK (RNN) t-1 t t+1 [Neural Network] [Recurrent Neural Network] http://www.dmi.usherb.ca/~larocheh/index_en.html
  • 49. Terry Taewoong Um (terry.t.um@gmail.com) 49 RECURRENT NEURAL NETWORK (RNN) [Neural Network] [Recurrent Neural Network] back propagation back propagation through time (BPTT) • Vanishing gradient problem : Can’t have long memory! “Training Recurrent Neural Networks, I. Sutskever, 2013
  • 50. Terry Taewoong Um (terry.t.um@gmail.com) 50 RNN + LSTM • Long Short-Term Memory (LSTM) (Hochreiter and Schmidhuber, 1997) “Training Recurrent Neural Networks, I. Sutskever, 2013
  • 51. Terry Taewoong Um (terry.t.um@gmail.com) 51 INTERESTING RESULTS FROM RNN http://pail.unist.ac.kr/carpedm20/poet/ http://cs.stanford.edu/people/karpathy/deepimagesent/ “generating sequences with RNN”, A.Graves, 2013
  • 52. Terry Taewoong Um (terry.t.um@gmail.com) 52 WHY IS DL SO SUCCESSFUL? http://t-robotics.blogspot.kr/2015/05/deep-learning.html • Pre-training with unsupervised learning • Convolutional Neural Network • Recurrent Neural Net • GPGPU (parallel processing) & big data • Advanced algorithms for optimization, activation, regularization • Huge research society (Vision, Speech, NLP, Biology, etc.)
  • 53. Terry Taewoong Um (terry.t.um@gmail.com) CONTENTS 53 Questions about Part 2?
  • 54. Terry Taewoong Um (terry.t.um@gmail.com) CONTENTS 54 3. Machine Learning in Motion Analysis
  • 55. Terry Taewoong Um (terry.t.um@gmail.com) 55 MOTION DATA “츄리닝”, 이상신 국중록
  • 56. Terry Taewoong Um (terry.t.um@gmail.com) 56 MOTION DATA We need to know the state not only at time t but also at time t-1, t-2, t-3, etc. 𝑓 = 𝑓(𝑥, 𝑡) “츄리닝”, 이상신 국중록
  • 57. Terry Taewoong Um (terry.t.um@gmail.com) 57 MOTION DATA • Why do motion data need special treatment? - In general, most machine learning techniques assume i.i.d. (independent & identically distributed) sampling condition. e.g.) coins tossing - However, motion data is temporally & spatially correlated http://goo.gl/LQulvcswing motion manipulability ellipsoid https://goo.gl/dHjFO9
  • 58. Terry Taewoong Um (terry.t.um@gmail.com) 58 MOTION DATA http://goo.gl/ll3sq6 We can infer the next state based on the temporal & spatial information But, how can we exploit those benefits in ML method?
  • 59. Terry Taewoong Um (terry.t.um@gmail.com) 59 WHAT CAN WE DO WITH MOTION DATA? • Learning the kinematic/dynamic model • Motion segmentation • Motion generation / synthesis • Motion imitation (Imitation learning) • Activity / Gesture recognition TASKS Data • Motion capture data • Vision Data • Dynamic-level data Applications • Biomechanics • Humanoid • Animation http://goo.gl/gFOVWL
  • 60. Terry Taewoong Um (terry.t.um@gmail.com) 60 HIDDEN MARKOV MODEL (HMM) Prob. of (n+1) state only depends on state at (n+1)
  • 61. Terry Taewoong Um (terry.t.um@gmail.com) 61 LIMITATIONS OF HMM 1. Extract features (e.g. PCA) 2. Define the HMM structure (e.g. using GMM) 3. Train a separate HMM per class (Baum-Welch algorithm) 4. Evaluate probability under each HMM (Fwd/Bwd algorithm) or 3. Choose most probable sequence (Viterbi algorithm) - HMM handle discrete states only! - HMM has short memory! (using just the previous state) - HMM has limited expressive power! - [Trend1] features-GMM → unsupervised learning methods - [Trend2] features-GMM-HMM → recurrent neural network • A common procedure of HMM for motion analysis • Limitations & trend change in speech recognition area
  • 62. Terry Taewoong Um (terry.t.um@gmail.com) 62 CAPTURE TEMPORAL INFORMATION • 3D ConvNet - “3D Convolutional Neural Network for Human Action Recognition” (Ji et al., 2010) - 3D convolution - Activity recognition / Pose estimation from video “Joint Training of a Convolutional Network and a Graphical Model for Human Pose Estimation”, Tompson et al., 2014
  • 63. Terry Taewoong Um (terry.t.um@gmail.com) 63 CAPTURE TEMPORAL INFORMATION • Recurrent Neural Network (RNN) “Hierarchical Recurrent Neural Network for Skeleton Based Action Recognition”, Y. Du et al., 2015 • However, how can we capture the spatial information about motions?
  • 64. Terry Taewoong Um (terry.t.um@gmail.com) 64 CHALLENGES We should connect the geometric information with deep neural network! • The link transformation from the i-1 th link to the i th link • Forward Kinematics constant, Mvariable, 𝜃 c.f.) 𝑋𝑖−1,𝑖 = 𝑅𝑜𝑡 𝑧, 𝜃𝑖 𝑇𝑟𝑎𝑛𝑠 𝑧, 𝑑𝑖 𝑇𝑟𝑎𝑛𝑠 𝑥, 𝑎𝑖 𝑅𝑜𝑡 𝑧, 𝛼𝑖 = 𝑒[𝐴 𝑖]𝜃 𝑖 𝑀𝑖−1,𝑖 𝑋0,𝑛 = 𝑒[𝐴1]𝜃1 𝑀0,1 𝑒[𝐴2]𝜃2 𝑀1,2 ⋯ 𝑒 𝐴 𝑛 𝜃 𝑛 𝑀 𝑛−1,𝑛 = 𝑒[𝑆1]𝜃1 𝑒[𝑆2]𝜃2 ⋯ 𝑒[𝑆 𝑛]𝜃 𝑛 𝑀0,𝑛 𝑆𝑖 = 𝐴𝑑 𝑀01⋯𝑀 𝑖−2,𝑖−1 𝐴𝑖 , 𝑖 = 1, ⋯ , 𝑛 propagated forces external force acting on the ith body where • Newton-Euler formulation for inverse dynamics Lie group & Lie algebra, http://goo.gl/uqilDV
  • 65. Terry Taewoong Um (terry.t.um@gmail.com) 65 CHALLENGES https://www.youtube.com/watch?v=oxA2O-tHftI
  • 66. Terry Taewoong Um (terry.t.um@gmail.com) 66 Thank you