SlideShare a Scribd company logo
1 of 17
Deep learning to the rescue
solving long standing problems of recommender systems
Balázs Hidasi
@balazshidasi
Budapest RecSys & Personalization meetup
12 May, 2016
What is deep learning?
• A class of machine learning algorithms
 that use a cascade of multiple non-linear
processing layers
 and complex model structures
 to learn different representations of the
data in each layer
 where higher level features are derived
from lower level features
 to form a hierarchical representation.
• Key component of recent technologies
 Speech recognition
 Personal assistants (e.g. Siri, Cortana)
 Computer vision, object recognition
 Machine translation
 Chatbot technology
 Face recognition
 Self driving cars
• An efficient tool for certain complex
problems
 Pattern recognition
 Computer vision
 Natural language processing
 Speech recognition
• Deep learning is NOT
 the true AI
o it may be a component of it when and
if AI is created
 how the human brain works
 the best solution to every machine
learning tasks
Deep learning in the news
Why is deep learning happening now?
• Actually it is not   first papers published in 1970s
• Third resurgence of neural networks
 Research breakthroughs
 Increase in computational power
 GP GPUs
Problem Solution
Vanishing
gradients
Sigmoid type activation functions easily saturate.
Gradients are small, in deeper layers updates become
almost zero.
Earlier: layer-by-layer pretraining
Recently: non-saturating
activation functions
Gradient
descent
First order methods (e.g. SGD) are easily stuck.
Second order methods are infeasible on larger data.
Adaptive training: adagrad, adam,
adadelta, RMSProp
Nesterov momentum
Regularization Networks easily overfit (even with L2 regularization). Dropout
ETC...
Challenges in RecSys
• Recommender systems ≠ Netflix challenge
 Rating prediction  Top-N recommendation (ranking)
 Explicit feedback  Implicit feedback
 Long user histories  Sessions
 Slowly changing taste  Goal oriented browsing
 Item-to-user only  Other scenarios
• Success of CF
 Human brain is a powerful feature extractor
 Cold-start
o CF can’t be used
o Decisions are rarely made on metadata
o But rather on what the user sees: e.g. product image, content itself
Domain
dependent
Session-based recommendations
• Permanent cold start
 User identification
o Possible but often not reliable
 Intent/theme
o What the user needs?
o Theme of the session
 Never/rarely returning users
• Workaround in practice
 Item-to-item recommendations
o Similar items
o Co-occurring items
 Non-personalized
 Not adaptive
Recurrent Neural Networks
• Hidden state
 Next hidden state depends on the input and the actual hidden state (recurrence)
 ℎ 𝑡 = tanh 𝑊𝑥 𝑡 + 𝑈ℎ 𝑡−1
• „Infinite depth”
• Backpropagation Through Time
• Exploding gradients
 Due to recurrence
 If the spectral radius of U > 1 (necessary)
• Lack of long term memory (vanishing gradients)
 Gradients of earlier states vanish
 If the spectral radius of U < 1 (sufficient)
ℎ
𝑥𝑡 ℎ 𝑡
ℎ
ℎ 𝑡 𝑥𝑡
ℎℎ 𝑡−1
𝑥𝑡−1
ℎℎ 𝑡−2
𝑥𝑡−2
ℎℎ 𝑡−3
𝑥𝑡−3
ℎ 𝑡−4
Advanced RNN units
• Long Short-Term Memory (LSTM)
 Memory cell (𝑐𝑡) is the mix of
o its previous value (governed by the
forget gate (𝑓𝑡))
o the cell value candidate (governed
by the input gate (𝑖 𝑡))
 Cell value candidate ( 𝑐𝑡) depends on
the input and the previous hidden state
 Hidden state is the memory cell
regulated by the output gate (𝑜𝑡)
 No vanishing/exploding gradients
• 𝑓𝑡 𝑖 𝑡 = 𝜎 𝑊 𝑓
𝑊 𝑖
𝑥𝑡 + 𝑈 𝑓
𝑈 𝑖
ℎ 𝑡−1 + 𝑉 𝑓
𝑉 𝑖
𝑐𝑡−1
• 𝑜𝑡 = 𝜎 𝑊 𝑜
𝑥𝑡 + 𝑈 𝑜
ℎ 𝑡−1 + 𝑉 𝑜
𝑐𝑡−1
• 𝑐𝑡 = tanh 𝑊𝑥𝑡 + 𝑈ℎ 𝑡−1
• 𝑐𝑡 = tanh 𝑓𝑡 𝑐𝑡−1 + 𝑖 𝑡 𝑐𝑡
• ℎ 𝑡 = 𝑜𝑡 𝑐𝑡
• Gated Recurrent Unit (GRU)
 Hidden state is the mix of
o the previous hidden state
o the hidden state candidate (ℎ 𝑡)
o governed by the update gate (𝑧𝑡)
– merged input+forget gate
 Hidden state candidate depends on the
input and the previous hidden state
through a reset gate (𝑟𝑡)
 Similar performance
 Less calculations
• 𝑧𝑡 = 𝜎 𝑊 𝑧
𝑥𝑡 + 𝑈 𝑧
ℎ 𝑡−1
• 𝑟𝑡 = 𝜎 𝑊 𝑟
𝑥𝑡 + 𝑈 𝑟
ℎ 𝑡−1
• ℎ 𝑡 = 𝜎 𝑊𝑥𝑡 + 𝑈(𝑟𝑡∘ ℎ 𝑡−1)
• ℎ 𝑡 = 1 − 𝑧𝑡 ℎ 𝑡−1 + 𝑧𝑡ℎ 𝑡
ℎℎ
𝑥𝑡
ℎ 𝑡
𝑧
ℎ𝑐
𝑐
𝑥𝑡
ℎ 𝑡
Powered by RNN
• Sequence labeling
 Document classification
 Speech recognition
• Sequence-to-sequence learning
 Machine translation
 Question answering
 Conversations
• Sequence generation
 Music
 Text
Session modeling with RNNs
• Input: actual item of session
• Output: score on items for being the
next in the event stream
• GRU based RNN
 RNN is worse
 LSTM is slower (same accuracy)
• Optional embedding and feedforward
layers
 Better results without
• Number of layers
 1 gave the best performance
 Sessions span over short timeframes
 No need for modeling on multiple scales
• Requires some adaptation
Feedforward layers
Embedding layers
…
Output: scores on all items
GRU layer
GRU layer
GRU layer
Input: actual item, 1-of-N coding
(optional)
(optional)
Adaptation: session parallel mini-batches
• Motivation
 High variance in the length of the sessions (from 2 to 100s of
events)
 The goal is to capture how sessions evolve
• Minibatch
 Input: current evets
 Output: next events
𝑖1,1 𝑖1,2 𝑖1,3 𝑖1,4
𝑖2,1 𝑖2,2 𝑖2,3
𝑖3,1 𝑖3,2 𝑖3,3 𝑖3,4 𝑖3,5 𝑖3,6
𝑖4,1 𝑖4,2
𝑖5,1 𝑖5,2 𝑖5,3
Session1
Session2
Session3
Session4
Session5
𝑖1,1 𝑖1,2 𝑖1,3
𝑖2,1 𝑖2,2
𝑖3,1 𝑖3,2 𝑖3,3 𝑖3,4 𝑖3,5
𝑖4,1
𝑖5,1 𝑖5,2
Input
(item of the
actual event)
Desired output
(next item in the
event stream)
…
…
…
Mini-batch1
Mini-batch3
Mini-batch2
𝑖1,2 𝑖1,3 𝑖1,4
𝑖2,2 𝑖2,3
𝑖3,2 𝑖3,3 𝑖3,4 𝑖3,5 𝑖3,6
𝑖4,2
𝑖5,2 𝑖5,3 …
…
…
• Active sessions
 First X
 Finished sessions
replaced by the next
available
Adaptation: pairwise loss function
• Motivation
 Goal of recommender: ranking
 Pairwise and pointwise ranking (listwise costly)
 Pairwise often better
• Pairwise loss functions
 Positive items compared to negatives
 BPR
o Bayesian personalized ranking
o 𝐿 = −
1
𝑁 𝑆
𝑗=1
𝑁 𝑆
log 𝜎 𝑟𝑠,𝑖 − 𝑟𝑠,𝑗
 TOP1
o Regularized approximation of the relative rank of the positive item
o 𝐿 =
1
𝑁 𝑆
𝑗=1
𝑁 𝑆
𝜎 𝑟𝑠,𝑖 − 𝑟𝑠,𝑗 + 𝜎 𝑟𝑠,𝑗
2
Adaptation: sampling the output
• Motivation
 Number of items is high  bottleneck
 Model needs to be trained frequently (should be quick)
• Sampling negative items
 Popularity based sampling
o Missing event on popular item  more likely sign of negative feedback
o Pop items often get large scores  faster learning
 Negative items for an example: examples of other sessions in the minibatch
o Technical benefits
o Follows data distribution (pop sampling)
𝑖1 𝑖5 𝑖8
Mini-batch
(desired items)
𝑦1
1
𝑦2
1
𝑦3
1 𝑦4
1
𝑦5
1
𝑦6
1 𝑦7
1
𝑦8
1
𝑦1
3
𝑦2
3
𝑦3
3
𝑦4
3
𝑦5
3
𝑦6
3
𝑦7
3
𝑦8
3
𝑦1
2
𝑦2
2
𝑦3
2
𝑦4
2 𝑦5
2
𝑦6
2
𝑦7
2 𝑦8
2
1 0 0 0 0 0 0 0
0 0 0 0 0 0 0 1
0 0 0 0 1 0 0 0
Network output
(scores)
Desired output scores
Positive
item
Sampled
negative items
Inactive outputs
(not computed)
Offline experiments
Data Description Items Train Test
Sessions Events Sessions Events
RSC15 RecSys Challenge 2015.
Clickstream data of a webshop.
37,483 7,966,257 31,637,239 15,324 71,222
VIDEO Video watch sequences. 327,929 2,954,816 13,180,128 48,746 178,637
+19.91% +19.82%
+15.55%+14.06%
+24.82%
+22.54%
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7 RSC15 - Recall@20
+18.65% +17.54%
+12.58% +5.16%
+20.47%
+31.49%
0
0.1
0.2
0.3 RSC15 - MRR@20
+15.69%
+8.92% +11.50%
N/A
+14.58%
+20.27%
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
VIDEO - Recall@20
+10.04%
-3.56% +3.84%
N/A
-7.23%
+15.08%
0
0.1
0.2
0.3
0.4
VIDEO - MRR@20
Pop
Sessionpop
Item-kNN
BPR-MF
GRU4Rec(100,unitsc.-entropy)
GRU4Rec(100units,BPR)
GRU4Rec(100units,TOP1)
GRU4Rec(1000units,c.-entropy)
GRU4Rec(1000units,BPR)
GRU4Rec(1000units,TOP1)
Pop
Sessionpop
Item-kNN
BPR-MF
GRU4Rec(100,unitsc.-entropy)
GRU4Rec(100units,BPR)
GRU4Rec(100units,TOP1)
GRU4Rec(1000units,c.-entropy)
GRU4Rec(1000units,BPR)
GRU4Rec(1000units,TOP1)
Pop
Sessionpop
Item-kNN
BPR-MF
GRU4Rec(100,unitsc.-entropy)
GRU4Rec(100units,BPR)
GRU4Rec(100units,TOP1)
GRU4Rec(1000units,TOP1)
GRU4Rec(1000units,BPR)
Pop
Sessionpop
Item-kNN
BPR-MF
GRU4Rec(100,unitsc.-entropy)
GRU4Rec(100units,BPR)
GRU4Rec(100units,TOP1)
GRU4Rec(1000units,TOP1)
GRU4Rec(1000units,BPR)
Online experiments
+17.09% +16.10%
+24.16% +23.69%
+5.52%
-3.21%
+7.05% +6.29
0
0.2
0.4
0.6
0.8
1
1.2
1.4
RelativeCTR
RNN Item-kNN Item-kNN-B
CTR
Default setup RNN
• Default trains:
 on ~10x events
 more frequently
• Absolute CTR increase: +0.9%±0.5%
 (p=0.01)
The next step in recsys technology
• is deep learning
• Besides session modelling
 Incorporating content into the model directly
 Modeling complex context-states based on sensory data
(IoT)
 Optimizing recommendations through deep reinforcement
learning
• Would you like to try something in this area?
 Submit to DLRS 2016
 dlrs-workshop.org
Thank you!
Detailed description of the RNN approach:
• B. Hidasi, A. Karatzoglou, L. Baltrunas, D. Tikk: Session-based recommendations with recurrent neural networks. ICLR 2016.
• http://arxiv.org/abs/1511.06939
• Public code: https://github.com/hidasib/GRU4Rec

More Related Content

What's hot

Deep Learning for Personalized Search and Recommender Systems
Deep Learning for Personalized Search and Recommender SystemsDeep Learning for Personalized Search and Recommender Systems
Deep Learning for Personalized Search and Recommender SystemsBenjamin Le
 
Artificial Intelligence Course: Linear models
Artificial Intelligence Course: Linear models Artificial Intelligence Course: Linear models
Artificial Intelligence Course: Linear models ananth
 
Deep Learning for Recommender Systems RecSys2017 Tutorial
Deep Learning for Recommender Systems RecSys2017 Tutorial Deep Learning for Recommender Systems RecSys2017 Tutorial
Deep Learning for Recommender Systems RecSys2017 Tutorial Alexandros Karatzoglou
 
Foundations: Artificial Neural Networks
Foundations: Artificial Neural NetworksFoundations: Artificial Neural Networks
Foundations: Artificial Neural Networksananth
 
Artificial Intelligence, Machine Learning and Deep Learning
Artificial Intelligence, Machine Learning and Deep LearningArtificial Intelligence, Machine Learning and Deep Learning
Artificial Intelligence, Machine Learning and Deep LearningSujit Pal
 
Introduction To Applied Machine Learning
Introduction To Applied Machine LearningIntroduction To Applied Machine Learning
Introduction To Applied Machine Learningananth
 
Deep Learning Models for Question Answering
Deep Learning Models for Question AnsweringDeep Learning Models for Question Answering
Deep Learning Models for Question AnsweringSujit Pal
 
An introduction to Machine Learning (and a little bit of Deep Learning)
An introduction to Machine Learning (and a little bit of Deep Learning)An introduction to Machine Learning (and a little bit of Deep Learning)
An introduction to Machine Learning (and a little bit of Deep Learning)Thomas da Silva Paula
 
Deep Learning For Practitioners, lecture 2: Selecting the right applications...
Deep Learning For Practitioners,  lecture 2: Selecting the right applications...Deep Learning For Practitioners,  lecture 2: Selecting the right applications...
Deep Learning For Practitioners, lecture 2: Selecting the right applications...ananth
 
Deep Learning for Natural Language Processing
Deep Learning for Natural Language ProcessingDeep Learning for Natural Language Processing
Deep Learning for Natural Language ProcessingSangwoo Mo
 
Deep learning and image analytics using Python by Dr Sanparit
Deep learning and image analytics using Python by Dr SanparitDeep learning and image analytics using Python by Dr Sanparit
Deep learning and image analytics using Python by Dr SanparitBAINIDA
 
Ranking and Diversity in Recommendations - RecSys Stammtisch at SoundCloud, B...
Ranking and Diversity in Recommendations - RecSys Stammtisch at SoundCloud, B...Ranking and Diversity in Recommendations - RecSys Stammtisch at SoundCloud, B...
Ranking and Diversity in Recommendations - RecSys Stammtisch at SoundCloud, B...Alexandros Karatzoglou
 
Embed, Encode, Attend, Predict – applying the 4 step NLP recipe for text clas...
Embed, Encode, Attend, Predict – applying the 4 step NLP recipe for text clas...Embed, Encode, Attend, Predict – applying the 4 step NLP recipe for text clas...
Embed, Encode, Attend, Predict – applying the 4 step NLP recipe for text clas...Sujit Pal
 
Challenging Common Assumptions in the Unsupervised Learning of Disentangled R...
Challenging Common Assumptions in the Unsupervised Learning of Disentangled R...Challenging Common Assumptions in the Unsupervised Learning of Disentangled R...
Challenging Common Assumptions in the Unsupervised Learning of Disentangled R...Sangwoo Mo
 
MaxEnt (Loglinear) Models - Overview
MaxEnt (Loglinear) Models - OverviewMaxEnt (Loglinear) Models - Overview
MaxEnt (Loglinear) Models - Overviewananth
 
Generative Models for General Audiences
Generative Models for General AudiencesGenerative Models for General Audiences
Generative Models for General AudiencesSangwoo Mo
 
Deep Learning: Chapter 11 Practical Methodology
Deep Learning: Chapter 11 Practical MethodologyDeep Learning: Chapter 11 Practical Methodology
Deep Learning: Chapter 11 Practical MethodologyJason Tsai
 
Deep Learning with Python (PyData Seattle 2015)
Deep Learning with Python (PyData Seattle 2015)Deep Learning with Python (PyData Seattle 2015)
Deep Learning with Python (PyData Seattle 2015)Alexander Korbonits
 
Deep Learning Enabled Question Answering System to Automate Corporate Helpdesk
Deep Learning Enabled Question Answering System to Automate Corporate HelpdeskDeep Learning Enabled Question Answering System to Automate Corporate Helpdesk
Deep Learning Enabled Question Answering System to Automate Corporate HelpdeskSaurabh Saxena
 

What's hot (20)

Deep Learning for Personalized Search and Recommender Systems
Deep Learning for Personalized Search and Recommender SystemsDeep Learning for Personalized Search and Recommender Systems
Deep Learning for Personalized Search and Recommender Systems
 
Artificial Intelligence Course: Linear models
Artificial Intelligence Course: Linear models Artificial Intelligence Course: Linear models
Artificial Intelligence Course: Linear models
 
Deep Learning for Recommender Systems RecSys2017 Tutorial
Deep Learning for Recommender Systems RecSys2017 Tutorial Deep Learning for Recommender Systems RecSys2017 Tutorial
Deep Learning for Recommender Systems RecSys2017 Tutorial
 
Foundations: Artificial Neural Networks
Foundations: Artificial Neural NetworksFoundations: Artificial Neural Networks
Foundations: Artificial Neural Networks
 
Artificial Intelligence, Machine Learning and Deep Learning
Artificial Intelligence, Machine Learning and Deep LearningArtificial Intelligence, Machine Learning and Deep Learning
Artificial Intelligence, Machine Learning and Deep Learning
 
Introduction To Applied Machine Learning
Introduction To Applied Machine LearningIntroduction To Applied Machine Learning
Introduction To Applied Machine Learning
 
Deep Learning Models for Question Answering
Deep Learning Models for Question AnsweringDeep Learning Models for Question Answering
Deep Learning Models for Question Answering
 
An introduction to Machine Learning (and a little bit of Deep Learning)
An introduction to Machine Learning (and a little bit of Deep Learning)An introduction to Machine Learning (and a little bit of Deep Learning)
An introduction to Machine Learning (and a little bit of Deep Learning)
 
Deeplearning in finance
Deeplearning in financeDeeplearning in finance
Deeplearning in finance
 
Deep Learning For Practitioners, lecture 2: Selecting the right applications...
Deep Learning For Practitioners,  lecture 2: Selecting the right applications...Deep Learning For Practitioners,  lecture 2: Selecting the right applications...
Deep Learning For Practitioners, lecture 2: Selecting the right applications...
 
Deep Learning for Natural Language Processing
Deep Learning for Natural Language ProcessingDeep Learning for Natural Language Processing
Deep Learning for Natural Language Processing
 
Deep learning and image analytics using Python by Dr Sanparit
Deep learning and image analytics using Python by Dr SanparitDeep learning and image analytics using Python by Dr Sanparit
Deep learning and image analytics using Python by Dr Sanparit
 
Ranking and Diversity in Recommendations - RecSys Stammtisch at SoundCloud, B...
Ranking and Diversity in Recommendations - RecSys Stammtisch at SoundCloud, B...Ranking and Diversity in Recommendations - RecSys Stammtisch at SoundCloud, B...
Ranking and Diversity in Recommendations - RecSys Stammtisch at SoundCloud, B...
 
Embed, Encode, Attend, Predict – applying the 4 step NLP recipe for text clas...
Embed, Encode, Attend, Predict – applying the 4 step NLP recipe for text clas...Embed, Encode, Attend, Predict – applying the 4 step NLP recipe for text clas...
Embed, Encode, Attend, Predict – applying the 4 step NLP recipe for text clas...
 
Challenging Common Assumptions in the Unsupervised Learning of Disentangled R...
Challenging Common Assumptions in the Unsupervised Learning of Disentangled R...Challenging Common Assumptions in the Unsupervised Learning of Disentangled R...
Challenging Common Assumptions in the Unsupervised Learning of Disentangled R...
 
MaxEnt (Loglinear) Models - Overview
MaxEnt (Loglinear) Models - OverviewMaxEnt (Loglinear) Models - Overview
MaxEnt (Loglinear) Models - Overview
 
Generative Models for General Audiences
Generative Models for General AudiencesGenerative Models for General Audiences
Generative Models for General Audiences
 
Deep Learning: Chapter 11 Practical Methodology
Deep Learning: Chapter 11 Practical MethodologyDeep Learning: Chapter 11 Practical Methodology
Deep Learning: Chapter 11 Practical Methodology
 
Deep Learning with Python (PyData Seattle 2015)
Deep Learning with Python (PyData Seattle 2015)Deep Learning with Python (PyData Seattle 2015)
Deep Learning with Python (PyData Seattle 2015)
 
Deep Learning Enabled Question Answering System to Automate Corporate Helpdesk
Deep Learning Enabled Question Answering System to Automate Corporate HelpdeskDeep Learning Enabled Question Answering System to Automate Corporate Helpdesk
Deep Learning Enabled Question Answering System to Automate Corporate Helpdesk
 

Viewers also liked

2017 How Deep Learning Changes the Design Process (1)
2017 How Deep Learning Changes the Design Process (1)2017 How Deep Learning Changes the Design Process (1)
2017 How Deep Learning Changes the Design Process (1)Alexander Meinhardt
 
Creativity through deep learning
Creativity through deep learningCreativity through deep learning
Creativity through deep learningAkin Osman Kazakci
 
쉬운 예제 중심의 HTML5 / CSS / MediaQuery / JQuery 강의
쉬운 예제 중심의 HTML5 / CSS / MediaQuery / JQuery 강의쉬운 예제 중심의 HTML5 / CSS / MediaQuery / JQuery 강의
쉬운 예제 중심의 HTML5 / CSS / MediaQuery / JQuery 강의Minha Yang
 
2017 How Deep Learning Changes the Design Process (2)
2017 How Deep Learning Changes the Design Process (2)2017 How Deep Learning Changes the Design Process (2)
2017 How Deep Learning Changes the Design Process (2)Alexander Meinhardt
 
Artificial Intelligence - Past, Present and Future
Artificial Intelligence - Past, Present and FutureArtificial Intelligence - Past, Present and Future
Artificial Intelligence - Past, Present and FutureGrigory Sapunov
 
How Deep Learning Changes the Design Process #NEXT17
How Deep Learning Changes the Design Process #NEXT17How Deep Learning Changes the Design Process #NEXT17
How Deep Learning Changes the Design Process #NEXT17Alexander Meinhardt
 

Viewers also liked (6)

2017 How Deep Learning Changes the Design Process (1)
2017 How Deep Learning Changes the Design Process (1)2017 How Deep Learning Changes the Design Process (1)
2017 How Deep Learning Changes the Design Process (1)
 
Creativity through deep learning
Creativity through deep learningCreativity through deep learning
Creativity through deep learning
 
쉬운 예제 중심의 HTML5 / CSS / MediaQuery / JQuery 강의
쉬운 예제 중심의 HTML5 / CSS / MediaQuery / JQuery 강의쉬운 예제 중심의 HTML5 / CSS / MediaQuery / JQuery 강의
쉬운 예제 중심의 HTML5 / CSS / MediaQuery / JQuery 강의
 
2017 How Deep Learning Changes the Design Process (2)
2017 How Deep Learning Changes the Design Process (2)2017 How Deep Learning Changes the Design Process (2)
2017 How Deep Learning Changes the Design Process (2)
 
Artificial Intelligence - Past, Present and Future
Artificial Intelligence - Past, Present and FutureArtificial Intelligence - Past, Present and Future
Artificial Intelligence - Past, Present and Future
 
How Deep Learning Changes the Design Process #NEXT17
How Deep Learning Changes the Design Process #NEXT17How Deep Learning Changes the Design Process #NEXT17
How Deep Learning Changes the Design Process #NEXT17
 

Similar to Deep learning to the rescue - solving long standing problems of recommender systems

Deep learning: what? how? why? How to win a Kaggle competition
Deep learning: what? how? why? How to win a Kaggle competitionDeep learning: what? how? why? How to win a Kaggle competition
Deep learning: what? how? why? How to win a Kaggle competition317070
 
Overcome the Reign of Chaos
Overcome the Reign of ChaosOvercome the Reign of Chaos
Overcome the Reign of ChaosMichael Stockerl
 
DEF CON 24 - Clarence Chio - machine duping 101
DEF CON 24 - Clarence Chio - machine duping 101DEF CON 24 - Clarence Chio - machine duping 101
DEF CON 24 - Clarence Chio - machine duping 101Felipe Prado
 
Yuandong Tian at AI Frontiers : Planning in Reinforcement Learning
Yuandong Tian at AI Frontiers : Planning in Reinforcement LearningYuandong Tian at AI Frontiers : Planning in Reinforcement Learning
Yuandong Tian at AI Frontiers : Planning in Reinforcement LearningAI Frontiers
 
DeepLearningLecture.pptx
DeepLearningLecture.pptxDeepLearningLecture.pptx
DeepLearningLecture.pptxssuserf07225
 
ODSC 2019: Sessionisation via stochastic periods for root event identification
ODSC 2019: Sessionisation via stochastic periods for root event identificationODSC 2019: Sessionisation via stochastic periods for root event identification
ODSC 2019: Sessionisation via stochastic periods for root event identificationKuldeep Jiwani
 
Using Deep Learning to do Real-Time Scoring in Practical Applications
Using Deep Learning to do Real-Time Scoring in Practical ApplicationsUsing Deep Learning to do Real-Time Scoring in Practical Applications
Using Deep Learning to do Real-Time Scoring in Practical ApplicationsGreg Makowski
 
Deep learning with TensorFlow
Deep learning with TensorFlowDeep learning with TensorFlow
Deep learning with TensorFlowBarbara Fusinska
 
Machine Learning with TensorFlow 2
Machine Learning with TensorFlow 2Machine Learning with TensorFlow 2
Machine Learning with TensorFlow 2Sarah Stemmler
 
Parismlmeetupfinalslides 151209190037-lva1-app6892
Parismlmeetupfinalslides 151209190037-lva1-app6892Parismlmeetupfinalslides 151209190037-lva1-app6892
Parismlmeetupfinalslides 151209190037-lva1-app6892mercedes calderon
 
Impatience is a Virtue: Revisiting Disorder in High-Performance Log Analytics
Impatience is a Virtue: Revisiting Disorder in High-Performance Log AnalyticsImpatience is a Virtue: Revisiting Disorder in High-Performance Log Analytics
Impatience is a Virtue: Revisiting Disorder in High-Performance Log AnalyticsBadrish Chandramouli
 
rsec2a-2016-jheaton-morning
rsec2a-2016-jheaton-morningrsec2a-2016-jheaton-morning
rsec2a-2016-jheaton-morningJeff Heaton
 
Visibility Optimization for Games
Visibility Optimization for GamesVisibility Optimization for Games
Visibility Optimization for GamesUmbra
 
Visibility Optimization for Games
Visibility Optimization for GamesVisibility Optimization for Games
Visibility Optimization for GamesSampo Lappalainen
 
Approximate "Now" is Better Than Accurate "Later"
Approximate "Now" is Better Than Accurate "Later"Approximate "Now" is Better Than Accurate "Later"
Approximate "Now" is Better Than Accurate "Later"NUS-ISS
 
Machine Duping 101: Pwning Deep Learning Systems
Machine Duping 101: Pwning Deep Learning SystemsMachine Duping 101: Pwning Deep Learning Systems
Machine Duping 101: Pwning Deep Learning SystemsClarence Chio
 

Similar to Deep learning to the rescue - solving long standing problems of recommender systems (20)

Deep learning: what? how? why? How to win a Kaggle competition
Deep learning: what? how? why? How to win a Kaggle competitionDeep learning: what? how? why? How to win a Kaggle competition
Deep learning: what? how? why? How to win a Kaggle competition
 
Overcome the Reign of Chaos
Overcome the Reign of ChaosOvercome the Reign of Chaos
Overcome the Reign of Chaos
 
DEF CON 24 - Clarence Chio - machine duping 101
DEF CON 24 - Clarence Chio - machine duping 101DEF CON 24 - Clarence Chio - machine duping 101
DEF CON 24 - Clarence Chio - machine duping 101
 
Deep learning - a primer
Deep learning - a primerDeep learning - a primer
Deep learning - a primer
 
Deep learning - a primer
Deep learning - a primerDeep learning - a primer
Deep learning - a primer
 
Yuandong Tian at AI Frontiers : Planning in Reinforcement Learning
Yuandong Tian at AI Frontiers : Planning in Reinforcement LearningYuandong Tian at AI Frontiers : Planning in Reinforcement Learning
Yuandong Tian at AI Frontiers : Planning in Reinforcement Learning
 
DeepLearningLecture.pptx
DeepLearningLecture.pptxDeepLearningLecture.pptx
DeepLearningLecture.pptx
 
ODSC 2019: Sessionisation via stochastic periods for root event identification
ODSC 2019: Sessionisation via stochastic periods for root event identificationODSC 2019: Sessionisation via stochastic periods for root event identification
ODSC 2019: Sessionisation via stochastic periods for root event identification
 
Using Deep Learning to do Real-Time Scoring in Practical Applications
Using Deep Learning to do Real-Time Scoring in Practical ApplicationsUsing Deep Learning to do Real-Time Scoring in Practical Applications
Using Deep Learning to do Real-Time Scoring in Practical Applications
 
Deep learning with TensorFlow
Deep learning with TensorFlowDeep learning with TensorFlow
Deep learning with TensorFlow
 
Machine Learning with TensorFlow 2
Machine Learning with TensorFlow 2Machine Learning with TensorFlow 2
Machine Learning with TensorFlow 2
 
Paris ML meetup
Paris ML meetupParis ML meetup
Paris ML meetup
 
Parismlmeetupfinalslides 151209190037-lva1-app6892
Parismlmeetupfinalslides 151209190037-lva1-app6892Parismlmeetupfinalslides 151209190037-lva1-app6892
Parismlmeetupfinalslides 151209190037-lva1-app6892
 
Impatience is a Virtue: Revisiting Disorder in High-Performance Log Analytics
Impatience is a Virtue: Revisiting Disorder in High-Performance Log AnalyticsImpatience is a Virtue: Revisiting Disorder in High-Performance Log Analytics
Impatience is a Virtue: Revisiting Disorder in High-Performance Log Analytics
 
Phx dl meetup
Phx dl meetupPhx dl meetup
Phx dl meetup
 
rsec2a-2016-jheaton-morning
rsec2a-2016-jheaton-morningrsec2a-2016-jheaton-morning
rsec2a-2016-jheaton-morning
 
Visibility Optimization for Games
Visibility Optimization for GamesVisibility Optimization for Games
Visibility Optimization for Games
 
Visibility Optimization for Games
Visibility Optimization for GamesVisibility Optimization for Games
Visibility Optimization for Games
 
Approximate "Now" is Better Than Accurate "Later"
Approximate "Now" is Better Than Accurate "Later"Approximate "Now" is Better Than Accurate "Later"
Approximate "Now" is Better Than Accurate "Later"
 
Machine Duping 101: Pwning Deep Learning Systems
Machine Duping 101: Pwning Deep Learning SystemsMachine Duping 101: Pwning Deep Learning Systems
Machine Duping 101: Pwning Deep Learning Systems
 

More from Balázs Hidasi

Egyedi termék kreatívok tömeges gyártása generatív AI segítségével
Egyedi termék kreatívok tömeges gyártása generatív AI segítségévelEgyedi termék kreatívok tömeges gyártása generatív AI segítségével
Egyedi termék kreatívok tömeges gyártása generatív AI segítségévelBalázs Hidasi
 
The Effect of Third Party Implementations on Reproducibility
The Effect of Third Party Implementations on ReproducibilityThe Effect of Third Party Implementations on Reproducibility
The Effect of Third Party Implementations on ReproducibilityBalázs Hidasi
 
Context aware factorization methods for implicit feedback based recommendatio...
Context aware factorization methods for implicit feedback based recommendatio...Context aware factorization methods for implicit feedback based recommendatio...
Context aware factorization methods for implicit feedback based recommendatio...Balázs Hidasi
 
Approximate modeling of continuous context in factorization algorithms (CaRR1...
Approximate modeling of continuous context in factorization algorithms (CaRR1...Approximate modeling of continuous context in factorization algorithms (CaRR1...
Approximate modeling of continuous context in factorization algorithms (CaRR1...Balázs Hidasi
 
Az implicit ajánlási probléma és néhány megoldása (BME TMIT szeminárium előad...
Az implicit ajánlási probléma és néhány megoldása (BME TMIT szeminárium előad...Az implicit ajánlási probléma és néhány megoldása (BME TMIT szeminárium előad...
Az implicit ajánlási probléma és néhány megoldása (BME TMIT szeminárium előad...Balázs Hidasi
 
Context-aware similarities within the factorization framework (CaRR 2013 pres...
Context-aware similarities within the factorization framework (CaRR 2013 pres...Context-aware similarities within the factorization framework (CaRR 2013 pres...
Context-aware similarities within the factorization framework (CaRR 2013 pres...Balázs Hidasi
 
iTALS: implicit tensor factorization for context-aware recommendations (ECML/...
iTALS: implicit tensor factorization for context-aware recommendations (ECML/...iTALS: implicit tensor factorization for context-aware recommendations (ECML/...
iTALS: implicit tensor factorization for context-aware recommendations (ECML/...Balázs Hidasi
 
Initialization of matrix factorization (CaRR 2012 presentation)
Initialization of matrix factorization (CaRR 2012 presentation)Initialization of matrix factorization (CaRR 2012 presentation)
Initialization of matrix factorization (CaRR 2012 presentation)Balázs Hidasi
 
ShiftTree: model alapú idősor-osztályozó (VK 2009 előadás)
ShiftTree: model alapú idősor-osztályozó (VK 2009 előadás)ShiftTree: model alapú idősor-osztályozó (VK 2009 előadás)
ShiftTree: model alapú idősor-osztályozó (VK 2009 előadás)Balázs Hidasi
 
ShiftTree: model alapú idősor-osztályozó (ML@BP előadás, 2012)
ShiftTree: model alapú idősor-osztályozó (ML@BP előadás, 2012)ShiftTree: model alapú idősor-osztályozó (ML@BP előadás, 2012)
ShiftTree: model alapú idősor-osztályozó (ML@BP előadás, 2012)Balázs Hidasi
 
ShiftTree: model based time series classifier (ECML/PKDD 2011 presentation)
ShiftTree: model based time series classifier (ECML/PKDD 2011 presentation)ShiftTree: model based time series classifier (ECML/PKDD 2011 presentation)
ShiftTree: model based time series classifier (ECML/PKDD 2011 presentation)Balázs Hidasi
 

More from Balázs Hidasi (11)

Egyedi termék kreatívok tömeges gyártása generatív AI segítségével
Egyedi termék kreatívok tömeges gyártása generatív AI segítségévelEgyedi termék kreatívok tömeges gyártása generatív AI segítségével
Egyedi termék kreatívok tömeges gyártása generatív AI segítségével
 
The Effect of Third Party Implementations on Reproducibility
The Effect of Third Party Implementations on ReproducibilityThe Effect of Third Party Implementations on Reproducibility
The Effect of Third Party Implementations on Reproducibility
 
Context aware factorization methods for implicit feedback based recommendatio...
Context aware factorization methods for implicit feedback based recommendatio...Context aware factorization methods for implicit feedback based recommendatio...
Context aware factorization methods for implicit feedback based recommendatio...
 
Approximate modeling of continuous context in factorization algorithms (CaRR1...
Approximate modeling of continuous context in factorization algorithms (CaRR1...Approximate modeling of continuous context in factorization algorithms (CaRR1...
Approximate modeling of continuous context in factorization algorithms (CaRR1...
 
Az implicit ajánlási probléma és néhány megoldása (BME TMIT szeminárium előad...
Az implicit ajánlási probléma és néhány megoldása (BME TMIT szeminárium előad...Az implicit ajánlási probléma és néhány megoldása (BME TMIT szeminárium előad...
Az implicit ajánlási probléma és néhány megoldása (BME TMIT szeminárium előad...
 
Context-aware similarities within the factorization framework (CaRR 2013 pres...
Context-aware similarities within the factorization framework (CaRR 2013 pres...Context-aware similarities within the factorization framework (CaRR 2013 pres...
Context-aware similarities within the factorization framework (CaRR 2013 pres...
 
iTALS: implicit tensor factorization for context-aware recommendations (ECML/...
iTALS: implicit tensor factorization for context-aware recommendations (ECML/...iTALS: implicit tensor factorization for context-aware recommendations (ECML/...
iTALS: implicit tensor factorization for context-aware recommendations (ECML/...
 
Initialization of matrix factorization (CaRR 2012 presentation)
Initialization of matrix factorization (CaRR 2012 presentation)Initialization of matrix factorization (CaRR 2012 presentation)
Initialization of matrix factorization (CaRR 2012 presentation)
 
ShiftTree: model alapú idősor-osztályozó (VK 2009 előadás)
ShiftTree: model alapú idősor-osztályozó (VK 2009 előadás)ShiftTree: model alapú idősor-osztályozó (VK 2009 előadás)
ShiftTree: model alapú idősor-osztályozó (VK 2009 előadás)
 
ShiftTree: model alapú idősor-osztályozó (ML@BP előadás, 2012)
ShiftTree: model alapú idősor-osztályozó (ML@BP előadás, 2012)ShiftTree: model alapú idősor-osztályozó (ML@BP előadás, 2012)
ShiftTree: model alapú idősor-osztályozó (ML@BP előadás, 2012)
 
ShiftTree: model based time series classifier (ECML/PKDD 2011 presentation)
ShiftTree: model based time series classifier (ECML/PKDD 2011 presentation)ShiftTree: model based time series classifier (ECML/PKDD 2011 presentation)
ShiftTree: model based time series classifier (ECML/PKDD 2011 presentation)
 

Recently uploaded

Base editing, prime editing, Cas13 & RNA editing and organelle base editing
Base editing, prime editing, Cas13 & RNA editing and organelle base editingBase editing, prime editing, Cas13 & RNA editing and organelle base editing
Base editing, prime editing, Cas13 & RNA editing and organelle base editingNetHelix
 
User Guide: Pulsar™ Weather Station (Columbia Weather Systems)
User Guide: Pulsar™ Weather Station (Columbia Weather Systems)User Guide: Pulsar™ Weather Station (Columbia Weather Systems)
User Guide: Pulsar™ Weather Station (Columbia Weather Systems)Columbia Weather Systems
 
FREE NURSING BUNDLE FOR NURSES.PDF by na
FREE NURSING BUNDLE FOR NURSES.PDF by naFREE NURSING BUNDLE FOR NURSES.PDF by na
FREE NURSING BUNDLE FOR NURSES.PDF by naJASISJULIANOELYNV
 
Good agricultural practices 3rd year bpharm. herbal drug technology .pptx
Good agricultural practices 3rd year bpharm. herbal drug technology .pptxGood agricultural practices 3rd year bpharm. herbal drug technology .pptx
Good agricultural practices 3rd year bpharm. herbal drug technology .pptxSimeonChristian
 
Vision and reflection on Mining Software Repositories research in 2024
Vision and reflection on Mining Software Repositories research in 2024Vision and reflection on Mining Software Repositories research in 2024
Vision and reflection on Mining Software Repositories research in 2024AyushiRastogi48
 
GenBio2 - Lesson 1 - Introduction to Genetics.pptx
GenBio2 - Lesson 1 - Introduction to Genetics.pptxGenBio2 - Lesson 1 - Introduction to Genetics.pptx
GenBio2 - Lesson 1 - Introduction to Genetics.pptxBerniceCayabyab1
 
Pests of castor_Binomics_Identification_Dr.UPR.pdf
Pests of castor_Binomics_Identification_Dr.UPR.pdfPests of castor_Binomics_Identification_Dr.UPR.pdf
Pests of castor_Binomics_Identification_Dr.UPR.pdfPirithiRaju
 
LIGHT-PHENOMENA-BY-CABUALDIONALDOPANOGANCADIENTE-CONDEZA (1).pptx
LIGHT-PHENOMENA-BY-CABUALDIONALDOPANOGANCADIENTE-CONDEZA (1).pptxLIGHT-PHENOMENA-BY-CABUALDIONALDOPANOGANCADIENTE-CONDEZA (1).pptx
LIGHT-PHENOMENA-BY-CABUALDIONALDOPANOGANCADIENTE-CONDEZA (1).pptxmalonesandreagweneth
 
Speech, hearing, noise, intelligibility.pptx
Speech, hearing, noise, intelligibility.pptxSpeech, hearing, noise, intelligibility.pptx
Speech, hearing, noise, intelligibility.pptxpriyankatabhane
 
User Guide: Capricorn FLX™ Weather Station
User Guide: Capricorn FLX™ Weather StationUser Guide: Capricorn FLX™ Weather Station
User Guide: Capricorn FLX™ Weather StationColumbia Weather Systems
 
OECD bibliometric indicators: Selected highlights, April 2024
OECD bibliometric indicators: Selected highlights, April 2024OECD bibliometric indicators: Selected highlights, April 2024
OECD bibliometric indicators: Selected highlights, April 2024innovationoecd
 
Environmental Biotechnology Topic:- Microbial Biosensor
Environmental Biotechnology Topic:- Microbial BiosensorEnvironmental Biotechnology Topic:- Microbial Biosensor
Environmental Biotechnology Topic:- Microbial Biosensorsonawaneprad
 
Pests of jatropha_Bionomics_identification_Dr.UPR.pdf
Pests of jatropha_Bionomics_identification_Dr.UPR.pdfPests of jatropha_Bionomics_identification_Dr.UPR.pdf
Pests of jatropha_Bionomics_identification_Dr.UPR.pdfPirithiRaju
 
Pests of safflower_Binomics_Identification_Dr.UPR.pdf
Pests of safflower_Binomics_Identification_Dr.UPR.pdfPests of safflower_Binomics_Identification_Dr.UPR.pdf
Pests of safflower_Binomics_Identification_Dr.UPR.pdfPirithiRaju
 
Pests of Blackgram, greengram, cowpea_Dr.UPR.pdf
Pests of Blackgram, greengram, cowpea_Dr.UPR.pdfPests of Blackgram, greengram, cowpea_Dr.UPR.pdf
Pests of Blackgram, greengram, cowpea_Dr.UPR.pdfPirithiRaju
 
Call Girls in Majnu Ka Tilla Delhi 🔝9711014705🔝 Genuine
Call Girls in Majnu Ka Tilla Delhi 🔝9711014705🔝 GenuineCall Girls in Majnu Ka Tilla Delhi 🔝9711014705🔝 Genuine
Call Girls in Majnu Ka Tilla Delhi 🔝9711014705🔝 Genuinethapagita
 
Microphone- characteristics,carbon microphone, dynamic microphone.pptx
Microphone- characteristics,carbon microphone, dynamic microphone.pptxMicrophone- characteristics,carbon microphone, dynamic microphone.pptx
Microphone- characteristics,carbon microphone, dynamic microphone.pptxpriyankatabhane
 
Pests of Bengal gram_Identification_Dr.UPR.pdf
Pests of Bengal gram_Identification_Dr.UPR.pdfPests of Bengal gram_Identification_Dr.UPR.pdf
Pests of Bengal gram_Identification_Dr.UPR.pdfPirithiRaju
 
THE ROLE OF PHARMACOGNOSY IN TRADITIONAL AND MODERN SYSTEM OF MEDICINE.pptx
THE ROLE OF PHARMACOGNOSY IN TRADITIONAL AND MODERN SYSTEM OF MEDICINE.pptxTHE ROLE OF PHARMACOGNOSY IN TRADITIONAL AND MODERN SYSTEM OF MEDICINE.pptx
THE ROLE OF PHARMACOGNOSY IN TRADITIONAL AND MODERN SYSTEM OF MEDICINE.pptxNandakishor Bhaurao Deshmukh
 
The dark energy paradox leads to a new structure of spacetime.pptx
The dark energy paradox leads to a new structure of spacetime.pptxThe dark energy paradox leads to a new structure of spacetime.pptx
The dark energy paradox leads to a new structure of spacetime.pptxEran Akiva Sinbar
 

Recently uploaded (20)

Base editing, prime editing, Cas13 & RNA editing and organelle base editing
Base editing, prime editing, Cas13 & RNA editing and organelle base editingBase editing, prime editing, Cas13 & RNA editing and organelle base editing
Base editing, prime editing, Cas13 & RNA editing and organelle base editing
 
User Guide: Pulsar™ Weather Station (Columbia Weather Systems)
User Guide: Pulsar™ Weather Station (Columbia Weather Systems)User Guide: Pulsar™ Weather Station (Columbia Weather Systems)
User Guide: Pulsar™ Weather Station (Columbia Weather Systems)
 
FREE NURSING BUNDLE FOR NURSES.PDF by na
FREE NURSING BUNDLE FOR NURSES.PDF by naFREE NURSING BUNDLE FOR NURSES.PDF by na
FREE NURSING BUNDLE FOR NURSES.PDF by na
 
Good agricultural practices 3rd year bpharm. herbal drug technology .pptx
Good agricultural practices 3rd year bpharm. herbal drug technology .pptxGood agricultural practices 3rd year bpharm. herbal drug technology .pptx
Good agricultural practices 3rd year bpharm. herbal drug technology .pptx
 
Vision and reflection on Mining Software Repositories research in 2024
Vision and reflection on Mining Software Repositories research in 2024Vision and reflection on Mining Software Repositories research in 2024
Vision and reflection on Mining Software Repositories research in 2024
 
GenBio2 - Lesson 1 - Introduction to Genetics.pptx
GenBio2 - Lesson 1 - Introduction to Genetics.pptxGenBio2 - Lesson 1 - Introduction to Genetics.pptx
GenBio2 - Lesson 1 - Introduction to Genetics.pptx
 
Pests of castor_Binomics_Identification_Dr.UPR.pdf
Pests of castor_Binomics_Identification_Dr.UPR.pdfPests of castor_Binomics_Identification_Dr.UPR.pdf
Pests of castor_Binomics_Identification_Dr.UPR.pdf
 
LIGHT-PHENOMENA-BY-CABUALDIONALDOPANOGANCADIENTE-CONDEZA (1).pptx
LIGHT-PHENOMENA-BY-CABUALDIONALDOPANOGANCADIENTE-CONDEZA (1).pptxLIGHT-PHENOMENA-BY-CABUALDIONALDOPANOGANCADIENTE-CONDEZA (1).pptx
LIGHT-PHENOMENA-BY-CABUALDIONALDOPANOGANCADIENTE-CONDEZA (1).pptx
 
Speech, hearing, noise, intelligibility.pptx
Speech, hearing, noise, intelligibility.pptxSpeech, hearing, noise, intelligibility.pptx
Speech, hearing, noise, intelligibility.pptx
 
User Guide: Capricorn FLX™ Weather Station
User Guide: Capricorn FLX™ Weather StationUser Guide: Capricorn FLX™ Weather Station
User Guide: Capricorn FLX™ Weather Station
 
OECD bibliometric indicators: Selected highlights, April 2024
OECD bibliometric indicators: Selected highlights, April 2024OECD bibliometric indicators: Selected highlights, April 2024
OECD bibliometric indicators: Selected highlights, April 2024
 
Environmental Biotechnology Topic:- Microbial Biosensor
Environmental Biotechnology Topic:- Microbial BiosensorEnvironmental Biotechnology Topic:- Microbial Biosensor
Environmental Biotechnology Topic:- Microbial Biosensor
 
Pests of jatropha_Bionomics_identification_Dr.UPR.pdf
Pests of jatropha_Bionomics_identification_Dr.UPR.pdfPests of jatropha_Bionomics_identification_Dr.UPR.pdf
Pests of jatropha_Bionomics_identification_Dr.UPR.pdf
 
Pests of safflower_Binomics_Identification_Dr.UPR.pdf
Pests of safflower_Binomics_Identification_Dr.UPR.pdfPests of safflower_Binomics_Identification_Dr.UPR.pdf
Pests of safflower_Binomics_Identification_Dr.UPR.pdf
 
Pests of Blackgram, greengram, cowpea_Dr.UPR.pdf
Pests of Blackgram, greengram, cowpea_Dr.UPR.pdfPests of Blackgram, greengram, cowpea_Dr.UPR.pdf
Pests of Blackgram, greengram, cowpea_Dr.UPR.pdf
 
Call Girls in Majnu Ka Tilla Delhi 🔝9711014705🔝 Genuine
Call Girls in Majnu Ka Tilla Delhi 🔝9711014705🔝 GenuineCall Girls in Majnu Ka Tilla Delhi 🔝9711014705🔝 Genuine
Call Girls in Majnu Ka Tilla Delhi 🔝9711014705🔝 Genuine
 
Microphone- characteristics,carbon microphone, dynamic microphone.pptx
Microphone- characteristics,carbon microphone, dynamic microphone.pptxMicrophone- characteristics,carbon microphone, dynamic microphone.pptx
Microphone- characteristics,carbon microphone, dynamic microphone.pptx
 
Pests of Bengal gram_Identification_Dr.UPR.pdf
Pests of Bengal gram_Identification_Dr.UPR.pdfPests of Bengal gram_Identification_Dr.UPR.pdf
Pests of Bengal gram_Identification_Dr.UPR.pdf
 
THE ROLE OF PHARMACOGNOSY IN TRADITIONAL AND MODERN SYSTEM OF MEDICINE.pptx
THE ROLE OF PHARMACOGNOSY IN TRADITIONAL AND MODERN SYSTEM OF MEDICINE.pptxTHE ROLE OF PHARMACOGNOSY IN TRADITIONAL AND MODERN SYSTEM OF MEDICINE.pptx
THE ROLE OF PHARMACOGNOSY IN TRADITIONAL AND MODERN SYSTEM OF MEDICINE.pptx
 
The dark energy paradox leads to a new structure of spacetime.pptx
The dark energy paradox leads to a new structure of spacetime.pptxThe dark energy paradox leads to a new structure of spacetime.pptx
The dark energy paradox leads to a new structure of spacetime.pptx
 

Deep learning to the rescue - solving long standing problems of recommender systems

  • 1. Deep learning to the rescue solving long standing problems of recommender systems Balázs Hidasi @balazshidasi Budapest RecSys & Personalization meetup 12 May, 2016
  • 2. What is deep learning? • A class of machine learning algorithms  that use a cascade of multiple non-linear processing layers  and complex model structures  to learn different representations of the data in each layer  where higher level features are derived from lower level features  to form a hierarchical representation. • Key component of recent technologies  Speech recognition  Personal assistants (e.g. Siri, Cortana)  Computer vision, object recognition  Machine translation  Chatbot technology  Face recognition  Self driving cars • An efficient tool for certain complex problems  Pattern recognition  Computer vision  Natural language processing  Speech recognition • Deep learning is NOT  the true AI o it may be a component of it when and if AI is created  how the human brain works  the best solution to every machine learning tasks
  • 3. Deep learning in the news
  • 4. Why is deep learning happening now? • Actually it is not   first papers published in 1970s • Third resurgence of neural networks  Research breakthroughs  Increase in computational power  GP GPUs Problem Solution Vanishing gradients Sigmoid type activation functions easily saturate. Gradients are small, in deeper layers updates become almost zero. Earlier: layer-by-layer pretraining Recently: non-saturating activation functions Gradient descent First order methods (e.g. SGD) are easily stuck. Second order methods are infeasible on larger data. Adaptive training: adagrad, adam, adadelta, RMSProp Nesterov momentum Regularization Networks easily overfit (even with L2 regularization). Dropout ETC...
  • 5. Challenges in RecSys • Recommender systems ≠ Netflix challenge  Rating prediction  Top-N recommendation (ranking)  Explicit feedback  Implicit feedback  Long user histories  Sessions  Slowly changing taste  Goal oriented browsing  Item-to-user only  Other scenarios • Success of CF  Human brain is a powerful feature extractor  Cold-start o CF can’t be used o Decisions are rarely made on metadata o But rather on what the user sees: e.g. product image, content itself Domain dependent
  • 6. Session-based recommendations • Permanent cold start  User identification o Possible but often not reliable  Intent/theme o What the user needs? o Theme of the session  Never/rarely returning users • Workaround in practice  Item-to-item recommendations o Similar items o Co-occurring items  Non-personalized  Not adaptive
  • 7. Recurrent Neural Networks • Hidden state  Next hidden state depends on the input and the actual hidden state (recurrence)  ℎ 𝑡 = tanh 𝑊𝑥 𝑡 + 𝑈ℎ 𝑡−1 • „Infinite depth” • Backpropagation Through Time • Exploding gradients  Due to recurrence  If the spectral radius of U > 1 (necessary) • Lack of long term memory (vanishing gradients)  Gradients of earlier states vanish  If the spectral radius of U < 1 (sufficient) ℎ 𝑥𝑡 ℎ 𝑡 ℎ ℎ 𝑡 𝑥𝑡 ℎℎ 𝑡−1 𝑥𝑡−1 ℎℎ 𝑡−2 𝑥𝑡−2 ℎℎ 𝑡−3 𝑥𝑡−3 ℎ 𝑡−4
  • 8. Advanced RNN units • Long Short-Term Memory (LSTM)  Memory cell (𝑐𝑡) is the mix of o its previous value (governed by the forget gate (𝑓𝑡)) o the cell value candidate (governed by the input gate (𝑖 𝑡))  Cell value candidate ( 𝑐𝑡) depends on the input and the previous hidden state  Hidden state is the memory cell regulated by the output gate (𝑜𝑡)  No vanishing/exploding gradients • 𝑓𝑡 𝑖 𝑡 = 𝜎 𝑊 𝑓 𝑊 𝑖 𝑥𝑡 + 𝑈 𝑓 𝑈 𝑖 ℎ 𝑡−1 + 𝑉 𝑓 𝑉 𝑖 𝑐𝑡−1 • 𝑜𝑡 = 𝜎 𝑊 𝑜 𝑥𝑡 + 𝑈 𝑜 ℎ 𝑡−1 + 𝑉 𝑜 𝑐𝑡−1 • 𝑐𝑡 = tanh 𝑊𝑥𝑡 + 𝑈ℎ 𝑡−1 • 𝑐𝑡 = tanh 𝑓𝑡 𝑐𝑡−1 + 𝑖 𝑡 𝑐𝑡 • ℎ 𝑡 = 𝑜𝑡 𝑐𝑡 • Gated Recurrent Unit (GRU)  Hidden state is the mix of o the previous hidden state o the hidden state candidate (ℎ 𝑡) o governed by the update gate (𝑧𝑡) – merged input+forget gate  Hidden state candidate depends on the input and the previous hidden state through a reset gate (𝑟𝑡)  Similar performance  Less calculations • 𝑧𝑡 = 𝜎 𝑊 𝑧 𝑥𝑡 + 𝑈 𝑧 ℎ 𝑡−1 • 𝑟𝑡 = 𝜎 𝑊 𝑟 𝑥𝑡 + 𝑈 𝑟 ℎ 𝑡−1 • ℎ 𝑡 = 𝜎 𝑊𝑥𝑡 + 𝑈(𝑟𝑡∘ ℎ 𝑡−1) • ℎ 𝑡 = 1 − 𝑧𝑡 ℎ 𝑡−1 + 𝑧𝑡ℎ 𝑡 ℎℎ 𝑥𝑡 ℎ 𝑡 𝑧 ℎ𝑐 𝑐 𝑥𝑡 ℎ 𝑡
  • 9. Powered by RNN • Sequence labeling  Document classification  Speech recognition • Sequence-to-sequence learning  Machine translation  Question answering  Conversations • Sequence generation  Music  Text
  • 10. Session modeling with RNNs • Input: actual item of session • Output: score on items for being the next in the event stream • GRU based RNN  RNN is worse  LSTM is slower (same accuracy) • Optional embedding and feedforward layers  Better results without • Number of layers  1 gave the best performance  Sessions span over short timeframes  No need for modeling on multiple scales • Requires some adaptation Feedforward layers Embedding layers … Output: scores on all items GRU layer GRU layer GRU layer Input: actual item, 1-of-N coding (optional) (optional)
  • 11. Adaptation: session parallel mini-batches • Motivation  High variance in the length of the sessions (from 2 to 100s of events)  The goal is to capture how sessions evolve • Minibatch  Input: current evets  Output: next events 𝑖1,1 𝑖1,2 𝑖1,3 𝑖1,4 𝑖2,1 𝑖2,2 𝑖2,3 𝑖3,1 𝑖3,2 𝑖3,3 𝑖3,4 𝑖3,5 𝑖3,6 𝑖4,1 𝑖4,2 𝑖5,1 𝑖5,2 𝑖5,3 Session1 Session2 Session3 Session4 Session5 𝑖1,1 𝑖1,2 𝑖1,3 𝑖2,1 𝑖2,2 𝑖3,1 𝑖3,2 𝑖3,3 𝑖3,4 𝑖3,5 𝑖4,1 𝑖5,1 𝑖5,2 Input (item of the actual event) Desired output (next item in the event stream) … … … Mini-batch1 Mini-batch3 Mini-batch2 𝑖1,2 𝑖1,3 𝑖1,4 𝑖2,2 𝑖2,3 𝑖3,2 𝑖3,3 𝑖3,4 𝑖3,5 𝑖3,6 𝑖4,2 𝑖5,2 𝑖5,3 … … … • Active sessions  First X  Finished sessions replaced by the next available
  • 12. Adaptation: pairwise loss function • Motivation  Goal of recommender: ranking  Pairwise and pointwise ranking (listwise costly)  Pairwise often better • Pairwise loss functions  Positive items compared to negatives  BPR o Bayesian personalized ranking o 𝐿 = − 1 𝑁 𝑆 𝑗=1 𝑁 𝑆 log 𝜎 𝑟𝑠,𝑖 − 𝑟𝑠,𝑗  TOP1 o Regularized approximation of the relative rank of the positive item o 𝐿 = 1 𝑁 𝑆 𝑗=1 𝑁 𝑆 𝜎 𝑟𝑠,𝑖 − 𝑟𝑠,𝑗 + 𝜎 𝑟𝑠,𝑗 2
  • 13. Adaptation: sampling the output • Motivation  Number of items is high  bottleneck  Model needs to be trained frequently (should be quick) • Sampling negative items  Popularity based sampling o Missing event on popular item  more likely sign of negative feedback o Pop items often get large scores  faster learning  Negative items for an example: examples of other sessions in the minibatch o Technical benefits o Follows data distribution (pop sampling) 𝑖1 𝑖5 𝑖8 Mini-batch (desired items) 𝑦1 1 𝑦2 1 𝑦3 1 𝑦4 1 𝑦5 1 𝑦6 1 𝑦7 1 𝑦8 1 𝑦1 3 𝑦2 3 𝑦3 3 𝑦4 3 𝑦5 3 𝑦6 3 𝑦7 3 𝑦8 3 𝑦1 2 𝑦2 2 𝑦3 2 𝑦4 2 𝑦5 2 𝑦6 2 𝑦7 2 𝑦8 2 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 1 0 0 0 Network output (scores) Desired output scores Positive item Sampled negative items Inactive outputs (not computed)
  • 14. Offline experiments Data Description Items Train Test Sessions Events Sessions Events RSC15 RecSys Challenge 2015. Clickstream data of a webshop. 37,483 7,966,257 31,637,239 15,324 71,222 VIDEO Video watch sequences. 327,929 2,954,816 13,180,128 48,746 178,637 +19.91% +19.82% +15.55%+14.06% +24.82% +22.54% 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 RSC15 - Recall@20 +18.65% +17.54% +12.58% +5.16% +20.47% +31.49% 0 0.1 0.2 0.3 RSC15 - MRR@20 +15.69% +8.92% +11.50% N/A +14.58% +20.27% 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 VIDEO - Recall@20 +10.04% -3.56% +3.84% N/A -7.23% +15.08% 0 0.1 0.2 0.3 0.4 VIDEO - MRR@20 Pop Sessionpop Item-kNN BPR-MF GRU4Rec(100,unitsc.-entropy) GRU4Rec(100units,BPR) GRU4Rec(100units,TOP1) GRU4Rec(1000units,c.-entropy) GRU4Rec(1000units,BPR) GRU4Rec(1000units,TOP1) Pop Sessionpop Item-kNN BPR-MF GRU4Rec(100,unitsc.-entropy) GRU4Rec(100units,BPR) GRU4Rec(100units,TOP1) GRU4Rec(1000units,c.-entropy) GRU4Rec(1000units,BPR) GRU4Rec(1000units,TOP1) Pop Sessionpop Item-kNN BPR-MF GRU4Rec(100,unitsc.-entropy) GRU4Rec(100units,BPR) GRU4Rec(100units,TOP1) GRU4Rec(1000units,TOP1) GRU4Rec(1000units,BPR) Pop Sessionpop Item-kNN BPR-MF GRU4Rec(100,unitsc.-entropy) GRU4Rec(100units,BPR) GRU4Rec(100units,TOP1) GRU4Rec(1000units,TOP1) GRU4Rec(1000units,BPR)
  • 15. Online experiments +17.09% +16.10% +24.16% +23.69% +5.52% -3.21% +7.05% +6.29 0 0.2 0.4 0.6 0.8 1 1.2 1.4 RelativeCTR RNN Item-kNN Item-kNN-B CTR Default setup RNN • Default trains:  on ~10x events  more frequently • Absolute CTR increase: +0.9%±0.5%  (p=0.01)
  • 16. The next step in recsys technology • is deep learning • Besides session modelling  Incorporating content into the model directly  Modeling complex context-states based on sensory data (IoT)  Optimizing recommendations through deep reinforcement learning • Would you like to try something in this area?  Submit to DLRS 2016  dlrs-workshop.org
  • 17. Thank you! Detailed description of the RNN approach: • B. Hidasi, A. Karatzoglou, L. Baltrunas, D. Tikk: Session-based recommendations with recurrent neural networks. ICLR 2016. • http://arxiv.org/abs/1511.06939 • Public code: https://github.com/hidasib/GRU4Rec