SlideShare a Scribd company logo
1 of 88
Download to read offline
Research on Recommender 
Systems: Beyond Ratings and Lists 
Denis Parra, Ph.D. Information Sciences 
Assistant Professor, CS Department 
School of Engineering 
Pontificia Universidad Católica de Chile 
Tuesday, November 11th of 2014
Outline 
• Personal Introduction 
• Quick Overview of Recommender Systems 
• My Work on Recommender Systems 
– Tag-Based Recommendation 
– Implicit-Feedback (time allowing …) 
– Visual Interactive Interfaces 
• Summary & Current & Future Work 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 2
Personal Introduction 
• I’m from Valdivia! 
• There are many reasons to love Valdivia 
The City 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 3
Personal Introduction 
• I’m from Valdivia! 
• There are many reasons to love Valdivia 
The Sports 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 4
Personal Introduction 
• I’m from Valdivia! 
• There are many reasons to love Valdivia 
The Animals 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 5
Personal Introduction 
• B.Eng. and professional title of Civil Engineer in 
Informatics from Universidad Austral de Chile 
(2004), Valdivia, Chile 
• Ph.D. in Information Sciences at University of 
Pittsburgh (2013), Pittsburgh, PA, USA 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 6
Personal Introduction 
• B.Eng. and professional title of Civil Engineer in 
Informatics from Universidad Austral de Chile 
(2004), Valdivia, Chile 
• Ph.D. in Information Sciences at University of 
Pittsburgh (2013), Pittsburgh, PA, USA 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 7
Recommender Systems 
INTRODUCTION 
* 
* Danboard (Danbo): Amazon’s cardboard robot, in these slides it represents a 
recommender system 
Nov 11th 2014 8
Recommender Systems (RecSys) 
Systems that help (groups of) people to find relevant items in 
a crowded item or information space (MacNee et al. 2006) 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 9
Why do we care about RecSys? 
• RecSys have gained popularity due to several 
domains & applications that require people to make 
decisions among a large set of items. 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 10
A lil’ bit of History 
• First recommender systems were built at the 
beginning of 90’s (Tapestry, GroupLens, Ringo) 
• Online contests, such as the Netflix prize, grew the 
attention on recommender systems beyond 
Computer Science 
(2006-2009) 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 11
The Recommendation Problem 
• The most popular way that the recommendation 
problem has been presented is about rating 
prediction: 
Predict! 
Item 1 Item 2 … Item m 
User 1 1 5 4 
User 2 5 1 ? 
… 
User n 2 5 ? 
• How good is my prediction? 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 12
Recommendation Methods 
• Without covering all possible methods, the two 
most typical classifications on recommender 
algorithms are 
Classification 1 Classification 2 
- Collaborative Filtering 
- Content-based Filtering 
- Hybrid 
- Memory-based 
- Model-based 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 13
Collaborative Filtering (User-based KNN) 
• Step 1: Finding Similar Users (Pearson Corr.) 
5 
4 
4 
1 
2 
1 
5 
4 
4 
1 
2 
5 
Active 
user 
User_1 
User_2 
User_3 
active user 
user_1 
user_2 
user_3 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 14
Collaborative Filtering (User-based KNN) 
• Step 1: Finding Similar Users (Pearson Corr.) 
5 
4 
4 
1 
2 
1 
5 
4 
4 
1 
2 
5 
Active 
user 
User_1 
User_2 
User_3 
Σ 
r r r r 
( )( ) 
i ⊂ 
CR ui u ni n 
r r r r 
− − 
u , 
n 
( )2 ( )2 
Σ Σ 
− − 
i ⊂ CR ui u i ⊂ 
CR ni n 
= 
u n u n 
Sim u n 
, , 
( , ) 
active user 
user_1 
0.4472136 
user_2 
0.49236596 
user_3 
-0.91520863 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 15
Collaborative Filtering (User-based KNN) 
• Step 2: Ranking the items to recommend 
5 
4 
4 
Item 1 
2 
1 
5 
4 
4 
Active 
user 
2 
User_1 
3 
User_2 
4 
2 
Item 2 
Item 3 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 16
Collaborative Filtering (User-based KNN) 
• Step 2: Ranking the items to recommend 
5 
4 
4 
Item 1 
2 
1 
5 
4 
4 
Active 
user 
2 pred u i r 
User_1 
3 
User_2 
userSim u n r r 
( , ) ( ) 
⊂ ⋅ − 
n neighbors u ni n 
u userSim u n 
Σ 
Σ 
n ⊂ 
neighbors u 
= + 
( ) 
( ) 
( , ) 
( , ) 
4 
2 
Item 2 
Item 3 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 17
Pros/Cons of CF 
PROS: 
• Very simple to implement 
• Content-agnostic 
• Compared to other techniques such as content-based, 
is more accurate 
CONS: 
• Sparsity 
• Cold-start 
• New Item 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 18
Content-Based Filtering 
• Can be traced back to techniques from IR, where 
the User Profile represents a query. 
Doc_1 = {w_1, w_2, …., w_3} 
5 
4 
5 
user_profile = {w_1, w_2, …., w_3} using TF-IDF, weighting 
Doc_2 = {w_1, w_2, …., w_3} 
Doc_3 = {w_1, w_2, …., w_3} 
Doc_n = {w_1, w_2, …., w_3} 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 19
PROS/CONS of Content-Based Filtering 
PROS: 
• New items can be matched without previous 
feedback 
• It can exploit also techniques such as LSA or LDA 
• It can use semantic data (ConceptNet, WordNet, 
etc.) 
CONS: 
• Less accurate than collaborative filtering 
• Tends to overspecialization 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 20
Hybridization 
• Combine previous methods to overcome their 
weaknesses (Burke, 2002) 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 21
C2. Model/Memory Classification 
• Memory-based methods use the whole dataset in 
training and prediction. User and Item-based CF are 
examples. 
• Model-based methods build a model during training 
and only use this model during prediction. This 
makes prediction performance way faster and 
scalable 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 22
Model-based: Matrix Factorization 
Latent vector of the item 
Latent vector of the user 
SVD ~ 
Singular Value Decomposition 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 23
PROS/CONS of MF and latent factors model 
PROS: 
• So far, state-of-the-art in terms of accuracy (these 
methods won the Netflix Prize) 
• Performance-wise, the best option nowadays: slow 
at training time O((m+n)3) compared to correlation 
O(m2n), but linear at prediction time O(m+n) 
CONS: 
• Recommendations are obscure: How to explain that 
certain “latent factors” produced the 
recommendation 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 24
Rethinking the Recommendation Problem 
• Ratings are scarce: need for exploiting other sources 
of user preference 
• User-centric recommendation takes the problem 
beyond ratings and ranked lists: evaluate user 
engagement and satisfaction, not only RMSE 
• Several other dimensions to consider in the 
evaluation: novelty of the results, diversity, 
coverage (user and catalog), serendipity 
• Study de effect of interface characteristics: user-control, 
explainability 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 25
My Take on RecSys Research 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 30
My Work on RecSys 
• Traditional RecSys: accurate prediction and TopN 
algorithms 
• In my research I have contributed to RecSys by: 
– Utilizing other sources of user preference (Social Tags) 
– Exploiting implicit feedback for recommendation and for 
mapping explicit feedback 
– Studying user-centric evaluation: the effect of user 
controllability on user satisfaction in interactive 
interfaces 
• And nowadays: Studying whether Virtual Worlds are a 
good proxy for real world recommendation tasks 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 31
This is not only My work J 
• Dr. Peter Brusilovsky 
University of Pittsburgh, PA, USA 
• Dr. Alexander Troussov 
IBM Dublin and TCD, Ireland 
• Dr. Xavier Amatriain 
TID / Netflix, CA, USA 
• Dr. Christoff Trattner 
NTNU, Norway 
• Dr. Katrien Verbert 
KU Leuven, Belgium 
• Dr. Leandro Balby-Marinho 
UFCG, Brasil 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 32
TAG-BASED RECOMMENDATION
Tag-based Recommendation 
• D. Parra, P. Brusilovsky. Improving Collaborative 
Filtering in Social Tagging Systems for the 
Recommendation of Scientific Articles. Web 
Intelligence 2010, Toronto, Canada 
• D. Parra, P. Brusilovsky. Collaborative Filtering 
for Social Tagging Systems: an Experiment with 
CiteULike. ACM Recsys 2009, New York, NY, USA 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 34
Motivation 
• Ratings are scarce. Find another source of user 
preference: Social Tagging Systems 
User 
Resource 
Tags 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 35
A Folksonomy 
• When a user u uses adds an item i using one or 
more tags t1,…, tn, there is a tagging instance. 
• The collection of tagging instances produces a 
folksonomy 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 36
Applying CF over the Folksonomy 
• In the first step: Calculate user similarity 
Traditional CF Tag-based CF 
Pearson Correlation 
over ratings 
BM25 over social tags 
• In the second step: incorporate the amount of raters 
to rank the items (NwCF) 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 37
Tag-based CF 
= 
Query 
Doc_1 
Doc_2 
Doc_3 
BM25 
Active 
User 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 38
Okapi BM25 
BM25: We obtain the similarity between 
users (neighbors) using their set of tags as 
“documents” and performing an Okapi BM25 
(probabilistic IR model) Retrieval Status 
Value calculation. 
Tag frequency in the neighbor (v) profile 
1 ( 1) 
RSV IDF k tf 
sim(u, v) = ( 1) 
Σ∈ 
+ 
td 
k tf 
3 
d k ((1 − b ) + b × ( L / L )) 
+ 
tf 
k tf 
= ⋅ 
t q 1 
d ave td 
tq 
Tag frequency in the active user (u) profile 
( , ) log (1 ( )) ( , ) 10 predʹ′ u i = + nbr i ⋅ pred u i 
+ 
+ 
⋅ 
tq 
3 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 39
Evaluation 
• Crawl during 38 days during June-July 2009 
Item # unique instances 
# users 784 
# items 26,599 
# tags 26,009 
# posts 71,413 
# annotations 218,930 
avg # items per user 91 
avg # users per item 2.68 
avg # tags per user 88.02 
avg # users per tag 2.65 
avg # tags per item 7.07 
avg # items per tag 7.23 
Item Phase 2 
dataset 
# users 5,849 
# articles 574,907 
# tags 139,993 
#tagging 
incidents 
2,337,571 
Filtering 
process 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 40
Cross-validation 
• Test-validation-train sets, 10-fold cross 
validation 
• Training to obtain parameter K: neighb. size 
• One run the experiment: ~12 hours 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 41
Results & Statistical Significance 
• BM25 is intended to bring more neighbors, at the 
cost of more noise (neighbors not so similar) 
• NwCF helps to decrease noise, so it was natural to 
combine them and try just that option 
CCF NwCF BM25+CCF BM25+NwCF 
MAP@10 0.12875 0.1432* 0.1876** 0.1942*** 
K (neigh.size) 20 22 21 29 
Ucov 81.12% 81.12% 99.23% 99.23% 
Significance over the baseline: *p < 0.236, ** p < 0.033, *** p < 0.001 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 42
Take-aways 
• We can exploit tags as a source for user similarity in 
recommendation algorithms 
• Tag-based (BM25) similarity can be considered as 
an alternative to Pearson Correlation to calculate 
user similarity in STS. 
• Incorporating the number of raters helped to 
decrease the noise produced by items with too few 
ratings 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 43
Work with Xavier Amatriain 
IMPLICIT FEEDBACK
Implicit-Feedback 
• Slides are based on two articles: 
– Parra-Santander, D., & Amatriain, X. (2011). Walk the 
Talk: Analyzing the relation between implicit and 
explicit feedback for preference elicitation. Proceedings 
of UMAP 2011, Girona, Spain 
– Parra, D., Karatzoglou, A., Amatriain, X., & Yavuz, I. 
(2011). Implicit feedback recommendation via implicit-to- 
explicit ordinal logistic regression mapping. 
Proceedings of the CARS Workshop, Chicago, IL, USA, 
2011. 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 45
Introduction (1/2) 
• Most of recommender system approaches rely on 
explicit information of the users, but… 
• Explicit feedback: scarce (people are not especially 
eager to rate or to provide personal info) 
• Implicit feedback: Is less scarce, but (Hu et al., 2008) 
There’s no negative feedback 
… and if you watch a TV program just 
once or twice? 
Noisy 
… but explicit feedback is also noisy 
(Amatriain et al., 2009) 
Preference & Confidence 
… we aim to map the I.F. to preference 
(our main goal) 
Lack of evaluation metrics 
… if we can map I.F. and E.F., we can 
have a comparable evaluation 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 46
Introduction (2/2) 
• Is it possible to map implicit behavior to explicit 
preference (ratings)? 
• Which variables better account for the amount of 
times a user listens to online albums? [Baltrunas & 
Amatriain CARS ‘09 workshop – RecSys 2009.] 
• OUR APPROACH: Study with Last.fm users 
– Part I: Ask users to rate 100 albums (how to sample) 
– Part II: Build a model to map collected implicit feedback 
and context to explicit feedback 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 47
Walk the Talk (2011) 
Albums they listened to during last: 
7days, 3months, 6months, year, 
overall For each album in the list we obtained: 
# user plays (in each period), # of 
global listeners and # of global plays 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 48
Walk the Talk - 2 
• Requirements: 18 y.o., scrobblings > 5000 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 49
Quantization of Data for Sampling 
• What items should they rate? Item (album) sampling: 
– Implicit Feedback (IF): playcount for a user on a given album. 
Changed to scale [1-3], 3 means being more listened to. 
– Global Popularity (GP): global playcount for all users on a given 
album [1-3]. Changed to scale [1-3], 3 means being more listened to. 
– Recentness (R) : time elapsed since user played a given album. 
Changed to scale [1-3], 3 means being listened to more recently. 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 50
4 Regression Analysis 
M1: implicit feedback 
M2: 
implicit 
feedback & 
recentness 
M4: 
Interaction of 
implicit 
feedback & 
recentness 
M3: implicit 
feedback, 
recentness, 
global 
popularity 
• Including Recentness increases R2 in more than 10% [ 1 -> 2] 
• Including GP increases R2, not much compared to RE + IF [ 1 -> 3] 
• Not Including GP, but including interaction between IF and RE 
improves the variance of the DV explained by the regression model. 
[ 2 -> 4 ] 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 51
4.1 Regression Analysis 
Model RMSE1 RMSE2 
User average 1.5308 1.1051 
M1: Implicit feedback 1.4206 1.0402 
M2: Implicit feedback + recentness 1.4136 1.034 
M3: Implicit feedback + recentness + global popularity 1.4130 1.0338 
M4: Interaction of Implicit feedback * recentness 1.4127 1.0332 
• We tested conclusions of regression analysis by 
predicting the score, checking RMSE in 10-fold 
cross validation. 
• Results of regression analysis are supported. 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 52
Conclusions of Part I 
• Using a linear model, Implicit feedback and 
recentness can help to predict explicit feedback (in 
the form of ratings) 
• Global popularity doesn’t show a significant 
improvement in the prediction task 
• Our model can help to relate implicit and explicit 
feedback, helping to evaluate and compare explicit 
and implicit recommender systems. 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 53
Part II: Extension of Walk the Talk 
• Implicit Feedback Recommendation via Implicit-to- 
Explicit OLR Mapping (Recsys 2011, CARS 
Workshop) 
– Consider ratings as ordinal variables 
– Use mixed-models to account for non-independence of 
observations 
– Compare with state-of-the-art implicit feedback 
algorithm 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 54
Recalling the 1st study (5/5) 
• Prediction of rating by multiple Linear Regression 
evaluated with RMSE. 
• Results showed that Implicit feedback (play count 
of the album by a specific user) and recentness (how 
recently an album was listened to) were important 
factors, global popularity had a weaker effect. 
• Results also showed that listening style (if user 
preferred to listen to single tracks, CDs, or either) 
was also an important factor, and not the other 
ones. 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 55
... but 
• Linear Regression didn’t account for the nested 
nature of ratings 
User 1 
1 3 5 3 0 4 5 2 2 1 5 4 3 2 
User n 
3 2 1 0 4 5 2 5 4 3 2 1 3 5 
. . . 
• And ratings were treated as continuous, when they 
are actually ordinal. 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 56
So, Ordinal Logistic Regression! 
• Actually Mixed-Efects Ordinal Multinomial Logistic 
Regression 
• Mixed-effects: Nested nature of ratings 
• We obtain a distribution over ratings (ordinal 
multinomial) per each pair USER, ITEM -> we 
predict the rating using the expected value. 
• … And we can compare the inferred ratings with 
a method that directly uses implicit information 
(playcounts) to recommend ( by Hu, Koren et al. 
2007) 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 57
Ordinal Regression for Mapping 
• Model 
• Predicted value 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 58
Datasets 
• D1: users, albums, if, re, gp, ratings, 
demographics/consumption 
• D2: users, albums, if, re, gp, NO RATINGS. 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 59
Results 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 60
Conclusions & Current Work 
Problem/ Challenge 
1. Ground truth: How many Playcounts to relevancy? 
> Sensibility Analysis needed 
2. Quantization of playcounts (implicit feedback), 
recentness, and overall number of listeners of an album 
(global popularity) [1-3] scale v/s raw playcounts > 
modifiy and compare 
3. Additional/Alternative metrics for evaluation [MAP 
and nDCG used in the paper] 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 61
Part of this work with Katrien Verbert 
VISUALIZATION + USER 
CONTROLLABILITY
Visualization & User Controllability 
• Motivation: Can user controllability and 
explainability improve user engagement and 
satisfaction with a recommender system? 
• Specific research question: How intersections of 
contexts of relevance (of recommendation 
algorithms) might be better represented for user 
experience with the recommender? 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 63
The Concept of Controllability 
MovieLens: example of traditional recommender list 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 64
Visualization & User Controllability 
• Motivation: Can user controllability and 
explainability improve user engagement and 
satisfaction with a recommender system? 
• Specific research question: How intersections of 
contexts of relevance (of recommendation 
algorithms) might be better represented for user 
experience with the recommender? 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 65
Research Platform 
• The studies were conducted using Conference 
Navigator, a Conference Support System 
• Our goal was recommending conference talks 
http://halley.exp.sis.pitt.edu/cn3/ 
Program 
Proceedings 
Author List 
Recommendations 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 66
Hybrid RecSys: Visualizing Intersections 
• Clustermap vs. Venn Diagram 
Clustermap Venn diagram 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 67
TalkExplorer – IUI 2013 
• Adaptation of Aduna Visualization to CN 
• Main research question: Does fusion (intersection) of 
contexts of relevance improve user experience? 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 68
TalkExplorer - I 
Entities 
Tags, Recommender Agents, 
Users 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 69
TalkExplorer - II 
• Canvas Area: Intersections of Different Entities 
Recommender 
Recommender 
Cluster 
with 
intersect 
ion of 
entities Cluster (of 
talks) 
associated 
to only one 
entity 
User 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 70
TalkExplorer - III 
Items 
Talks explored by the 
user 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 71
Our Assumptions 
• Items which are relevant in more that 
one aspect could be more valuable to the 
users 
• Displaying multiple aspects of relevance 
visually is important for the users in the 
process of item’s exploration 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 72
TalkExplorer Studies I & II 
• Study I 
– Controlled Experiment: Users were asked to discover 
relevant talks by exploring the three types of entities: tags, 
recommender agents and users. 
– Conducted at Hypertext and UMAP 2012 (21 users) 
– Subjects familiar with Visualizations and Recsys 
• Study II 
– Field Study: Users were left free to explore the interface. 
– Conducted at LAK 2012 and ECTEL 2013 (18 users) 
– Subjects familiar with visualizations, but not much with 
RecSys 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 73
Evaluation: Intersections & Effectiveness 
• What do we call an “Intersection”? 
• We used #explorations on intersections and their 
effectiveness, defined as: 
Effectiveness = |푏표표푘푚푎푟푘푒푑 푖푡푒푚푠|/| 
푖푛푡푒푟푒푠푒푐푡푖표푛푠 푒푥푝푙표푟푒푑|  
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 74
Results of Studies I & II 
• Effectiveness increases 
with intersections of 
more entities 
• Effectiveness wasn’t 
affected in the field 
study (study 2) 
• … but exploration 
distribution was 
affected 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 75
SETFUSION: VENN DIAGRAM FOR 
USER-CONTROLLABLE INTERFACE 
76
SetFusion – IUI 2014 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 77
SetFusion I 
Traditional 
Ranked List 
Papers sorted by 
Relevance. 
It combines 3 
recommendation 
approaches. 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 78
SetFusion - II 
Sliders 
Allow the user to control the importance of 
each data source or recommendation method 
Interactive Venn Diagram 
Allows the user to inspect and to filter papers 
recommended. Actions available: 
- Filter item list by clicking on an area 
- Highlight a paper by mouse-over on a circle 
- Scroll to paper by clicking on a circle 
- Indicate bookmarked papers 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 79
SetFusion – UMAP 2012 
• Field Study: let users freely explore the interface 
- ~50% (50 users) tried the 
SetFusion recommender 
- 28% (14 users) bookmarked at 
least one paper 
- Users explored in average 14.9 
talks and bookmarked 7.36 
talks in average. 
A AB ABC AC B BC C 
15 7 9 26 18 4 17 
16% 7% 9% 27% 19% 4% 18% 
Distribution of bookmarks per method or combination of methods 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 80
TalkExplorer vs. SetFusion 
• Comparing distributions of explorations 
In studies 1 and 2 over 
talkExplorer we observed an 
important change in the 
distribution of explorations. 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 81
TalkExplorer vs. SetFusion 
• Comparing distributions of explorations 
Comparing the field studies: 
- In TalkExplorer, 84% of 
the explorations over 
intersections were 
performed over clusters of 
1 item 
- In SetFusion, was only 
52%, compared to 48% 
(18% + 30%) of multiple 
intersections, diff. not 
statistically significant 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 82
Summary & Conclusions 
• We presented two implementations of visual 
interactive interfaces that tackle exploration on a 
recommendation setting 
• We showed that intersections of several contexts of 
relevance help to discover relevant items 
• The visual paradigm used can have a strong effect on 
user behavior: we need to keep working on visual 
representation that promote exploration without 
increasing the cognitive load over the users 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 83
Limitations & Future Work 
• Apply our approach to other domains (fusion of 
data sources or recommendation algorithms) 
• For SetFusion, find alternatives to scale the 
approach to more than 3 sets, potential alternatives: 
– Clustering and 
– Radial sets 
• Consider other factors that interact with the user 
satisfaction: 
– Controllability by itself vs. minimum level of accuracy 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 84
More Details on SetFusion? 
• Effect of other variables: gender, age, experience 
with in the domain, or familiarity with the system 
• Check our upcoming paper in the IJHCS “User-controllable 
Personalization: A Case Study with 
SetFusion”: Controlled Laboratory study with 
SetFusion versus traditional ranked list 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 85
CONCLUSIONS (& CURRENT) & 
FUTURE WORK
Challenges in Recommender Systems 
• Recommendation to groups 
• Cross-Domain recommendation 
• User-centric evaluation 
• Interactive interfaces and visualization 
• Improve Evaluation for comparison (P. Campos of U. 
of Bio-Bio on doing fair evaluations considering time) 
• ML: Active learning, multi-armed bandits (exploration, 
exploitation) 
• Prevent the “Filter Bubble” 
• Make algorithms resistant to attacks 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 87
Are Virtual Worlds Good Proxies for Real 
World ? 
• Why? We have a Second Life dataset with 3 
connected dimensions of information 
Social Network 
Marketplace 
Virtual World 
• 2 undergoing projects: Entrepreneurship and LBSN 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 88
Entrepreneurship 
• Can we predict whether a user will create a store 
and how successful will she/he be? Literature on 
this area is extremely scarce. 
Social Network Marketplace 
James Gaskin 
SEM, Causal models 
BYU, USA 
Stephen Zhang 
Entrepreneurship 
PUC Chile 
Christoph Trattner 
Social Networks 
NTNU, Norway 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 89
Location-Based Social Networks (LBSN) 
• How similar are the patterns of mobility in real 
world and virtual world ? 
Social Network 
Virtual World 
Christoph Trattner 
Social Networks 
NTNU, Norway 
Leandro Balby-Marinho 
LBSN and RecSys 
UFCG, Brasil 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 90
Other RecSys Activities 
• I am part of the Program Committee of the 2015 
RecSys challenge. Don’t miss it! 
» Is the user going to buy items in this session? Yes|No 
» If yes, what are the items that are going to be bought? 
• Part of team creating the upcoming RecSys Forum 
(like SIGIR Forum). Coming Soon! (Alan Said, 
Cataldo Musto, Alejandro Bellogin, etc.) 
Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 92
dparra@ing.puc.cl 
THANKS!

More Related Content

What's hot

An Evolution of Deep Learning Models for AI2 Reasoning Challenge
An Evolution of Deep Learning Models for AI2 Reasoning ChallengeAn Evolution of Deep Learning Models for AI2 Reasoning Challenge
An Evolution of Deep Learning Models for AI2 Reasoning ChallengeTraian Rebedea
 
Investigating the Effects of Personality on Second Language Learning through ...
Investigating the Effects of Personality on Second Language Learning through ...Investigating the Effects of Personality on Second Language Learning through ...
Investigating the Effects of Personality on Second Language Learning through ...CSCJournals
 
Open domain Question Answering System - Research project in NLP
Open domain  Question Answering System - Research project in NLPOpen domain  Question Answering System - Research project in NLP
Open domain Question Answering System - Research project in NLPGVS Chaitanya
 
Paper Presentation: Data Mining User Preference in Interactive Multimedia
Paper Presentation: Data Mining User Preference in Interactive MultimediaPaper Presentation: Data Mining User Preference in Interactive Multimedia
Paper Presentation: Data Mining User Preference in Interactive MultimediaJeanette Howe
 
Setting Up a Qualitative or Mixed Methods Research Project in NVivo 10 to Cod...
Setting Up a Qualitative or Mixed Methods Research Project in NVivo 10 to Cod...Setting Up a Qualitative or Mixed Methods Research Project in NVivo 10 to Cod...
Setting Up a Qualitative or Mixed Methods Research Project in NVivo 10 to Cod...Shalin Hai-Jew
 
Not Good Enough but Try Again! Mitigating the Impact of Rejections on New Con...
Not Good Enough but Try Again! Mitigating the Impact of Rejections on New Con...Not Good Enough but Try Again! Mitigating the Impact of Rejections on New Con...
Not Good Enough but Try Again! Mitigating the Impact of Rejections on New Con...Aleksi Aaltonen
 
Admixture of Poisson MRFs: A New Topic Model with Word Dependencies
Admixture of Poisson MRFs: A New Topic Model with Word DependenciesAdmixture of Poisson MRFs: A New Topic Model with Word Dependencies
Admixture of Poisson MRFs: A New Topic Model with Word DependenciesDavid Inouye
 
Question Retrieval in Community Question Answering via NON-Negative Matrix Fa...
Question Retrieval in Community Question Answering via NON-Negative Matrix Fa...Question Retrieval in Community Question Answering via NON-Negative Matrix Fa...
Question Retrieval in Community Question Answering via NON-Negative Matrix Fa...IRJET Journal
 
Combining IR with Relevance Feedback for Concept Location
Combining IR with Relevance Feedback for Concept LocationCombining IR with Relevance Feedback for Concept Location
Combining IR with Relevance Feedback for Concept LocationSonia Haiduc
 
Query formulation process
Query formulation processQuery formulation process
Query formulation processmalathimurugan
 
A hybrid recommender system user profiling from keywords and ratings
A hybrid recommender system user profiling from keywords and ratingsA hybrid recommender system user profiling from keywords and ratings
A hybrid recommender system user profiling from keywords and ratingsAravindharamanan S
 
Question Answering - Application and Challenges
Question Answering - Application and ChallengesQuestion Answering - Application and Challenges
Question Answering - Application and ChallengesJens Lehmann
 
Penetrating the black box of time-on-task estimation
Penetrating the black box of time-on-task estimationPenetrating the black box of time-on-task estimation
Penetrating the black box of time-on-task estimationVitomir Kovanovic
 
Recommendation systems
Recommendation systems  Recommendation systems
Recommendation systems Badr Hirchoua
 
Query formulation (chapter 1)
Query formulation (chapter 1)Query formulation (chapter 1)
Query formulation (chapter 1)Mohamed Rafique
 
Effective Cross-Domain Collaborative Filtering using Temporal Domain – A Brie...
Effective Cross-Domain Collaborative Filtering using Temporal Domain – A Brie...Effective Cross-Domain Collaborative Filtering using Temporal Domain – A Brie...
Effective Cross-Domain Collaborative Filtering using Temporal Domain – A Brie...IJMTST Journal
 
Dynamic Topic Modeling via Non-negative Matrix Factorization (Dr. Derek Greene)
Dynamic Topic Modeling via Non-negative Matrix Factorization (Dr. Derek Greene)Dynamic Topic Modeling via Non-negative Matrix Factorization (Dr. Derek Greene)
Dynamic Topic Modeling via Non-negative Matrix Factorization (Dr. Derek Greene)Sebastian Ruder
 
From TREC to Watson: is open domain question answering a solved problem?
From TREC to Watson: is open domain question answering a solved problem?From TREC to Watson: is open domain question answering a solved problem?
From TREC to Watson: is open domain question answering a solved problem?Constantin Orasan
 
LinkedUp kickoff meeting session 4
LinkedUp kickoff meeting session 4LinkedUp kickoff meeting session 4
LinkedUp kickoff meeting session 4Hendrik Drachsler
 
SelQA: A New Benchmark for Selection-based Question Answering
SelQA: A New Benchmark for Selection-based Question AnsweringSelQA: A New Benchmark for Selection-based Question Answering
SelQA: A New Benchmark for Selection-based Question AnsweringJinho Choi
 

What's hot (20)

An Evolution of Deep Learning Models for AI2 Reasoning Challenge
An Evolution of Deep Learning Models for AI2 Reasoning ChallengeAn Evolution of Deep Learning Models for AI2 Reasoning Challenge
An Evolution of Deep Learning Models for AI2 Reasoning Challenge
 
Investigating the Effects of Personality on Second Language Learning through ...
Investigating the Effects of Personality on Second Language Learning through ...Investigating the Effects of Personality on Second Language Learning through ...
Investigating the Effects of Personality on Second Language Learning through ...
 
Open domain Question Answering System - Research project in NLP
Open domain  Question Answering System - Research project in NLPOpen domain  Question Answering System - Research project in NLP
Open domain Question Answering System - Research project in NLP
 
Paper Presentation: Data Mining User Preference in Interactive Multimedia
Paper Presentation: Data Mining User Preference in Interactive MultimediaPaper Presentation: Data Mining User Preference in Interactive Multimedia
Paper Presentation: Data Mining User Preference in Interactive Multimedia
 
Setting Up a Qualitative or Mixed Methods Research Project in NVivo 10 to Cod...
Setting Up a Qualitative or Mixed Methods Research Project in NVivo 10 to Cod...Setting Up a Qualitative or Mixed Methods Research Project in NVivo 10 to Cod...
Setting Up a Qualitative or Mixed Methods Research Project in NVivo 10 to Cod...
 
Not Good Enough but Try Again! Mitigating the Impact of Rejections on New Con...
Not Good Enough but Try Again! Mitigating the Impact of Rejections on New Con...Not Good Enough but Try Again! Mitigating the Impact of Rejections on New Con...
Not Good Enough but Try Again! Mitigating the Impact of Rejections on New Con...
 
Admixture of Poisson MRFs: A New Topic Model with Word Dependencies
Admixture of Poisson MRFs: A New Topic Model with Word DependenciesAdmixture of Poisson MRFs: A New Topic Model with Word Dependencies
Admixture of Poisson MRFs: A New Topic Model with Word Dependencies
 
Question Retrieval in Community Question Answering via NON-Negative Matrix Fa...
Question Retrieval in Community Question Answering via NON-Negative Matrix Fa...Question Retrieval in Community Question Answering via NON-Negative Matrix Fa...
Question Retrieval in Community Question Answering via NON-Negative Matrix Fa...
 
Combining IR with Relevance Feedback for Concept Location
Combining IR with Relevance Feedback for Concept LocationCombining IR with Relevance Feedback for Concept Location
Combining IR with Relevance Feedback for Concept Location
 
Query formulation process
Query formulation processQuery formulation process
Query formulation process
 
A hybrid recommender system user profiling from keywords and ratings
A hybrid recommender system user profiling from keywords and ratingsA hybrid recommender system user profiling from keywords and ratings
A hybrid recommender system user profiling from keywords and ratings
 
Question Answering - Application and Challenges
Question Answering - Application and ChallengesQuestion Answering - Application and Challenges
Question Answering - Application and Challenges
 
Penetrating the black box of time-on-task estimation
Penetrating the black box of time-on-task estimationPenetrating the black box of time-on-task estimation
Penetrating the black box of time-on-task estimation
 
Recommendation systems
Recommendation systems  Recommendation systems
Recommendation systems
 
Query formulation (chapter 1)
Query formulation (chapter 1)Query formulation (chapter 1)
Query formulation (chapter 1)
 
Effective Cross-Domain Collaborative Filtering using Temporal Domain – A Brie...
Effective Cross-Domain Collaborative Filtering using Temporal Domain – A Brie...Effective Cross-Domain Collaborative Filtering using Temporal Domain – A Brie...
Effective Cross-Domain Collaborative Filtering using Temporal Domain – A Brie...
 
Dynamic Topic Modeling via Non-negative Matrix Factorization (Dr. Derek Greene)
Dynamic Topic Modeling via Non-negative Matrix Factorization (Dr. Derek Greene)Dynamic Topic Modeling via Non-negative Matrix Factorization (Dr. Derek Greene)
Dynamic Topic Modeling via Non-negative Matrix Factorization (Dr. Derek Greene)
 
From TREC to Watson: is open domain question answering a solved problem?
From TREC to Watson: is open domain question answering a solved problem?From TREC to Watson: is open domain question answering a solved problem?
From TREC to Watson: is open domain question answering a solved problem?
 
LinkedUp kickoff meeting session 4
LinkedUp kickoff meeting session 4LinkedUp kickoff meeting session 4
LinkedUp kickoff meeting session 4
 
SelQA: A New Benchmark for Selection-based Question Answering
SelQA: A New Benchmark for Selection-based Question AnsweringSelQA: A New Benchmark for Selection-based Question Answering
SelQA: A New Benchmark for Selection-based Question Answering
 

Viewers also liked

Recommender Systems, Part 1 - Introduction to approaches and algorithms
Recommender Systems, Part 1 - Introduction to approaches and algorithmsRecommender Systems, Part 1 - Introduction to approaches and algorithms
Recommender Systems, Part 1 - Introduction to approaches and algorithmsRamzi Alqrainy
 
Similarity functions in Lucene 4.0
Similarity functions in Lucene 4.0Similarity functions in Lucene 4.0
Similarity functions in Lucene 4.0Koji Sekiguchi
 
A Hybrid Peer Recommender System for a Online Community Teachers
A Hybrid Peer Recommender System for a Online Community TeachersA Hybrid Peer Recommender System for a Online Community Teachers
A Hybrid Peer Recommender System for a Online Community TeachersDenis Parra Santander
 
Evaluation of Collaborative Filtering Algorithms for Recommending Articles on...
Evaluation of Collaborative Filtering Algorithms for Recommending Articles on...Evaluation of Collaborative Filtering Algorithms for Recommending Articles on...
Evaluation of Collaborative Filtering Algorithms for Recommending Articles on...Denis Parra Santander
 
Identifying Relevant Messages in a Twitter-based Citizen Channel for Natural ...
Identifying Relevant Messages in a Twitter-based Citizen Channel for Natural ...Identifying Relevant Messages in a Twitter-based Citizen Channel for Natural ...
Identifying Relevant Messages in a Twitter-based Citizen Channel for Natural ...Denis Parra Santander
 
Currents steps to be a researcher and faculty
Currents steps to be a researcher and facultyCurrents steps to be a researcher and faculty
Currents steps to be a researcher and facultyDenis Parra Santander
 
Walk the Talk: Analyzing the relation between implicit and explicit feedback ...
Walk the Talk: Analyzing the relation between implicit and explicit feedback ...Walk the Talk: Analyzing the relation between implicit and explicit feedback ...
Walk the Talk: Analyzing the relation between implicit and explicit feedback ...Denis Parra Santander
 
Implicit Feedback Recommendation via Implicit-to-Explicit Ordinal Logistic Re...
Implicit Feedback Recommendation via Implicit-to-Explicit Ordinal Logistic Re...Implicit Feedback Recommendation via Implicit-to-Explicit Ordinal Logistic Re...
Implicit Feedback Recommendation via Implicit-to-Explicit Ordinal Logistic Re...Denis Parra Santander
 
Network Visualization guest lecture at #DataVizQMSS at @Columbia / #SNA at PU...
Network Visualization guest lecture at #DataVizQMSS at @Columbia / #SNA at PU...Network Visualization guest lecture at #DataVizQMSS at @Columbia / #SNA at PU...
Network Visualization guest lecture at #DataVizQMSS at @Columbia / #SNA at PU...Denis Parra Santander
 
Recommender Systems, Matrices and Graphs
Recommender Systems, Matrices and GraphsRecommender Systems, Matrices and Graphs
Recommender Systems, Matrices and GraphsRoelof Pieters
 
Latent factor models for Collaborative Filtering
Latent factor models for Collaborative FilteringLatent factor models for Collaborative Filtering
Latent factor models for Collaborative Filteringsscdotopen
 
Recommender Systems (Machine Learning Summer School 2014 @ CMU)
Recommender Systems (Machine Learning Summer School 2014 @ CMU)Recommender Systems (Machine Learning Summer School 2014 @ CMU)
Recommender Systems (Machine Learning Summer School 2014 @ CMU)Xavier Amatriain
 

Viewers also liked (14)

Recommender Systems, Part 1 - Introduction to approaches and algorithms
Recommender Systems, Part 1 - Introduction to approaches and algorithmsRecommender Systems, Part 1 - Introduction to approaches and algorithms
Recommender Systems, Part 1 - Introduction to approaches and algorithms
 
Similarity functions in Lucene 4.0
Similarity functions in Lucene 4.0Similarity functions in Lucene 4.0
Similarity functions in Lucene 4.0
 
A Hybrid Peer Recommender System for a Online Community Teachers
A Hybrid Peer Recommender System for a Online Community TeachersA Hybrid Peer Recommender System for a Online Community Teachers
A Hybrid Peer Recommender System for a Online Community Teachers
 
Evaluation of Collaborative Filtering Algorithms for Recommending Articles on...
Evaluation of Collaborative Filtering Algorithms for Recommending Articles on...Evaluation of Collaborative Filtering Algorithms for Recommending Articles on...
Evaluation of Collaborative Filtering Algorithms for Recommending Articles on...
 
Identifying Relevant Messages in a Twitter-based Citizen Channel for Natural ...
Identifying Relevant Messages in a Twitter-based Citizen Channel for Natural ...Identifying Relevant Messages in a Twitter-based Citizen Channel for Natural ...
Identifying Relevant Messages in a Twitter-based Citizen Channel for Natural ...
 
Currents steps to be a researcher and faculty
Currents steps to be a researcher and facultyCurrents steps to be a researcher and faculty
Currents steps to be a researcher and faculty
 
Twitter in Academic Conferences
Twitter in Academic ConferencesTwitter in Academic Conferences
Twitter in Academic Conferences
 
Walk the Talk: Analyzing the relation between implicit and explicit feedback ...
Walk the Talk: Analyzing the relation between implicit and explicit feedback ...Walk the Talk: Analyzing the relation between implicit and explicit feedback ...
Walk the Talk: Analyzing the relation between implicit and explicit feedback ...
 
Implicit Feedback Recommendation via Implicit-to-Explicit Ordinal Logistic Re...
Implicit Feedback Recommendation via Implicit-to-Explicit Ordinal Logistic Re...Implicit Feedback Recommendation via Implicit-to-Explicit Ordinal Logistic Re...
Implicit Feedback Recommendation via Implicit-to-Explicit Ordinal Logistic Re...
 
LDA on social bookmarking systems
LDA on social bookmarking systemsLDA on social bookmarking systems
LDA on social bookmarking systems
 
Network Visualization guest lecture at #DataVizQMSS at @Columbia / #SNA at PU...
Network Visualization guest lecture at #DataVizQMSS at @Columbia / #SNA at PU...Network Visualization guest lecture at #DataVizQMSS at @Columbia / #SNA at PU...
Network Visualization guest lecture at #DataVizQMSS at @Columbia / #SNA at PU...
 
Recommender Systems, Matrices and Graphs
Recommender Systems, Matrices and GraphsRecommender Systems, Matrices and Graphs
Recommender Systems, Matrices and Graphs
 
Latent factor models for Collaborative Filtering
Latent factor models for Collaborative FilteringLatent factor models for Collaborative Filtering
Latent factor models for Collaborative Filtering
 
Recommender Systems (Machine Learning Summer School 2014 @ CMU)
Recommender Systems (Machine Learning Summer School 2014 @ CMU)Recommender Systems (Machine Learning Summer School 2014 @ CMU)
Recommender Systems (Machine Learning Summer School 2014 @ CMU)
 

Similar to Research on Recommender Systems: Beyond Ratings and Lists

Recommandation systems -
Recommandation systems - Recommandation systems -
Recommandation systems - Yousef Fadila
 
User experience at Imperial: a case study of qualitative approaches to Primo ...
User experience at Imperial: a case study of qualitative approaches to Primo ...User experience at Imperial: a case study of qualitative approaches to Primo ...
User experience at Imperial: a case study of qualitative approaches to Primo ...Andrew Preater
 
Search and Hyperlinking Overview @MediaEval2014
Search and Hyperlinking Overview @MediaEval2014Search and Hyperlinking Overview @MediaEval2014
Search and Hyperlinking Overview @MediaEval2014Maria Eskevich
 
[系列活動] 人工智慧與機器學習在推薦系統上的應用
[系列活動] 人工智慧與機器學習在推薦系統上的應用[系列活動] 人工智慧與機器學習在推薦系統上的應用
[系列活動] 人工智慧與機器學習在推薦系統上的應用台灣資料科學年會
 
Tutorial: Context In Recommender Systems
Tutorial: Context In Recommender SystemsTutorial: Context In Recommender Systems
Tutorial: Context In Recommender SystemsYONG ZHENG
 
Project Credit: Clifford Lynch - Developing a contributor role taxonomy for s...
Project Credit: Clifford Lynch - Developing a contributor role taxonomy for s...Project Credit: Clifford Lynch - Developing a contributor role taxonomy for s...
Project Credit: Clifford Lynch - Developing a contributor role taxonomy for s...CASRAI
 
10.2.14 Slides: “Doing It: Research Results on Non-ARL Academic Libraries Man...
10.2.14 Slides: “Doing It: Research Results on Non-ARL Academic Libraries Man...10.2.14 Slides: “Doing It: Research Results on Non-ARL Academic Libraries Man...
10.2.14 Slides: “Doing It: Research Results on Non-ARL Academic Libraries Man...DuraSpace
 
Recsys 2014 Tutorial - The Recommender Problem Revisited
Recsys 2014 Tutorial - The Recommender Problem RevisitedRecsys 2014 Tutorial - The Recommender Problem Revisited
Recsys 2014 Tutorial - The Recommender Problem RevisitedXavier Amatriain
 
Scalable Exploration of Relevance Prospects to Support Decision Making
Scalable Exploration of Relevance Prospects to Support Decision MakingScalable Exploration of Relevance Prospects to Support Decision Making
Scalable Exploration of Relevance Prospects to Support Decision MakingKatrien Verbert
 
11.12.14 Slides: “Doing It: Trends Toward Hosted Service Adoption and Impleme...
11.12.14 Slides: “Doing It: Trends Toward Hosted Service Adoption and Impleme...11.12.14 Slides: “Doing It: Trends Toward Hosted Service Adoption and Impleme...
11.12.14 Slides: “Doing It: Trends Toward Hosted Service Adoption and Impleme...DuraSpace
 
Supporting Exploration and Serendipity in Information Retrieval
Supporting Exploration and Serendipity in Information RetrievalSupporting Exploration and Serendipity in Information Retrieval
Supporting Exploration and Serendipity in Information RetrievalNattiya Kanhabua
 
Benchmarking Domain-specific Expert Search using Workshop Program Committees
Benchmarking Domain-specific Expert Search using Workshop Program CommitteesBenchmarking Domain-specific Expert Search using Workshop Program Committees
Benchmarking Domain-specific Expert Search using Workshop Program CommitteesToine Bogers
 
Models for Information Retrieval and Recommendation
Models for Information Retrieval and RecommendationModels for Information Retrieval and Recommendation
Models for Information Retrieval and RecommendationArjen de Vries
 
[UMAP 2016] User-Oriented Context Suggestion
[UMAP 2016] User-Oriented Context Suggestion[UMAP 2016] User-Oriented Context Suggestion
[UMAP 2016] User-Oriented Context SuggestionYONG ZHENG
 
See to believe: capturing insights using contextual inquiry
See to believe: capturing insights using contextual inquirySee to believe: capturing insights using contextual inquiry
See to believe: capturing insights using contextual inquiryDeirdre Costello
 
Highlights from the 8th ACM Conference on Recommender Systems (RecSys 2014)
Highlights from the 8th ACM Conference on Recommender Systems (RecSys 2014)Highlights from the 8th ACM Conference on Recommender Systems (RecSys 2014)
Highlights from the 8th ACM Conference on Recommender Systems (RecSys 2014)David Zibriczky
 
Who's Afraid of Qualitative Analysis?
Who's Afraid of Qualitative Analysis?Who's Afraid of Qualitative Analysis?
Who's Afraid of Qualitative Analysis?BrigitteScott
 
SFScon18 - Ludovik Coba - rrecsys: an R library for prototyping and assessing...
SFScon18 - Ludovik Coba - rrecsys: an R library for prototyping and assessing...SFScon18 - Ludovik Coba - rrecsys: an R library for prototyping and assessing...
SFScon18 - Ludovik Coba - rrecsys: an R library for prototyping and assessing...South Tyrol Free Software Conference
 

Similar to Research on Recommender Systems: Beyond Ratings and Lists (20)

Recommandation systems -
Recommandation systems - Recommandation systems -
Recommandation systems -
 
User experience at Imperial: a case study of qualitative approaches to Primo ...
User experience at Imperial: a case study of qualitative approaches to Primo ...User experience at Imperial: a case study of qualitative approaches to Primo ...
User experience at Imperial: a case study of qualitative approaches to Primo ...
 
Search and Hyperlinking Overview @MediaEval2014
Search and Hyperlinking Overview @MediaEval2014Search and Hyperlinking Overview @MediaEval2014
Search and Hyperlinking Overview @MediaEval2014
 
[系列活動] 人工智慧與機器學習在推薦系統上的應用
[系列活動] 人工智慧與機器學習在推薦系統上的應用[系列活動] 人工智慧與機器學習在推薦系統上的應用
[系列活動] 人工智慧與機器學習在推薦系統上的應用
 
Tutorial: Context In Recommender Systems
Tutorial: Context In Recommender SystemsTutorial: Context In Recommender Systems
Tutorial: Context In Recommender Systems
 
Project Credit: Clifford Lynch - Developing a contributor role taxonomy for s...
Project Credit: Clifford Lynch - Developing a contributor role taxonomy for s...Project Credit: Clifford Lynch - Developing a contributor role taxonomy for s...
Project Credit: Clifford Lynch - Developing a contributor role taxonomy for s...
 
10.2.14 Slides: “Doing It: Research Results on Non-ARL Academic Libraries Man...
10.2.14 Slides: “Doing It: Research Results on Non-ARL Academic Libraries Man...10.2.14 Slides: “Doing It: Research Results on Non-ARL Academic Libraries Man...
10.2.14 Slides: “Doing It: Research Results on Non-ARL Academic Libraries Man...
 
The current oer search dilemma
The current oer search dilemmaThe current oer search dilemma
The current oer search dilemma
 
Recsys 2014 Tutorial - The Recommender Problem Revisited
Recsys 2014 Tutorial - The Recommender Problem RevisitedRecsys 2014 Tutorial - The Recommender Problem Revisited
Recsys 2014 Tutorial - The Recommender Problem Revisited
 
Scalable Exploration of Relevance Prospects to Support Decision Making
Scalable Exploration of Relevance Prospects to Support Decision MakingScalable Exploration of Relevance Prospects to Support Decision Making
Scalable Exploration of Relevance Prospects to Support Decision Making
 
11.12.14 Slides: “Doing It: Trends Toward Hosted Service Adoption and Impleme...
11.12.14 Slides: “Doing It: Trends Toward Hosted Service Adoption and Impleme...11.12.14 Slides: “Doing It: Trends Toward Hosted Service Adoption and Impleme...
11.12.14 Slides: “Doing It: Trends Toward Hosted Service Adoption and Impleme...
 
Supporting Exploration and Serendipity in Information Retrieval
Supporting Exploration and Serendipity in Information RetrievalSupporting Exploration and Serendipity in Information Retrieval
Supporting Exploration and Serendipity in Information Retrieval
 
Benchmarking Domain-specific Expert Search using Workshop Program Committees
Benchmarking Domain-specific Expert Search using Workshop Program CommitteesBenchmarking Domain-specific Expert Search using Workshop Program Committees
Benchmarking Domain-specific Expert Search using Workshop Program Committees
 
Models for Information Retrieval and Recommendation
Models for Information Retrieval and RecommendationModels for Information Retrieval and Recommendation
Models for Information Retrieval and Recommendation
 
[UMAP 2016] User-Oriented Context Suggestion
[UMAP 2016] User-Oriented Context Suggestion[UMAP 2016] User-Oriented Context Suggestion
[UMAP 2016] User-Oriented Context Suggestion
 
See to believe: capturing insights using contextual inquiry
See to believe: capturing insights using contextual inquirySee to believe: capturing insights using contextual inquiry
See to believe: capturing insights using contextual inquiry
 
Highlights from the 8th ACM Conference on Recommender Systems (RecSys 2014)
Highlights from the 8th ACM Conference on Recommender Systems (RecSys 2014)Highlights from the 8th ACM Conference on Recommender Systems (RecSys 2014)
Highlights from the 8th ACM Conference on Recommender Systems (RecSys 2014)
 
Who's Afraid of Qualitative Analysis?
Who's Afraid of Qualitative Analysis?Who's Afraid of Qualitative Analysis?
Who's Afraid of Qualitative Analysis?
 
Exploratory Analysis of User Data
Exploratory Analysis of User DataExploratory Analysis of User Data
Exploratory Analysis of User Data
 
SFScon18 - Ludovik Coba - rrecsys: an R library for prototyping and assessing...
SFScon18 - Ludovik Coba - rrecsys: an R library for prototyping and assessing...SFScon18 - Ludovik Coba - rrecsys: an R library for prototyping and assessing...
SFScon18 - Ludovik Coba - rrecsys: an R library for prototyping and assessing...
 

Recently uploaded

Unraveling Hypertext_ Analyzing Postmodern Elements in Literature.pptx
Unraveling Hypertext_ Analyzing  Postmodern Elements in  Literature.pptxUnraveling Hypertext_ Analyzing  Postmodern Elements in  Literature.pptx
Unraveling Hypertext_ Analyzing Postmodern Elements in Literature.pptxDhatriParmar
 
Grade 9 Quarter 4 Dll Grade 9 Quarter 4 DLL.pdf
Grade 9 Quarter 4 Dll Grade 9 Quarter 4 DLL.pdfGrade 9 Quarter 4 Dll Grade 9 Quarter 4 DLL.pdf
Grade 9 Quarter 4 Dll Grade 9 Quarter 4 DLL.pdfJemuel Francisco
 
Concurrency Control in Database Management system
Concurrency Control in Database Management systemConcurrency Control in Database Management system
Concurrency Control in Database Management systemChristalin Nelson
 
CHEST Proprioceptive neuromuscular facilitation.pptx
CHEST Proprioceptive neuromuscular facilitation.pptxCHEST Proprioceptive neuromuscular facilitation.pptx
CHEST Proprioceptive neuromuscular facilitation.pptxAneriPatwari
 
Beauty Amidst the Bytes_ Unearthing Unexpected Advantages of the Digital Wast...
Beauty Amidst the Bytes_ Unearthing Unexpected Advantages of the Digital Wast...Beauty Amidst the Bytes_ Unearthing Unexpected Advantages of the Digital Wast...
Beauty Amidst the Bytes_ Unearthing Unexpected Advantages of the Digital Wast...DhatriParmar
 
ClimART Action | eTwinning Project
ClimART Action    |    eTwinning ProjectClimART Action    |    eTwinning Project
ClimART Action | eTwinning Projectjordimapav
 
Q-Factor HISPOL Quiz-6th April 2024, Quiz Club NITW
Q-Factor HISPOL Quiz-6th April 2024, Quiz Club NITWQ-Factor HISPOL Quiz-6th April 2024, Quiz Club NITW
Q-Factor HISPOL Quiz-6th April 2024, Quiz Club NITWQuiz Club NITW
 
Mythology Quiz-4th April 2024, Quiz Club NITW
Mythology Quiz-4th April 2024, Quiz Club NITWMythology Quiz-4th April 2024, Quiz Club NITW
Mythology Quiz-4th April 2024, Quiz Club NITWQuiz Club NITW
 
Indexing Structures in Database Management system.pdf
Indexing Structures in Database Management system.pdfIndexing Structures in Database Management system.pdf
Indexing Structures in Database Management system.pdfChristalin Nelson
 
ESP 4-EDITED.pdfmmcncncncmcmmnmnmncnmncmnnjvnnv
ESP 4-EDITED.pdfmmcncncncmcmmnmnmncnmncmnnjvnnvESP 4-EDITED.pdfmmcncncncmcmmnmnmncnmncmnnjvnnv
ESP 4-EDITED.pdfmmcncncncmcmmnmnmncnmncmnnjvnnvRicaMaeCastro1
 
Scientific Writing :Research Discourse
Scientific  Writing :Research  DiscourseScientific  Writing :Research  Discourse
Scientific Writing :Research DiscourseAnita GoswamiGiri
 
Daily Lesson Plan in Mathematics Quarter 4
Daily Lesson Plan in Mathematics Quarter 4Daily Lesson Plan in Mathematics Quarter 4
Daily Lesson Plan in Mathematics Quarter 4JOYLYNSAMANIEGO
 
Decoding the Tweet _ Practical Criticism in the Age of Hashtag.pptx
Decoding the Tweet _ Practical Criticism in the Age of Hashtag.pptxDecoding the Tweet _ Practical Criticism in the Age of Hashtag.pptx
Decoding the Tweet _ Practical Criticism in the Age of Hashtag.pptxDhatriParmar
 
4.16.24 Poverty and Precarity--Desmond.pptx
4.16.24 Poverty and Precarity--Desmond.pptx4.16.24 Poverty and Precarity--Desmond.pptx
4.16.24 Poverty and Precarity--Desmond.pptxmary850239
 
Active Learning Strategies (in short ALS).pdf
Active Learning Strategies (in short ALS).pdfActive Learning Strategies (in short ALS).pdf
Active Learning Strategies (in short ALS).pdfPatidar M
 
MS4 level being good citizen -imperative- (1) (1).pdf
MS4 level   being good citizen -imperative- (1) (1).pdfMS4 level   being good citizen -imperative- (1) (1).pdf
MS4 level being good citizen -imperative- (1) (1).pdfMr Bounab Samir
 
Oppenheimer Film Discussion for Philosophy and Film
Oppenheimer Film Discussion for Philosophy and FilmOppenheimer Film Discussion for Philosophy and Film
Oppenheimer Film Discussion for Philosophy and FilmStan Meyer
 
Transaction Management in Database Management System
Transaction Management in Database Management SystemTransaction Management in Database Management System
Transaction Management in Database Management SystemChristalin Nelson
 

Recently uploaded (20)

Unraveling Hypertext_ Analyzing Postmodern Elements in Literature.pptx
Unraveling Hypertext_ Analyzing  Postmodern Elements in  Literature.pptxUnraveling Hypertext_ Analyzing  Postmodern Elements in  Literature.pptx
Unraveling Hypertext_ Analyzing Postmodern Elements in Literature.pptx
 
Grade 9 Quarter 4 Dll Grade 9 Quarter 4 DLL.pdf
Grade 9 Quarter 4 Dll Grade 9 Quarter 4 DLL.pdfGrade 9 Quarter 4 Dll Grade 9 Quarter 4 DLL.pdf
Grade 9 Quarter 4 Dll Grade 9 Quarter 4 DLL.pdf
 
Concurrency Control in Database Management system
Concurrency Control in Database Management systemConcurrency Control in Database Management system
Concurrency Control in Database Management system
 
CHEST Proprioceptive neuromuscular facilitation.pptx
CHEST Proprioceptive neuromuscular facilitation.pptxCHEST Proprioceptive neuromuscular facilitation.pptx
CHEST Proprioceptive neuromuscular facilitation.pptx
 
Beauty Amidst the Bytes_ Unearthing Unexpected Advantages of the Digital Wast...
Beauty Amidst the Bytes_ Unearthing Unexpected Advantages of the Digital Wast...Beauty Amidst the Bytes_ Unearthing Unexpected Advantages of the Digital Wast...
Beauty Amidst the Bytes_ Unearthing Unexpected Advantages of the Digital Wast...
 
ClimART Action | eTwinning Project
ClimART Action    |    eTwinning ProjectClimART Action    |    eTwinning Project
ClimART Action | eTwinning Project
 
Q-Factor HISPOL Quiz-6th April 2024, Quiz Club NITW
Q-Factor HISPOL Quiz-6th April 2024, Quiz Club NITWQ-Factor HISPOL Quiz-6th April 2024, Quiz Club NITW
Q-Factor HISPOL Quiz-6th April 2024, Quiz Club NITW
 
Faculty Profile prashantha K EEE dept Sri Sairam college of Engineering
Faculty Profile prashantha K EEE dept Sri Sairam college of EngineeringFaculty Profile prashantha K EEE dept Sri Sairam college of Engineering
Faculty Profile prashantha K EEE dept Sri Sairam college of Engineering
 
Mythology Quiz-4th April 2024, Quiz Club NITW
Mythology Quiz-4th April 2024, Quiz Club NITWMythology Quiz-4th April 2024, Quiz Club NITW
Mythology Quiz-4th April 2024, Quiz Club NITW
 
Indexing Structures in Database Management system.pdf
Indexing Structures in Database Management system.pdfIndexing Structures in Database Management system.pdf
Indexing Structures in Database Management system.pdf
 
ESP 4-EDITED.pdfmmcncncncmcmmnmnmncnmncmnnjvnnv
ESP 4-EDITED.pdfmmcncncncmcmmnmnmncnmncmnnjvnnvESP 4-EDITED.pdfmmcncncncmcmmnmnmncnmncmnnjvnnv
ESP 4-EDITED.pdfmmcncncncmcmmnmnmncnmncmnnjvnnv
 
Scientific Writing :Research Discourse
Scientific  Writing :Research  DiscourseScientific  Writing :Research  Discourse
Scientific Writing :Research Discourse
 
Daily Lesson Plan in Mathematics Quarter 4
Daily Lesson Plan in Mathematics Quarter 4Daily Lesson Plan in Mathematics Quarter 4
Daily Lesson Plan in Mathematics Quarter 4
 
Decoding the Tweet _ Practical Criticism in the Age of Hashtag.pptx
Decoding the Tweet _ Practical Criticism in the Age of Hashtag.pptxDecoding the Tweet _ Practical Criticism in the Age of Hashtag.pptx
Decoding the Tweet _ Practical Criticism in the Age of Hashtag.pptx
 
4.16.24 Poverty and Precarity--Desmond.pptx
4.16.24 Poverty and Precarity--Desmond.pptx4.16.24 Poverty and Precarity--Desmond.pptx
4.16.24 Poverty and Precarity--Desmond.pptx
 
prashanth updated resume 2024 for Teaching Profession
prashanth updated resume 2024 for Teaching Professionprashanth updated resume 2024 for Teaching Profession
prashanth updated resume 2024 for Teaching Profession
 
Active Learning Strategies (in short ALS).pdf
Active Learning Strategies (in short ALS).pdfActive Learning Strategies (in short ALS).pdf
Active Learning Strategies (in short ALS).pdf
 
MS4 level being good citizen -imperative- (1) (1).pdf
MS4 level   being good citizen -imperative- (1) (1).pdfMS4 level   being good citizen -imperative- (1) (1).pdf
MS4 level being good citizen -imperative- (1) (1).pdf
 
Oppenheimer Film Discussion for Philosophy and Film
Oppenheimer Film Discussion for Philosophy and FilmOppenheimer Film Discussion for Philosophy and Film
Oppenheimer Film Discussion for Philosophy and Film
 
Transaction Management in Database Management System
Transaction Management in Database Management SystemTransaction Management in Database Management System
Transaction Management in Database Management System
 

Research on Recommender Systems: Beyond Ratings and Lists

  • 1. Research on Recommender Systems: Beyond Ratings and Lists Denis Parra, Ph.D. Information Sciences Assistant Professor, CS Department School of Engineering Pontificia Universidad Católica de Chile Tuesday, November 11th of 2014
  • 2. Outline • Personal Introduction • Quick Overview of Recommender Systems • My Work on Recommender Systems – Tag-Based Recommendation – Implicit-Feedback (time allowing …) – Visual Interactive Interfaces • Summary & Current & Future Work Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 2
  • 3. Personal Introduction • I’m from Valdivia! • There are many reasons to love Valdivia The City Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 3
  • 4. Personal Introduction • I’m from Valdivia! • There are many reasons to love Valdivia The Sports Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 4
  • 5. Personal Introduction • I’m from Valdivia! • There are many reasons to love Valdivia The Animals Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 5
  • 6. Personal Introduction • B.Eng. and professional title of Civil Engineer in Informatics from Universidad Austral de Chile (2004), Valdivia, Chile • Ph.D. in Information Sciences at University of Pittsburgh (2013), Pittsburgh, PA, USA Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 6
  • 7. Personal Introduction • B.Eng. and professional title of Civil Engineer in Informatics from Universidad Austral de Chile (2004), Valdivia, Chile • Ph.D. in Information Sciences at University of Pittsburgh (2013), Pittsburgh, PA, USA Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 7
  • 8. Recommender Systems INTRODUCTION * * Danboard (Danbo): Amazon’s cardboard robot, in these slides it represents a recommender system Nov 11th 2014 8
  • 9. Recommender Systems (RecSys) Systems that help (groups of) people to find relevant items in a crowded item or information space (MacNee et al. 2006) Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 9
  • 10. Why do we care about RecSys? • RecSys have gained popularity due to several domains & applications that require people to make decisions among a large set of items. Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 10
  • 11. A lil’ bit of History • First recommender systems were built at the beginning of 90’s (Tapestry, GroupLens, Ringo) • Online contests, such as the Netflix prize, grew the attention on recommender systems beyond Computer Science (2006-2009) Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 11
  • 12. The Recommendation Problem • The most popular way that the recommendation problem has been presented is about rating prediction: Predict! Item 1 Item 2 … Item m User 1 1 5 4 User 2 5 1 ? … User n 2 5 ? • How good is my prediction? Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 12
  • 13. Recommendation Methods • Without covering all possible methods, the two most typical classifications on recommender algorithms are Classification 1 Classification 2 - Collaborative Filtering - Content-based Filtering - Hybrid - Memory-based - Model-based Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 13
  • 14. Collaborative Filtering (User-based KNN) • Step 1: Finding Similar Users (Pearson Corr.) 5 4 4 1 2 1 5 4 4 1 2 5 Active user User_1 User_2 User_3 active user user_1 user_2 user_3 Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 14
  • 15. Collaborative Filtering (User-based KNN) • Step 1: Finding Similar Users (Pearson Corr.) 5 4 4 1 2 1 5 4 4 1 2 5 Active user User_1 User_2 User_3 Σ r r r r ( )( ) i ⊂ CR ui u ni n r r r r − − u , n ( )2 ( )2 Σ Σ − − i ⊂ CR ui u i ⊂ CR ni n = u n u n Sim u n , , ( , ) active user user_1 0.4472136 user_2 0.49236596 user_3 -0.91520863 Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 15
  • 16. Collaborative Filtering (User-based KNN) • Step 2: Ranking the items to recommend 5 4 4 Item 1 2 1 5 4 4 Active user 2 User_1 3 User_2 4 2 Item 2 Item 3 Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 16
  • 17. Collaborative Filtering (User-based KNN) • Step 2: Ranking the items to recommend 5 4 4 Item 1 2 1 5 4 4 Active user 2 pred u i r User_1 3 User_2 userSim u n r r ( , ) ( ) ⊂ ⋅ − n neighbors u ni n u userSim u n Σ Σ n ⊂ neighbors u = + ( ) ( ) ( , ) ( , ) 4 2 Item 2 Item 3 Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 17
  • 18. Pros/Cons of CF PROS: • Very simple to implement • Content-agnostic • Compared to other techniques such as content-based, is more accurate CONS: • Sparsity • Cold-start • New Item Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 18
  • 19. Content-Based Filtering • Can be traced back to techniques from IR, where the User Profile represents a query. Doc_1 = {w_1, w_2, …., w_3} 5 4 5 user_profile = {w_1, w_2, …., w_3} using TF-IDF, weighting Doc_2 = {w_1, w_2, …., w_3} Doc_3 = {w_1, w_2, …., w_3} Doc_n = {w_1, w_2, …., w_3} Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 19
  • 20. PROS/CONS of Content-Based Filtering PROS: • New items can be matched without previous feedback • It can exploit also techniques such as LSA or LDA • It can use semantic data (ConceptNet, WordNet, etc.) CONS: • Less accurate than collaborative filtering • Tends to overspecialization Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 20
  • 21. Hybridization • Combine previous methods to overcome their weaknesses (Burke, 2002) Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 21
  • 22. C2. Model/Memory Classification • Memory-based methods use the whole dataset in training and prediction. User and Item-based CF are examples. • Model-based methods build a model during training and only use this model during prediction. This makes prediction performance way faster and scalable Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 22
  • 23. Model-based: Matrix Factorization Latent vector of the item Latent vector of the user SVD ~ Singular Value Decomposition Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 23
  • 24. PROS/CONS of MF and latent factors model PROS: • So far, state-of-the-art in terms of accuracy (these methods won the Netflix Prize) • Performance-wise, the best option nowadays: slow at training time O((m+n)3) compared to correlation O(m2n), but linear at prediction time O(m+n) CONS: • Recommendations are obscure: How to explain that certain “latent factors” produced the recommendation Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 24
  • 25. Rethinking the Recommendation Problem • Ratings are scarce: need for exploiting other sources of user preference • User-centric recommendation takes the problem beyond ratings and ranked lists: evaluate user engagement and satisfaction, not only RMSE • Several other dimensions to consider in the evaluation: novelty of the results, diversity, coverage (user and catalog), serendipity • Study de effect of interface characteristics: user-control, explainability Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 25
  • 26. My Take on RecSys Research Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 30
  • 27. My Work on RecSys • Traditional RecSys: accurate prediction and TopN algorithms • In my research I have contributed to RecSys by: – Utilizing other sources of user preference (Social Tags) – Exploiting implicit feedback for recommendation and for mapping explicit feedback – Studying user-centric evaluation: the effect of user controllability on user satisfaction in interactive interfaces • And nowadays: Studying whether Virtual Worlds are a good proxy for real world recommendation tasks Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 31
  • 28. This is not only My work J • Dr. Peter Brusilovsky University of Pittsburgh, PA, USA • Dr. Alexander Troussov IBM Dublin and TCD, Ireland • Dr. Xavier Amatriain TID / Netflix, CA, USA • Dr. Christoff Trattner NTNU, Norway • Dr. Katrien Verbert KU Leuven, Belgium • Dr. Leandro Balby-Marinho UFCG, Brasil Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 32
  • 30. Tag-based Recommendation • D. Parra, P. Brusilovsky. Improving Collaborative Filtering in Social Tagging Systems for the Recommendation of Scientific Articles. Web Intelligence 2010, Toronto, Canada • D. Parra, P. Brusilovsky. Collaborative Filtering for Social Tagging Systems: an Experiment with CiteULike. ACM Recsys 2009, New York, NY, USA Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 34
  • 31. Motivation • Ratings are scarce. Find another source of user preference: Social Tagging Systems User Resource Tags Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 35
  • 32. A Folksonomy • When a user u uses adds an item i using one or more tags t1,…, tn, there is a tagging instance. • The collection of tagging instances produces a folksonomy Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 36
  • 33. Applying CF over the Folksonomy • In the first step: Calculate user similarity Traditional CF Tag-based CF Pearson Correlation over ratings BM25 over social tags • In the second step: incorporate the amount of raters to rank the items (NwCF) Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 37
  • 34. Tag-based CF = Query Doc_1 Doc_2 Doc_3 BM25 Active User Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 38
  • 35. Okapi BM25 BM25: We obtain the similarity between users (neighbors) using their set of tags as “documents” and performing an Okapi BM25 (probabilistic IR model) Retrieval Status Value calculation. Tag frequency in the neighbor (v) profile 1 ( 1) RSV IDF k tf sim(u, v) = ( 1) Σ∈ + td k tf 3 d k ((1 − b ) + b × ( L / L )) + tf k tf = ⋅ t q 1 d ave td tq Tag frequency in the active user (u) profile ( , ) log (1 ( )) ( , ) 10 predʹ′ u i = + nbr i ⋅ pred u i + + ⋅ tq 3 Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 39
  • 36. Evaluation • Crawl during 38 days during June-July 2009 Item # unique instances # users 784 # items 26,599 # tags 26,009 # posts 71,413 # annotations 218,930 avg # items per user 91 avg # users per item 2.68 avg # tags per user 88.02 avg # users per tag 2.65 avg # tags per item 7.07 avg # items per tag 7.23 Item Phase 2 dataset # users 5,849 # articles 574,907 # tags 139,993 #tagging incidents 2,337,571 Filtering process Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 40
  • 37. Cross-validation • Test-validation-train sets, 10-fold cross validation • Training to obtain parameter K: neighb. size • One run the experiment: ~12 hours Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 41
  • 38. Results & Statistical Significance • BM25 is intended to bring more neighbors, at the cost of more noise (neighbors not so similar) • NwCF helps to decrease noise, so it was natural to combine them and try just that option CCF NwCF BM25+CCF BM25+NwCF MAP@10 0.12875 0.1432* 0.1876** 0.1942*** K (neigh.size) 20 22 21 29 Ucov 81.12% 81.12% 99.23% 99.23% Significance over the baseline: *p < 0.236, ** p < 0.033, *** p < 0.001 Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 42
  • 39. Take-aways • We can exploit tags as a source for user similarity in recommendation algorithms • Tag-based (BM25) similarity can be considered as an alternative to Pearson Correlation to calculate user similarity in STS. • Incorporating the number of raters helped to decrease the noise produced by items with too few ratings Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 43
  • 40. Work with Xavier Amatriain IMPLICIT FEEDBACK
  • 41. Implicit-Feedback • Slides are based on two articles: – Parra-Santander, D., & Amatriain, X. (2011). Walk the Talk: Analyzing the relation between implicit and explicit feedback for preference elicitation. Proceedings of UMAP 2011, Girona, Spain – Parra, D., Karatzoglou, A., Amatriain, X., & Yavuz, I. (2011). Implicit feedback recommendation via implicit-to- explicit ordinal logistic regression mapping. Proceedings of the CARS Workshop, Chicago, IL, USA, 2011. Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 45
  • 42. Introduction (1/2) • Most of recommender system approaches rely on explicit information of the users, but… • Explicit feedback: scarce (people are not especially eager to rate or to provide personal info) • Implicit feedback: Is less scarce, but (Hu et al., 2008) There’s no negative feedback … and if you watch a TV program just once or twice? Noisy … but explicit feedback is also noisy (Amatriain et al., 2009) Preference & Confidence … we aim to map the I.F. to preference (our main goal) Lack of evaluation metrics … if we can map I.F. and E.F., we can have a comparable evaluation Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 46
  • 43. Introduction (2/2) • Is it possible to map implicit behavior to explicit preference (ratings)? • Which variables better account for the amount of times a user listens to online albums? [Baltrunas & Amatriain CARS ‘09 workshop – RecSys 2009.] • OUR APPROACH: Study with Last.fm users – Part I: Ask users to rate 100 albums (how to sample) – Part II: Build a model to map collected implicit feedback and context to explicit feedback Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 47
  • 44. Walk the Talk (2011) Albums they listened to during last: 7days, 3months, 6months, year, overall For each album in the list we obtained: # user plays (in each period), # of global listeners and # of global plays Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 48
  • 45. Walk the Talk - 2 • Requirements: 18 y.o., scrobblings > 5000 Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 49
  • 46. Quantization of Data for Sampling • What items should they rate? Item (album) sampling: – Implicit Feedback (IF): playcount for a user on a given album. Changed to scale [1-3], 3 means being more listened to. – Global Popularity (GP): global playcount for all users on a given album [1-3]. Changed to scale [1-3], 3 means being more listened to. – Recentness (R) : time elapsed since user played a given album. Changed to scale [1-3], 3 means being listened to more recently. Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 50
  • 47. 4 Regression Analysis M1: implicit feedback M2: implicit feedback & recentness M4: Interaction of implicit feedback & recentness M3: implicit feedback, recentness, global popularity • Including Recentness increases R2 in more than 10% [ 1 -> 2] • Including GP increases R2, not much compared to RE + IF [ 1 -> 3] • Not Including GP, but including interaction between IF and RE improves the variance of the DV explained by the regression model. [ 2 -> 4 ] Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 51
  • 48. 4.1 Regression Analysis Model RMSE1 RMSE2 User average 1.5308 1.1051 M1: Implicit feedback 1.4206 1.0402 M2: Implicit feedback + recentness 1.4136 1.034 M3: Implicit feedback + recentness + global popularity 1.4130 1.0338 M4: Interaction of Implicit feedback * recentness 1.4127 1.0332 • We tested conclusions of regression analysis by predicting the score, checking RMSE in 10-fold cross validation. • Results of regression analysis are supported. Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 52
  • 49. Conclusions of Part I • Using a linear model, Implicit feedback and recentness can help to predict explicit feedback (in the form of ratings) • Global popularity doesn’t show a significant improvement in the prediction task • Our model can help to relate implicit and explicit feedback, helping to evaluate and compare explicit and implicit recommender systems. Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 53
  • 50. Part II: Extension of Walk the Talk • Implicit Feedback Recommendation via Implicit-to- Explicit OLR Mapping (Recsys 2011, CARS Workshop) – Consider ratings as ordinal variables – Use mixed-models to account for non-independence of observations – Compare with state-of-the-art implicit feedback algorithm Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 54
  • 51. Recalling the 1st study (5/5) • Prediction of rating by multiple Linear Regression evaluated with RMSE. • Results showed that Implicit feedback (play count of the album by a specific user) and recentness (how recently an album was listened to) were important factors, global popularity had a weaker effect. • Results also showed that listening style (if user preferred to listen to single tracks, CDs, or either) was also an important factor, and not the other ones. Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 55
  • 52. ... but • Linear Regression didn’t account for the nested nature of ratings User 1 1 3 5 3 0 4 5 2 2 1 5 4 3 2 User n 3 2 1 0 4 5 2 5 4 3 2 1 3 5 . . . • And ratings were treated as continuous, when they are actually ordinal. Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 56
  • 53. So, Ordinal Logistic Regression! • Actually Mixed-Efects Ordinal Multinomial Logistic Regression • Mixed-effects: Nested nature of ratings • We obtain a distribution over ratings (ordinal multinomial) per each pair USER, ITEM -> we predict the rating using the expected value. • … And we can compare the inferred ratings with a method that directly uses implicit information (playcounts) to recommend ( by Hu, Koren et al. 2007) Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 57
  • 54. Ordinal Regression for Mapping • Model • Predicted value Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 58
  • 55. Datasets • D1: users, albums, if, re, gp, ratings, demographics/consumption • D2: users, albums, if, re, gp, NO RATINGS. Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 59
  • 56. Results Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 60
  • 57. Conclusions & Current Work Problem/ Challenge 1. Ground truth: How many Playcounts to relevancy? > Sensibility Analysis needed 2. Quantization of playcounts (implicit feedback), recentness, and overall number of listeners of an album (global popularity) [1-3] scale v/s raw playcounts > modifiy and compare 3. Additional/Alternative metrics for evaluation [MAP and nDCG used in the paper] Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 61
  • 58. Part of this work with Katrien Verbert VISUALIZATION + USER CONTROLLABILITY
  • 59. Visualization & User Controllability • Motivation: Can user controllability and explainability improve user engagement and satisfaction with a recommender system? • Specific research question: How intersections of contexts of relevance (of recommendation algorithms) might be better represented for user experience with the recommender? Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 63
  • 60. The Concept of Controllability MovieLens: example of traditional recommender list Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 64
  • 61. Visualization & User Controllability • Motivation: Can user controllability and explainability improve user engagement and satisfaction with a recommender system? • Specific research question: How intersections of contexts of relevance (of recommendation algorithms) might be better represented for user experience with the recommender? Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 65
  • 62. Research Platform • The studies were conducted using Conference Navigator, a Conference Support System • Our goal was recommending conference talks http://halley.exp.sis.pitt.edu/cn3/ Program Proceedings Author List Recommendations Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 66
  • 63. Hybrid RecSys: Visualizing Intersections • Clustermap vs. Venn Diagram Clustermap Venn diagram Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 67
  • 64. TalkExplorer – IUI 2013 • Adaptation of Aduna Visualization to CN • Main research question: Does fusion (intersection) of contexts of relevance improve user experience? Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 68
  • 65. TalkExplorer - I Entities Tags, Recommender Agents, Users Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 69
  • 66. TalkExplorer - II • Canvas Area: Intersections of Different Entities Recommender Recommender Cluster with intersect ion of entities Cluster (of talks) associated to only one entity User Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 70
  • 67. TalkExplorer - III Items Talks explored by the user Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 71
  • 68. Our Assumptions • Items which are relevant in more that one aspect could be more valuable to the users • Displaying multiple aspects of relevance visually is important for the users in the process of item’s exploration Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 72
  • 69. TalkExplorer Studies I & II • Study I – Controlled Experiment: Users were asked to discover relevant talks by exploring the three types of entities: tags, recommender agents and users. – Conducted at Hypertext and UMAP 2012 (21 users) – Subjects familiar with Visualizations and Recsys • Study II – Field Study: Users were left free to explore the interface. – Conducted at LAK 2012 and ECTEL 2013 (18 users) – Subjects familiar with visualizations, but not much with RecSys Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 73
  • 70. Evaluation: Intersections & Effectiveness • What do we call an “Intersection”? • We used #explorations on intersections and their effectiveness, defined as: Effectiveness = |푏표표푘푚푎푟푘푒푑 푖푡푒푚푠|/| 푖푛푡푒푟푒푠푒푐푡푖표푛푠 푒푥푝푙표푟푒푑|  Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 74
  • 71. Results of Studies I & II • Effectiveness increases with intersections of more entities • Effectiveness wasn’t affected in the field study (study 2) • … but exploration distribution was affected Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 75
  • 72. SETFUSION: VENN DIAGRAM FOR USER-CONTROLLABLE INTERFACE 76
  • 73. SetFusion – IUI 2014 Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 77
  • 74. SetFusion I Traditional Ranked List Papers sorted by Relevance. It combines 3 recommendation approaches. Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 78
  • 75. SetFusion - II Sliders Allow the user to control the importance of each data source or recommendation method Interactive Venn Diagram Allows the user to inspect and to filter papers recommended. Actions available: - Filter item list by clicking on an area - Highlight a paper by mouse-over on a circle - Scroll to paper by clicking on a circle - Indicate bookmarked papers Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 79
  • 76. SetFusion – UMAP 2012 • Field Study: let users freely explore the interface - ~50% (50 users) tried the SetFusion recommender - 28% (14 users) bookmarked at least one paper - Users explored in average 14.9 talks and bookmarked 7.36 talks in average. A AB ABC AC B BC C 15 7 9 26 18 4 17 16% 7% 9% 27% 19% 4% 18% Distribution of bookmarks per method or combination of methods Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 80
  • 77. TalkExplorer vs. SetFusion • Comparing distributions of explorations In studies 1 and 2 over talkExplorer we observed an important change in the distribution of explorations. Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 81
  • 78. TalkExplorer vs. SetFusion • Comparing distributions of explorations Comparing the field studies: - In TalkExplorer, 84% of the explorations over intersections were performed over clusters of 1 item - In SetFusion, was only 52%, compared to 48% (18% + 30%) of multiple intersections, diff. not statistically significant Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 82
  • 79. Summary & Conclusions • We presented two implementations of visual interactive interfaces that tackle exploration on a recommendation setting • We showed that intersections of several contexts of relevance help to discover relevant items • The visual paradigm used can have a strong effect on user behavior: we need to keep working on visual representation that promote exploration without increasing the cognitive load over the users Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 83
  • 80. Limitations & Future Work • Apply our approach to other domains (fusion of data sources or recommendation algorithms) • For SetFusion, find alternatives to scale the approach to more than 3 sets, potential alternatives: – Clustering and – Radial sets • Consider other factors that interact with the user satisfaction: – Controllability by itself vs. minimum level of accuracy Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 84
  • 81. More Details on SetFusion? • Effect of other variables: gender, age, experience with in the domain, or familiarity with the system • Check our upcoming paper in the IJHCS “User-controllable Personalization: A Case Study with SetFusion”: Controlled Laboratory study with SetFusion versus traditional ranked list Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 85
  • 82. CONCLUSIONS (& CURRENT) & FUTURE WORK
  • 83. Challenges in Recommender Systems • Recommendation to groups • Cross-Domain recommendation • User-centric evaluation • Interactive interfaces and visualization • Improve Evaluation for comparison (P. Campos of U. of Bio-Bio on doing fair evaluations considering time) • ML: Active learning, multi-armed bandits (exploration, exploitation) • Prevent the “Filter Bubble” • Make algorithms resistant to attacks Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 87
  • 84. Are Virtual Worlds Good Proxies for Real World ? • Why? We have a Second Life dataset with 3 connected dimensions of information Social Network Marketplace Virtual World • 2 undergoing projects: Entrepreneurship and LBSN Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 88
  • 85. Entrepreneurship • Can we predict whether a user will create a store and how successful will she/he be? Literature on this area is extremely scarce. Social Network Marketplace James Gaskin SEM, Causal models BYU, USA Stephen Zhang Entrepreneurship PUC Chile Christoph Trattner Social Networks NTNU, Norway Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 89
  • 86. Location-Based Social Networks (LBSN) • How similar are the patterns of mobility in real world and virtual world ? Social Network Virtual World Christoph Trattner Social Networks NTNU, Norway Leandro Balby-Marinho LBSN and RecSys UFCG, Brasil Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 90
  • 87. Other RecSys Activities • I am part of the Program Committee of the 2015 RecSys challenge. Don’t miss it! » Is the user going to buy items in this session? Yes|No » If yes, what are the items that are going to be bought? • Part of team creating the upcoming RecSys Forum (like SIGIR Forum). Coming Soon! (Alan Said, Cataldo Musto, Alejandro Bellogin, etc.) Nov 11th 2014 D.Parra ~ JCC 2014 ~ Invited Talk 92