SlideShare a Scribd company logo
1 of 19
Download to read offline
ACM RecSys 2013, Hong Kong

FABIO AIOLLI
UNIVERSITY OF PADOVA (ITALY)

Efficient Top-N Recommendation for
Very Large Scale Binary Rated Datasets

16/10/2013

F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets

1
Abstract

 Very large datasets: n users, m items, both in the order of millions
 Top-N type of prediction
 Implicit feedback: only information about what people have already rated

Efficiency:
 Efficient MB-like scoring function tailored to implicit feedback that avoids
the computation of the whole m x m (n x n) similarity matrix
Effectiveness:
 Asymmetric similarity matrix
 Asymmetric scoring function
 Calibration
 Ranking Aggregation
16/10/2013

F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets

2
The MSD Challenge @kaggle
 Very large scale, music recommendation challenge
 Predict which songs a user will listen to given the listening history of the user
 Based on the MSD (Million Song Dataset), a freely available collection of

meta-data for one million of contemporary songs
 The challenge was actually based on a subset (Taste Profile Subset) of more

than 48 million rating pairs (user,song). Data consists of about 1.2 million
users and covers more than 380.000 songs
 User-Song matrix is very sparse (density 0.01%)
 153 teams participating
 We had full listening history for about 1M users, plus half of the listening

history for 110K users, for which we were required to predict the missing half
16/10/2013

F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets

3
Why do not use Matrix Factorization?
MF is recognized as a state-of-the-art technique in CF but…
 Model building is very expansive
 Regression setting does not match exactly the implicit setting
 Gradient descent issues: local minima and slow convergence rate

 Too many parameters to optimize (n + m) x k, very sparse matrix, and

no priori knowledge used -> overfitting
 …

MF based solutions provided by organizers at the beginning of the
challenge and MF based entries by other teams have shown really
poor results on this task.
16/10/2013

F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets

4
Memory-based Models
In standard memory based NN models the entire
matrix R is used to generate a prediction
Prediction is performed on-the-fly and no models have to
be constructed
 Independent predictions, can be easily parallelized!

cons

pros



Only few external parameters (lacks of flexibility)
 Needs the complete computation of similarities for every
user-user (or item-item) pair in order to compute NNs


16/10/2013

F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets

5
Memory-based Collaborative Filtering
We first define a modified version of the MB standard model tailored to
CF with implicit feedback as it uses rated information only
 User based:
 Item based:

q represents a locality parameter whose role is similar to taking the NNs. A
bigger q corresponds to a fewer nearest neighbors considered.
Note that, like in MF case, we can write

For each user, the N top-score items are recommended.
 User based: only U similarity computations (U = avg # of users x item)
 Item based: only I similarity computations (I = avg # of items x user)
16/10/2013

F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets

6
Asymmetric Cosine based CF
Given two variables and their (binary) vector representation, we
define:

AsymC has a probabilistic interpretation as an asymmetric
product of conditionals

16/10/2013

F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets

7
Locality effect: Item-based
IS (α=0)

mAP@500

q=1

0.12224

q=2

0.16581

q=3

0.17144

q=4

0.17004

q=5

0.16830

IS (α=1/2)

mAP@500

q=1

0.16439

q=2

0.16214

q=3

0.15587

q=4

0.15021

q=5

0.14621

16/10/2013

F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets

8
Locality effect: User-based
US (α=0)

mAP@500

q=3

0.12479

q=4

0.13289

q=5

0.13400

q=6

0.13187

q=7

0.12878

US (α=1/2)

mAP@500

q=3

0.12532

q=4

0.13779

q=5

0.14355

q=6

0.14487

q=7

0.14352

16/10/2013

F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets

9
AsymC similarity effect

item-based
varying α

16/10/2013

user-based
varying α

F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets

10
Asymmetric Scoring function

User
based

Item
based

Unfortunately, the norm of the weights term is inefficient to compute exactly and,
whenever the number of items is very large, we suggest to estimated it from data
16/10/2013

F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets

11
AsymC scoring effect on user based
recommandation

US, α=0. 5

Best β

mAP@500

q=1

0.07679

0.3

0.14890

q=2

0.10436

0.5

0.15801

q=3

0.12532

0.6

0.16132

q=4

0.13779

0.7

0.16229

q=5

0.14355

0.8

0.16152

q=6

0.14487

0.9

0.15975

q=7

16/10/2013

mAP@500

0.14352

0.9

0.15658

F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets

12
Calibration
Analyses of predicted score when items are actually rated on the
training set
Different items could be map on different scales

16/10/2013

F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets

13
Calibration
The scores are calibrated by a simple piece-wise linear function
1.0

0.5

type

parameters

uncalibrated

calibrated

IS

mAP@500=0.1773 mAP@500=0.1811*

US

mAP@500=0.1623 mAP@500=0.1649

16/10/2013

F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets

14
Ranking Aggregation
 Assuming that



the strategies are precision oriented, meaning that each one tends to make
good recommendations for songs on which they are more confident
different strategies are diverse and can recommend different songs

.. then aggregating different rankings can improve the results
 Aggregating item-based and user-based strategies





Stochastic aggregation: recommended items are chosen stochastically
from the lists
Linear aggregation: recommended items are chosen based on a
combination of scores on different lists
Borda aggregation: recommended items are chosen based on a variant of
the Borda Count algorithm

 More details in the paper!
16/10/2013

F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets

15
Ranking aggregation results
IS, α=0.15,q=3

mAP@500

0.0

1.0

0.14098

0.1

0.9

0.14813

0.2

0.8

0.15559

0.3

0.7

0.16248

0.4

0.6

0.16859

0.5

0.5

0.17362

0.6

0.4

0.17684

0.7

0.3

0.17870

0.8

0.2

0.17896

0.9

0.1

0.17813

1.0
16/10/2013

US, α=0.3,q=5

0.0

1.17732

F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets

16
Final MSD Challenge Results
RANK

TEAM NAME

mAP@500

1

Aio

0.17910

2

Learner

0.17196

3

Nohair

0.15892

4

Team Ubuntu

0.15695

5

TheMiner

0.15639

…

…

…

135

Songs by
Popularity

0.02079

Random

0.00002

…
151

16/10/2013

F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets

17
Final discussion
 Best ranked teams all used approaches based on CF
 The 2-nd ranked team used an approach similar to ours to create a set of features

to use in a learning to rank algorithm
 The 5-th ranked team used the Absorption algorithm by YouTube (graph based,
random walks) to get their best pubblic score
 Based on mine and other participant’s opinion and experiments



Metadata did not help (given the very large implicit info contained in user history is
much more than explicit info in metadata)
Matrix factorization did not help too

 Additional experiments on the MovieLens1M dataset can be found in the paper
 In the future we want to study more on how to exploit rich metadata

information, especially in a cold start setting
16/10/2013

F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets

18
Thank you! Questions are welcome

The MSD competition (info and data):
http://www.kaggle.com/c/msdchallenge
Python code I used for the challenge can be found in
http://www.math.unipd.it/~aiolli/CODE/MSD/

More Related Content

Similar to Efficient Top-N Recommendation for Very Large Binary Datasets

Deep_Learning__INAF_baroncelli.pdf
Deep_Learning__INAF_baroncelli.pdfDeep_Learning__INAF_baroncelli.pdf
Deep_Learning__INAF_baroncelli.pdfasdfasdf214078
 
Airfare prediction using Machine Learning with Apache Spark on 1 billion obse...
Airfare prediction using Machine Learning with Apache Spark on 1 billion obse...Airfare prediction using Machine Learning with Apache Spark on 1 billion obse...
Airfare prediction using Machine Learning with Apache Spark on 1 billion obse...Josef A. Habdank
 
A scalable collaborative filtering framework based on co clustering
A scalable collaborative filtering framework based on co clusteringA scalable collaborative filtering framework based on co clustering
A scalable collaborative filtering framework based on co clusteringAllenWu
 
Combining Lazy Learning, Racing and Subsampling for Effective Feature Selection
Combining Lazy Learning, Racing and Subsampling for Effective Feature SelectionCombining Lazy Learning, Racing and Subsampling for Effective Feature Selection
Combining Lazy Learning, Racing and Subsampling for Effective Feature SelectionGianluca Bontempi
 
Development Infographic
Development InfographicDevelopment Infographic
Development InfographicRealMassive
 
Object Tracking By Online Discriminative Feature Selection Algorithm
Object Tracking By Online Discriminative Feature Selection AlgorithmObject Tracking By Online Discriminative Feature Selection Algorithm
Object Tracking By Online Discriminative Feature Selection AlgorithmIRJET Journal
 
IDA 2015: Efficient model selection for regularized classification by exploit...
IDA 2015: Efficient model selection for regularized classification by exploit...IDA 2015: Efficient model selection for regularized classification by exploit...
IDA 2015: Efficient model selection for regularized classification by exploit...George Balikas
 
Artificial intelligence and IoT
Artificial intelligence and IoTArtificial intelligence and IoT
Artificial intelligence and IoTVeselin Pizurica
 
Developing Web-scale Machine Learning at LinkedIn - From Soup to Nuts
Developing Web-scale Machine Learning at LinkedIn - From Soup to NutsDeveloping Web-scale Machine Learning at LinkedIn - From Soup to Nuts
Developing Web-scale Machine Learning at LinkedIn - From Soup to NutsKun Liu
 
[DOLAP2020] Towards Conversational OLAP
[DOLAP2020] Towards Conversational OLAP[DOLAP2020] Towards Conversational OLAP
[DOLAP2020] Towards Conversational OLAPUniversity of Bologna
 
IRJET- Online Course Recommendation System
IRJET- Online Course Recommendation SystemIRJET- Online Course Recommendation System
IRJET- Online Course Recommendation SystemIRJET Journal
 
Collaborative Filtering Survey
Collaborative Filtering SurveyCollaborative Filtering Survey
Collaborative Filtering Surveymobilizer1000
 
Effective data pre-processing for AutoML
Effective data pre-processing for AutoMLEffective data pre-processing for AutoML
Effective data pre-processing for AutoMLJoseph Giovanelli
 
Stock Price Trend Forecasting using Supervised Learning
Stock Price Trend Forecasting using Supervised LearningStock Price Trend Forecasting using Supervised Learning
Stock Price Trend Forecasting using Supervised LearningSharvil Katariya
 
Real Estate Investment Advising Using Machine Learning
Real Estate Investment Advising Using Machine LearningReal Estate Investment Advising Using Machine Learning
Real Estate Investment Advising Using Machine LearningIRJET Journal
 
Fast Feature Selection for Learning to Rank - ACM International Conference on...
Fast Feature Selection for Learning to Rank - ACM International Conference on...Fast Feature Selection for Learning to Rank - ACM International Conference on...
Fast Feature Selection for Learning to Rank - ACM International Conference on...Andrea Gigli
 

Similar to Efficient Top-N Recommendation for Very Large Binary Datasets (20)

Deep_Learning__INAF_baroncelli.pdf
Deep_Learning__INAF_baroncelli.pdfDeep_Learning__INAF_baroncelli.pdf
Deep_Learning__INAF_baroncelli.pdf
 
Airfare prediction using Machine Learning with Apache Spark on 1 billion obse...
Airfare prediction using Machine Learning with Apache Spark on 1 billion obse...Airfare prediction using Machine Learning with Apache Spark on 1 billion obse...
Airfare prediction using Machine Learning with Apache Spark on 1 billion obse...
 
A scalable collaborative filtering framework based on co clustering
A scalable collaborative filtering framework based on co clusteringA scalable collaborative filtering framework based on co clustering
A scalable collaborative filtering framework based on co clustering
 
Combining Lazy Learning, Racing and Subsampling for Effective Feature Selection
Combining Lazy Learning, Racing and Subsampling for Effective Feature SelectionCombining Lazy Learning, Racing and Subsampling for Effective Feature Selection
Combining Lazy Learning, Racing and Subsampling for Effective Feature Selection
 
AggNet: Deep Learning from Crowds
AggNet: Deep Learning from CrowdsAggNet: Deep Learning from Crowds
AggNet: Deep Learning from Crowds
 
Development Infographic
Development InfographicDevelopment Infographic
Development Infographic
 
GRAPH-BASED RECOMMENDATION SYSTEM
GRAPH-BASED RECOMMENDATION SYSTEMGRAPH-BASED RECOMMENDATION SYSTEM
GRAPH-BASED RECOMMENDATION SYSTEM
 
Object Tracking By Online Discriminative Feature Selection Algorithm
Object Tracking By Online Discriminative Feature Selection AlgorithmObject Tracking By Online Discriminative Feature Selection Algorithm
Object Tracking By Online Discriminative Feature Selection Algorithm
 
IDA 2015: Efficient model selection for regularized classification by exploit...
IDA 2015: Efficient model selection for regularized classification by exploit...IDA 2015: Efficient model selection for regularized classification by exploit...
IDA 2015: Efficient model selection for regularized classification by exploit...
 
Artificial intelligence and IoT
Artificial intelligence and IoTArtificial intelligence and IoT
Artificial intelligence and IoT
 
Developing Web-scale Machine Learning at LinkedIn - From Soup to Nuts
Developing Web-scale Machine Learning at LinkedIn - From Soup to NutsDeveloping Web-scale Machine Learning at LinkedIn - From Soup to Nuts
Developing Web-scale Machine Learning at LinkedIn - From Soup to Nuts
 
[DOLAP2020] Towards Conversational OLAP
[DOLAP2020] Towards Conversational OLAP[DOLAP2020] Towards Conversational OLAP
[DOLAP2020] Towards Conversational OLAP
 
IRJET- Online Course Recommendation System
IRJET- Online Course Recommendation SystemIRJET- Online Course Recommendation System
IRJET- Online Course Recommendation System
 
Collaborative Filtering Survey
Collaborative Filtering SurveyCollaborative Filtering Survey
Collaborative Filtering Survey
 
Effective data pre-processing for AutoML
Effective data pre-processing for AutoMLEffective data pre-processing for AutoML
Effective data pre-processing for AutoML
 
IPL Data Analysis using Data Science
IPL Data Analysis using Data ScienceIPL Data Analysis using Data Science
IPL Data Analysis using Data Science
 
Francesco Serafin
Francesco Serafin Francesco Serafin
Francesco Serafin
 
Stock Price Trend Forecasting using Supervised Learning
Stock Price Trend Forecasting using Supervised LearningStock Price Trend Forecasting using Supervised Learning
Stock Price Trend Forecasting using Supervised Learning
 
Real Estate Investment Advising Using Machine Learning
Real Estate Investment Advising Using Machine LearningReal Estate Investment Advising Using Machine Learning
Real Estate Investment Advising Using Machine Learning
 
Fast Feature Selection for Learning to Rank - ACM International Conference on...
Fast Feature Selection for Learning to Rank - ACM International Conference on...Fast Feature Selection for Learning to Rank - ACM International Conference on...
Fast Feature Selection for Learning to Rank - ACM International Conference on...
 

Recently uploaded

The State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptxThe State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptxLoriGlavin3
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubKalema Edgar
 
Dev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio WebDev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio WebUiPathCommunity
 
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdfHyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdfPrecisely
 
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptxUse of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptxLoriGlavin3
 
A Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptxA Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptxLoriGlavin3
 
Developer Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLDeveloper Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLScyllaDB
 
Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!Commit University
 
WordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your BrandWordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your Brandgvaughan
 
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek SchlawackFwdays
 
Generative AI for Technical Writer or Information Developers
Generative AI for Technical Writer or Information DevelopersGenerative AI for Technical Writer or Information Developers
Generative AI for Technical Writer or Information DevelopersRaghuram Pandurangan
 
Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 3652toLead Limited
 
Moving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdfMoving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdfLoriGlavin3
 
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptxThe Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptxLoriGlavin3
 
Take control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test SuiteTake control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test SuiteDianaGray10
 
"Debugging python applications inside k8s environment", Andrii Soldatenko
"Debugging python applications inside k8s environment", Andrii Soldatenko"Debugging python applications inside k8s environment", Andrii Soldatenko
"Debugging python applications inside k8s environment", Andrii SoldatenkoFwdays
 
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxMerck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxLoriGlavin3
 
TrustArc Webinar - How to Build Consumer Trust Through Data Privacy
TrustArc Webinar - How to Build Consumer Trust Through Data PrivacyTrustArc Webinar - How to Build Consumer Trust Through Data Privacy
TrustArc Webinar - How to Build Consumer Trust Through Data PrivacyTrustArc
 
DevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache MavenDevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache MavenHervé Boutemy
 

Recently uploaded (20)

The State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptxThe State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptx
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding Club
 
Dev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio WebDev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio Web
 
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdfHyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
 
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptxUse of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
 
A Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptxA Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptx
 
Developer Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLDeveloper Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQL
 
Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!
 
WordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your BrandWordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your Brand
 
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
 
Generative AI for Technical Writer or Information Developers
Generative AI for Technical Writer or Information DevelopersGenerative AI for Technical Writer or Information Developers
Generative AI for Technical Writer or Information Developers
 
Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365
 
DMCC Future of Trade Web3 - Special Edition
DMCC Future of Trade Web3 - Special EditionDMCC Future of Trade Web3 - Special Edition
DMCC Future of Trade Web3 - Special Edition
 
Moving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdfMoving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdf
 
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptxThe Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
 
Take control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test SuiteTake control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test Suite
 
"Debugging python applications inside k8s environment", Andrii Soldatenko
"Debugging python applications inside k8s environment", Andrii Soldatenko"Debugging python applications inside k8s environment", Andrii Soldatenko
"Debugging python applications inside k8s environment", Andrii Soldatenko
 
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxMerck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
 
TrustArc Webinar - How to Build Consumer Trust Through Data Privacy
TrustArc Webinar - How to Build Consumer Trust Through Data PrivacyTrustArc Webinar - How to Build Consumer Trust Through Data Privacy
TrustArc Webinar - How to Build Consumer Trust Through Data Privacy
 
DevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache MavenDevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache Maven
 

Efficient Top-N Recommendation for Very Large Binary Datasets

  • 1. ACM RecSys 2013, Hong Kong FABIO AIOLLI UNIVERSITY OF PADOVA (ITALY) Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets 16/10/2013 F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets 1
  • 2. Abstract  Very large datasets: n users, m items, both in the order of millions  Top-N type of prediction  Implicit feedback: only information about what people have already rated Efficiency:  Efficient MB-like scoring function tailored to implicit feedback that avoids the computation of the whole m x m (n x n) similarity matrix Effectiveness:  Asymmetric similarity matrix  Asymmetric scoring function  Calibration  Ranking Aggregation 16/10/2013 F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets 2
  • 3. The MSD Challenge @kaggle  Very large scale, music recommendation challenge  Predict which songs a user will listen to given the listening history of the user  Based on the MSD (Million Song Dataset), a freely available collection of meta-data for one million of contemporary songs  The challenge was actually based on a subset (Taste Profile Subset) of more than 48 million rating pairs (user,song). Data consists of about 1.2 million users and covers more than 380.000 songs  User-Song matrix is very sparse (density 0.01%)  153 teams participating  We had full listening history for about 1M users, plus half of the listening history for 110K users, for which we were required to predict the missing half 16/10/2013 F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets 3
  • 4. Why do not use Matrix Factorization? MF is recognized as a state-of-the-art technique in CF but…  Model building is very expansive  Regression setting does not match exactly the implicit setting  Gradient descent issues: local minima and slow convergence rate  Too many parameters to optimize (n + m) x k, very sparse matrix, and no priori knowledge used -> overfitting  … MF based solutions provided by organizers at the beginning of the challenge and MF based entries by other teams have shown really poor results on this task. 16/10/2013 F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets 4
  • 5. Memory-based Models In standard memory based NN models the entire matrix R is used to generate a prediction Prediction is performed on-the-fly and no models have to be constructed  Independent predictions, can be easily parallelized! cons pros  Only few external parameters (lacks of flexibility)  Needs the complete computation of similarities for every user-user (or item-item) pair in order to compute NNs  16/10/2013 F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets 5
  • 6. Memory-based Collaborative Filtering We first define a modified version of the MB standard model tailored to CF with implicit feedback as it uses rated information only  User based:  Item based: q represents a locality parameter whose role is similar to taking the NNs. A bigger q corresponds to a fewer nearest neighbors considered. Note that, like in MF case, we can write For each user, the N top-score items are recommended.  User based: only U similarity computations (U = avg # of users x item)  Item based: only I similarity computations (I = avg # of items x user) 16/10/2013 F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets 6
  • 7. Asymmetric Cosine based CF Given two variables and their (binary) vector representation, we define: AsymC has a probabilistic interpretation as an asymmetric product of conditionals 16/10/2013 F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets 7
  • 8. Locality effect: Item-based IS (α=0) mAP@500 q=1 0.12224 q=2 0.16581 q=3 0.17144 q=4 0.17004 q=5 0.16830 IS (α=1/2) mAP@500 q=1 0.16439 q=2 0.16214 q=3 0.15587 q=4 0.15021 q=5 0.14621 16/10/2013 F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets 8
  • 9. Locality effect: User-based US (α=0) mAP@500 q=3 0.12479 q=4 0.13289 q=5 0.13400 q=6 0.13187 q=7 0.12878 US (α=1/2) mAP@500 q=3 0.12532 q=4 0.13779 q=5 0.14355 q=6 0.14487 q=7 0.14352 16/10/2013 F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets 9
  • 10. AsymC similarity effect item-based varying α 16/10/2013 user-based varying α F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets 10
  • 11. Asymmetric Scoring function User based Item based Unfortunately, the norm of the weights term is inefficient to compute exactly and, whenever the number of items is very large, we suggest to estimated it from data 16/10/2013 F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets 11
  • 12. AsymC scoring effect on user based recommandation US, α=0. 5 Best β mAP@500 q=1 0.07679 0.3 0.14890 q=2 0.10436 0.5 0.15801 q=3 0.12532 0.6 0.16132 q=4 0.13779 0.7 0.16229 q=5 0.14355 0.8 0.16152 q=6 0.14487 0.9 0.15975 q=7 16/10/2013 mAP@500 0.14352 0.9 0.15658 F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets 12
  • 13. Calibration Analyses of predicted score when items are actually rated on the training set Different items could be map on different scales 16/10/2013 F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets 13
  • 14. Calibration The scores are calibrated by a simple piece-wise linear function 1.0 0.5 type parameters uncalibrated calibrated IS mAP@500=0.1773 mAP@500=0.1811* US mAP@500=0.1623 mAP@500=0.1649 16/10/2013 F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets 14
  • 15. Ranking Aggregation  Assuming that   the strategies are precision oriented, meaning that each one tends to make good recommendations for songs on which they are more confident different strategies are diverse and can recommend different songs .. then aggregating different rankings can improve the results  Aggregating item-based and user-based strategies    Stochastic aggregation: recommended items are chosen stochastically from the lists Linear aggregation: recommended items are chosen based on a combination of scores on different lists Borda aggregation: recommended items are chosen based on a variant of the Borda Count algorithm  More details in the paper! 16/10/2013 F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets 15
  • 16. Ranking aggregation results IS, α=0.15,q=3 mAP@500 0.0 1.0 0.14098 0.1 0.9 0.14813 0.2 0.8 0.15559 0.3 0.7 0.16248 0.4 0.6 0.16859 0.5 0.5 0.17362 0.6 0.4 0.17684 0.7 0.3 0.17870 0.8 0.2 0.17896 0.9 0.1 0.17813 1.0 16/10/2013 US, α=0.3,q=5 0.0 1.17732 F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets 16
  • 17. Final MSD Challenge Results RANK TEAM NAME mAP@500 1 Aio 0.17910 2 Learner 0.17196 3 Nohair 0.15892 4 Team Ubuntu 0.15695 5 TheMiner 0.15639 … … … 135 Songs by Popularity 0.02079 Random 0.00002 … 151 16/10/2013 F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets 17
  • 18. Final discussion  Best ranked teams all used approaches based on CF  The 2-nd ranked team used an approach similar to ours to create a set of features to use in a learning to rank algorithm  The 5-th ranked team used the Absorption algorithm by YouTube (graph based, random walks) to get their best pubblic score  Based on mine and other participant’s opinion and experiments   Metadata did not help (given the very large implicit info contained in user history is much more than explicit info in metadata) Matrix factorization did not help too  Additional experiments on the MovieLens1M dataset can be found in the paper  In the future we want to study more on how to exploit rich metadata information, especially in a cold start setting 16/10/2013 F. Aiolli – Efficient Top-N Recommendation for Very Large Scale Binary Rated Datasets 18
  • 19. Thank you! Questions are welcome The MSD competition (info and data): http://www.kaggle.com/c/msdchallenge Python code I used for the challenge can be found in http://www.math.unipd.it/~aiolli/CODE/MSD/