evaluating r e c o m m e n d e r systems
play

EVALUATING R E C O M M E N D E R SYSTEMS A C C U R A C Y A N D B - PowerPoint PPT Presentation

EVALUATING R E C O M M E N D E R SYSTEMS A C C U R A C Y A N D B E Y O N D GITHUB.COM/HCORONA/AICS-2016 H U M B E R TO C O R O N A @ TO TO PA M P I N 2 4 -1 0 - 2 0 1 6 A B O U T M E 2 R E F E R E N C E S [1] Humberto Jess Corona


  1. EVALUATING R E C O M M E N D E R SYSTEMS A C C U R A C Y A N D B E Y O N D GITHUB.COM/HCORONA/AICS-2016 H U M B E R TO C O R O N A @ TO TO PA M P I N 2 4 -1 0 - 2 0 1 6

  2. A B O U T M E 2

  3. R E F E R E N C E S [1] Humberto Jesús Corona Pampín, Houssem Jerbi, and Michael P. O’Mahony. "Evaluating the Relative Performance of Neighbourhood-Based Recommender Systems." Spanish Conference of Information Retrieval, 2014 [2] Humberto Jesús Corona Pampín, Houssem Jerbi, and Michael P. O’Mahony. "Evaluating the Relative Performance of Collaborative Filtering Recommender Systems." Journal of Universal Computer Science 21.13 (2015): 1849-1868. 3

  4. ZALANDO https://www.zalando.co.uk/women-street-style/ https://www.zalando.co.uk/men-street-style/ 4

  5. R E C O M M E N D E R S Y S T E M S Enable content discovery by learning the user preferences and exploiting the wisdom of the crowd. 5

  6. E VA L U AT I O N 6

  7. E VA L U AT I O N M E T R I C S R M S E P RE CI S I O N R E C A L L F-1 PER USER C ATA L O G D I V E R S I T Y P O P ULARI TY ITEM U N I Q U E N E S S C O V E R A G E C O V E R A G E 7

  8. E VA L U AT I O N M E T R I C S , A C C U R A C Y R M S E P RE CI S I O N R E C A L L F-1 8

  9. E VA L U AT I O N M E T R I C S , B E Y O N D A C C U R A C Y PER USER C ATA L O G D I V E R S I T Y P O P ULARI TY ITEM U N I Q U E N E S S C O V E R A G E C O V E R A G E 9

  10. E VA L U AT I O N M E T R I C S D I V E R S I T Y 1 0

  11. E VA L U AT I O N M E T R I C S P O P U L A R ITY 11

  12. E VA L U AT I O N M E T R I C S The proportion of items, across the catalog, which are P E R U SER ITEM candidates for recommendations. C O V E R A G E Proportion of items which ever get recommended. C ATA L O G C O V E R A G E 1 2

  13. E VA L U AT I O N M E T R I C S U N I Q U E N E S S 1 3

  14. E VA L U AT I O N M E T R I C S R M S E P RE CI S I O N R E C A L L F-1 PER USER C ATA L O G D I V E R S I T Y P O P ULARI TY ITEM U N I Q U E N E S S C O V E R A G E C O V E R A G E 1 4

  15. E VA L U AT I O N M E T R I C S R M S E P RE CI S I O N R E C A L L F-1 PER USER C ATA L O G D I V E R S I T Y P O P ULARI TY ITEM U N I Q U E N E S S C O V E R A G E C O V E R A G E 1 5

  16. A R E U K N N A N D I K N N R E A L LY T H AT D I F F E R E N T ? A C O M PA R AT I V E A N A LY S I S 1 6

  17. E X P E R I M E N T D E S I G N THE MODELS THE DATA E VA L U AT I O N U K N N A C C U R A C Y M O V I E L E N S - 1 0 0 K IKNN B E Y O N D A C C U R A C Y M O V I E L E N S - 1 M U K N N [ 2 0 , 2 0 0 ] IKNN FIXED TRAINING DATA TESTING DATA 1 0 I T E M S T E S T S E T 1 7

  18. THE ALGORITHMS U S E R B A S E D ITEM-BASED C O L L A B O R AT I V E C O L L A B O R AT I V E FILTERING FILTERING ( U K N N ) ( I K N N ) • Find similar users • Find similar items • word of mouth • Scalable • The neighbours paradigm • Widely used • Scales with number of users 1 8

  19. R E S U LT S 1 9 Insert footnote

  20. R E S U LT S 2 0 Insert footnote

  21. R E S U LT S 2 1 Insert footnote

  22. S U M M A RY 2 2

  23. LESSONS LEARNED • One size fits all is not true, never, ever! • Use many metrics, even if you don’t optimise for them • They help understanding what is the model doing • Use various datasets (if you want to publish a paper) - Do results generalise? • Understand what is the best proxy or dataset for your evaluation goal. 2 3

  24. C O N C L U S I O N S • User-based (UKNN) and item-based (UKNN) collaborative filtering algorithms have a high inverse correlation between popularity and diversity. • Smaller neighbourhood sizes (for UKNN) lead to more unique, less popular, and more diverse recommendations. • Recommend a common set of items at large neighbourhood sizes. • Matrix factorisation approach (WMF) leads to more accurate and diverse recommendations, while being less biased toward popularity. • item-based collaborative filtering (IKNN) has significantly better catalog coverage. 2 4

  25. EVALUATING R E C O M M E N D E R SYSTEMS A C C U R A C Y A N D B E Y O N D GITHUB.COM/HCORONA/AICS-2016 H U M B E R TO C O R O N A @ TO TO PA M P I N 2 4 -1 0 - 2 0 1 6

  26. E X P E R I M E N T I I 2 6

  27. A B I A S A N A LY S I S 2 7

  28. E X P E R I M E N T D E S I G N THE MODELS THE DATA E VA L U AT I O N U K N N A C C U R A C Y FACEBOOK D ATA S E T IKNN B E Y O N D A C C U R A C Y M O V I E L E N S - WMF H E T R E C S I G N I F I C A N C E LASTFM - HETREC A C C U R A C Y OPTIMISATION TRAINING DATA TESTING DATA 1 0 F O L D C R O S S VA L I D AT I O N 2 8

  29. THE DATASETS FACEBOOK M U S I C / B A N D S D ATA S E T LASTFM - HETREC M U S I C / B A N D S M O V I E L E N S - M O V I E S H E T R E C 2 9

  30. THE ALGORITHMS U S E R B A S E D ITEM-BASED M AT R I X C O L L A B O R AT I V E C O L L A B O R AT I V E FACTORISATION FILTERING FILTERING ( W E I G H T E D ) ( U K N N ) ( I K N N ) • Latent Factors • Find similar users • Find similar items • Really good accuracy • word of mouth • Scalable • Scalable • The neighbours paradigm • Widely used • Parallel computing • Scales with number of users • Very accurate 3 0

  31. E VA L U AT I O N M E T R I C S • PRECISION : Out of the items recommended, how many are good recommendations? • RECALL : How many of the items the user likes are being recommended? • F-1 : Mixes the properties of Precision and Recall into a single metric • DIVERSITY : How different are the items in the list of the recommendations? • POPULARITY: How popular are the items recommended • (PER USER) ITEM COVERAGE: Proportion of items that are candidates for recommendations • CATALOG COVERAGE: The proportion of items of the catalog that ever get recommended • UNIQUENESS: How many items in two recommendation lists are different from each other? 3 1

  32. R E S U LT S 3 2

  33. R E S U LT S - P O P U L A R I T Y B I A S 3 3

  34. R E S U LT S - O T H E R P R O P E R T I E S • Accuracy : WMF performs best in terms of F-1 for the Facebook and MovieLens datasets, while the accuracy of the UKNN and IKNN algorithms are similar. • Per-user item coverage •WMF algorithm considers almost every item as a candidate (UICov > 98%). •The UKNN algorithm (by definition) only items which are in the user’s neighbourhood can be considered as recommendation candidates. IKNN was seen to outperform UKNN in all datasets in terms of • Coverage : the IKNN algorithm, performs significantly better than the other algorithms, covering up to 30% of the item catalog - Up to 6 times more items than the UKNN and WMF algorithms. • Diversity: the WMF algorithm performs better, with a performance around 9% higher on average than the best neighbourhood-based approach 3 4

  35. R E S U LT S - C O N S I S T E N C Y •Important to evaluate in different datasets. • MovieLens dataset, (3 times more dense than the Facebook and LastFM datasets), the catalog coverage of the IKNN algorithm is ∼ 10 times smaller than for the LastFM and Facebook datasets. 3 5

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend