SlideShare a Scribd company logo
1 of 22
Download to read offline
User-Centric Evaluation of a K-Furthest Neighbor
    Collaborative Filtering Recommender Algorithm
    Alan Said*1, Ben Fields#2, Brijnesh J. Jain*, Sahin Albayrak*
*
    Technische Universität Berlin
# musicmetric
    1   @alansaid
    2
                                      CSCW 2013, San Antonio, TX, USA
        @alsothings

                                                  February 27th, 2013
Abstract

• New recommendation algorithm for diverse recommendations
• Based on the k-nearest neighbor algorithm
• Two types of evaluation
   o standard offline evaluation

   o user-centric online evaluation




• Proposed algorithm performs worse than baseline in offline
  evaluation but has higher perceived usefulness from the users
  in online evaluation


                        February 27th, 2013                  2
Outline

•   Background
•   Recommendation
•   K-Nearest Neighbors (knn)
•   K-Furthest Neighbors (kfn)
•   Evaluation & Results
•   Conclusions




                         February 27th, 2013   3
Background and Acknowledgements

Started as a (not very serious) discussion at IJCAI & ICWSM 2011

•   Ben Fields - @alsothings
•   Òscar Celma - @ocelma
•   Markus Schedl - @m_schedl
•   Mohamed Sordo - @neomoha




                         February 27th, 2013                   4
Recommendation

• What is it?
   o Personalized information filtering




• What is the difference to search?
   o Implicit

   o Passively finds most interesting items




• How?



                         February 27th, 2013   5
Recommendation - An example (knn)

Recommending a movie to Bert:

                                                     Cookie            Herry
          what/who     Bert   Ernie       Big Bird             Elmo
                                                     Monster          Monster


          Toy Story     4      4             5         1        4


            E.T.               2             5                          2


         Beetlejuice    4      4             5                  2       3


           Shrek                                       1        3       1


         Zoolander             4                       1




                              February 27th, 2013                               6
Recommendation - An example (knn)

  Recommending a movie to Bert:
                                               Similar to Bert

                                                                      Cookie                 Herry
                 what/who      Bert            Ernie       Big Bird                Elmo
                                                                      Monster               Monster


                 Toy Story      4               4             5         1   K-Nearest Neighbor
                                                                                    4

                                      poor
                    E.T.                        2             5                               2
                                      rating

                 Beetlejuice    4               4             5
                                                                                Potential movies
                                                                                     2         3
                                                                                to recommend
                   Shrek                                                1           3         1

recommendation   Zoolander                      4                       1




                                               February 27th, 2013                                    6
Recommendation - A counter example



       What happens if we flip it?


                              Can we recommend movies
                               disliked by those who are
                                    dissimilar to Bert?


                                               Yes!




                                     February 27th, 2013   7
Recommendation - A counter example (kfn)

      Recommending a movie to Bert:

1. Who is dissimilar to Bert?                                                 Cookie            Herry
                                what/who      Bert     Ernie       Big Bird             Elmo
                                                                              Monster          Monster


                                Toy Story      4         4            5         1        4


                                   E.T.                  2            5                          2


                                Beetlejuice    4         4            5         2        2       3


                                  Shrek                                         1        3       1


                                Zoolander                4




                                                     February 27th, 2013                                 8
Recommendation - A counter example (kfn)

      Recommending a movie to Bert:

1. Who is dissimilar to Bert?                         Cookie
                                what/who      Bert
                                                      Monster
2. What do they dislike?

                                Toy Story      4         1
                                                                                 K-Furthest Neighbor

                                   E.T.


                                Beetlejuice    4         2

                                                                           Disliked by Cookie Monster -
                                  Shrek                  1
                                                                           Liked by Bert?

                                Zoolander




                                                     February 27th, 2013                                  9
Evaluation

What are the effects of this?



                        Diversity :
                        •     Less popular items
                        •     Items the users are not familiar with
                        •     Non standard items




                            February 27th, 2013                       10
Evaluation - Recommendation Accuracy
Traditional - Offline Evaluation
 • Movielens 10M, 70k users
 • Precision@N for users with
   >2N ratings
 • Furthest performs at ~60% of
   Nearest neighbor (for N=100)                             <0.001

However
 • lists of recommended items are practically
    disjoint




                                      February 27th, 2013            11
Evaluation

•Are we missing something?
     Yes



train



test




                       February 27th, 2013   12
Evaluation - Online User Study




                       February 27th, 2013   13
Evaluation - Online User Study




10 recommended
     movies




                                               7 questions




                         February 27th, 2013                 13
Evaluation – Recommendation Utility

Data                               Questions
 • 132 users                        • Novel?
 • 10 recommended                   • Obvious?
   movies each                      • Recognizable?
                                    • Serendipitous?
 • knn: 47 users
                                    • Useful?
 • kfn: 43 users                    • Best movie?
 • random: 42 users                 • Worst movie?
 • training set: Movielens          • Rate each seen movie
   10M                              • State whether movie is familiar
                                    • State whether you would see it

                         February 27th, 2013                      14
Evaluation – Recommendation Utility




                  Do you know the movie?




                      February 27th, 2013   15
Evaluation – Recommendation Utility




                     Have you seen it?



                      February 27th, 2013   15
Evaluation – Recommendation Utility




                                   Would you watch it?



                     February 27th, 2013                 15
Evaluation – Recommendation Utility
                                                                                                  Likert scale
                                                                               1: least agree; 5: most agree


                   rating       novelty       obviousness       recognizable       serendipity     usefulness


      knn          3.64          3.83            2.27              2.69               2.71            2.69


       kfn         3.65          3.95            1.79              2.07               2.65            2.63


     random        3.07          4.17            1.64              1.81               2.48            2.24




 highest rating
                               less obvious/recognizable                          comparable serendipity and
                                                                                  usefulness


 remember: knn and kfn recommend different items, still the experienced quality is similar (or higher)




                                          February 27th, 2013                                                    16
Conclusion

Recommending what your anti-peers do not like creates:
 • more diverse recommendations,
 • with comparable overall usefulness,
 • even though standard offline evaluation says otherwise




                        February 27th, 2013                 17
Questions?




         Thank you for listening!


         For more RecSys stuff, check out:
         www.recsyswiki.com




                            February 27th, 2013   18

More Related Content

More from Alan Said

More from Alan Said (10)

RecSysChallenge Opening
RecSysChallenge OpeningRecSysChallenge Opening
RecSysChallenge Opening
 
Best Practices in Recommender System Challenges
Best Practices in Recommender System ChallengesBest Practices in Recommender System Challenges
Best Practices in Recommender System Challenges
 
Estimating the Magic Barrier of Recommender Systems: A User Study
Estimating the Magic Barrier of Recommender Systems: A User StudyEstimating the Magic Barrier of Recommender Systems: A User Study
Estimating the Magic Barrier of Recommender Systems: A User Study
 
Users and Noise: The Magic Barrier of Recommender Systems
Users and Noise: The Magic Barrier of Recommender SystemsUsers and Noise: The Magic Barrier of Recommender Systems
Users and Noise: The Magic Barrier of Recommender Systems
 
Analyzing Weighting Schemes in Collaborative Filtering: Cold Start, Post Cold...
Analyzing Weighting Schemes in Collaborative Filtering: Cold Start, Post Cold...Analyzing Weighting Schemes in Collaborative Filtering: Cold Start, Post Cold...
Analyzing Weighting Schemes in Collaborative Filtering: Cold Start, Post Cold...
 
CaRR 2012 Opening Presentation
CaRR 2012 Opening PresentationCaRR 2012 Opening Presentation
CaRR 2012 Opening Presentation
 
Personalizing Tags: A Folksonomy-like Approach for Recommending Movies
Personalizing Tags: A Folksonomy-like Approach for Recommending MoviesPersonalizing Tags: A Folksonomy-like Approach for Recommending Movies
Personalizing Tags: A Folksonomy-like Approach for Recommending Movies
 
Inferring Contextual User Profiles - Improving Recommender Performance
Inferring Contextual User Profiles - Improving Recommender PerformanceInferring Contextual User Profiles - Improving Recommender Performance
Inferring Contextual User Profiles - Improving Recommender Performance
 
Using Social- and Pseudo-Social Networks to Improve Recommendation Quality
Using Social- and Pseudo-Social Networks to Improve Recommendation QualityUsing Social- and Pseudo-Social Networks to Improve Recommendation Quality
Using Social- and Pseudo-Social Networks to Improve Recommendation Quality
 
Recommender Systems
Recommender SystemsRecommender Systems
Recommender Systems
 

Recently uploaded

Architecting Cloud Native Applications
Architecting Cloud Native ApplicationsArchitecting Cloud Native Applications
Architecting Cloud Native Applications
WSO2
 
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers:  A Deep Dive into Serverless Spatial Data and FMECloud Frontiers:  A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
Safe Software
 
Finding Java's Hidden Performance Traps @ DevoxxUK 2024
Finding Java's Hidden Performance Traps @ DevoxxUK 2024Finding Java's Hidden Performance Traps @ DevoxxUK 2024
Finding Java's Hidden Performance Traps @ DevoxxUK 2024
Victor Rentea
 

Recently uploaded (20)

Architecting Cloud Native Applications
Architecting Cloud Native ApplicationsArchitecting Cloud Native Applications
Architecting Cloud Native Applications
 
Artificial Intelligence Chap.5 : Uncertainty
Artificial Intelligence Chap.5 : UncertaintyArtificial Intelligence Chap.5 : Uncertainty
Artificial Intelligence Chap.5 : Uncertainty
 
Mcleodganj Call Girls 🥰 8617370543 Service Offer VIP Hot Model
Mcleodganj Call Girls 🥰 8617370543 Service Offer VIP Hot ModelMcleodganj Call Girls 🥰 8617370543 Service Offer VIP Hot Model
Mcleodganj Call Girls 🥰 8617370543 Service Offer VIP Hot Model
 
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
 
Introduction to Multilingual Retrieval Augmented Generation (RAG)
Introduction to Multilingual Retrieval Augmented Generation (RAG)Introduction to Multilingual Retrieval Augmented Generation (RAG)
Introduction to Multilingual Retrieval Augmented Generation (RAG)
 
Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...
Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...
Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...
 
Corporate and higher education May webinar.pptx
Corporate and higher education May webinar.pptxCorporate and higher education May webinar.pptx
Corporate and higher education May webinar.pptx
 
CNIC Information System with Pakdata Cf In Pakistan
CNIC Information System with Pakdata Cf In PakistanCNIC Information System with Pakdata Cf In Pakistan
CNIC Information System with Pakdata Cf In Pakistan
 
Strategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a FresherStrategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a Fresher
 
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers:  A Deep Dive into Serverless Spatial Data and FMECloud Frontiers:  A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
 
"I see eyes in my soup": How Delivery Hero implemented the safety system for ...
"I see eyes in my soup": How Delivery Hero implemented the safety system for ..."I see eyes in my soup": How Delivery Hero implemented the safety system for ...
"I see eyes in my soup": How Delivery Hero implemented the safety system for ...
 
Platformless Horizons for Digital Adaptability
Platformless Horizons for Digital AdaptabilityPlatformless Horizons for Digital Adaptability
Platformless Horizons for Digital Adaptability
 
Apidays New York 2024 - Passkeys: Developing APIs to enable passwordless auth...
Apidays New York 2024 - Passkeys: Developing APIs to enable passwordless auth...Apidays New York 2024 - Passkeys: Developing APIs to enable passwordless auth...
Apidays New York 2024 - Passkeys: Developing APIs to enable passwordless auth...
 
Strategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
Strategize a Smooth Tenant-to-tenant Migration and Copilot TakeoffStrategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
Strategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
 
Finding Java's Hidden Performance Traps @ DevoxxUK 2024
Finding Java's Hidden Performance Traps @ DevoxxUK 2024Finding Java's Hidden Performance Traps @ DevoxxUK 2024
Finding Java's Hidden Performance Traps @ DevoxxUK 2024
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected Worker
 
Vector Search -An Introduction in Oracle Database 23ai.pptx
Vector Search -An Introduction in Oracle Database 23ai.pptxVector Search -An Introduction in Oracle Database 23ai.pptx
Vector Search -An Introduction in Oracle Database 23ai.pptx
 
MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024
 
AI+A11Y 11MAY2024 HYDERBAD GAAD 2024 - HelloA11Y (11 May 2024)
AI+A11Y 11MAY2024 HYDERBAD GAAD 2024 - HelloA11Y (11 May 2024)AI+A11Y 11MAY2024 HYDERBAD GAAD 2024 - HelloA11Y (11 May 2024)
AI+A11Y 11MAY2024 HYDERBAD GAAD 2024 - HelloA11Y (11 May 2024)
 
JohnPollard-hybrid-app-RailsConf2024.pptx
JohnPollard-hybrid-app-RailsConf2024.pptxJohnPollard-hybrid-app-RailsConf2024.pptx
JohnPollard-hybrid-app-RailsConf2024.pptx
 

User-Centric Evaluation of a K-Furthest Neighbor Collaborative Filtering Recommender Algorithm

  • 1. User-Centric Evaluation of a K-Furthest Neighbor Collaborative Filtering Recommender Algorithm Alan Said*1, Ben Fields#2, Brijnesh J. Jain*, Sahin Albayrak* * Technische Universität Berlin # musicmetric 1 @alansaid 2 CSCW 2013, San Antonio, TX, USA @alsothings February 27th, 2013
  • 2. Abstract • New recommendation algorithm for diverse recommendations • Based on the k-nearest neighbor algorithm • Two types of evaluation o standard offline evaluation o user-centric online evaluation • Proposed algorithm performs worse than baseline in offline evaluation but has higher perceived usefulness from the users in online evaluation February 27th, 2013 2
  • 3. Outline • Background • Recommendation • K-Nearest Neighbors (knn) • K-Furthest Neighbors (kfn) • Evaluation & Results • Conclusions February 27th, 2013 3
  • 4. Background and Acknowledgements Started as a (not very serious) discussion at IJCAI & ICWSM 2011 • Ben Fields - @alsothings • Òscar Celma - @ocelma • Markus Schedl - @m_schedl • Mohamed Sordo - @neomoha February 27th, 2013 4
  • 5. Recommendation • What is it? o Personalized information filtering • What is the difference to search? o Implicit o Passively finds most interesting items • How? February 27th, 2013 5
  • 6. Recommendation - An example (knn) Recommending a movie to Bert: Cookie Herry what/who Bert Ernie Big Bird Elmo Monster Monster Toy Story 4 4 5 1 4 E.T. 2 5 2 Beetlejuice 4 4 5 2 3 Shrek 1 3 1 Zoolander 4 1 February 27th, 2013 6
  • 7. Recommendation - An example (knn) Recommending a movie to Bert: Similar to Bert Cookie Herry what/who Bert Ernie Big Bird Elmo Monster Monster Toy Story 4 4 5 1 K-Nearest Neighbor 4 poor E.T. 2 5 2 rating Beetlejuice 4 4 5 Potential movies 2 3 to recommend Shrek 1 3 1 recommendation Zoolander 4 1 February 27th, 2013 6
  • 8. Recommendation - A counter example What happens if we flip it? Can we recommend movies disliked by those who are dissimilar to Bert? Yes! February 27th, 2013 7
  • 9. Recommendation - A counter example (kfn) Recommending a movie to Bert: 1. Who is dissimilar to Bert? Cookie Herry what/who Bert Ernie Big Bird Elmo Monster Monster Toy Story 4 4 5 1 4 E.T. 2 5 2 Beetlejuice 4 4 5 2 2 3 Shrek 1 3 1 Zoolander 4 February 27th, 2013 8
  • 10. Recommendation - A counter example (kfn) Recommending a movie to Bert: 1. Who is dissimilar to Bert? Cookie what/who Bert Monster 2. What do they dislike? Toy Story 4 1 K-Furthest Neighbor E.T. Beetlejuice 4 2 Disliked by Cookie Monster - Shrek 1 Liked by Bert? Zoolander February 27th, 2013 9
  • 11. Evaluation What are the effects of this? Diversity : • Less popular items • Items the users are not familiar with • Non standard items February 27th, 2013 10
  • 12. Evaluation - Recommendation Accuracy Traditional - Offline Evaluation • Movielens 10M, 70k users • Precision@N for users with >2N ratings • Furthest performs at ~60% of Nearest neighbor (for N=100) <0.001 However • lists of recommended items are practically disjoint February 27th, 2013 11
  • 13. Evaluation •Are we missing something? Yes train test February 27th, 2013 12
  • 14. Evaluation - Online User Study February 27th, 2013 13
  • 15. Evaluation - Online User Study 10 recommended movies 7 questions February 27th, 2013 13
  • 16. Evaluation – Recommendation Utility Data Questions • 132 users • Novel? • 10 recommended • Obvious? movies each • Recognizable? • Serendipitous? • knn: 47 users • Useful? • kfn: 43 users • Best movie? • random: 42 users • Worst movie? • training set: Movielens • Rate each seen movie 10M • State whether movie is familiar • State whether you would see it February 27th, 2013 14
  • 17. Evaluation – Recommendation Utility Do you know the movie? February 27th, 2013 15
  • 18. Evaluation – Recommendation Utility Have you seen it? February 27th, 2013 15
  • 19. Evaluation – Recommendation Utility Would you watch it? February 27th, 2013 15
  • 20. Evaluation – Recommendation Utility Likert scale 1: least agree; 5: most agree rating novelty obviousness recognizable serendipity usefulness knn 3.64 3.83 2.27 2.69 2.71 2.69 kfn 3.65 3.95 1.79 2.07 2.65 2.63 random 3.07 4.17 1.64 1.81 2.48 2.24 highest rating less obvious/recognizable comparable serendipity and usefulness remember: knn and kfn recommend different items, still the experienced quality is similar (or higher) February 27th, 2013 16
  • 21. Conclusion Recommending what your anti-peers do not like creates: • more diverse recommendations, • with comparable overall usefulness, • even though standard offline evaluation says otherwise February 27th, 2013 17
  • 22. Questions? Thank you for listening! For more RecSys stuff, check out: www.recsyswiki.com February 27th, 2013 18