skip to main content
10.1145/3397271.3401230acmconferencesArticle/Chapter ViewAbstractPublication PagesirConference Proceedingsconference-collections
short-paper

Using Exploration to Alleviate Closed Loop Effects in Recommender Systems

Published: 25 July 2020 Publication History

Abstract

Recommendation systems are often trained and evaluated based on users' interactions obtained through the use of an existing, already deployed, recommendation system. Hence the deployed recommendation systems will recommend some items and not others, and items will have varying levels of exposure to users. As a result, the collected feedback dataset (including most public datasets) can be skewed towards the particular items favored by the deployed model. In this manner, training new recommender systems from interaction data obtained from a previous model creates a feedback loop, i.e. a closed loop feedback. In this paper, we first introduce the closed loop feedback and then investigate the effect of closed loop feedback in both the training and offline evaluation of recommendation models, in contrast to a further exploration of the users' preferences (obtained from the randomly presented items). To achieve this, we make use of open loop datasets, where randomly selected items are presented to users for feedback. Our experiments using an open loop Yahoo! dataset reveal that there is a strong correlation between the deployed model and a new model that is trained based on the closed loop feedback. Moreover, with the aid of exploration we can decrease the effect of closed loop feedback and obtain new and better generalizable models.

Supplementary Material

MP4 File (3397271.3401230.mp4)
Recommenders are an instance of dynamic systems -- i.e. both the RecSys model and the User have a direct influence on each other in a closed loop scenario. The current public datasets are just a snapshot of this dynamic system. Offline Evaluation based on closed loop feedback is problematic without injecting exploration into the deployed RecSys model.

References

[1]
K. Astrom and R. Murray. 2008. Feedback Systems: An Introduction for Scientists and Engineers .Princeton University Press.
[2]
R. Ca namares, M. Redondo, and P. Castells. 2019. Multi-Armed Recommender System Bandit Ensembles. In Proceedings of RecSys.
[3]
A. Chaney, B. Stewart, and B. Engelhardt. 2018. How Algorithmic Confounding in Recommendation Systems Increases Homogeneity and Decreases Utility. In Proceedings of RecSys.
[4]
A. Jadidinejad, C. Macdonald, and I. Ounis. 2019. How Sensitive is Recommendation Systems' Offline Evaluation to Popularity?. In Workshop on Reinforcement and Robust Estimators for Recommendation (REVEAL).
[5]
R. Jiang, S. Chiappa, T. Lattimore, A. György, and P. Kohli. 2019. Degenerate Feedback Loops in Recommender Systems. In Proceedings of AIES.
[6]
T. Joachims, A. Swaminathan, Y. Raimond, O. Koch, and F. Vasile. 2018. REVEAL 2018: Offline Evaluation for Recommender Systems. In Proceedings of RecSys.
[7]
T. Joachims, A. Swaminathan, and T. Schnabel. 2017. Unbiased Learning-to-Rank with Biased Feedback. In Proceedings of WSDM.
[8]
L. Li, W. Chu, J. Langford, and X. Wang. 2011. Unbiased Offline Evaluation of Contextual-bandit-based News Article Recommendation Algorithms. In Proceedings of WSDM.
[9]
J. McInerney, B. Lacker, S. Hansen, K. Higley, H. Bouchard, A. Gruson, and R. Mehrotra. 2018. Explore, Exploit, and Explain: Personalizing Explainable Recommendations with Bandits. In Proceedings of RecSys.
[10]
T. Nguyen, P. Hui, F.M. Harper, L. Terveen, and J. Konstan. 2014. Exploring the Filter Bubble: The Effect of Using Recommender Systems on Content Diversity. In Proceedings of WWW.
[11]
F. Radlinski, M. Kurup, and T. Joachims. 2008. How Does Clickthrough Data Reflect Retrieval Quality?. In Proceedings of CIKM.
[12]
S. Rendle, C. Freudenthaler, Z. Gantner, and L. Schmidt-Thieme. 2009. BPR: Bayesian Personalized Ranking from Implicit Feedback. In Proceedings of UAI.
[13]
T. Schnabel, P. Bennett, S. Dumais, and T. Joachims. 2018. Short-Term Satisfaction and Long-Term Coverage: Understanding How Users Tolerate Algorithmic Exploration. In Proceedings of WSDM.
[14]
T. Schnabel, A. Swaminathan, A. Singh, N. Chandak, and T. Joachims. 2016. Recommendations as Treatments: Debiasing Learning and Evaluation. In Proceedings of ICML.
[15]
W. Sun, S. Khenissi, O. Nasraoui, and P. Shafto. 2019. Debiasing the Human-Recommender System Feedback Loop in Collaborative Filtering. In Companion Proceedings of WWW.
[16]
W. Sun, O. Nasraoui, and P. Shafto. 2018. Iterated Algorithmic Bias in the Interactive Machine Learning Process of Information Filtering. In Proceedings of KDIR.
[17]
X. Wang, M. Bendersky, D. Metzler, and M. Najork. 2016. Learning to Rank with Selection Bias in Personal Search. In Proceedings of SIGIR.
[18]
J. Weston, H. Yee, and R. Weiss. 2013. Learning to Rank Recommendations with the K-order Statistic Loss. In Proceedings of RecSys.
[19]
L. Yang, Y. Cui, Y. Xuan, C. Wang, S. Belongie, and D. Estrin. 2018. Unbiased Offline Recommender Evaluation for Missing-not-at-random Implicit Feedback. In Proceedings of RecSys.

Cited By

View all

Index Terms

  1. Using Exploration to Alleviate Closed Loop Effects in Recommender Systems

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    SIGIR '20: Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval
    July 2020
    2548 pages
    ISBN:9781450380164
    DOI:10.1145/3397271
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 25 July 2020

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. algorithmic confounding
    2. bandit feedback
    3. closed loop feedback
    4. exploration vs. exploitation
    5. off-policy evaluation
    6. recommendation systems

    Qualifiers

    • Short-paper

    Funding Sources

    • EPSRC

    Conference

    SIGIR '20
    Sponsor:

    Acceptance Rates

    Overall Acceptance Rate 792 of 3,983 submissions, 20%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)28
    • Downloads (Last 6 weeks)3
    Reflects downloads up to 06 Nov 2024

    Other Metrics

    Citations

    Cited By

    View all

    View Options

    Get Access

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media