Abstract
This tutorial provides a common ground for both researchers and practitioners interested in data and algorithmic bias in recommender systems. Guided by real-world examples in various domains, we introduce problem space and concepts underlying bias investigation in recommendation. Then, we practically show two use cases, addressing biases that lead to disparate exposure of items based on their popularity and to systematically discriminate against a legally-protected class of users. Finally, we cover a range of techniques for evaluating and mitigating the impact of these biases on the recommended lists, including pre-, in-, and post-processing procedures. This tutorial is accompanied by Jupyter notebooks putting into practice core concepts in data from real-world platforms.
Scope
Recommender systems support individuals at filtering the overwhelming alternatives our daily life offers. However, while learning patterns from historical data, such systems may capture biases conveyed by the underlying data in terms of imbalances and inequalities. In many cases, recommender systems can even end up to emphasize these biases in the results they provide, leading to unintended ethical and societal consequences.
Experiments conducted under offline settings with historical data are being frequently used in this field. However, they often fail in providing evidence on biased situations hidden in the recommended lists, as their main focus has been for years related to the accuracy perspective only. With this in mind, recent works showed that some recommenders tend to focus on a limited part of the item catalog mainly due to a popularity bias. Other studies demonstrated that recommenders may generate a category-wise bias due to imbalanced rating distributions across categories. Biased patterns may also impact on users characterized by a specific common sensitive attribute or on users as individuals. With the ever-increasing adoption of recommendation services, it is therefore essential for researchers and practitioners to understand how to face biases and mitigate their impact on these systems.
Even though biases and their consequences are receiving more and more attention by the machine learning community, the complex ecosystems wherein recommender systems are being integrated and the specificity of the modern approaches are requiring solutions tailored for the recommendation task. This tutorial is a natural continuation of our work in this field, and will allow the UMAP community to delve into this area even more, with concrete and practical expertise. We cover the components of modern recommender systems based on neural collaborative filtering, including how they work and their training and testing processes, with pointers to relevant literature. Then, we introduce problem space and definitions of bias in recommendation, and we describe countermeasures at different levels, along with their limitations.
Audience
This tutorial will be particularly useful for early-stage researchers and practitioners. Intermediate and experienced figures who work on user modeling and personalization techniques will also find handy and practical content to deepen their knowledge in data and algorithmic bias. Participants who are not familiar with modern collaborative filtering and recommender systems in general will benefit from an introductory part including foundations and key concepts on such topics. No prior knowledge on data and algorithmic bias by attendees is assumed in order to attend this tutorial.
Objectives
This tutorial will aim to:
- Raise awareness on the importance and the relevance of considering data and algorithmic bias issues in recommendation.
- Play with recommendation pipelines and conduct exploratory analysis aimed at uncovering sources of bias along them.
- Showcase approaches that mitigate bias along with the recommendation pipeline and assess their influence on stakeholders.
- Provide an overview on the current trends and challenges in bias-aware research and identify new research directions.
Structure
All times are displayed in conference local time (UTC+00:00)
Timing | Content |
---|---|
10:00 - 10:10 | Welcome Message and Connection Setup |
10:10 - 11:30 | Session I: Foundations |
10:10 - 10:20 | Recommendation Principles |
10:20 - 11:30 | Data and Algorithmic Bias Foudamentals |
11:30 - 12:00 | Zoom Breakout Room + Q&A |
12:00 - 14:00 | Session II: Hands-on Case Studies |
12:00 - 12:15 | Recommender Systems in Practice |
12:15 - 12:40 | Investigation on Item Popularity Bias |
12:40 - 13:10 | Investigation on Item Provider Fairness |
13:10 - 13:20 | Research Challenges and Emerging Opportunities |
13:20 - 14:00 | Open Discussion + Q&A |
Organizers
- Ludovico Boratto, Eurecat - Centre Tecnológic de Catalunya (Spain)
- Mirko Marras, University of Cagliari (Italy)
Contacts
Please, reaching out to us at ludovico.boratto@acm.org and mirko.marras@unica.it.