One common characteristic of research works focused on fairness evaluation (in machine learning) is that they call for some form of parity (equality) either in treatment—meaning they ignore the information about users’ memberships in protected classes during training—or in impact—by enforcing proportional beneficial outcomes to users in different protected classes. In the recommender systems community, fairness has been studied with respect to both users’ and items’ memberships in protected classes defined by some sensitive attributes (e.g., gender or race for users, revenue in a multi-stakeholder setting for items). Again here, the concept has been commonly interpreted as some form of equality—i.e., the degree to which the system is meeting the information needs of all its users in an equal sense. In this work, we propose a probabilistic framework based on generalized cross entropy (GCE) to measure fairness of a given recommendation model. The framework comes with a suite of advantages: first, it allows the system designer to define and measure fairness for both users and items and can be applied to any classification task; second, it can incorporate various notions of fairness as it does not rely on specific and predefined probability distributions and they can be defined at design time; finally, in its design it uses a gain factor, which can be flexibly defined to contemplate different accuracy-related metrics to measure fairness upon decision-support metrics (e.g., precision, recall) or rank-based measures (e.g., NDCG, MAP). An experimental evaluation on four real-world datasets shows the nuances captured by our proposed metric regarding fairness on different user and item attributes, where nearest-neighbor recommenders tend to obtain good results under equality constraints. We observed that when the users are clustered based on both their interaction with the system and other sensitive attributes, such as age or gender, algorithms with similar performance values get different behaviors with respect to user fairness due to the different way they process data for each user cluster.

A flexible framework for evaluating user and item fairness in recommender systems / Deldjoo, Yashar; Anelli, Vito Walter; Zamani, Hamed; Bellogin, Alejandro; Di Noia, Tommaso. - In: USER MODELING AND USER-ADAPTED INTERACTION. - ISSN 0924-1868. - STAMPA. - 31:3(2021), pp. 457-511. [10.1007/s11257-020-09285-1]

A flexible framework for evaluating user and item fairness in recommender systems

Deldjoo, Yashar;Anelli, Vito Walter;Di Noia, Tommaso
2021-01-01

Abstract

One common characteristic of research works focused on fairness evaluation (in machine learning) is that they call for some form of parity (equality) either in treatment—meaning they ignore the information about users’ memberships in protected classes during training—or in impact—by enforcing proportional beneficial outcomes to users in different protected classes. In the recommender systems community, fairness has been studied with respect to both users’ and items’ memberships in protected classes defined by some sensitive attributes (e.g., gender or race for users, revenue in a multi-stakeholder setting for items). Again here, the concept has been commonly interpreted as some form of equality—i.e., the degree to which the system is meeting the information needs of all its users in an equal sense. In this work, we propose a probabilistic framework based on generalized cross entropy (GCE) to measure fairness of a given recommendation model. The framework comes with a suite of advantages: first, it allows the system designer to define and measure fairness for both users and items and can be applied to any classification task; second, it can incorporate various notions of fairness as it does not rely on specific and predefined probability distributions and they can be defined at design time; finally, in its design it uses a gain factor, which can be flexibly defined to contemplate different accuracy-related metrics to measure fairness upon decision-support metrics (e.g., precision, recall) or rank-based measures (e.g., NDCG, MAP). An experimental evaluation on four real-world datasets shows the nuances captured by our proposed metric regarding fairness on different user and item attributes, where nearest-neighbor recommenders tend to obtain good results under equality constraints. We observed that when the users are clustered based on both their interaction with the system and other sensitive attributes, such as age or gender, algorithms with similar performance values get different behaviors with respect to user fairness due to the different way they process data for each user cluster.
2021
A flexible framework for evaluating user and item fairness in recommender systems / Deldjoo, Yashar; Anelli, Vito Walter; Zamani, Hamed; Bellogin, Alejandro; Di Noia, Tommaso. - In: USER MODELING AND USER-ADAPTED INTERACTION. - ISSN 0924-1868. - STAMPA. - 31:3(2021), pp. 457-511. [10.1007/s11257-020-09285-1]
File in questo prodotto:
File Dimensione Formato  
2021_A_flexible_framework_for_evaluating_user_and_item_fairness_in_recommender_systems_pdfeditoriale.pdf

solo gestori catalogo

Tipologia: Versione editoriale
Licenza: Tutti i diritti riservati
Dimensione 896.93 kB
Formato Adobe PDF
896.93 kB Adobe PDF   Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11589/223966
Citazioni
  • Scopus 51
  • ???jsp.display-item.citation.isi??? 41
social impact