E Pluribus Unum: Guidelines on Multi-Objective Evaluation of Recommender Systems

Recommender Systems today are still mostly evaluated in terms of accuracy, with other aspects beyond the immediate relevance of recommendations, such as diversity, long-term user retention and fairness, often taking a back seat. Moreover, reconciling multiple performance perspectives is by definitio...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Chia, Patrick John, Attanasio, Giuseppe, Tagliabue, Jacopo, Bianchi, Federico, Greco, Ciro, Moreira, Gabriel de Souza P, Eynard, Davide, Husain, Fahd
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Chia, Patrick John
Attanasio, Giuseppe
Tagliabue, Jacopo
Bianchi, Federico
Greco, Ciro
Moreira, Gabriel de Souza P
Eynard, Davide
Husain, Fahd
description Recommender Systems today are still mostly evaluated in terms of accuracy, with other aspects beyond the immediate relevance of recommendations, such as diversity, long-term user retention and fairness, often taking a back seat. Moreover, reconciling multiple performance perspectives is by definition indeterminate, presenting a stumbling block to those in the pursuit of rounded evaluation of Recommender Systems. EvalRS 2022 -- a data challenge designed around Multi-Objective Evaluation -- was a first practical endeavour, providing many insights into the requirements and challenges of balancing multiple objectives in evaluation. In this work, we reflect on EvalRS 2022 and expound upon crucial learnings to formulate a first-principles approach toward Multi-Objective model selection, and outline a set of guidelines for carrying out a Multi-Objective Evaluation challenge, with potential applicability to the problem of rounded evaluation of competing models in real-world deployments.
doi_str_mv 10.48550/arxiv.2304.10621
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2304_10621</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2304_10621</sourcerecordid><originalsourceid>FETCH-LOGICAL-a671-a8e6660d921d89110b84526fdf9c3bb313df7ec5b62ad3cb3ee1c4db3ea7976e3</originalsourceid><addsrcrecordid>eNotz71OwzAYhWEvDKhwAUz4BhL8kzgJG6pCQSpqBWWO_PNZMrITZMcRvXtK6fQORzrSg9AdJWXV1jV5kPHHLSXjpCopEYxeo32P9z5Hp3LCn2MOj3iTnQHvRkh4GvFb9rMrduoL9OwWwP0ifZazO02Txe-gpxBgNBDxxzHNENINurLSJ7i9dIUOz_1h_VJsd5vX9dO2kKKhhWxBCEFMx6hpO0qJaquaCWtsp7lSnHJjG9C1EkwarhUHoLoyp8qmawTwFbr_vz2Thu_ogozH4Y82nGn8Fzb6Sig</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>E Pluribus Unum: Guidelines on Multi-Objective Evaluation of Recommender Systems</title><source>arXiv.org</source><creator>Chia, Patrick John ; Attanasio, Giuseppe ; Tagliabue, Jacopo ; Bianchi, Federico ; Greco, Ciro ; Moreira, Gabriel de Souza P ; Eynard, Davide ; Husain, Fahd</creator><creatorcontrib>Chia, Patrick John ; Attanasio, Giuseppe ; Tagliabue, Jacopo ; Bianchi, Federico ; Greco, Ciro ; Moreira, Gabriel de Souza P ; Eynard, Davide ; Husain, Fahd</creatorcontrib><description>Recommender Systems today are still mostly evaluated in terms of accuracy, with other aspects beyond the immediate relevance of recommendations, such as diversity, long-term user retention and fairness, often taking a back seat. Moreover, reconciling multiple performance perspectives is by definition indeterminate, presenting a stumbling block to those in the pursuit of rounded evaluation of Recommender Systems. EvalRS 2022 -- a data challenge designed around Multi-Objective Evaluation -- was a first practical endeavour, providing many insights into the requirements and challenges of balancing multiple objectives in evaluation. In this work, we reflect on EvalRS 2022 and expound upon crucial learnings to formulate a first-principles approach toward Multi-Objective model selection, and outline a set of guidelines for carrying out a Multi-Objective Evaluation challenge, with potential applicability to the problem of rounded evaluation of competing models in real-world deployments.</description><identifier>DOI: 10.48550/arxiv.2304.10621</identifier><language>eng</language><subject>Computer Science - Information Retrieval</subject><creationdate>2023-04</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2304.10621$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2304.10621$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Chia, Patrick John</creatorcontrib><creatorcontrib>Attanasio, Giuseppe</creatorcontrib><creatorcontrib>Tagliabue, Jacopo</creatorcontrib><creatorcontrib>Bianchi, Federico</creatorcontrib><creatorcontrib>Greco, Ciro</creatorcontrib><creatorcontrib>Moreira, Gabriel de Souza P</creatorcontrib><creatorcontrib>Eynard, Davide</creatorcontrib><creatorcontrib>Husain, Fahd</creatorcontrib><title>E Pluribus Unum: Guidelines on Multi-Objective Evaluation of Recommender Systems</title><description>Recommender Systems today are still mostly evaluated in terms of accuracy, with other aspects beyond the immediate relevance of recommendations, such as diversity, long-term user retention and fairness, often taking a back seat. Moreover, reconciling multiple performance perspectives is by definition indeterminate, presenting a stumbling block to those in the pursuit of rounded evaluation of Recommender Systems. EvalRS 2022 -- a data challenge designed around Multi-Objective Evaluation -- was a first practical endeavour, providing many insights into the requirements and challenges of balancing multiple objectives in evaluation. In this work, we reflect on EvalRS 2022 and expound upon crucial learnings to formulate a first-principles approach toward Multi-Objective model selection, and outline a set of guidelines for carrying out a Multi-Objective Evaluation challenge, with potential applicability to the problem of rounded evaluation of competing models in real-world deployments.</description><subject>Computer Science - Information Retrieval</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotz71OwzAYhWEvDKhwAUz4BhL8kzgJG6pCQSpqBWWO_PNZMrITZMcRvXtK6fQORzrSg9AdJWXV1jV5kPHHLSXjpCopEYxeo32P9z5Hp3LCn2MOj3iTnQHvRkh4GvFb9rMrduoL9OwWwP0ifZazO02Txe-gpxBgNBDxxzHNENINurLSJ7i9dIUOz_1h_VJsd5vX9dO2kKKhhWxBCEFMx6hpO0qJaquaCWtsp7lSnHJjG9C1EkwarhUHoLoyp8qmawTwFbr_vz2Thu_ogozH4Y82nGn8Fzb6Sig</recordid><startdate>20230420</startdate><enddate>20230420</enddate><creator>Chia, Patrick John</creator><creator>Attanasio, Giuseppe</creator><creator>Tagliabue, Jacopo</creator><creator>Bianchi, Federico</creator><creator>Greco, Ciro</creator><creator>Moreira, Gabriel de Souza P</creator><creator>Eynard, Davide</creator><creator>Husain, Fahd</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20230420</creationdate><title>E Pluribus Unum: Guidelines on Multi-Objective Evaluation of Recommender Systems</title><author>Chia, Patrick John ; Attanasio, Giuseppe ; Tagliabue, Jacopo ; Bianchi, Federico ; Greco, Ciro ; Moreira, Gabriel de Souza P ; Eynard, Davide ; Husain, Fahd</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a671-a8e6660d921d89110b84526fdf9c3bb313df7ec5b62ad3cb3ee1c4db3ea7976e3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Computer Science - Information Retrieval</topic><toplevel>online_resources</toplevel><creatorcontrib>Chia, Patrick John</creatorcontrib><creatorcontrib>Attanasio, Giuseppe</creatorcontrib><creatorcontrib>Tagliabue, Jacopo</creatorcontrib><creatorcontrib>Bianchi, Federico</creatorcontrib><creatorcontrib>Greco, Ciro</creatorcontrib><creatorcontrib>Moreira, Gabriel de Souza P</creatorcontrib><creatorcontrib>Eynard, Davide</creatorcontrib><creatorcontrib>Husain, Fahd</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Chia, Patrick John</au><au>Attanasio, Giuseppe</au><au>Tagliabue, Jacopo</au><au>Bianchi, Federico</au><au>Greco, Ciro</au><au>Moreira, Gabriel de Souza P</au><au>Eynard, Davide</au><au>Husain, Fahd</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>E Pluribus Unum: Guidelines on Multi-Objective Evaluation of Recommender Systems</atitle><date>2023-04-20</date><risdate>2023</risdate><abstract>Recommender Systems today are still mostly evaluated in terms of accuracy, with other aspects beyond the immediate relevance of recommendations, such as diversity, long-term user retention and fairness, often taking a back seat. Moreover, reconciling multiple performance perspectives is by definition indeterminate, presenting a stumbling block to those in the pursuit of rounded evaluation of Recommender Systems. EvalRS 2022 -- a data challenge designed around Multi-Objective Evaluation -- was a first practical endeavour, providing many insights into the requirements and challenges of balancing multiple objectives in evaluation. In this work, we reflect on EvalRS 2022 and expound upon crucial learnings to formulate a first-principles approach toward Multi-Objective model selection, and outline a set of guidelines for carrying out a Multi-Objective Evaluation challenge, with potential applicability to the problem of rounded evaluation of competing models in real-world deployments.</abstract><doi>10.48550/arxiv.2304.10621</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2304.10621
ispartof
issn
language eng
recordid cdi_arxiv_primary_2304_10621
source arXiv.org
subjects Computer Science - Information Retrieval
title E Pluribus Unum: Guidelines on Multi-Objective Evaluation of Recommender Systems
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-21T11%3A35%3A14IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=E%20Pluribus%20Unum:%20Guidelines%20on%20Multi-Objective%20Evaluation%20of%20Recommender%20Systems&rft.au=Chia,%20Patrick%20John&rft.date=2023-04-20&rft_id=info:doi/10.48550/arxiv.2304.10621&rft_dat=%3Carxiv_GOX%3E2304_10621%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true