A quantitative approach for the comparison of additive local explanation methods

Local additive explanation methods are increasingly used to understand the predictions of complex Machine Learning (ML) models. The most used additive methods, SHAP and LIME, suffer from limitations that are rarely measured in the literature. This paper aims to measure these limitations on a wide ra...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Information systems (Oxford) 2023-03, Vol.114 (Special issue on DOLAP 2022: Design, Optimization, Languages and Analytical Processing of Big Data), p.102162, Article 102162
Hauptverfasser: Doumard, Emmanuel, Aligon, Julien, Escriva, Elodie, Excoffier, Jean-Baptiste, Monsarrat, Paul, Soulé-Dupuy, Chantal
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue Special issue on DOLAP 2022: Design, Optimization, Languages and Analytical Processing of Big Data
container_start_page 102162
container_title Information systems (Oxford)
container_volume 114
creator Doumard, Emmanuel
Aligon, Julien
Escriva, Elodie
Excoffier, Jean-Baptiste
Monsarrat, Paul
Soulé-Dupuy, Chantal
description Local additive explanation methods are increasingly used to understand the predictions of complex Machine Learning (ML) models. The most used additive methods, SHAP and LIME, suffer from limitations that are rarely measured in the literature. This paper aims to measure these limitations on a wide range (304) of OpenML datasets using six quantitative metrics, and also evaluate emergent coalitional-based methods to tackle the weaknesses of other methods. We illustrate and validate results on a specific medical dataset, SA-Heart. Our findings reveal that LIME and SHAP’s approximations are particularly efficient in high dimension and generate intelligible global explanations, but they suffer from a lack of precision regarding local explanations and possibly unwanted behavior when changing the method’s parameters. Coalitional-based methods are computationally expensive in high dimension, but offer higher quality local explanations. Finally, we present a roadmap summarizing our work by pointing out the most appropriate method depending on dataset dimensionality and user’s objectives. •A methodology to compare local explanation methods is proposed (including new metrics).•Machine Learning models complexity have an impact on explanations.•Additive local explanation methods are complementary.•Trade-offs exist between the desirable characteristics of local explanations.•A roadmap is proposed to choose the most appropriate explanation method.
doi_str_mv 10.1016/j.is.2022.102162
format Article
fullrecord <record><control><sourceid>hal_cross</sourceid><recordid>TN_cdi_hal_primary_oai_HAL_hal_04239122v1</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0306437922001405</els_id><sourcerecordid>oai_HAL_hal_04239122v1</sourcerecordid><originalsourceid>FETCH-LOGICAL-c370t-b4367904e942ff2e9e66a038afe4ce42bc39cb7eafd67c8a448cf8a52fb52c93</originalsourceid><addsrcrecordid>eNp1kDFPwzAQRi0EEqWwM3plSHFs14nZqgooUiUYulsX56y4SuNghwr-PSlBbEynu_veSfcIuc3ZIme5ut8vfFpwxvnY8lzxMzLLy0JkihXqnMyYYCqTotCX5CqlPWOML7WekbcVff-AbvADDP6IFPo-BrANdSHSoUFqw6GH6FPoaHAU6tr_5NpgoaX42bfQjeS4PeDQhDpdkwsHbcKb3zonu6fH3XqTbV-fX9arbWZFwYaskkIVmknUkjvHUaNSwEQJDqVFySsrtK0KBFerwpYgZWldCUvuqiW3WszJ3XS2gdb00R8gfpkA3mxWW3OaMcmFzjk_5mOWTVkbQ0oR3R-QM3OSZ_bGJ3OSZyZ5I_IwITi-cPQYTbIeO4u1j2gHUwf_P_wNDrt3VA</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>A quantitative approach for the comparison of additive local explanation methods</title><source>Elsevier ScienceDirect Journals</source><creator>Doumard, Emmanuel ; Aligon, Julien ; Escriva, Elodie ; Excoffier, Jean-Baptiste ; Monsarrat, Paul ; Soulé-Dupuy, Chantal</creator><creatorcontrib>Doumard, Emmanuel ; Aligon, Julien ; Escriva, Elodie ; Excoffier, Jean-Baptiste ; Monsarrat, Paul ; Soulé-Dupuy, Chantal</creatorcontrib><description>Local additive explanation methods are increasingly used to understand the predictions of complex Machine Learning (ML) models. The most used additive methods, SHAP and LIME, suffer from limitations that are rarely measured in the literature. This paper aims to measure these limitations on a wide range (304) of OpenML datasets using six quantitative metrics, and also evaluate emergent coalitional-based methods to tackle the weaknesses of other methods. We illustrate and validate results on a specific medical dataset, SA-Heart. Our findings reveal that LIME and SHAP’s approximations are particularly efficient in high dimension and generate intelligible global explanations, but they suffer from a lack of precision regarding local explanations and possibly unwanted behavior when changing the method’s parameters. Coalitional-based methods are computationally expensive in high dimension, but offer higher quality local explanations. Finally, we present a roadmap summarizing our work by pointing out the most appropriate method depending on dataset dimensionality and user’s objectives. •A methodology to compare local explanation methods is proposed (including new metrics).•Machine Learning models complexity have an impact on explanations.•Additive local explanation methods are complementary.•Trade-offs exist between the desirable characteristics of local explanations.•A roadmap is proposed to choose the most appropriate explanation method.</description><identifier>ISSN: 0306-4379</identifier><identifier>EISSN: 1873-6076</identifier><identifier>DOI: 10.1016/j.is.2022.102162</identifier><language>eng</language><publisher>Elsevier Ltd</publisher><subject>Computer Science ; Explainable artificial intelligence (XAI) ; Machine Learning ; Prediction explanation</subject><ispartof>Information systems (Oxford), 2023-03, Vol.114 (Special issue on DOLAP 2022: Design, Optimization, Languages and Analytical Processing of Big Data), p.102162, Article 102162</ispartof><rights>2023 Elsevier Ltd</rights><rights>Distributed under a Creative Commons Attribution 4.0 International License</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c370t-b4367904e942ff2e9e66a038afe4ce42bc39cb7eafd67c8a448cf8a52fb52c93</citedby><cites>FETCH-LOGICAL-c370t-b4367904e942ff2e9e66a038afe4ce42bc39cb7eafd67c8a448cf8a52fb52c93</cites><orcidid>0000-0003-1185-9630 ; 0000-0002-9313-2429 ; 0000-0002-2637-724X ; 0000-0003-3618-967X ; 0000-0002-5473-6035 ; 0000-0002-1954-8733</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://dx.doi.org/10.1016/j.is.2022.102162$$EHTML$$P50$$Gelsevier$$H</linktohtml><link.rule.ids>230,314,777,781,882,3537,27905,27906,45976</link.rule.ids><backlink>$$Uhttps://hal.science/hal-04239122$$DView record in HAL$$Hfree_for_read</backlink></links><search><creatorcontrib>Doumard, Emmanuel</creatorcontrib><creatorcontrib>Aligon, Julien</creatorcontrib><creatorcontrib>Escriva, Elodie</creatorcontrib><creatorcontrib>Excoffier, Jean-Baptiste</creatorcontrib><creatorcontrib>Monsarrat, Paul</creatorcontrib><creatorcontrib>Soulé-Dupuy, Chantal</creatorcontrib><title>A quantitative approach for the comparison of additive local explanation methods</title><title>Information systems (Oxford)</title><description>Local additive explanation methods are increasingly used to understand the predictions of complex Machine Learning (ML) models. The most used additive methods, SHAP and LIME, suffer from limitations that are rarely measured in the literature. This paper aims to measure these limitations on a wide range (304) of OpenML datasets using six quantitative metrics, and also evaluate emergent coalitional-based methods to tackle the weaknesses of other methods. We illustrate and validate results on a specific medical dataset, SA-Heart. Our findings reveal that LIME and SHAP’s approximations are particularly efficient in high dimension and generate intelligible global explanations, but they suffer from a lack of precision regarding local explanations and possibly unwanted behavior when changing the method’s parameters. Coalitional-based methods are computationally expensive in high dimension, but offer higher quality local explanations. Finally, we present a roadmap summarizing our work by pointing out the most appropriate method depending on dataset dimensionality and user’s objectives. •A methodology to compare local explanation methods is proposed (including new metrics).•Machine Learning models complexity have an impact on explanations.•Additive local explanation methods are complementary.•Trade-offs exist between the desirable characteristics of local explanations.•A roadmap is proposed to choose the most appropriate explanation method.</description><subject>Computer Science</subject><subject>Explainable artificial intelligence (XAI)</subject><subject>Machine Learning</subject><subject>Prediction explanation</subject><issn>0306-4379</issn><issn>1873-6076</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><recordid>eNp1kDFPwzAQRi0EEqWwM3plSHFs14nZqgooUiUYulsX56y4SuNghwr-PSlBbEynu_veSfcIuc3ZIme5ut8vfFpwxvnY8lzxMzLLy0JkihXqnMyYYCqTotCX5CqlPWOML7WekbcVff-AbvADDP6IFPo-BrANdSHSoUFqw6GH6FPoaHAU6tr_5NpgoaX42bfQjeS4PeDQhDpdkwsHbcKb3zonu6fH3XqTbV-fX9arbWZFwYaskkIVmknUkjvHUaNSwEQJDqVFySsrtK0KBFerwpYgZWldCUvuqiW3WszJ3XS2gdb00R8gfpkA3mxWW3OaMcmFzjk_5mOWTVkbQ0oR3R-QM3OSZ_bGJ3OSZyZ5I_IwITi-cPQYTbIeO4u1j2gHUwf_P_wNDrt3VA</recordid><startdate>202303</startdate><enddate>202303</enddate><creator>Doumard, Emmanuel</creator><creator>Aligon, Julien</creator><creator>Escriva, Elodie</creator><creator>Excoffier, Jean-Baptiste</creator><creator>Monsarrat, Paul</creator><creator>Soulé-Dupuy, Chantal</creator><general>Elsevier Ltd</general><general>Elsevier</general><scope>AAYXX</scope><scope>CITATION</scope><scope>1XC</scope><scope>VOOES</scope><orcidid>https://orcid.org/0000-0003-1185-9630</orcidid><orcidid>https://orcid.org/0000-0002-9313-2429</orcidid><orcidid>https://orcid.org/0000-0002-2637-724X</orcidid><orcidid>https://orcid.org/0000-0003-3618-967X</orcidid><orcidid>https://orcid.org/0000-0002-5473-6035</orcidid><orcidid>https://orcid.org/0000-0002-1954-8733</orcidid></search><sort><creationdate>202303</creationdate><title>A quantitative approach for the comparison of additive local explanation methods</title><author>Doumard, Emmanuel ; Aligon, Julien ; Escriva, Elodie ; Excoffier, Jean-Baptiste ; Monsarrat, Paul ; Soulé-Dupuy, Chantal</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c370t-b4367904e942ff2e9e66a038afe4ce42bc39cb7eafd67c8a448cf8a52fb52c93</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Computer Science</topic><topic>Explainable artificial intelligence (XAI)</topic><topic>Machine Learning</topic><topic>Prediction explanation</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Doumard, Emmanuel</creatorcontrib><creatorcontrib>Aligon, Julien</creatorcontrib><creatorcontrib>Escriva, Elodie</creatorcontrib><creatorcontrib>Excoffier, Jean-Baptiste</creatorcontrib><creatorcontrib>Monsarrat, Paul</creatorcontrib><creatorcontrib>Soulé-Dupuy, Chantal</creatorcontrib><collection>CrossRef</collection><collection>Hyper Article en Ligne (HAL)</collection><collection>Hyper Article en Ligne (HAL) (Open Access)</collection><jtitle>Information systems (Oxford)</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Doumard, Emmanuel</au><au>Aligon, Julien</au><au>Escriva, Elodie</au><au>Excoffier, Jean-Baptiste</au><au>Monsarrat, Paul</au><au>Soulé-Dupuy, Chantal</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>A quantitative approach for the comparison of additive local explanation methods</atitle><jtitle>Information systems (Oxford)</jtitle><date>2023-03</date><risdate>2023</risdate><volume>114</volume><issue>Special issue on DOLAP 2022: Design, Optimization, Languages and Analytical Processing of Big Data</issue><spage>102162</spage><pages>102162-</pages><artnum>102162</artnum><issn>0306-4379</issn><eissn>1873-6076</eissn><abstract>Local additive explanation methods are increasingly used to understand the predictions of complex Machine Learning (ML) models. The most used additive methods, SHAP and LIME, suffer from limitations that are rarely measured in the literature. This paper aims to measure these limitations on a wide range (304) of OpenML datasets using six quantitative metrics, and also evaluate emergent coalitional-based methods to tackle the weaknesses of other methods. We illustrate and validate results on a specific medical dataset, SA-Heart. Our findings reveal that LIME and SHAP’s approximations are particularly efficient in high dimension and generate intelligible global explanations, but they suffer from a lack of precision regarding local explanations and possibly unwanted behavior when changing the method’s parameters. Coalitional-based methods are computationally expensive in high dimension, but offer higher quality local explanations. Finally, we present a roadmap summarizing our work by pointing out the most appropriate method depending on dataset dimensionality and user’s objectives. •A methodology to compare local explanation methods is proposed (including new metrics).•Machine Learning models complexity have an impact on explanations.•Additive local explanation methods are complementary.•Trade-offs exist between the desirable characteristics of local explanations.•A roadmap is proposed to choose the most appropriate explanation method.</abstract><pub>Elsevier Ltd</pub><doi>10.1016/j.is.2022.102162</doi><orcidid>https://orcid.org/0000-0003-1185-9630</orcidid><orcidid>https://orcid.org/0000-0002-9313-2429</orcidid><orcidid>https://orcid.org/0000-0002-2637-724X</orcidid><orcidid>https://orcid.org/0000-0003-3618-967X</orcidid><orcidid>https://orcid.org/0000-0002-5473-6035</orcidid><orcidid>https://orcid.org/0000-0002-1954-8733</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 0306-4379
ispartof Information systems (Oxford), 2023-03, Vol.114 (Special issue on DOLAP 2022: Design, Optimization, Languages and Analytical Processing of Big Data), p.102162, Article 102162
issn 0306-4379
1873-6076
language eng
recordid cdi_hal_primary_oai_HAL_hal_04239122v1
source Elsevier ScienceDirect Journals
subjects Computer Science
Explainable artificial intelligence (XAI)
Machine Learning
Prediction explanation
title A quantitative approach for the comparison of additive local explanation methods
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-20T08%3A30%3A51IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-hal_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=A%20quantitative%20approach%20for%20the%20comparison%20of%20additive%20local%20explanation%20methods&rft.jtitle=Information%20systems%20(Oxford)&rft.au=Doumard,%20Emmanuel&rft.date=2023-03&rft.volume=114&rft.issue=Special%20issue%20on%20DOLAP%202022:%20Design,%20Optimization,%20Languages%20and%20Analytical%20Processing%20of%20Big%20Data&rft.spage=102162&rft.pages=102162-&rft.artnum=102162&rft.issn=0306-4379&rft.eissn=1873-6076&rft_id=info:doi/10.1016/j.is.2022.102162&rft_dat=%3Chal_cross%3Eoai_HAL_hal_04239122v1%3C/hal_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rft_els_id=S0306437922001405&rfr_iscdi=true