Explanations based on the Missing: Towards Contrastive Explanations with Pertinent Negatives

In this paper we propose a novel method that provides contrastive explanations justifying the classification of an input by a black box classifier such as a deep neural network. Given an input we find what should be %necessarily and minimally and sufficiently present (viz. important object pixels in...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Dhurandhar, Amit, Chen, Pin-Yu, Luss, Ronny, Tu, Chun-Chen, Ting, Paishun, Shanmugam, Karthikeyan, Das, Payel
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Dhurandhar, Amit
Chen, Pin-Yu
Luss, Ronny
Tu, Chun-Chen
Ting, Paishun
Shanmugam, Karthikeyan
Das, Payel
description In this paper we propose a novel method that provides contrastive explanations justifying the classification of an input by a black box classifier such as a deep neural network. Given an input we find what should be %necessarily and minimally and sufficiently present (viz. important object pixels in an image) to justify its classification and analogously what should be minimally and necessarily \emph{absent} (viz. certain background pixels). We argue that such explanations are natural for humans and are used commonly in domains such as health care and criminology. What is minimally but critically \emph{absent} is an important part of an explanation, which to the best of our knowledge, has not been explicitly identified by current explanation methods that explain predictions of neural networks. We validate our approach on three real datasets obtained from diverse domains; namely, a handwritten digits dataset MNIST, a large procurement fraud dataset and a brain activity strength dataset. In all three cases, we witness the power of our approach in generating precise explanations that are also easy for human experts to understand and evaluate.
doi_str_mv 10.48550/arxiv.1802.07623
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_1802_07623</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>1802_07623</sourcerecordid><originalsourceid>FETCH-LOGICAL-a673-8e9281b16e7807d387bd0c0a23225fea7cb950495c32a018a85693a99ccc29c43</originalsourceid><addsrcrecordid>eNpVz7tuwjAYBWAvDBXtA3TCL5DUlzi2u1URvUi0MGREiv44P2AJHGRbQN--gnbpdJZzjvQR8shZWRml2BPEiz-V3DBRMl0LeUfW88txDwGyH0OiPSQc6Bho3iH99Cn5sH2m7XiGOCTajCFHSNmfkP6bnX3e0RXG7AOGTL9wC9dSuieTDewTPvzllLSv87Z5LxbLt4_mZVFArWVh0ArDe16jNkwP0uh-YI6BkEKoDYJ2vVWssspJAYwbMKq2Eqx1zgnrKjkls9_bG687Rn-A-N1dmd2NKX8A1YxOQw</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Explanations based on the Missing: Towards Contrastive Explanations with Pertinent Negatives</title><source>arXiv.org</source><creator>Dhurandhar, Amit ; Chen, Pin-Yu ; Luss, Ronny ; Tu, Chun-Chen ; Ting, Paishun ; Shanmugam, Karthikeyan ; Das, Payel</creator><creatorcontrib>Dhurandhar, Amit ; Chen, Pin-Yu ; Luss, Ronny ; Tu, Chun-Chen ; Ting, Paishun ; Shanmugam, Karthikeyan ; Das, Payel</creatorcontrib><description>In this paper we propose a novel method that provides contrastive explanations justifying the classification of an input by a black box classifier such as a deep neural network. Given an input we find what should be %necessarily and minimally and sufficiently present (viz. important object pixels in an image) to justify its classification and analogously what should be minimally and necessarily \emph{absent} (viz. certain background pixels). We argue that such explanations are natural for humans and are used commonly in domains such as health care and criminology. What is minimally but critically \emph{absent} is an important part of an explanation, which to the best of our knowledge, has not been explicitly identified by current explanation methods that explain predictions of neural networks. We validate our approach on three real datasets obtained from diverse domains; namely, a handwritten digits dataset MNIST, a large procurement fraud dataset and a brain activity strength dataset. In all three cases, we witness the power of our approach in generating precise explanations that are also easy for human experts to understand and evaluate.</description><identifier>DOI: 10.48550/arxiv.1802.07623</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Computer Vision and Pattern Recognition ; Computer Science - Learning</subject><creationdate>2018-02</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/1802.07623$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.1802.07623$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Dhurandhar, Amit</creatorcontrib><creatorcontrib>Chen, Pin-Yu</creatorcontrib><creatorcontrib>Luss, Ronny</creatorcontrib><creatorcontrib>Tu, Chun-Chen</creatorcontrib><creatorcontrib>Ting, Paishun</creatorcontrib><creatorcontrib>Shanmugam, Karthikeyan</creatorcontrib><creatorcontrib>Das, Payel</creatorcontrib><title>Explanations based on the Missing: Towards Contrastive Explanations with Pertinent Negatives</title><description>In this paper we propose a novel method that provides contrastive explanations justifying the classification of an input by a black box classifier such as a deep neural network. Given an input we find what should be %necessarily and minimally and sufficiently present (viz. important object pixels in an image) to justify its classification and analogously what should be minimally and necessarily \emph{absent} (viz. certain background pixels). We argue that such explanations are natural for humans and are used commonly in domains such as health care and criminology. What is minimally but critically \emph{absent} is an important part of an explanation, which to the best of our knowledge, has not been explicitly identified by current explanation methods that explain predictions of neural networks. We validate our approach on three real datasets obtained from diverse domains; namely, a handwritten digits dataset MNIST, a large procurement fraud dataset and a brain activity strength dataset. In all three cases, we witness the power of our approach in generating precise explanations that are also easy for human experts to understand and evaluate.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Computer Vision and Pattern Recognition</subject><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2018</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNpVz7tuwjAYBWAvDBXtA3TCL5DUlzi2u1URvUi0MGREiv44P2AJHGRbQN--gnbpdJZzjvQR8shZWRml2BPEiz-V3DBRMl0LeUfW88txDwGyH0OiPSQc6Bho3iH99Cn5sH2m7XiGOCTajCFHSNmfkP6bnX3e0RXG7AOGTL9wC9dSuieTDewTPvzllLSv87Z5LxbLt4_mZVFArWVh0ArDe16jNkwP0uh-YI6BkEKoDYJ2vVWssspJAYwbMKq2Eqx1zgnrKjkls9_bG687Rn-A-N1dmd2NKX8A1YxOQw</recordid><startdate>20180221</startdate><enddate>20180221</enddate><creator>Dhurandhar, Amit</creator><creator>Chen, Pin-Yu</creator><creator>Luss, Ronny</creator><creator>Tu, Chun-Chen</creator><creator>Ting, Paishun</creator><creator>Shanmugam, Karthikeyan</creator><creator>Das, Payel</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20180221</creationdate><title>Explanations based on the Missing: Towards Contrastive Explanations with Pertinent Negatives</title><author>Dhurandhar, Amit ; Chen, Pin-Yu ; Luss, Ronny ; Tu, Chun-Chen ; Ting, Paishun ; Shanmugam, Karthikeyan ; Das, Payel</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a673-8e9281b16e7807d387bd0c0a23225fea7cb950495c32a018a85693a99ccc29c43</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2018</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Computer Vision and Pattern Recognition</topic><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Dhurandhar, Amit</creatorcontrib><creatorcontrib>Chen, Pin-Yu</creatorcontrib><creatorcontrib>Luss, Ronny</creatorcontrib><creatorcontrib>Tu, Chun-Chen</creatorcontrib><creatorcontrib>Ting, Paishun</creatorcontrib><creatorcontrib>Shanmugam, Karthikeyan</creatorcontrib><creatorcontrib>Das, Payel</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Dhurandhar, Amit</au><au>Chen, Pin-Yu</au><au>Luss, Ronny</au><au>Tu, Chun-Chen</au><au>Ting, Paishun</au><au>Shanmugam, Karthikeyan</au><au>Das, Payel</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Explanations based on the Missing: Towards Contrastive Explanations with Pertinent Negatives</atitle><date>2018-02-21</date><risdate>2018</risdate><abstract>In this paper we propose a novel method that provides contrastive explanations justifying the classification of an input by a black box classifier such as a deep neural network. Given an input we find what should be %necessarily and minimally and sufficiently present (viz. important object pixels in an image) to justify its classification and analogously what should be minimally and necessarily \emph{absent} (viz. certain background pixels). We argue that such explanations are natural for humans and are used commonly in domains such as health care and criminology. What is minimally but critically \emph{absent} is an important part of an explanation, which to the best of our knowledge, has not been explicitly identified by current explanation methods that explain predictions of neural networks. We validate our approach on three real datasets obtained from diverse domains; namely, a handwritten digits dataset MNIST, a large procurement fraud dataset and a brain activity strength dataset. In all three cases, we witness the power of our approach in generating precise explanations that are also easy for human experts to understand and evaluate.</abstract><doi>10.48550/arxiv.1802.07623</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.1802.07623
ispartof
issn
language eng
recordid cdi_arxiv_primary_1802_07623
source arXiv.org
subjects Computer Science - Artificial Intelligence
Computer Science - Computer Vision and Pattern Recognition
Computer Science - Learning
title Explanations based on the Missing: Towards Contrastive Explanations with Pertinent Negatives
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-04T18%3A25%3A29IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Explanations%20based%20on%20the%20Missing:%20Towards%20Contrastive%20Explanations%20with%20Pertinent%20Negatives&rft.au=Dhurandhar,%20Amit&rft.date=2018-02-21&rft_id=info:doi/10.48550/arxiv.1802.07623&rft_dat=%3Carxiv_GOX%3E1802_07623%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true