Neural Reasoning Networks: Efficient Interpretable Neural Networks With Automatic Textual Explanations

Recent advances in machine learning have led to a surge in adoption of neural networks for various tasks, but lack of interpretability remains an issue for many others in which an understanding of the features influencing the prediction is necessary to ensure fairness, safety, and legal compliance....

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Carrow, Stephen, Erwin, Kyle Harper, Vilenskaia, Olga, Ram, Parikshit, Klinger, Tim, Khan, Naweed Aghmad, Makondo, Ndivhuwo, Gray, Alexander
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Carrow, Stephen
Erwin, Kyle Harper
Vilenskaia, Olga
Ram, Parikshit
Klinger, Tim
Khan, Naweed Aghmad
Makondo, Ndivhuwo
Gray, Alexander
description Recent advances in machine learning have led to a surge in adoption of neural networks for various tasks, but lack of interpretability remains an issue for many others in which an understanding of the features influencing the prediction is necessary to ensure fairness, safety, and legal compliance. In this paper we consider one class of such tasks, tabular dataset classification, and propose a novel neuro-symbolic architecture, Neural Reasoning Networks (NRN), that is scalable and generates logically sound textual explanations for its predictions. NRNs are connected layers of logical neurons which implement a form of real valued logic. A training algorithm (R-NRN) learns the weights of the network as usual using gradient descent optimization with backprop, but also learns the network structure itself using a bandit-based optimization. Both are implemented in an extension to PyTorch (https://github.com/IBM/torchlogic) that takes full advantage of GPU scaling and batched training. Evaluation on a diverse set of 22 open-source datasets for tabular classification demonstrates performance (measured by ROC AUC) which improves over multi-layer perceptron (MLP) and is statistically similar to other state-of-the-art approaches such as Random Forest, XGBoost and Gradient Boosted Trees, while offering 43% faster training and a more than 2 orders of magnitude reduction in the number of parameters required, on average. Furthermore, R-NRN explanations are shorter than the compared approaches while producing more accurate feature importance scores.
doi_str_mv 10.48550/arxiv.2410.07966
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2410_07966</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2410_07966</sourcerecordid><originalsourceid>FETCH-arxiv_primary_2410_079663</originalsourceid><addsrcrecordid>eNpjYJA0NNAzsTA1NdBPLKrILNMzMgEKGJhbmplxMqT5pZYWJeYoBKUmFufnZealK_illpTnF2UXWym4pqVlJmem5pUoeOaVpBYVFKWWJCblpCpAtcAUKoRnlmQoOJaW5OcmlmQmK4SkVpSUAuVdKwpyEvOAQvl5xTwMrGmJOcWpvFCam0HezTXE2UMX7KL4gqLM3MSiyniQy-LBLjMmrAIAkBxFqQ</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Neural Reasoning Networks: Efficient Interpretable Neural Networks With Automatic Textual Explanations</title><source>arXiv.org</source><creator>Carrow, Stephen ; Erwin, Kyle Harper ; Vilenskaia, Olga ; Ram, Parikshit ; Klinger, Tim ; Khan, Naweed Aghmad ; Makondo, Ndivhuwo ; Gray, Alexander</creator><creatorcontrib>Carrow, Stephen ; Erwin, Kyle Harper ; Vilenskaia, Olga ; Ram, Parikshit ; Klinger, Tim ; Khan, Naweed Aghmad ; Makondo, Ndivhuwo ; Gray, Alexander</creatorcontrib><description>Recent advances in machine learning have led to a surge in adoption of neural networks for various tasks, but lack of interpretability remains an issue for many others in which an understanding of the features influencing the prediction is necessary to ensure fairness, safety, and legal compliance. In this paper we consider one class of such tasks, tabular dataset classification, and propose a novel neuro-symbolic architecture, Neural Reasoning Networks (NRN), that is scalable and generates logically sound textual explanations for its predictions. NRNs are connected layers of logical neurons which implement a form of real valued logic. A training algorithm (R-NRN) learns the weights of the network as usual using gradient descent optimization with backprop, but also learns the network structure itself using a bandit-based optimization. Both are implemented in an extension to PyTorch (https://github.com/IBM/torchlogic) that takes full advantage of GPU scaling and batched training. Evaluation on a diverse set of 22 open-source datasets for tabular classification demonstrates performance (measured by ROC AUC) which improves over multi-layer perceptron (MLP) and is statistically similar to other state-of-the-art approaches such as Random Forest, XGBoost and Gradient Boosted Trees, while offering 43% faster training and a more than 2 orders of magnitude reduction in the number of parameters required, on average. Furthermore, R-NRN explanations are shorter than the compared approaches while producing more accurate feature importance scores.</description><identifier>DOI: 10.48550/arxiv.2410.07966</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Learning</subject><creationdate>2024-10</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,781,886</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2410.07966$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2410.07966$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Carrow, Stephen</creatorcontrib><creatorcontrib>Erwin, Kyle Harper</creatorcontrib><creatorcontrib>Vilenskaia, Olga</creatorcontrib><creatorcontrib>Ram, Parikshit</creatorcontrib><creatorcontrib>Klinger, Tim</creatorcontrib><creatorcontrib>Khan, Naweed Aghmad</creatorcontrib><creatorcontrib>Makondo, Ndivhuwo</creatorcontrib><creatorcontrib>Gray, Alexander</creatorcontrib><title>Neural Reasoning Networks: Efficient Interpretable Neural Networks With Automatic Textual Explanations</title><description>Recent advances in machine learning have led to a surge in adoption of neural networks for various tasks, but lack of interpretability remains an issue for many others in which an understanding of the features influencing the prediction is necessary to ensure fairness, safety, and legal compliance. In this paper we consider one class of such tasks, tabular dataset classification, and propose a novel neuro-symbolic architecture, Neural Reasoning Networks (NRN), that is scalable and generates logically sound textual explanations for its predictions. NRNs are connected layers of logical neurons which implement a form of real valued logic. A training algorithm (R-NRN) learns the weights of the network as usual using gradient descent optimization with backprop, but also learns the network structure itself using a bandit-based optimization. Both are implemented in an extension to PyTorch (https://github.com/IBM/torchlogic) that takes full advantage of GPU scaling and batched training. Evaluation on a diverse set of 22 open-source datasets for tabular classification demonstrates performance (measured by ROC AUC) which improves over multi-layer perceptron (MLP) and is statistically similar to other state-of-the-art approaches such as Random Forest, XGBoost and Gradient Boosted Trees, while offering 43% faster training and a more than 2 orders of magnitude reduction in the number of parameters required, on average. Furthermore, R-NRN explanations are shorter than the compared approaches while producing more accurate feature importance scores.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNpjYJA0NNAzsTA1NdBPLKrILNMzMgEKGJhbmplxMqT5pZYWJeYoBKUmFufnZealK_illpTnF2UXWym4pqVlJmem5pUoeOaVpBYVFKWWJCblpCpAtcAUKoRnlmQoOJaW5OcmlmQmK4SkVpSUAuVdKwpyEvOAQvl5xTwMrGmJOcWpvFCam0HezTXE2UMX7KL4gqLM3MSiyniQy-LBLjMmrAIAkBxFqQ</recordid><startdate>20241010</startdate><enddate>20241010</enddate><creator>Carrow, Stephen</creator><creator>Erwin, Kyle Harper</creator><creator>Vilenskaia, Olga</creator><creator>Ram, Parikshit</creator><creator>Klinger, Tim</creator><creator>Khan, Naweed Aghmad</creator><creator>Makondo, Ndivhuwo</creator><creator>Gray, Alexander</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20241010</creationdate><title>Neural Reasoning Networks: Efficient Interpretable Neural Networks With Automatic Textual Explanations</title><author>Carrow, Stephen ; Erwin, Kyle Harper ; Vilenskaia, Olga ; Ram, Parikshit ; Klinger, Tim ; Khan, Naweed Aghmad ; Makondo, Ndivhuwo ; Gray, Alexander</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-arxiv_primary_2410_079663</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Carrow, Stephen</creatorcontrib><creatorcontrib>Erwin, Kyle Harper</creatorcontrib><creatorcontrib>Vilenskaia, Olga</creatorcontrib><creatorcontrib>Ram, Parikshit</creatorcontrib><creatorcontrib>Klinger, Tim</creatorcontrib><creatorcontrib>Khan, Naweed Aghmad</creatorcontrib><creatorcontrib>Makondo, Ndivhuwo</creatorcontrib><creatorcontrib>Gray, Alexander</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Carrow, Stephen</au><au>Erwin, Kyle Harper</au><au>Vilenskaia, Olga</au><au>Ram, Parikshit</au><au>Klinger, Tim</au><au>Khan, Naweed Aghmad</au><au>Makondo, Ndivhuwo</au><au>Gray, Alexander</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Neural Reasoning Networks: Efficient Interpretable Neural Networks With Automatic Textual Explanations</atitle><date>2024-10-10</date><risdate>2024</risdate><abstract>Recent advances in machine learning have led to a surge in adoption of neural networks for various tasks, but lack of interpretability remains an issue for many others in which an understanding of the features influencing the prediction is necessary to ensure fairness, safety, and legal compliance. In this paper we consider one class of such tasks, tabular dataset classification, and propose a novel neuro-symbolic architecture, Neural Reasoning Networks (NRN), that is scalable and generates logically sound textual explanations for its predictions. NRNs are connected layers of logical neurons which implement a form of real valued logic. A training algorithm (R-NRN) learns the weights of the network as usual using gradient descent optimization with backprop, but also learns the network structure itself using a bandit-based optimization. Both are implemented in an extension to PyTorch (https://github.com/IBM/torchlogic) that takes full advantage of GPU scaling and batched training. Evaluation on a diverse set of 22 open-source datasets for tabular classification demonstrates performance (measured by ROC AUC) which improves over multi-layer perceptron (MLP) and is statistically similar to other state-of-the-art approaches such as Random Forest, XGBoost and Gradient Boosted Trees, while offering 43% faster training and a more than 2 orders of magnitude reduction in the number of parameters required, on average. Furthermore, R-NRN explanations are shorter than the compared approaches while producing more accurate feature importance scores.</abstract><doi>10.48550/arxiv.2410.07966</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2410.07966
ispartof
issn
language eng
recordid cdi_arxiv_primary_2410_07966
source arXiv.org
subjects Computer Science - Artificial Intelligence
Computer Science - Learning
title Neural Reasoning Networks: Efficient Interpretable Neural Networks With Automatic Textual Explanations
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-12T06%3A04%3A43IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Neural%20Reasoning%20Networks:%20Efficient%20Interpretable%20Neural%20Networks%20With%20Automatic%20Textual%20Explanations&rft.au=Carrow,%20Stephen&rft.date=2024-10-10&rft_id=info:doi/10.48550/arxiv.2410.07966&rft_dat=%3Carxiv_GOX%3E2410_07966%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true