Neural Reasoning Networks: Efficient Interpretable Neural Networks With Automatic Textual Explanations
Recent advances in machine learning have led to a surge in adoption of neural networks for various tasks, but lack of interpretability remains an issue for many others in which an understanding of the features influencing the prediction is necessary to ensure fairness, safety, and legal compliance....
Gespeichert in:
Hauptverfasser: | , , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Carrow, Stephen Erwin, Kyle Harper Vilenskaia, Olga Ram, Parikshit Klinger, Tim Khan, Naweed Aghmad Makondo, Ndivhuwo Gray, Alexander |
description | Recent advances in machine learning have led to a surge in adoption of neural
networks for various tasks, but lack of interpretability remains an issue for
many others in which an understanding of the features influencing the
prediction is necessary to ensure fairness, safety, and legal compliance. In
this paper we consider one class of such tasks, tabular dataset classification,
and propose a novel neuro-symbolic architecture, Neural Reasoning Networks
(NRN), that is scalable and generates logically sound textual explanations for
its predictions. NRNs are connected layers of logical neurons which implement a
form of real valued logic. A training algorithm (R-NRN) learns the weights of
the network as usual using gradient descent optimization with backprop, but
also learns the network structure itself using a bandit-based optimization.
Both are implemented in an extension to PyTorch
(https://github.com/IBM/torchlogic) that takes full advantage of GPU scaling
and batched training. Evaluation on a diverse set of 22 open-source datasets
for tabular classification demonstrates performance (measured by ROC AUC) which
improves over multi-layer perceptron (MLP) and is statistically similar to
other state-of-the-art approaches such as Random Forest, XGBoost and Gradient
Boosted Trees, while offering 43% faster training and a more than 2 orders of
magnitude reduction in the number of parameters required, on average.
Furthermore, R-NRN explanations are shorter than the compared approaches while
producing more accurate feature importance scores. |
doi_str_mv | 10.48550/arxiv.2410.07966 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2410_07966</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2410_07966</sourcerecordid><originalsourceid>FETCH-arxiv_primary_2410_079663</originalsourceid><addsrcrecordid>eNpjYJA0NNAzsTA1NdBPLKrILNMzMgEKGJhbmplxMqT5pZYWJeYoBKUmFufnZealK_illpTnF2UXWym4pqVlJmem5pUoeOaVpBYVFKWWJCblpCpAtcAUKoRnlmQoOJaW5OcmlmQmK4SkVpSUAuVdKwpyEvOAQvl5xTwMrGmJOcWpvFCam0HezTXE2UMX7KL4gqLM3MSiyniQy-LBLjMmrAIAkBxFqQ</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Neural Reasoning Networks: Efficient Interpretable Neural Networks With Automatic Textual Explanations</title><source>arXiv.org</source><creator>Carrow, Stephen ; Erwin, Kyle Harper ; Vilenskaia, Olga ; Ram, Parikshit ; Klinger, Tim ; Khan, Naweed Aghmad ; Makondo, Ndivhuwo ; Gray, Alexander</creator><creatorcontrib>Carrow, Stephen ; Erwin, Kyle Harper ; Vilenskaia, Olga ; Ram, Parikshit ; Klinger, Tim ; Khan, Naweed Aghmad ; Makondo, Ndivhuwo ; Gray, Alexander</creatorcontrib><description>Recent advances in machine learning have led to a surge in adoption of neural
networks for various tasks, but lack of interpretability remains an issue for
many others in which an understanding of the features influencing the
prediction is necessary to ensure fairness, safety, and legal compliance. In
this paper we consider one class of such tasks, tabular dataset classification,
and propose a novel neuro-symbolic architecture, Neural Reasoning Networks
(NRN), that is scalable and generates logically sound textual explanations for
its predictions. NRNs are connected layers of logical neurons which implement a
form of real valued logic. A training algorithm (R-NRN) learns the weights of
the network as usual using gradient descent optimization with backprop, but
also learns the network structure itself using a bandit-based optimization.
Both are implemented in an extension to PyTorch
(https://github.com/IBM/torchlogic) that takes full advantage of GPU scaling
and batched training. Evaluation on a diverse set of 22 open-source datasets
for tabular classification demonstrates performance (measured by ROC AUC) which
improves over multi-layer perceptron (MLP) and is statistically similar to
other state-of-the-art approaches such as Random Forest, XGBoost and Gradient
Boosted Trees, while offering 43% faster training and a more than 2 orders of
magnitude reduction in the number of parameters required, on average.
Furthermore, R-NRN explanations are shorter than the compared approaches while
producing more accurate feature importance scores.</description><identifier>DOI: 10.48550/arxiv.2410.07966</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Learning</subject><creationdate>2024-10</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,781,886</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2410.07966$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2410.07966$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Carrow, Stephen</creatorcontrib><creatorcontrib>Erwin, Kyle Harper</creatorcontrib><creatorcontrib>Vilenskaia, Olga</creatorcontrib><creatorcontrib>Ram, Parikshit</creatorcontrib><creatorcontrib>Klinger, Tim</creatorcontrib><creatorcontrib>Khan, Naweed Aghmad</creatorcontrib><creatorcontrib>Makondo, Ndivhuwo</creatorcontrib><creatorcontrib>Gray, Alexander</creatorcontrib><title>Neural Reasoning Networks: Efficient Interpretable Neural Networks With Automatic Textual Explanations</title><description>Recent advances in machine learning have led to a surge in adoption of neural
networks for various tasks, but lack of interpretability remains an issue for
many others in which an understanding of the features influencing the
prediction is necessary to ensure fairness, safety, and legal compliance. In
this paper we consider one class of such tasks, tabular dataset classification,
and propose a novel neuro-symbolic architecture, Neural Reasoning Networks
(NRN), that is scalable and generates logically sound textual explanations for
its predictions. NRNs are connected layers of logical neurons which implement a
form of real valued logic. A training algorithm (R-NRN) learns the weights of
the network as usual using gradient descent optimization with backprop, but
also learns the network structure itself using a bandit-based optimization.
Both are implemented in an extension to PyTorch
(https://github.com/IBM/torchlogic) that takes full advantage of GPU scaling
and batched training. Evaluation on a diverse set of 22 open-source datasets
for tabular classification demonstrates performance (measured by ROC AUC) which
improves over multi-layer perceptron (MLP) and is statistically similar to
other state-of-the-art approaches such as Random Forest, XGBoost and Gradient
Boosted Trees, while offering 43% faster training and a more than 2 orders of
magnitude reduction in the number of parameters required, on average.
Furthermore, R-NRN explanations are shorter than the compared approaches while
producing more accurate feature importance scores.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNpjYJA0NNAzsTA1NdBPLKrILNMzMgEKGJhbmplxMqT5pZYWJeYoBKUmFufnZealK_illpTnF2UXWym4pqVlJmem5pUoeOaVpBYVFKWWJCblpCpAtcAUKoRnlmQoOJaW5OcmlmQmK4SkVpSUAuVdKwpyEvOAQvl5xTwMrGmJOcWpvFCam0HezTXE2UMX7KL4gqLM3MSiyniQy-LBLjMmrAIAkBxFqQ</recordid><startdate>20241010</startdate><enddate>20241010</enddate><creator>Carrow, Stephen</creator><creator>Erwin, Kyle Harper</creator><creator>Vilenskaia, Olga</creator><creator>Ram, Parikshit</creator><creator>Klinger, Tim</creator><creator>Khan, Naweed Aghmad</creator><creator>Makondo, Ndivhuwo</creator><creator>Gray, Alexander</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20241010</creationdate><title>Neural Reasoning Networks: Efficient Interpretable Neural Networks With Automatic Textual Explanations</title><author>Carrow, Stephen ; Erwin, Kyle Harper ; Vilenskaia, Olga ; Ram, Parikshit ; Klinger, Tim ; Khan, Naweed Aghmad ; Makondo, Ndivhuwo ; Gray, Alexander</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-arxiv_primary_2410_079663</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Carrow, Stephen</creatorcontrib><creatorcontrib>Erwin, Kyle Harper</creatorcontrib><creatorcontrib>Vilenskaia, Olga</creatorcontrib><creatorcontrib>Ram, Parikshit</creatorcontrib><creatorcontrib>Klinger, Tim</creatorcontrib><creatorcontrib>Khan, Naweed Aghmad</creatorcontrib><creatorcontrib>Makondo, Ndivhuwo</creatorcontrib><creatorcontrib>Gray, Alexander</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Carrow, Stephen</au><au>Erwin, Kyle Harper</au><au>Vilenskaia, Olga</au><au>Ram, Parikshit</au><au>Klinger, Tim</au><au>Khan, Naweed Aghmad</au><au>Makondo, Ndivhuwo</au><au>Gray, Alexander</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Neural Reasoning Networks: Efficient Interpretable Neural Networks With Automatic Textual Explanations</atitle><date>2024-10-10</date><risdate>2024</risdate><abstract>Recent advances in machine learning have led to a surge in adoption of neural
networks for various tasks, but lack of interpretability remains an issue for
many others in which an understanding of the features influencing the
prediction is necessary to ensure fairness, safety, and legal compliance. In
this paper we consider one class of such tasks, tabular dataset classification,
and propose a novel neuro-symbolic architecture, Neural Reasoning Networks
(NRN), that is scalable and generates logically sound textual explanations for
its predictions. NRNs are connected layers of logical neurons which implement a
form of real valued logic. A training algorithm (R-NRN) learns the weights of
the network as usual using gradient descent optimization with backprop, but
also learns the network structure itself using a bandit-based optimization.
Both are implemented in an extension to PyTorch
(https://github.com/IBM/torchlogic) that takes full advantage of GPU scaling
and batched training. Evaluation on a diverse set of 22 open-source datasets
for tabular classification demonstrates performance (measured by ROC AUC) which
improves over multi-layer perceptron (MLP) and is statistically similar to
other state-of-the-art approaches such as Random Forest, XGBoost and Gradient
Boosted Trees, while offering 43% faster training and a more than 2 orders of
magnitude reduction in the number of parameters required, on average.
Furthermore, R-NRN explanations are shorter than the compared approaches while
producing more accurate feature importance scores.</abstract><doi>10.48550/arxiv.2410.07966</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2410.07966 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2410_07966 |
source | arXiv.org |
subjects | Computer Science - Artificial Intelligence Computer Science - Learning |
title | Neural Reasoning Networks: Efficient Interpretable Neural Networks With Automatic Textual Explanations |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-12T06%3A04%3A43IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Neural%20Reasoning%20Networks:%20Efficient%20Interpretable%20Neural%20Networks%20With%20Automatic%20Textual%20Explanations&rft.au=Carrow,%20Stephen&rft.date=2024-10-10&rft_id=info:doi/10.48550/arxiv.2410.07966&rft_dat=%3Carxiv_GOX%3E2410_07966%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |