Interpreting neural-network results: a simulation study

Artificial neural networks (ANN) seem very promising for regression and classification, especially for large covariate spaces. Yet, their usefulness for medical and social research is limited because they present only prediction results and do not present features of the underlying process relating...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Computational statistics & data analysis 2001-09, Vol.37 (3), p.373-393
Hauptverfasser: Intrator, Orna, Intrator, Nathan
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 393
container_issue 3
container_start_page 373
container_title Computational statistics & data analysis
container_volume 37
creator Intrator, Orna
Intrator, Nathan
description Artificial neural networks (ANN) seem very promising for regression and classification, especially for large covariate spaces. Yet, their usefulness for medical and social research is limited because they present only prediction results and do not present features of the underlying process relating the inputs to the output. ANNs approximate a non-linear function by a composition of low-dimensional ridge functions, and therefore appear to be less sensitive to the dimensionality of the covariate space. However, due to non-uniqueness of a global minimum and the existence of (possibly) many local minima, the model revealed by the network is non-stable. We introduce a method that demonstrates the effects of inputs on output of ANNs by using novel robustification techniques. Simulated data from known models are used to demonstrate the interpretability results of the ANNs. Graphical tools are used for studying the interpretation results, and for detecting interactions between covariates. The effects of different regularization methods on the robustness of the interpretation are discussed; in particular we note that ANNs must include skip layer connections. An application to an ANN model predicting 5-yr mortality following breast cancer diagnosis is presented. We conclude that neural networks estimated with sufficient regularization can be reliably interpreted using the method presented in this paper.
doi_str_mv 10.1016/S0167-9473(01)00016-0
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_miscellaneous_21477598</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0167947301000160</els_id><sourcerecordid>582235</sourcerecordid><originalsourceid>FETCH-LOGICAL-c482t-78dad95bd2b89bd179ca1f7a356e8883fd22b6f5d650f1443603740468b5c5a03</originalsourceid><addsrcrecordid>eNqFkMtO3DAUhi1UpA7QR6iUDQgWAV9jhw1CiKuQumi7thz7hLpknGA7VPP2eGYQLLs4F9vfOf71I_Sd4FOCSXP2syRZt1yyY0xOMC7HGu-gBVGS1pIJ-gUtPpCvaC-lvwWiXKoFkvchQ5wiZB-eqgBzNEMdIP8b43MVIc1DTueVqZJfzoPJfgxVyrNbHaDd3gwJvr3XffT75vrX1V39-OP2_urysbZc0VxL5YxrRedop9rOEdlaQ3ppmGhAKcV6R2nX9MI1AveEc9ZgJjnmjeqEFQazfXS03TvF8WWGlPXSJwvDYAKMc9KUcClFqwootqCNY0oRej1FvzRxpQnWa5v0xia99kBjojc26fUHD9u5CBPYjyEAsMmZYPSrZobJklYlaJkqxa_vSkybt9K1TP_Jy7Ls8F2tSdYMfTTB-vSphJOmkYQW7mLLQbHu1UPUyXoIFpyPYLN2o_-P7DfBqpSR</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>21477598</pqid></control><display><type>article</type><title>Interpreting neural-network results: a simulation study</title><source>RePEc</source><source>ScienceDirect Journals (5 years ago - present)</source><creator>Intrator, Orna ; Intrator, Nathan</creator><creatorcontrib>Intrator, Orna ; Intrator, Nathan</creatorcontrib><description>Artificial neural networks (ANN) seem very promising for regression and classification, especially for large covariate spaces. Yet, their usefulness for medical and social research is limited because they present only prediction results and do not present features of the underlying process relating the inputs to the output. ANNs approximate a non-linear function by a composition of low-dimensional ridge functions, and therefore appear to be less sensitive to the dimensionality of the covariate space. However, due to non-uniqueness of a global minimum and the existence of (possibly) many local minima, the model revealed by the network is non-stable. We introduce a method that demonstrates the effects of inputs on output of ANNs by using novel robustification techniques. Simulated data from known models are used to demonstrate the interpretability results of the ANNs. Graphical tools are used for studying the interpretation results, and for detecting interactions between covariates. The effects of different regularization methods on the robustness of the interpretation are discussed; in particular we note that ANNs must include skip layer connections. An application to an ANN model predicting 5-yr mortality following breast cancer diagnosis is presented. We conclude that neural networks estimated with sufficient regularization can be reliably interpreted using the method presented in this paper.</description><identifier>ISSN: 0167-9473</identifier><identifier>EISSN: 1872-7352</identifier><identifier>DOI: 10.1016/S0167-9473(01)00016-0</identifier><language>eng</language><publisher>Amsterdam: Elsevier B.V</publisher><subject>Applications ; Data mining tools ; Diagnosis ; Exact sciences and technology ; Inference from stochastic processes; time series analysis ; Interaction effects ; Linear inference, regression ; Logistic regression ; Mathematics ; Multivariate analysis ; Nonlinear models ; Probability and statistics ; Regression analysis ; Reliability, life testing, quality control ; Sciences and techniques of general use ; Splitlevel plots ; Statistics ; Tumors</subject><ispartof>Computational statistics &amp; data analysis, 2001-09, Vol.37 (3), p.373-393</ispartof><rights>2001 Elsevier Science B.V.</rights><rights>2002 INIST-CNRS</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c482t-78dad95bd2b89bd179ca1f7a356e8883fd22b6f5d650f1443603740468b5c5a03</citedby><cites>FETCH-LOGICAL-c482t-78dad95bd2b89bd179ca1f7a356e8883fd22b6f5d650f1443603740468b5c5a03</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://dx.doi.org/10.1016/S0167-9473(01)00016-0$$EHTML$$P50$$Gelsevier$$H</linktohtml><link.rule.ids>314,780,784,3550,4008,27924,27925,45995</link.rule.ids><backlink>$$Uhttp://pascal-francis.inist.fr/vibad/index.php?action=getRecordDetail&amp;idt=14166712$$DView record in Pascal Francis$$Hfree_for_read</backlink><backlink>$$Uhttp://econpapers.repec.org/article/eeecsdana/v_3a37_3ay_3a2001_3ai_3a3_3ap_3a373-393.htm$$DView record in RePEc$$Hfree_for_read</backlink></links><search><creatorcontrib>Intrator, Orna</creatorcontrib><creatorcontrib>Intrator, Nathan</creatorcontrib><title>Interpreting neural-network results: a simulation study</title><title>Computational statistics &amp; data analysis</title><description>Artificial neural networks (ANN) seem very promising for regression and classification, especially for large covariate spaces. Yet, their usefulness for medical and social research is limited because they present only prediction results and do not present features of the underlying process relating the inputs to the output. ANNs approximate a non-linear function by a composition of low-dimensional ridge functions, and therefore appear to be less sensitive to the dimensionality of the covariate space. However, due to non-uniqueness of a global minimum and the existence of (possibly) many local minima, the model revealed by the network is non-stable. We introduce a method that demonstrates the effects of inputs on output of ANNs by using novel robustification techniques. Simulated data from known models are used to demonstrate the interpretability results of the ANNs. Graphical tools are used for studying the interpretation results, and for detecting interactions between covariates. The effects of different regularization methods on the robustness of the interpretation are discussed; in particular we note that ANNs must include skip layer connections. An application to an ANN model predicting 5-yr mortality following breast cancer diagnosis is presented. We conclude that neural networks estimated with sufficient regularization can be reliably interpreted using the method presented in this paper.</description><subject>Applications</subject><subject>Data mining tools</subject><subject>Diagnosis</subject><subject>Exact sciences and technology</subject><subject>Inference from stochastic processes; time series analysis</subject><subject>Interaction effects</subject><subject>Linear inference, regression</subject><subject>Logistic regression</subject><subject>Mathematics</subject><subject>Multivariate analysis</subject><subject>Nonlinear models</subject><subject>Probability and statistics</subject><subject>Regression analysis</subject><subject>Reliability, life testing, quality control</subject><subject>Sciences and techniques of general use</subject><subject>Splitlevel plots</subject><subject>Statistics</subject><subject>Tumors</subject><issn>0167-9473</issn><issn>1872-7352</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2001</creationdate><recordtype>article</recordtype><sourceid>X2L</sourceid><recordid>eNqFkMtO3DAUhi1UpA7QR6iUDQgWAV9jhw1CiKuQumi7thz7hLpknGA7VPP2eGYQLLs4F9vfOf71I_Sd4FOCSXP2syRZt1yyY0xOMC7HGu-gBVGS1pIJ-gUtPpCvaC-lvwWiXKoFkvchQ5wiZB-eqgBzNEMdIP8b43MVIc1DTueVqZJfzoPJfgxVyrNbHaDd3gwJvr3XffT75vrX1V39-OP2_urysbZc0VxL5YxrRedop9rOEdlaQ3ppmGhAKcV6R2nX9MI1AveEc9ZgJjnmjeqEFQazfXS03TvF8WWGlPXSJwvDYAKMc9KUcClFqwootqCNY0oRej1FvzRxpQnWa5v0xia99kBjojc26fUHD9u5CBPYjyEAsMmZYPSrZobJklYlaJkqxa_vSkybt9K1TP_Jy7Ls8F2tSdYMfTTB-vSphJOmkYQW7mLLQbHu1UPUyXoIFpyPYLN2o_-P7DfBqpSR</recordid><startdate>20010928</startdate><enddate>20010928</enddate><creator>Intrator, Orna</creator><creator>Intrator, Nathan</creator><general>Elsevier B.V</general><general>Elsevier Science</general><general>Elsevier</general><scope>IQODW</scope><scope>DKI</scope><scope>X2L</scope><scope>AAYXX</scope><scope>CITATION</scope></search><sort><creationdate>20010928</creationdate><title>Interpreting neural-network results: a simulation study</title><author>Intrator, Orna ; Intrator, Nathan</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c482t-78dad95bd2b89bd179ca1f7a356e8883fd22b6f5d650f1443603740468b5c5a03</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2001</creationdate><topic>Applications</topic><topic>Data mining tools</topic><topic>Diagnosis</topic><topic>Exact sciences and technology</topic><topic>Inference from stochastic processes; time series analysis</topic><topic>Interaction effects</topic><topic>Linear inference, regression</topic><topic>Logistic regression</topic><topic>Mathematics</topic><topic>Multivariate analysis</topic><topic>Nonlinear models</topic><topic>Probability and statistics</topic><topic>Regression analysis</topic><topic>Reliability, life testing, quality control</topic><topic>Sciences and techniques of general use</topic><topic>Splitlevel plots</topic><topic>Statistics</topic><topic>Tumors</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Intrator, Orna</creatorcontrib><creatorcontrib>Intrator, Nathan</creatorcontrib><collection>Pascal-Francis</collection><collection>RePEc IDEAS</collection><collection>RePEc</collection><collection>CrossRef</collection><jtitle>Computational statistics &amp; data analysis</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Intrator, Orna</au><au>Intrator, Nathan</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Interpreting neural-network results: a simulation study</atitle><jtitle>Computational statistics &amp; data analysis</jtitle><date>2001-09-28</date><risdate>2001</risdate><volume>37</volume><issue>3</issue><spage>373</spage><epage>393</epage><pages>373-393</pages><issn>0167-9473</issn><eissn>1872-7352</eissn><abstract>Artificial neural networks (ANN) seem very promising for regression and classification, especially for large covariate spaces. Yet, their usefulness for medical and social research is limited because they present only prediction results and do not present features of the underlying process relating the inputs to the output. ANNs approximate a non-linear function by a composition of low-dimensional ridge functions, and therefore appear to be less sensitive to the dimensionality of the covariate space. However, due to non-uniqueness of a global minimum and the existence of (possibly) many local minima, the model revealed by the network is non-stable. We introduce a method that demonstrates the effects of inputs on output of ANNs by using novel robustification techniques. Simulated data from known models are used to demonstrate the interpretability results of the ANNs. Graphical tools are used for studying the interpretation results, and for detecting interactions between covariates. The effects of different regularization methods on the robustness of the interpretation are discussed; in particular we note that ANNs must include skip layer connections. An application to an ANN model predicting 5-yr mortality following breast cancer diagnosis is presented. We conclude that neural networks estimated with sufficient regularization can be reliably interpreted using the method presented in this paper.</abstract><cop>Amsterdam</cop><pub>Elsevier B.V</pub><doi>10.1016/S0167-9473(01)00016-0</doi><tpages>21</tpages><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 0167-9473
ispartof Computational statistics & data analysis, 2001-09, Vol.37 (3), p.373-393
issn 0167-9473
1872-7352
language eng
recordid cdi_proquest_miscellaneous_21477598
source RePEc; ScienceDirect Journals (5 years ago - present)
subjects Applications
Data mining tools
Diagnosis
Exact sciences and technology
Inference from stochastic processes
time series analysis
Interaction effects
Linear inference, regression
Logistic regression
Mathematics
Multivariate analysis
Nonlinear models
Probability and statistics
Regression analysis
Reliability, life testing, quality control
Sciences and techniques of general use
Splitlevel plots
Statistics
Tumors
title Interpreting neural-network results: a simulation study
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-06T22%3A46%3A38IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Interpreting%20neural-network%20results:%20a%20simulation%20study&rft.jtitle=Computational%20statistics%20&%20data%20analysis&rft.au=Intrator,%20Orna&rft.date=2001-09-28&rft.volume=37&rft.issue=3&rft.spage=373&rft.epage=393&rft.pages=373-393&rft.issn=0167-9473&rft.eissn=1872-7352&rft_id=info:doi/10.1016/S0167-9473(01)00016-0&rft_dat=%3Cproquest_cross%3E582235%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=21477598&rft_id=info:pmid/&rft_els_id=S0167947301000160&rfr_iscdi=true