Interpreting neural-network results: a simulation study
Artificial neural networks (ANN) seem very promising for regression and classification, especially for large covariate spaces. Yet, their usefulness for medical and social research is limited because they present only prediction results and do not present features of the underlying process relating...
Gespeichert in:
Veröffentlicht in: | Computational statistics & data analysis 2001-09, Vol.37 (3), p.373-393 |
---|---|
Hauptverfasser: | , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 393 |
---|---|
container_issue | 3 |
container_start_page | 373 |
container_title | Computational statistics & data analysis |
container_volume | 37 |
creator | Intrator, Orna Intrator, Nathan |
description | Artificial neural networks (ANN) seem very promising for regression and classification, especially for large covariate spaces. Yet, their usefulness for medical and social research is limited because they present only prediction results and do not present features of the underlying process relating the inputs to the output. ANNs approximate a non-linear function by a composition of low-dimensional ridge functions, and therefore appear to be less sensitive to the dimensionality of the covariate space. However, due to non-uniqueness of a global minimum and the existence of (possibly) many local minima, the model revealed by the network is non-stable. We introduce a method that demonstrates the effects of inputs on output of ANNs by using novel robustification techniques. Simulated data from known models are used to demonstrate the interpretability results of the ANNs. Graphical tools are used for studying the interpretation results, and for detecting interactions between covariates. The effects of different regularization methods on the robustness of the interpretation are discussed; in particular we note that ANNs must include skip layer connections. An application to an ANN model predicting 5-yr mortality following breast cancer diagnosis is presented. We conclude that neural networks estimated with sufficient regularization can be reliably interpreted using the method presented in this paper. |
doi_str_mv | 10.1016/S0167-9473(01)00016-0 |
format | Article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_miscellaneous_21477598</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0167947301000160</els_id><sourcerecordid>582235</sourcerecordid><originalsourceid>FETCH-LOGICAL-c482t-78dad95bd2b89bd179ca1f7a356e8883fd22b6f5d650f1443603740468b5c5a03</originalsourceid><addsrcrecordid>eNqFkMtO3DAUhi1UpA7QR6iUDQgWAV9jhw1CiKuQumi7thz7hLpknGA7VPP2eGYQLLs4F9vfOf71I_Sd4FOCSXP2syRZt1yyY0xOMC7HGu-gBVGS1pIJ-gUtPpCvaC-lvwWiXKoFkvchQ5wiZB-eqgBzNEMdIP8b43MVIc1DTueVqZJfzoPJfgxVyrNbHaDd3gwJvr3XffT75vrX1V39-OP2_urysbZc0VxL5YxrRedop9rOEdlaQ3ppmGhAKcV6R2nX9MI1AveEc9ZgJjnmjeqEFQazfXS03TvF8WWGlPXSJwvDYAKMc9KUcClFqwootqCNY0oRej1FvzRxpQnWa5v0xia99kBjojc26fUHD9u5CBPYjyEAsMmZYPSrZobJklYlaJkqxa_vSkybt9K1TP_Jy7Ls8F2tSdYMfTTB-vSphJOmkYQW7mLLQbHu1UPUyXoIFpyPYLN2o_-P7DfBqpSR</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>21477598</pqid></control><display><type>article</type><title>Interpreting neural-network results: a simulation study</title><source>RePEc</source><source>ScienceDirect Journals (5 years ago - present)</source><creator>Intrator, Orna ; Intrator, Nathan</creator><creatorcontrib>Intrator, Orna ; Intrator, Nathan</creatorcontrib><description>Artificial neural networks (ANN) seem very promising for regression and classification, especially for large covariate spaces. Yet, their usefulness for medical and social research is limited because they present only prediction results and do not present features of the underlying process relating the inputs to the output. ANNs approximate a non-linear function by a composition of low-dimensional ridge functions, and therefore appear to be less sensitive to the dimensionality of the covariate space. However, due to non-uniqueness of a global minimum and the existence of (possibly) many local minima, the model revealed by the network is non-stable. We introduce a method that demonstrates the effects of inputs on output of ANNs by using novel robustification techniques. Simulated data from known models are used to demonstrate the interpretability results of the ANNs. Graphical tools are used for studying the interpretation results, and for detecting interactions between covariates. The effects of different regularization methods on the robustness of the interpretation are discussed; in particular we note that ANNs must include skip layer connections. An application to an ANN model predicting 5-yr mortality following breast cancer diagnosis is presented. We conclude that neural networks estimated with sufficient regularization can be reliably interpreted using the method presented in this paper.</description><identifier>ISSN: 0167-9473</identifier><identifier>EISSN: 1872-7352</identifier><identifier>DOI: 10.1016/S0167-9473(01)00016-0</identifier><language>eng</language><publisher>Amsterdam: Elsevier B.V</publisher><subject>Applications ; Data mining tools ; Diagnosis ; Exact sciences and technology ; Inference from stochastic processes; time series analysis ; Interaction effects ; Linear inference, regression ; Logistic regression ; Mathematics ; Multivariate analysis ; Nonlinear models ; Probability and statistics ; Regression analysis ; Reliability, life testing, quality control ; Sciences and techniques of general use ; Splitlevel plots ; Statistics ; Tumors</subject><ispartof>Computational statistics & data analysis, 2001-09, Vol.37 (3), p.373-393</ispartof><rights>2001 Elsevier Science B.V.</rights><rights>2002 INIST-CNRS</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c482t-78dad95bd2b89bd179ca1f7a356e8883fd22b6f5d650f1443603740468b5c5a03</citedby><cites>FETCH-LOGICAL-c482t-78dad95bd2b89bd179ca1f7a356e8883fd22b6f5d650f1443603740468b5c5a03</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://dx.doi.org/10.1016/S0167-9473(01)00016-0$$EHTML$$P50$$Gelsevier$$H</linktohtml><link.rule.ids>314,780,784,3550,4008,27924,27925,45995</link.rule.ids><backlink>$$Uhttp://pascal-francis.inist.fr/vibad/index.php?action=getRecordDetail&idt=14166712$$DView record in Pascal Francis$$Hfree_for_read</backlink><backlink>$$Uhttp://econpapers.repec.org/article/eeecsdana/v_3a37_3ay_3a2001_3ai_3a3_3ap_3a373-393.htm$$DView record in RePEc$$Hfree_for_read</backlink></links><search><creatorcontrib>Intrator, Orna</creatorcontrib><creatorcontrib>Intrator, Nathan</creatorcontrib><title>Interpreting neural-network results: a simulation study</title><title>Computational statistics & data analysis</title><description>Artificial neural networks (ANN) seem very promising for regression and classification, especially for large covariate spaces. Yet, their usefulness for medical and social research is limited because they present only prediction results and do not present features of the underlying process relating the inputs to the output. ANNs approximate a non-linear function by a composition of low-dimensional ridge functions, and therefore appear to be less sensitive to the dimensionality of the covariate space. However, due to non-uniqueness of a global minimum and the existence of (possibly) many local minima, the model revealed by the network is non-stable. We introduce a method that demonstrates the effects of inputs on output of ANNs by using novel robustification techniques. Simulated data from known models are used to demonstrate the interpretability results of the ANNs. Graphical tools are used for studying the interpretation results, and for detecting interactions between covariates. The effects of different regularization methods on the robustness of the interpretation are discussed; in particular we note that ANNs must include skip layer connections. An application to an ANN model predicting 5-yr mortality following breast cancer diagnosis is presented. We conclude that neural networks estimated with sufficient regularization can be reliably interpreted using the method presented in this paper.</description><subject>Applications</subject><subject>Data mining tools</subject><subject>Diagnosis</subject><subject>Exact sciences and technology</subject><subject>Inference from stochastic processes; time series analysis</subject><subject>Interaction effects</subject><subject>Linear inference, regression</subject><subject>Logistic regression</subject><subject>Mathematics</subject><subject>Multivariate analysis</subject><subject>Nonlinear models</subject><subject>Probability and statistics</subject><subject>Regression analysis</subject><subject>Reliability, life testing, quality control</subject><subject>Sciences and techniques of general use</subject><subject>Splitlevel plots</subject><subject>Statistics</subject><subject>Tumors</subject><issn>0167-9473</issn><issn>1872-7352</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2001</creationdate><recordtype>article</recordtype><sourceid>X2L</sourceid><recordid>eNqFkMtO3DAUhi1UpA7QR6iUDQgWAV9jhw1CiKuQumi7thz7hLpknGA7VPP2eGYQLLs4F9vfOf71I_Sd4FOCSXP2syRZt1yyY0xOMC7HGu-gBVGS1pIJ-gUtPpCvaC-lvwWiXKoFkvchQ5wiZB-eqgBzNEMdIP8b43MVIc1DTueVqZJfzoPJfgxVyrNbHaDd3gwJvr3XffT75vrX1V39-OP2_urysbZc0VxL5YxrRedop9rOEdlaQ3ppmGhAKcV6R2nX9MI1AveEc9ZgJjnmjeqEFQazfXS03TvF8WWGlPXSJwvDYAKMc9KUcClFqwootqCNY0oRej1FvzRxpQnWa5v0xia99kBjojc26fUHD9u5CBPYjyEAsMmZYPSrZobJklYlaJkqxa_vSkybt9K1TP_Jy7Ls8F2tSdYMfTTB-vSphJOmkYQW7mLLQbHu1UPUyXoIFpyPYLN2o_-P7DfBqpSR</recordid><startdate>20010928</startdate><enddate>20010928</enddate><creator>Intrator, Orna</creator><creator>Intrator, Nathan</creator><general>Elsevier B.V</general><general>Elsevier Science</general><general>Elsevier</general><scope>IQODW</scope><scope>DKI</scope><scope>X2L</scope><scope>AAYXX</scope><scope>CITATION</scope></search><sort><creationdate>20010928</creationdate><title>Interpreting neural-network results: a simulation study</title><author>Intrator, Orna ; Intrator, Nathan</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c482t-78dad95bd2b89bd179ca1f7a356e8883fd22b6f5d650f1443603740468b5c5a03</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2001</creationdate><topic>Applications</topic><topic>Data mining tools</topic><topic>Diagnosis</topic><topic>Exact sciences and technology</topic><topic>Inference from stochastic processes; time series analysis</topic><topic>Interaction effects</topic><topic>Linear inference, regression</topic><topic>Logistic regression</topic><topic>Mathematics</topic><topic>Multivariate analysis</topic><topic>Nonlinear models</topic><topic>Probability and statistics</topic><topic>Regression analysis</topic><topic>Reliability, life testing, quality control</topic><topic>Sciences and techniques of general use</topic><topic>Splitlevel plots</topic><topic>Statistics</topic><topic>Tumors</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Intrator, Orna</creatorcontrib><creatorcontrib>Intrator, Nathan</creatorcontrib><collection>Pascal-Francis</collection><collection>RePEc IDEAS</collection><collection>RePEc</collection><collection>CrossRef</collection><jtitle>Computational statistics & data analysis</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Intrator, Orna</au><au>Intrator, Nathan</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Interpreting neural-network results: a simulation study</atitle><jtitle>Computational statistics & data analysis</jtitle><date>2001-09-28</date><risdate>2001</risdate><volume>37</volume><issue>3</issue><spage>373</spage><epage>393</epage><pages>373-393</pages><issn>0167-9473</issn><eissn>1872-7352</eissn><abstract>Artificial neural networks (ANN) seem very promising for regression and classification, especially for large covariate spaces. Yet, their usefulness for medical and social research is limited because they present only prediction results and do not present features of the underlying process relating the inputs to the output. ANNs approximate a non-linear function by a composition of low-dimensional ridge functions, and therefore appear to be less sensitive to the dimensionality of the covariate space. However, due to non-uniqueness of a global minimum and the existence of (possibly) many local minima, the model revealed by the network is non-stable. We introduce a method that demonstrates the effects of inputs on output of ANNs by using novel robustification techniques. Simulated data from known models are used to demonstrate the interpretability results of the ANNs. Graphical tools are used for studying the interpretation results, and for detecting interactions between covariates. The effects of different regularization methods on the robustness of the interpretation are discussed; in particular we note that ANNs must include skip layer connections. An application to an ANN model predicting 5-yr mortality following breast cancer diagnosis is presented. We conclude that neural networks estimated with sufficient regularization can be reliably interpreted using the method presented in this paper.</abstract><cop>Amsterdam</cop><pub>Elsevier B.V</pub><doi>10.1016/S0167-9473(01)00016-0</doi><tpages>21</tpages><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0167-9473 |
ispartof | Computational statistics & data analysis, 2001-09, Vol.37 (3), p.373-393 |
issn | 0167-9473 1872-7352 |
language | eng |
recordid | cdi_proquest_miscellaneous_21477598 |
source | RePEc; ScienceDirect Journals (5 years ago - present) |
subjects | Applications Data mining tools Diagnosis Exact sciences and technology Inference from stochastic processes time series analysis Interaction effects Linear inference, regression Logistic regression Mathematics Multivariate analysis Nonlinear models Probability and statistics Regression analysis Reliability, life testing, quality control Sciences and techniques of general use Splitlevel plots Statistics Tumors |
title | Interpreting neural-network results: a simulation study |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-06T22%3A46%3A38IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Interpreting%20neural-network%20results:%20a%20simulation%20study&rft.jtitle=Computational%20statistics%20&%20data%20analysis&rft.au=Intrator,%20Orna&rft.date=2001-09-28&rft.volume=37&rft.issue=3&rft.spage=373&rft.epage=393&rft.pages=373-393&rft.issn=0167-9473&rft.eissn=1872-7352&rft_id=info:doi/10.1016/S0167-9473(01)00016-0&rft_dat=%3Cproquest_cross%3E582235%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=21477598&rft_id=info:pmid/&rft_els_id=S0167947301000160&rfr_iscdi=true |