Training a neural network with conjugate gradient methods
This study investigates the use of several variants of conjugate gradient (CG) optimisation and line search methods to accelerate the convergence of an MLP neural network learning two medical signal classification problems. Much of the previous work has been done with artificial problems which have...
Gespeichert in:
Hauptverfasser: | , , |
---|---|
Format: | Tagungsbericht |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 378 vol.1 |
---|---|
container_issue | |
container_start_page | 373 |
container_title | |
container_volume | 1 |
creator | Towsey, M. Alpsan, D. Sztriha, L. |
description | This study investigates the use of several variants of conjugate gradient (CG) optimisation and line search methods to accelerate the convergence of an MLP neural network learning two medical signal classification problems. Much of the previous work has been done with artificial problems which have little relevance to real world problems and results on real world problems have been variable. The effectiveness of CG compared to standard backpropagation (BP) depended on the degree to which the learning task required finding a global minimum. If learning was stopped when the training set had been learned to an acceptable degree of error tolerance (the typical pattern classification problem), standard BP was faster than CG and did not display the convergence difficulties usually attributed to it. If learning required finding a global minimum (as in function minimisation or function estimation tasks), CG methods were faster but performance was very much dependent on careful selection of 'tuning' parameters and line search. This requirement for meta-optimisation was more difficult for CG than for BP because of the larger number of parameters. |
doi_str_mv | 10.1109/ICNN.1995.488128 |
format | Conference Proceeding |
fullrecord | <record><control><sourceid>ieee_6IE</sourceid><recordid>TN_cdi_ieee_primary_488128</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>488128</ieee_id><sourcerecordid>488128</sourcerecordid><originalsourceid>FETCH-LOGICAL-i104t-de1ef6b9161580a3703b5e27cf0d800e16aff4e6e313d794fae988c0b7e4d7e53</originalsourceid><addsrcrecordid>eNotj8tqwzAQRQWl0JJ6X7rSD9idsSRLWhbTRyCkm3QdZGvkKE3sIiuE_n0D6eHC2V04jD0iVIhgn5ftel2htaqSxmBtblhhtYHLRK0bY-9YMc97uCCVElrdM7tJLo5xHLjjI52SO1yUz1P65ueYd7yfxv1pcJn4kJyPNGZ-pLyb_PzAboM7zFT8e8G-3l437Ue5-nxfti-rMiLIXHpCCk1nsUFlwAkNolNU6z6ANwCEjQtBUkMChddWBkfWmB46TdJrUmLBnq6_kYi2PykeXfrdXvvEHxrnRhc</addsrcrecordid><sourcetype>Publisher</sourcetype><iscdi>true</iscdi><recordtype>conference_proceeding</recordtype></control><display><type>conference_proceeding</type><title>Training a neural network with conjugate gradient methods</title><source>IEEE Electronic Library (IEL) Conference Proceedings</source><creator>Towsey, M. ; Alpsan, D. ; Sztriha, L.</creator><creatorcontrib>Towsey, M. ; Alpsan, D. ; Sztriha, L.</creatorcontrib><description>This study investigates the use of several variants of conjugate gradient (CG) optimisation and line search methods to accelerate the convergence of an MLP neural network learning two medical signal classification problems. Much of the previous work has been done with artificial problems which have little relevance to real world problems and results on real world problems have been variable. The effectiveness of CG compared to standard backpropagation (BP) depended on the degree to which the learning task required finding a global minimum. If learning was stopped when the training set had been learned to an acceptable degree of error tolerance (the typical pattern classification problem), standard BP was faster than CG and did not display the convergence difficulties usually attributed to it. If learning required finding a global minimum (as in function minimisation or function estimation tasks), CG methods were faster but performance was very much dependent on careful selection of 'tuning' parameters and line search. This requirement for meta-optimisation was more difficult for CG than for BP because of the larger number of parameters.</description><identifier>ISBN: 9780780327689</identifier><identifier>ISBN: 0780327683</identifier><identifier>DOI: 10.1109/ICNN.1995.488128</identifier><language>eng</language><publisher>IEEE</publisher><subject>Acceleration ; Artificial neural networks ; Backpropagation ; Character generation ; Convergence ; Gradient methods ; Neural networks ; Optimization methods ; Pattern classification ; Search methods</subject><ispartof>Proceedings of ICNN'95 - International Conference on Neural Networks, 1995, Vol.1, p.373-378 vol.1</ispartof><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/488128$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>309,310,780,784,789,790,2058,4050,4051,27925,54920</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/488128$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc></links><search><creatorcontrib>Towsey, M.</creatorcontrib><creatorcontrib>Alpsan, D.</creatorcontrib><creatorcontrib>Sztriha, L.</creatorcontrib><title>Training a neural network with conjugate gradient methods</title><title>Proceedings of ICNN'95 - International Conference on Neural Networks</title><addtitle>ICNN</addtitle><description>This study investigates the use of several variants of conjugate gradient (CG) optimisation and line search methods to accelerate the convergence of an MLP neural network learning two medical signal classification problems. Much of the previous work has been done with artificial problems which have little relevance to real world problems and results on real world problems have been variable. The effectiveness of CG compared to standard backpropagation (BP) depended on the degree to which the learning task required finding a global minimum. If learning was stopped when the training set had been learned to an acceptable degree of error tolerance (the typical pattern classification problem), standard BP was faster than CG and did not display the convergence difficulties usually attributed to it. If learning required finding a global minimum (as in function minimisation or function estimation tasks), CG methods were faster but performance was very much dependent on careful selection of 'tuning' parameters and line search. This requirement for meta-optimisation was more difficult for CG than for BP because of the larger number of parameters.</description><subject>Acceleration</subject><subject>Artificial neural networks</subject><subject>Backpropagation</subject><subject>Character generation</subject><subject>Convergence</subject><subject>Gradient methods</subject><subject>Neural networks</subject><subject>Optimization methods</subject><subject>Pattern classification</subject><subject>Search methods</subject><isbn>9780780327689</isbn><isbn>0780327683</isbn><fulltext>true</fulltext><rsrctype>conference_proceeding</rsrctype><creationdate>1995</creationdate><recordtype>conference_proceeding</recordtype><sourceid>6IE</sourceid><sourceid>RIE</sourceid><recordid>eNotj8tqwzAQRQWl0JJ6X7rSD9idsSRLWhbTRyCkm3QdZGvkKE3sIiuE_n0D6eHC2V04jD0iVIhgn5ftel2htaqSxmBtblhhtYHLRK0bY-9YMc97uCCVElrdM7tJLo5xHLjjI52SO1yUz1P65ueYd7yfxv1pcJn4kJyPNGZ-pLyb_PzAboM7zFT8e8G-3l437Ue5-nxfti-rMiLIXHpCCk1nsUFlwAkNolNU6z6ANwCEjQtBUkMChddWBkfWmB46TdJrUmLBnq6_kYi2PykeXfrdXvvEHxrnRhc</recordid><startdate>1995</startdate><enddate>1995</enddate><creator>Towsey, M.</creator><creator>Alpsan, D.</creator><creator>Sztriha, L.</creator><general>IEEE</general><scope>6IE</scope><scope>6IL</scope><scope>CBEJK</scope><scope>RIE</scope><scope>RIL</scope></search><sort><creationdate>1995</creationdate><title>Training a neural network with conjugate gradient methods</title><author>Towsey, M. ; Alpsan, D. ; Sztriha, L.</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-i104t-de1ef6b9161580a3703b5e27cf0d800e16aff4e6e313d794fae988c0b7e4d7e53</frbrgroupid><rsrctype>conference_proceedings</rsrctype><prefilter>conference_proceedings</prefilter><language>eng</language><creationdate>1995</creationdate><topic>Acceleration</topic><topic>Artificial neural networks</topic><topic>Backpropagation</topic><topic>Character generation</topic><topic>Convergence</topic><topic>Gradient methods</topic><topic>Neural networks</topic><topic>Optimization methods</topic><topic>Pattern classification</topic><topic>Search methods</topic><toplevel>online_resources</toplevel><creatorcontrib>Towsey, M.</creatorcontrib><creatorcontrib>Alpsan, D.</creatorcontrib><creatorcontrib>Sztriha, L.</creatorcontrib><collection>IEEE Electronic Library (IEL) Conference Proceedings</collection><collection>IEEE Proceedings Order Plan All Online (POP All Online) 1998-present by volume</collection><collection>IEEE Xplore All Conference Proceedings</collection><collection>IEEE Electronic Library (IEL)</collection><collection>IEEE Proceedings Order Plans (POP All) 1998-Present</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Towsey, M.</au><au>Alpsan, D.</au><au>Sztriha, L.</au><format>book</format><genre>proceeding</genre><ristype>CONF</ristype><atitle>Training a neural network with conjugate gradient methods</atitle><btitle>Proceedings of ICNN'95 - International Conference on Neural Networks</btitle><stitle>ICNN</stitle><date>1995</date><risdate>1995</risdate><volume>1</volume><spage>373</spage><epage>378 vol.1</epage><pages>373-378 vol.1</pages><isbn>9780780327689</isbn><isbn>0780327683</isbn><abstract>This study investigates the use of several variants of conjugate gradient (CG) optimisation and line search methods to accelerate the convergence of an MLP neural network learning two medical signal classification problems. Much of the previous work has been done with artificial problems which have little relevance to real world problems and results on real world problems have been variable. The effectiveness of CG compared to standard backpropagation (BP) depended on the degree to which the learning task required finding a global minimum. If learning was stopped when the training set had been learned to an acceptable degree of error tolerance (the typical pattern classification problem), standard BP was faster than CG and did not display the convergence difficulties usually attributed to it. If learning required finding a global minimum (as in function minimisation or function estimation tasks), CG methods were faster but performance was very much dependent on careful selection of 'tuning' parameters and line search. This requirement for meta-optimisation was more difficult for CG than for BP because of the larger number of parameters.</abstract><pub>IEEE</pub><doi>10.1109/ICNN.1995.488128</doi></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | ISBN: 9780780327689 |
ispartof | Proceedings of ICNN'95 - International Conference on Neural Networks, 1995, Vol.1, p.373-378 vol.1 |
issn | |
language | eng |
recordid | cdi_ieee_primary_488128 |
source | IEEE Electronic Library (IEL) Conference Proceedings |
subjects | Acceleration Artificial neural networks Backpropagation Character generation Convergence Gradient methods Neural networks Optimization methods Pattern classification Search methods |
title | Training a neural network with conjugate gradient methods |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-07T04%3A11%3A27IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-ieee_6IE&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=proceeding&rft.atitle=Training%20a%20neural%20network%20with%20conjugate%20gradient%20methods&rft.btitle=Proceedings%20of%20ICNN'95%20-%20International%20Conference%20on%20Neural%20Networks&rft.au=Towsey,%20M.&rft.date=1995&rft.volume=1&rft.spage=373&rft.epage=378%20vol.1&rft.pages=373-378%20vol.1&rft.isbn=9780780327689&rft.isbn_list=0780327683&rft_id=info:doi/10.1109/ICNN.1995.488128&rft_dat=%3Cieee_6IE%3E488128%3C/ieee_6IE%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rft_ieee_id=488128&rfr_iscdi=true |