Training a neural network with conjugate gradient methods

This study investigates the use of several variants of conjugate gradient (CG) optimisation and line search methods to accelerate the convergence of an MLP neural network learning two medical signal classification problems. Much of the previous work has been done with artificial problems which have...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Towsey, M., Alpsan, D., Sztriha, L.
Format: Tagungsbericht
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 378 vol.1
container_issue
container_start_page 373
container_title
container_volume 1
creator Towsey, M.
Alpsan, D.
Sztriha, L.
description This study investigates the use of several variants of conjugate gradient (CG) optimisation and line search methods to accelerate the convergence of an MLP neural network learning two medical signal classification problems. Much of the previous work has been done with artificial problems which have little relevance to real world problems and results on real world problems have been variable. The effectiveness of CG compared to standard backpropagation (BP) depended on the degree to which the learning task required finding a global minimum. If learning was stopped when the training set had been learned to an acceptable degree of error tolerance (the typical pattern classification problem), standard BP was faster than CG and did not display the convergence difficulties usually attributed to it. If learning required finding a global minimum (as in function minimisation or function estimation tasks), CG methods were faster but performance was very much dependent on careful selection of 'tuning' parameters and line search. This requirement for meta-optimisation was more difficult for CG than for BP because of the larger number of parameters.
doi_str_mv 10.1109/ICNN.1995.488128
format Conference Proceeding
fullrecord <record><control><sourceid>ieee_6IE</sourceid><recordid>TN_cdi_ieee_primary_488128</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>488128</ieee_id><sourcerecordid>488128</sourcerecordid><originalsourceid>FETCH-LOGICAL-i104t-de1ef6b9161580a3703b5e27cf0d800e16aff4e6e313d794fae988c0b7e4d7e53</originalsourceid><addsrcrecordid>eNotj8tqwzAQRQWl0JJ6X7rSD9idsSRLWhbTRyCkm3QdZGvkKE3sIiuE_n0D6eHC2V04jD0iVIhgn5ftel2htaqSxmBtblhhtYHLRK0bY-9YMc97uCCVElrdM7tJLo5xHLjjI52SO1yUz1P65ueYd7yfxv1pcJn4kJyPNGZ-pLyb_PzAboM7zFT8e8G-3l437Ue5-nxfti-rMiLIXHpCCk1nsUFlwAkNolNU6z6ANwCEjQtBUkMChddWBkfWmB46TdJrUmLBnq6_kYi2PykeXfrdXvvEHxrnRhc</addsrcrecordid><sourcetype>Publisher</sourcetype><iscdi>true</iscdi><recordtype>conference_proceeding</recordtype></control><display><type>conference_proceeding</type><title>Training a neural network with conjugate gradient methods</title><source>IEEE Electronic Library (IEL) Conference Proceedings</source><creator>Towsey, M. ; Alpsan, D. ; Sztriha, L.</creator><creatorcontrib>Towsey, M. ; Alpsan, D. ; Sztriha, L.</creatorcontrib><description>This study investigates the use of several variants of conjugate gradient (CG) optimisation and line search methods to accelerate the convergence of an MLP neural network learning two medical signal classification problems. Much of the previous work has been done with artificial problems which have little relevance to real world problems and results on real world problems have been variable. The effectiveness of CG compared to standard backpropagation (BP) depended on the degree to which the learning task required finding a global minimum. If learning was stopped when the training set had been learned to an acceptable degree of error tolerance (the typical pattern classification problem), standard BP was faster than CG and did not display the convergence difficulties usually attributed to it. If learning required finding a global minimum (as in function minimisation or function estimation tasks), CG methods were faster but performance was very much dependent on careful selection of 'tuning' parameters and line search. This requirement for meta-optimisation was more difficult for CG than for BP because of the larger number of parameters.</description><identifier>ISBN: 9780780327689</identifier><identifier>ISBN: 0780327683</identifier><identifier>DOI: 10.1109/ICNN.1995.488128</identifier><language>eng</language><publisher>IEEE</publisher><subject>Acceleration ; Artificial neural networks ; Backpropagation ; Character generation ; Convergence ; Gradient methods ; Neural networks ; Optimization methods ; Pattern classification ; Search methods</subject><ispartof>Proceedings of ICNN'95 - International Conference on Neural Networks, 1995, Vol.1, p.373-378 vol.1</ispartof><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/488128$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>309,310,780,784,789,790,2058,4050,4051,27925,54920</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/488128$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc></links><search><creatorcontrib>Towsey, M.</creatorcontrib><creatorcontrib>Alpsan, D.</creatorcontrib><creatorcontrib>Sztriha, L.</creatorcontrib><title>Training a neural network with conjugate gradient methods</title><title>Proceedings of ICNN'95 - International Conference on Neural Networks</title><addtitle>ICNN</addtitle><description>This study investigates the use of several variants of conjugate gradient (CG) optimisation and line search methods to accelerate the convergence of an MLP neural network learning two medical signal classification problems. Much of the previous work has been done with artificial problems which have little relevance to real world problems and results on real world problems have been variable. The effectiveness of CG compared to standard backpropagation (BP) depended on the degree to which the learning task required finding a global minimum. If learning was stopped when the training set had been learned to an acceptable degree of error tolerance (the typical pattern classification problem), standard BP was faster than CG and did not display the convergence difficulties usually attributed to it. If learning required finding a global minimum (as in function minimisation or function estimation tasks), CG methods were faster but performance was very much dependent on careful selection of 'tuning' parameters and line search. This requirement for meta-optimisation was more difficult for CG than for BP because of the larger number of parameters.</description><subject>Acceleration</subject><subject>Artificial neural networks</subject><subject>Backpropagation</subject><subject>Character generation</subject><subject>Convergence</subject><subject>Gradient methods</subject><subject>Neural networks</subject><subject>Optimization methods</subject><subject>Pattern classification</subject><subject>Search methods</subject><isbn>9780780327689</isbn><isbn>0780327683</isbn><fulltext>true</fulltext><rsrctype>conference_proceeding</rsrctype><creationdate>1995</creationdate><recordtype>conference_proceeding</recordtype><sourceid>6IE</sourceid><sourceid>RIE</sourceid><recordid>eNotj8tqwzAQRQWl0JJ6X7rSD9idsSRLWhbTRyCkm3QdZGvkKE3sIiuE_n0D6eHC2V04jD0iVIhgn5ftel2htaqSxmBtblhhtYHLRK0bY-9YMc97uCCVElrdM7tJLo5xHLjjI52SO1yUz1P65ueYd7yfxv1pcJn4kJyPNGZ-pLyb_PzAboM7zFT8e8G-3l437Ue5-nxfti-rMiLIXHpCCk1nsUFlwAkNolNU6z6ANwCEjQtBUkMChddWBkfWmB46TdJrUmLBnq6_kYi2PykeXfrdXvvEHxrnRhc</recordid><startdate>1995</startdate><enddate>1995</enddate><creator>Towsey, M.</creator><creator>Alpsan, D.</creator><creator>Sztriha, L.</creator><general>IEEE</general><scope>6IE</scope><scope>6IL</scope><scope>CBEJK</scope><scope>RIE</scope><scope>RIL</scope></search><sort><creationdate>1995</creationdate><title>Training a neural network with conjugate gradient methods</title><author>Towsey, M. ; Alpsan, D. ; Sztriha, L.</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-i104t-de1ef6b9161580a3703b5e27cf0d800e16aff4e6e313d794fae988c0b7e4d7e53</frbrgroupid><rsrctype>conference_proceedings</rsrctype><prefilter>conference_proceedings</prefilter><language>eng</language><creationdate>1995</creationdate><topic>Acceleration</topic><topic>Artificial neural networks</topic><topic>Backpropagation</topic><topic>Character generation</topic><topic>Convergence</topic><topic>Gradient methods</topic><topic>Neural networks</topic><topic>Optimization methods</topic><topic>Pattern classification</topic><topic>Search methods</topic><toplevel>online_resources</toplevel><creatorcontrib>Towsey, M.</creatorcontrib><creatorcontrib>Alpsan, D.</creatorcontrib><creatorcontrib>Sztriha, L.</creatorcontrib><collection>IEEE Electronic Library (IEL) Conference Proceedings</collection><collection>IEEE Proceedings Order Plan All Online (POP All Online) 1998-present by volume</collection><collection>IEEE Xplore All Conference Proceedings</collection><collection>IEEE Electronic Library (IEL)</collection><collection>IEEE Proceedings Order Plans (POP All) 1998-Present</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Towsey, M.</au><au>Alpsan, D.</au><au>Sztriha, L.</au><format>book</format><genre>proceeding</genre><ristype>CONF</ristype><atitle>Training a neural network with conjugate gradient methods</atitle><btitle>Proceedings of ICNN'95 - International Conference on Neural Networks</btitle><stitle>ICNN</stitle><date>1995</date><risdate>1995</risdate><volume>1</volume><spage>373</spage><epage>378 vol.1</epage><pages>373-378 vol.1</pages><isbn>9780780327689</isbn><isbn>0780327683</isbn><abstract>This study investigates the use of several variants of conjugate gradient (CG) optimisation and line search methods to accelerate the convergence of an MLP neural network learning two medical signal classification problems. Much of the previous work has been done with artificial problems which have little relevance to real world problems and results on real world problems have been variable. The effectiveness of CG compared to standard backpropagation (BP) depended on the degree to which the learning task required finding a global minimum. If learning was stopped when the training set had been learned to an acceptable degree of error tolerance (the typical pattern classification problem), standard BP was faster than CG and did not display the convergence difficulties usually attributed to it. If learning required finding a global minimum (as in function minimisation or function estimation tasks), CG methods were faster but performance was very much dependent on careful selection of 'tuning' parameters and line search. This requirement for meta-optimisation was more difficult for CG than for BP because of the larger number of parameters.</abstract><pub>IEEE</pub><doi>10.1109/ICNN.1995.488128</doi></addata></record>
fulltext fulltext_linktorsrc
identifier ISBN: 9780780327689
ispartof Proceedings of ICNN'95 - International Conference on Neural Networks, 1995, Vol.1, p.373-378 vol.1
issn
language eng
recordid cdi_ieee_primary_488128
source IEEE Electronic Library (IEL) Conference Proceedings
subjects Acceleration
Artificial neural networks
Backpropagation
Character generation
Convergence
Gradient methods
Neural networks
Optimization methods
Pattern classification
Search methods
title Training a neural network with conjugate gradient methods
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-07T04%3A11%3A27IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-ieee_6IE&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=proceeding&rft.atitle=Training%20a%20neural%20network%20with%20conjugate%20gradient%20methods&rft.btitle=Proceedings%20of%20ICNN'95%20-%20International%20Conference%20on%20Neural%20Networks&rft.au=Towsey,%20M.&rft.date=1995&rft.volume=1&rft.spage=373&rft.epage=378%20vol.1&rft.pages=373-378%20vol.1&rft.isbn=9780780327689&rft.isbn_list=0780327683&rft_id=info:doi/10.1109/ICNN.1995.488128&rft_dat=%3Cieee_6IE%3E488128%3C/ieee_6IE%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rft_ieee_id=488128&rfr_iscdi=true