Improving the Performance of Deep Neural Networks Using Two Proposed Activation Functions
In artificial neural networks, activation functions play a significant role in the learning process. Choosing the proper activation function is a major factor in achieving a successful learning performance. Many activation functions are sufficient universal approximators, but their performance is la...
Gespeichert in:
Veröffentlicht in: | IEEE access 2021, Vol.9, p.82249-82271 |
---|---|
Hauptverfasser: | , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 82271 |
---|---|
container_issue | |
container_start_page | 82249 |
container_title | IEEE access |
container_volume | 9 |
creator | Alkhouly, Asmaa A. Mohammed, Ammar Hefny, Hesham A. |
description | In artificial neural networks, activation functions play a significant role in the learning process. Choosing the proper activation function is a major factor in achieving a successful learning performance. Many activation functions are sufficient universal approximators, but their performance is lacking. Thus, many efforts have been directed toward activation functions to improve the learning performance of artificial neural networks. However, the learning process involves many challenges, such as saturation, dying, and exploding/vanishing the gradient problems. The contribution of this work resides in several axes. First, we introduce two novel activation functions: absolute linear units and inverse polynomial linear units. Both activation functions are augmented by an adjustable parameter that controls the slope of the gradient. Second, we present a comprehensive study and a taxonomy of various types of activation functions. Third, we conduct a broad range of experiments on several deep neural architecture models with consideration of network type and depth. Fourth, we evaluate the proposed activation functions' performance in image and text classification tasks. For this purpose, several public benchmark datasets are utilized to evaluate and compare the performance of the proposed functions with that of a group of common activation functions. Finally, we deeply analyze the impact of several common activation functions on deep network architectures. Results reveal that the proposed functions outperform most of the popular activation functions in several benchmarks. The statistical study of the overall experiments on both classification categories indicates that the proposed activation functions are robust and superior among all the competitive activation functions in terms of average accuracy. |
doi_str_mv | 10.1109/ACCESS.2021.3085855 |
format | Article |
fullrecord | <record><control><sourceid>proquest_ieee_</sourceid><recordid>TN_cdi_ieee_primary_9446115</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>9446115</ieee_id><doaj_id>oai_doaj_org_article_ae2f8b6683294953a519080f06d57016</doaj_id><sourcerecordid>2541467548</sourcerecordid><originalsourceid>FETCH-LOGICAL-c408t-9ee896938a521a2fc030bd691ea96684b2aa5c23c4bda0451568124662157e833</originalsourceid><addsrcrecordid>eNpNUU1PAjEQ3RhNJMov4NLEM9hv2iNBUBKiJMDBU1O6s7gIW2wXiP_e4hLjXGby8t6bybws6xDcIwTrx8FwOJrPexRT0mNYCSXEVdaiROouE0xe_5tvs3aMG5xKJUj0W9n7ZLcP_lhWa1R_AJpBKHzY2coB8gV6AtijVzgEu02tPvnwGdEyntmLk0ez4Pc-Qo4Gri6Pti59hcaHyp2HeJ_dFHYboX3pd9lyPFoMX7rTt-fJcDDtOo5V3dUASkvNlBWUWFo4zPAql5qA1VIqvqLWCkeZ46vcYi6IkIpQLiUlog-Ksbts0vjm3m7MPpQ7G76Nt6X5BXxYGxvq0m3BWKCFWiVXRjXXgllBNFa4wDIXfUxk8npovNJPvg4Qa7Pxh1Cl8w0VnHDZF1wlFmtYLvgYAxR_Wwk250hME4k5R2IukSRVp1GVAPCn0JxLQgT7Abg7hW0</addsrcrecordid><sourcetype>Open Website</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2541467548</pqid></control><display><type>article</type><title>Improving the Performance of Deep Neural Networks Using Two Proposed Activation Functions</title><source>IEEE Open Access Journals</source><source>DOAJ Directory of Open Access Journals</source><source>Elektronische Zeitschriftenbibliothek - Frei zugängliche E-Journals</source><creator>Alkhouly, Asmaa A. ; Mohammed, Ammar ; Hefny, Hesham A.</creator><creatorcontrib>Alkhouly, Asmaa A. ; Mohammed, Ammar ; Hefny, Hesham A.</creatorcontrib><description>In artificial neural networks, activation functions play a significant role in the learning process. Choosing the proper activation function is a major factor in achieving a successful learning performance. Many activation functions are sufficient universal approximators, but their performance is lacking. Thus, many efforts have been directed toward activation functions to improve the learning performance of artificial neural networks. However, the learning process involves many challenges, such as saturation, dying, and exploding/vanishing the gradient problems. The contribution of this work resides in several axes. First, we introduce two novel activation functions: absolute linear units and inverse polynomial linear units. Both activation functions are augmented by an adjustable parameter that controls the slope of the gradient. Second, we present a comprehensive study and a taxonomy of various types of activation functions. Third, we conduct a broad range of experiments on several deep neural architecture models with consideration of network type and depth. Fourth, we evaluate the proposed activation functions' performance in image and text classification tasks. For this purpose, several public benchmark datasets are utilized to evaluate and compare the performance of the proposed functions with that of a group of common activation functions. Finally, we deeply analyze the impact of several common activation functions on deep network architectures. Results reveal that the proposed functions outperform most of the popular activation functions in several benchmarks. The statistical study of the overall experiments on both classification categories indicates that the proposed activation functions are robust and superior among all the competitive activation functions in terms of average accuracy.</description><identifier>ISSN: 2169-3536</identifier><identifier>EISSN: 2169-3536</identifier><identifier>DOI: 10.1109/ACCESS.2021.3085855</identifier><identifier>CODEN: IAECCG</identifier><language>eng</language><publisher>Piscataway: IEEE</publisher><subject>activation function ; Artificial neural network ; Artificial neural networks ; Benchmarks ; Computer architecture ; Computer science ; Convergence ; deep neural network ; Functions (mathematics) ; Image classification ; Impact analysis ; Learning ; learning challenges ; Logistics ; Machine learning ; Neural networks ; Neurons ; Polynomials ; Slope gradients ; Taxonomy</subject><ispartof>IEEE access, 2021, Vol.9, p.82249-82271</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2021</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c408t-9ee896938a521a2fc030bd691ea96684b2aa5c23c4bda0451568124662157e833</citedby><cites>FETCH-LOGICAL-c408t-9ee896938a521a2fc030bd691ea96684b2aa5c23c4bda0451568124662157e833</cites><orcidid>0000-0001-6844-9451 ; 0000-0003-4851-8688 ; 0000-0001-5862-675X</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/9446115$$EHTML$$P50$$Gieee$$Hfree_for_read</linktohtml><link.rule.ids>314,776,780,860,2095,4009,27612,27902,27903,27904,54912</link.rule.ids></links><search><creatorcontrib>Alkhouly, Asmaa A.</creatorcontrib><creatorcontrib>Mohammed, Ammar</creatorcontrib><creatorcontrib>Hefny, Hesham A.</creatorcontrib><title>Improving the Performance of Deep Neural Networks Using Two Proposed Activation Functions</title><title>IEEE access</title><addtitle>Access</addtitle><description>In artificial neural networks, activation functions play a significant role in the learning process. Choosing the proper activation function is a major factor in achieving a successful learning performance. Many activation functions are sufficient universal approximators, but their performance is lacking. Thus, many efforts have been directed toward activation functions to improve the learning performance of artificial neural networks. However, the learning process involves many challenges, such as saturation, dying, and exploding/vanishing the gradient problems. The contribution of this work resides in several axes. First, we introduce two novel activation functions: absolute linear units and inverse polynomial linear units. Both activation functions are augmented by an adjustable parameter that controls the slope of the gradient. Second, we present a comprehensive study and a taxonomy of various types of activation functions. Third, we conduct a broad range of experiments on several deep neural architecture models with consideration of network type and depth. Fourth, we evaluate the proposed activation functions' performance in image and text classification tasks. For this purpose, several public benchmark datasets are utilized to evaluate and compare the performance of the proposed functions with that of a group of common activation functions. Finally, we deeply analyze the impact of several common activation functions on deep network architectures. Results reveal that the proposed functions outperform most of the popular activation functions in several benchmarks. The statistical study of the overall experiments on both classification categories indicates that the proposed activation functions are robust and superior among all the competitive activation functions in terms of average accuracy.</description><subject>activation function</subject><subject>Artificial neural network</subject><subject>Artificial neural networks</subject><subject>Benchmarks</subject><subject>Computer architecture</subject><subject>Computer science</subject><subject>Convergence</subject><subject>deep neural network</subject><subject>Functions (mathematics)</subject><subject>Image classification</subject><subject>Impact analysis</subject><subject>Learning</subject><subject>learning challenges</subject><subject>Logistics</subject><subject>Machine learning</subject><subject>Neural networks</subject><subject>Neurons</subject><subject>Polynomials</subject><subject>Slope gradients</subject><subject>Taxonomy</subject><issn>2169-3536</issn><issn>2169-3536</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>ESBDL</sourceid><sourceid>RIE</sourceid><sourceid>DOA</sourceid><recordid>eNpNUU1PAjEQ3RhNJMov4NLEM9hv2iNBUBKiJMDBU1O6s7gIW2wXiP_e4hLjXGby8t6bybws6xDcIwTrx8FwOJrPexRT0mNYCSXEVdaiROouE0xe_5tvs3aMG5xKJUj0W9n7ZLcP_lhWa1R_AJpBKHzY2coB8gV6AtijVzgEu02tPvnwGdEyntmLk0ez4Pc-Qo4Gri6Pti59hcaHyp2HeJ_dFHYboX3pd9lyPFoMX7rTt-fJcDDtOo5V3dUASkvNlBWUWFo4zPAql5qA1VIqvqLWCkeZ46vcYi6IkIpQLiUlog-Ksbts0vjm3m7MPpQ7G76Nt6X5BXxYGxvq0m3BWKCFWiVXRjXXgllBNFa4wDIXfUxk8npovNJPvg4Qa7Pxh1Cl8w0VnHDZF1wlFmtYLvgYAxR_Wwk250hME4k5R2IukSRVp1GVAPCn0JxLQgT7Abg7hW0</recordid><startdate>2021</startdate><enddate>2021</enddate><creator>Alkhouly, Asmaa A.</creator><creator>Mohammed, Ammar</creator><creator>Hefny, Hesham A.</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>ESBDL</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>7SR</scope><scope>8BQ</scope><scope>8FD</scope><scope>JG9</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>DOA</scope><orcidid>https://orcid.org/0000-0001-6844-9451</orcidid><orcidid>https://orcid.org/0000-0003-4851-8688</orcidid><orcidid>https://orcid.org/0000-0001-5862-675X</orcidid></search><sort><creationdate>2021</creationdate><title>Improving the Performance of Deep Neural Networks Using Two Proposed Activation Functions</title><author>Alkhouly, Asmaa A. ; Mohammed, Ammar ; Hefny, Hesham A.</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c408t-9ee896938a521a2fc030bd691ea96684b2aa5c23c4bda0451568124662157e833</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>activation function</topic><topic>Artificial neural network</topic><topic>Artificial neural networks</topic><topic>Benchmarks</topic><topic>Computer architecture</topic><topic>Computer science</topic><topic>Convergence</topic><topic>deep neural network</topic><topic>Functions (mathematics)</topic><topic>Image classification</topic><topic>Impact analysis</topic><topic>Learning</topic><topic>learning challenges</topic><topic>Logistics</topic><topic>Machine learning</topic><topic>Neural networks</topic><topic>Neurons</topic><topic>Polynomials</topic><topic>Slope gradients</topic><topic>Taxonomy</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Alkhouly, Asmaa A.</creatorcontrib><creatorcontrib>Mohammed, Ammar</creatorcontrib><creatorcontrib>Hefny, Hesham A.</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE Open Access Journals</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics & Communications Abstracts</collection><collection>Engineered Materials Abstracts</collection><collection>METADEX</collection><collection>Technology Research Database</collection><collection>Materials Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>DOAJ Directory of Open Access Journals</collection><jtitle>IEEE access</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Alkhouly, Asmaa A.</au><au>Mohammed, Ammar</au><au>Hefny, Hesham A.</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Improving the Performance of Deep Neural Networks Using Two Proposed Activation Functions</atitle><jtitle>IEEE access</jtitle><stitle>Access</stitle><date>2021</date><risdate>2021</risdate><volume>9</volume><spage>82249</spage><epage>82271</epage><pages>82249-82271</pages><issn>2169-3536</issn><eissn>2169-3536</eissn><coden>IAECCG</coden><abstract>In artificial neural networks, activation functions play a significant role in the learning process. Choosing the proper activation function is a major factor in achieving a successful learning performance. Many activation functions are sufficient universal approximators, but their performance is lacking. Thus, many efforts have been directed toward activation functions to improve the learning performance of artificial neural networks. However, the learning process involves many challenges, such as saturation, dying, and exploding/vanishing the gradient problems. The contribution of this work resides in several axes. First, we introduce two novel activation functions: absolute linear units and inverse polynomial linear units. Both activation functions are augmented by an adjustable parameter that controls the slope of the gradient. Second, we present a comprehensive study and a taxonomy of various types of activation functions. Third, we conduct a broad range of experiments on several deep neural architecture models with consideration of network type and depth. Fourth, we evaluate the proposed activation functions' performance in image and text classification tasks. For this purpose, several public benchmark datasets are utilized to evaluate and compare the performance of the proposed functions with that of a group of common activation functions. Finally, we deeply analyze the impact of several common activation functions on deep network architectures. Results reveal that the proposed functions outperform most of the popular activation functions in several benchmarks. The statistical study of the overall experiments on both classification categories indicates that the proposed activation functions are robust and superior among all the competitive activation functions in terms of average accuracy.</abstract><cop>Piscataway</cop><pub>IEEE</pub><doi>10.1109/ACCESS.2021.3085855</doi><tpages>23</tpages><orcidid>https://orcid.org/0000-0001-6844-9451</orcidid><orcidid>https://orcid.org/0000-0003-4851-8688</orcidid><orcidid>https://orcid.org/0000-0001-5862-675X</orcidid><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 2169-3536 |
ispartof | IEEE access, 2021, Vol.9, p.82249-82271 |
issn | 2169-3536 2169-3536 |
language | eng |
recordid | cdi_ieee_primary_9446115 |
source | IEEE Open Access Journals; DOAJ Directory of Open Access Journals; Elektronische Zeitschriftenbibliothek - Frei zugängliche E-Journals |
subjects | activation function Artificial neural network Artificial neural networks Benchmarks Computer architecture Computer science Convergence deep neural network Functions (mathematics) Image classification Impact analysis Learning learning challenges Logistics Machine learning Neural networks Neurons Polynomials Slope gradients Taxonomy |
title | Improving the Performance of Deep Neural Networks Using Two Proposed Activation Functions |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-23T05%3A13%3A29IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_ieee_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Improving%20the%20Performance%20of%20Deep%20Neural%20Networks%20Using%20Two%20Proposed%20Activation%20Functions&rft.jtitle=IEEE%20access&rft.au=Alkhouly,%20Asmaa%20A.&rft.date=2021&rft.volume=9&rft.spage=82249&rft.epage=82271&rft.pages=82249-82271&rft.issn=2169-3536&rft.eissn=2169-3536&rft.coden=IAECCG&rft_id=info:doi/10.1109/ACCESS.2021.3085855&rft_dat=%3Cproquest_ieee_%3E2541467548%3C/proquest_ieee_%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2541467548&rft_id=info:pmid/&rft_ieee_id=9446115&rft_doaj_id=oai_doaj_org_article_ae2f8b6683294953a519080f06d57016&rfr_iscdi=true |