Improving the Performance of Deep Neural Networks Using Two Proposed Activation Functions

In artificial neural networks, activation functions play a significant role in the learning process. Choosing the proper activation function is a major factor in achieving a successful learning performance. Many activation functions are sufficient universal approximators, but their performance is la...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE access 2021, Vol.9, p.82249-82271
Hauptverfasser: Alkhouly, Asmaa A., Mohammed, Ammar, Hefny, Hesham A.
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 82271
container_issue
container_start_page 82249
container_title IEEE access
container_volume 9
creator Alkhouly, Asmaa A.
Mohammed, Ammar
Hefny, Hesham A.
description In artificial neural networks, activation functions play a significant role in the learning process. Choosing the proper activation function is a major factor in achieving a successful learning performance. Many activation functions are sufficient universal approximators, but their performance is lacking. Thus, many efforts have been directed toward activation functions to improve the learning performance of artificial neural networks. However, the learning process involves many challenges, such as saturation, dying, and exploding/vanishing the gradient problems. The contribution of this work resides in several axes. First, we introduce two novel activation functions: absolute linear units and inverse polynomial linear units. Both activation functions are augmented by an adjustable parameter that controls the slope of the gradient. Second, we present a comprehensive study and a taxonomy of various types of activation functions. Third, we conduct a broad range of experiments on several deep neural architecture models with consideration of network type and depth. Fourth, we evaluate the proposed activation functions' performance in image and text classification tasks. For this purpose, several public benchmark datasets are utilized to evaluate and compare the performance of the proposed functions with that of a group of common activation functions. Finally, we deeply analyze the impact of several common activation functions on deep network architectures. Results reveal that the proposed functions outperform most of the popular activation functions in several benchmarks. The statistical study of the overall experiments on both classification categories indicates that the proposed activation functions are robust and superior among all the competitive activation functions in terms of average accuracy.
doi_str_mv 10.1109/ACCESS.2021.3085855
format Article
fullrecord <record><control><sourceid>proquest_ieee_</sourceid><recordid>TN_cdi_ieee_primary_9446115</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>9446115</ieee_id><doaj_id>oai_doaj_org_article_ae2f8b6683294953a519080f06d57016</doaj_id><sourcerecordid>2541467548</sourcerecordid><originalsourceid>FETCH-LOGICAL-c408t-9ee896938a521a2fc030bd691ea96684b2aa5c23c4bda0451568124662157e833</originalsourceid><addsrcrecordid>eNpNUU1PAjEQ3RhNJMov4NLEM9hv2iNBUBKiJMDBU1O6s7gIW2wXiP_e4hLjXGby8t6bybws6xDcIwTrx8FwOJrPexRT0mNYCSXEVdaiROouE0xe_5tvs3aMG5xKJUj0W9n7ZLcP_lhWa1R_AJpBKHzY2coB8gV6AtijVzgEu02tPvnwGdEyntmLk0ez4Pc-Qo4Gri6Pti59hcaHyp2HeJ_dFHYboX3pd9lyPFoMX7rTt-fJcDDtOo5V3dUASkvNlBWUWFo4zPAql5qA1VIqvqLWCkeZ46vcYi6IkIpQLiUlog-Ksbts0vjm3m7MPpQ7G76Nt6X5BXxYGxvq0m3BWKCFWiVXRjXXgllBNFa4wDIXfUxk8npovNJPvg4Qa7Pxh1Cl8w0VnHDZF1wlFmtYLvgYAxR_Wwk250hME4k5R2IukSRVp1GVAPCn0JxLQgT7Abg7hW0</addsrcrecordid><sourcetype>Open Website</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2541467548</pqid></control><display><type>article</type><title>Improving the Performance of Deep Neural Networks Using Two Proposed Activation Functions</title><source>IEEE Open Access Journals</source><source>DOAJ Directory of Open Access Journals</source><source>Elektronische Zeitschriftenbibliothek - Frei zugängliche E-Journals</source><creator>Alkhouly, Asmaa A. ; Mohammed, Ammar ; Hefny, Hesham A.</creator><creatorcontrib>Alkhouly, Asmaa A. ; Mohammed, Ammar ; Hefny, Hesham A.</creatorcontrib><description>In artificial neural networks, activation functions play a significant role in the learning process. Choosing the proper activation function is a major factor in achieving a successful learning performance. Many activation functions are sufficient universal approximators, but their performance is lacking. Thus, many efforts have been directed toward activation functions to improve the learning performance of artificial neural networks. However, the learning process involves many challenges, such as saturation, dying, and exploding/vanishing the gradient problems. The contribution of this work resides in several axes. First, we introduce two novel activation functions: absolute linear units and inverse polynomial linear units. Both activation functions are augmented by an adjustable parameter that controls the slope of the gradient. Second, we present a comprehensive study and a taxonomy of various types of activation functions. Third, we conduct a broad range of experiments on several deep neural architecture models with consideration of network type and depth. Fourth, we evaluate the proposed activation functions' performance in image and text classification tasks. For this purpose, several public benchmark datasets are utilized to evaluate and compare the performance of the proposed functions with that of a group of common activation functions. Finally, we deeply analyze the impact of several common activation functions on deep network architectures. Results reveal that the proposed functions outperform most of the popular activation functions in several benchmarks. The statistical study of the overall experiments on both classification categories indicates that the proposed activation functions are robust and superior among all the competitive activation functions in terms of average accuracy.</description><identifier>ISSN: 2169-3536</identifier><identifier>EISSN: 2169-3536</identifier><identifier>DOI: 10.1109/ACCESS.2021.3085855</identifier><identifier>CODEN: IAECCG</identifier><language>eng</language><publisher>Piscataway: IEEE</publisher><subject>activation function ; Artificial neural network ; Artificial neural networks ; Benchmarks ; Computer architecture ; Computer science ; Convergence ; deep neural network ; Functions (mathematics) ; Image classification ; Impact analysis ; Learning ; learning challenges ; Logistics ; Machine learning ; Neural networks ; Neurons ; Polynomials ; Slope gradients ; Taxonomy</subject><ispartof>IEEE access, 2021, Vol.9, p.82249-82271</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2021</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c408t-9ee896938a521a2fc030bd691ea96684b2aa5c23c4bda0451568124662157e833</citedby><cites>FETCH-LOGICAL-c408t-9ee896938a521a2fc030bd691ea96684b2aa5c23c4bda0451568124662157e833</cites><orcidid>0000-0001-6844-9451 ; 0000-0003-4851-8688 ; 0000-0001-5862-675X</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/9446115$$EHTML$$P50$$Gieee$$Hfree_for_read</linktohtml><link.rule.ids>314,776,780,860,2095,4009,27612,27902,27903,27904,54912</link.rule.ids></links><search><creatorcontrib>Alkhouly, Asmaa A.</creatorcontrib><creatorcontrib>Mohammed, Ammar</creatorcontrib><creatorcontrib>Hefny, Hesham A.</creatorcontrib><title>Improving the Performance of Deep Neural Networks Using Two Proposed Activation Functions</title><title>IEEE access</title><addtitle>Access</addtitle><description>In artificial neural networks, activation functions play a significant role in the learning process. Choosing the proper activation function is a major factor in achieving a successful learning performance. Many activation functions are sufficient universal approximators, but their performance is lacking. Thus, many efforts have been directed toward activation functions to improve the learning performance of artificial neural networks. However, the learning process involves many challenges, such as saturation, dying, and exploding/vanishing the gradient problems. The contribution of this work resides in several axes. First, we introduce two novel activation functions: absolute linear units and inverse polynomial linear units. Both activation functions are augmented by an adjustable parameter that controls the slope of the gradient. Second, we present a comprehensive study and a taxonomy of various types of activation functions. Third, we conduct a broad range of experiments on several deep neural architecture models with consideration of network type and depth. Fourth, we evaluate the proposed activation functions' performance in image and text classification tasks. For this purpose, several public benchmark datasets are utilized to evaluate and compare the performance of the proposed functions with that of a group of common activation functions. Finally, we deeply analyze the impact of several common activation functions on deep network architectures. Results reveal that the proposed functions outperform most of the popular activation functions in several benchmarks. The statistical study of the overall experiments on both classification categories indicates that the proposed activation functions are robust and superior among all the competitive activation functions in terms of average accuracy.</description><subject>activation function</subject><subject>Artificial neural network</subject><subject>Artificial neural networks</subject><subject>Benchmarks</subject><subject>Computer architecture</subject><subject>Computer science</subject><subject>Convergence</subject><subject>deep neural network</subject><subject>Functions (mathematics)</subject><subject>Image classification</subject><subject>Impact analysis</subject><subject>Learning</subject><subject>learning challenges</subject><subject>Logistics</subject><subject>Machine learning</subject><subject>Neural networks</subject><subject>Neurons</subject><subject>Polynomials</subject><subject>Slope gradients</subject><subject>Taxonomy</subject><issn>2169-3536</issn><issn>2169-3536</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>ESBDL</sourceid><sourceid>RIE</sourceid><sourceid>DOA</sourceid><recordid>eNpNUU1PAjEQ3RhNJMov4NLEM9hv2iNBUBKiJMDBU1O6s7gIW2wXiP_e4hLjXGby8t6bybws6xDcIwTrx8FwOJrPexRT0mNYCSXEVdaiROouE0xe_5tvs3aMG5xKJUj0W9n7ZLcP_lhWa1R_AJpBKHzY2coB8gV6AtijVzgEu02tPvnwGdEyntmLk0ez4Pc-Qo4Gri6Pti59hcaHyp2HeJ_dFHYboX3pd9lyPFoMX7rTt-fJcDDtOo5V3dUASkvNlBWUWFo4zPAql5qA1VIqvqLWCkeZ46vcYi6IkIpQLiUlog-Ksbts0vjm3m7MPpQ7G76Nt6X5BXxYGxvq0m3BWKCFWiVXRjXXgllBNFa4wDIXfUxk8npovNJPvg4Qa7Pxh1Cl8w0VnHDZF1wlFmtYLvgYAxR_Wwk250hME4k5R2IukSRVp1GVAPCn0JxLQgT7Abg7hW0</recordid><startdate>2021</startdate><enddate>2021</enddate><creator>Alkhouly, Asmaa A.</creator><creator>Mohammed, Ammar</creator><creator>Hefny, Hesham A.</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>ESBDL</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>7SR</scope><scope>8BQ</scope><scope>8FD</scope><scope>JG9</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>DOA</scope><orcidid>https://orcid.org/0000-0001-6844-9451</orcidid><orcidid>https://orcid.org/0000-0003-4851-8688</orcidid><orcidid>https://orcid.org/0000-0001-5862-675X</orcidid></search><sort><creationdate>2021</creationdate><title>Improving the Performance of Deep Neural Networks Using Two Proposed Activation Functions</title><author>Alkhouly, Asmaa A. ; Mohammed, Ammar ; Hefny, Hesham A.</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c408t-9ee896938a521a2fc030bd691ea96684b2aa5c23c4bda0451568124662157e833</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>activation function</topic><topic>Artificial neural network</topic><topic>Artificial neural networks</topic><topic>Benchmarks</topic><topic>Computer architecture</topic><topic>Computer science</topic><topic>Convergence</topic><topic>deep neural network</topic><topic>Functions (mathematics)</topic><topic>Image classification</topic><topic>Impact analysis</topic><topic>Learning</topic><topic>learning challenges</topic><topic>Logistics</topic><topic>Machine learning</topic><topic>Neural networks</topic><topic>Neurons</topic><topic>Polynomials</topic><topic>Slope gradients</topic><topic>Taxonomy</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Alkhouly, Asmaa A.</creatorcontrib><creatorcontrib>Mohammed, Ammar</creatorcontrib><creatorcontrib>Hefny, Hesham A.</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE Open Access Journals</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>Engineered Materials Abstracts</collection><collection>METADEX</collection><collection>Technology Research Database</collection><collection>Materials Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>DOAJ Directory of Open Access Journals</collection><jtitle>IEEE access</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Alkhouly, Asmaa A.</au><au>Mohammed, Ammar</au><au>Hefny, Hesham A.</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Improving the Performance of Deep Neural Networks Using Two Proposed Activation Functions</atitle><jtitle>IEEE access</jtitle><stitle>Access</stitle><date>2021</date><risdate>2021</risdate><volume>9</volume><spage>82249</spage><epage>82271</epage><pages>82249-82271</pages><issn>2169-3536</issn><eissn>2169-3536</eissn><coden>IAECCG</coden><abstract>In artificial neural networks, activation functions play a significant role in the learning process. Choosing the proper activation function is a major factor in achieving a successful learning performance. Many activation functions are sufficient universal approximators, but their performance is lacking. Thus, many efforts have been directed toward activation functions to improve the learning performance of artificial neural networks. However, the learning process involves many challenges, such as saturation, dying, and exploding/vanishing the gradient problems. The contribution of this work resides in several axes. First, we introduce two novel activation functions: absolute linear units and inverse polynomial linear units. Both activation functions are augmented by an adjustable parameter that controls the slope of the gradient. Second, we present a comprehensive study and a taxonomy of various types of activation functions. Third, we conduct a broad range of experiments on several deep neural architecture models with consideration of network type and depth. Fourth, we evaluate the proposed activation functions' performance in image and text classification tasks. For this purpose, several public benchmark datasets are utilized to evaluate and compare the performance of the proposed functions with that of a group of common activation functions. Finally, we deeply analyze the impact of several common activation functions on deep network architectures. Results reveal that the proposed functions outperform most of the popular activation functions in several benchmarks. The statistical study of the overall experiments on both classification categories indicates that the proposed activation functions are robust and superior among all the competitive activation functions in terms of average accuracy.</abstract><cop>Piscataway</cop><pub>IEEE</pub><doi>10.1109/ACCESS.2021.3085855</doi><tpages>23</tpages><orcidid>https://orcid.org/0000-0001-6844-9451</orcidid><orcidid>https://orcid.org/0000-0003-4851-8688</orcidid><orcidid>https://orcid.org/0000-0001-5862-675X</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 2169-3536
ispartof IEEE access, 2021, Vol.9, p.82249-82271
issn 2169-3536
2169-3536
language eng
recordid cdi_ieee_primary_9446115
source IEEE Open Access Journals; DOAJ Directory of Open Access Journals; Elektronische Zeitschriftenbibliothek - Frei zugängliche E-Journals
subjects activation function
Artificial neural network
Artificial neural networks
Benchmarks
Computer architecture
Computer science
Convergence
deep neural network
Functions (mathematics)
Image classification
Impact analysis
Learning
learning challenges
Logistics
Machine learning
Neural networks
Neurons
Polynomials
Slope gradients
Taxonomy
title Improving the Performance of Deep Neural Networks Using Two Proposed Activation Functions
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-23T05%3A13%3A29IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_ieee_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Improving%20the%20Performance%20of%20Deep%20Neural%20Networks%20Using%20Two%20Proposed%20Activation%20Functions&rft.jtitle=IEEE%20access&rft.au=Alkhouly,%20Asmaa%20A.&rft.date=2021&rft.volume=9&rft.spage=82249&rft.epage=82271&rft.pages=82249-82271&rft.issn=2169-3536&rft.eissn=2169-3536&rft.coden=IAECCG&rft_id=info:doi/10.1109/ACCESS.2021.3085855&rft_dat=%3Cproquest_ieee_%3E2541467548%3C/proquest_ieee_%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2541467548&rft_id=info:pmid/&rft_ieee_id=9446115&rft_doaj_id=oai_doaj_org_article_ae2f8b6683294953a519080f06d57016&rfr_iscdi=true