Forward propagation dropout in deep neural networks using Jensen–Shannon and random forest feature importance ranking

Dropout is a mechanism to prevent deep neural networks from overfitting and improving their generalization. Random dropout is the simplest method, where nodes are randomly terminated at each step of the training phase, which may lead to network accuracy reduction. In dynamic dropout, the importance...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Neural networks 2023-08, Vol.165, p.238-247
Hauptverfasser: Heidari, Mohsen, Moattar, Mohammad Hossein, Ghaffari, Hamidreza
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 247
container_issue
container_start_page 238
container_title Neural networks
container_volume 165
creator Heidari, Mohsen
Moattar, Mohammad Hossein
Ghaffari, Hamidreza
description Dropout is a mechanism to prevent deep neural networks from overfitting and improving their generalization. Random dropout is the simplest method, where nodes are randomly terminated at each step of the training phase, which may lead to network accuracy reduction. In dynamic dropout, the importance of each node and its impact on the network performance is calculated, and the important nodes do not participate in the dropout. But the problem is that the importance of the nodes is not calculated consistently. A node may be considered less important and be dropped in one training epoch and on a batch of data before entering the next epoch, in which it may be an important node. On the other hand, calculating the importance of each unit in every training step is costly. In the proposed method, using random forest and Jensen–Shannon divergence, the importance of each node is calculated once. Then, in the forward propagation steps, the importance of the nodes is propagated and used in the dropout mechanism. This method is evaluated and compared with some previously proposed dropout approaches using two different deep neural network architectures on the MNIST, NorB, CIFAR10, CIFAR100, SVHN, and ImageNet datasets. The results suggest that the proposed method has better accuracy with fewer nodes and better generalizability. Also, the evaluations show that the approach has comparable complexity with other approaches and its convergence time is low as compared with state-of-the-art methods.
doi_str_mv 10.1016/j.neunet.2023.05.044
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_miscellaneous_2825500900</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0893608023002885</els_id><sourcerecordid>2825500900</sourcerecordid><originalsourceid>FETCH-LOGICAL-c362t-2415c7f6f9d7f945f2d3254b246989c397a07f25e709b2f1006c5f9f6bb615b33</originalsourceid><addsrcrecordid>eNp9kM1u1DAUhS1ERaeFN0DISzYJN3ZsxxskVLX8qFIXwNpykuvi6cQOdtIRO96hb9gnwaMpLNn43sU5Pvd8hLxuoG6gke-2dcA14FIzYLwGUUPbPiObplO6Yqpjz8kGOs0rCR2ckrOctwAgu5a_IKdccVBSqg3ZX8W0t2mkc4qzvbWLj4GOZY_rQn1ZEWdagpLdlbHsY7rLdM0-3NIvGDKGx98PX3_YEIrNhpGm8sSJupgwL9ShXdaE1E9zTIsNAx4Ed8X9kpw4u8v46mmek-9Xl98uPlXXNx8_X3y4rgYu2VKxthGDctLpUTndCsdGzkTbs1bqTg9cKwvKMYEKdM9cUxoOwmkn-142ouf8nLw9_lv6_VzLTWbyecDdzgaMazasY0IAaIAibY_SIcWcEzozJz_Z9Ms0YA7IzdYckZsDcgPCFOTF9uYpYe0nHP-Z_jIugvdHAZae9x6TyYPHwmL0CYfFjNH_P-EP2qqXWw</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2825500900</pqid></control><display><type>article</type><title>Forward propagation dropout in deep neural networks using Jensen–Shannon and random forest feature importance ranking</title><source>Elsevier ScienceDirect Journals</source><creator>Heidari, Mohsen ; Moattar, Mohammad Hossein ; Ghaffari, Hamidreza</creator><creatorcontrib>Heidari, Mohsen ; Moattar, Mohammad Hossein ; Ghaffari, Hamidreza</creatorcontrib><description>Dropout is a mechanism to prevent deep neural networks from overfitting and improving their generalization. Random dropout is the simplest method, where nodes are randomly terminated at each step of the training phase, which may lead to network accuracy reduction. In dynamic dropout, the importance of each node and its impact on the network performance is calculated, and the important nodes do not participate in the dropout. But the problem is that the importance of the nodes is not calculated consistently. A node may be considered less important and be dropped in one training epoch and on a batch of data before entering the next epoch, in which it may be an important node. On the other hand, calculating the importance of each unit in every training step is costly. In the proposed method, using random forest and Jensen–Shannon divergence, the importance of each node is calculated once. Then, in the forward propagation steps, the importance of the nodes is propagated and used in the dropout mechanism. This method is evaluated and compared with some previously proposed dropout approaches using two different deep neural network architectures on the MNIST, NorB, CIFAR10, CIFAR100, SVHN, and ImageNet datasets. The results suggest that the proposed method has better accuracy with fewer nodes and better generalizability. Also, the evaluations show that the approach has comparable complexity with other approaches and its convergence time is low as compared with state-of-the-art methods.</description><identifier>ISSN: 0893-6080</identifier><identifier>EISSN: 1879-2782</identifier><identifier>DOI: 10.1016/j.neunet.2023.05.044</identifier><identifier>PMID: 37307667</identifier><language>eng</language><publisher>United States: Elsevier Ltd</publisher><subject>Convolutional neural networks ; Deep neural networks ; Dropout ; Jensen–Shannon divergence ; Overfitting ; Random forest</subject><ispartof>Neural networks, 2023-08, Vol.165, p.238-247</ispartof><rights>2023 Elsevier Ltd</rights><rights>Copyright © 2023 Elsevier Ltd. All rights reserved.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c362t-2415c7f6f9d7f945f2d3254b246989c397a07f25e709b2f1006c5f9f6bb615b33</citedby><cites>FETCH-LOGICAL-c362t-2415c7f6f9d7f945f2d3254b246989c397a07f25e709b2f1006c5f9f6bb615b33</cites><orcidid>0000-0002-8968-6744</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.sciencedirect.com/science/article/pii/S0893608023002885$$EHTML$$P50$$Gelsevier$$H</linktohtml><link.rule.ids>314,776,780,3537,27901,27902,65306</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/37307667$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Heidari, Mohsen</creatorcontrib><creatorcontrib>Moattar, Mohammad Hossein</creatorcontrib><creatorcontrib>Ghaffari, Hamidreza</creatorcontrib><title>Forward propagation dropout in deep neural networks using Jensen–Shannon and random forest feature importance ranking</title><title>Neural networks</title><addtitle>Neural Netw</addtitle><description>Dropout is a mechanism to prevent deep neural networks from overfitting and improving their generalization. Random dropout is the simplest method, where nodes are randomly terminated at each step of the training phase, which may lead to network accuracy reduction. In dynamic dropout, the importance of each node and its impact on the network performance is calculated, and the important nodes do not participate in the dropout. But the problem is that the importance of the nodes is not calculated consistently. A node may be considered less important and be dropped in one training epoch and on a batch of data before entering the next epoch, in which it may be an important node. On the other hand, calculating the importance of each unit in every training step is costly. In the proposed method, using random forest and Jensen–Shannon divergence, the importance of each node is calculated once. Then, in the forward propagation steps, the importance of the nodes is propagated and used in the dropout mechanism. This method is evaluated and compared with some previously proposed dropout approaches using two different deep neural network architectures on the MNIST, NorB, CIFAR10, CIFAR100, SVHN, and ImageNet datasets. The results suggest that the proposed method has better accuracy with fewer nodes and better generalizability. Also, the evaluations show that the approach has comparable complexity with other approaches and its convergence time is low as compared with state-of-the-art methods.</description><subject>Convolutional neural networks</subject><subject>Deep neural networks</subject><subject>Dropout</subject><subject>Jensen–Shannon divergence</subject><subject>Overfitting</subject><subject>Random forest</subject><issn>0893-6080</issn><issn>1879-2782</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><recordid>eNp9kM1u1DAUhS1ERaeFN0DISzYJN3ZsxxskVLX8qFIXwNpykuvi6cQOdtIRO96hb9gnwaMpLNn43sU5Pvd8hLxuoG6gke-2dcA14FIzYLwGUUPbPiObplO6Yqpjz8kGOs0rCR2ckrOctwAgu5a_IKdccVBSqg3ZX8W0t2mkc4qzvbWLj4GOZY_rQn1ZEWdagpLdlbHsY7rLdM0-3NIvGDKGx98PX3_YEIrNhpGm8sSJupgwL9ShXdaE1E9zTIsNAx4Ed8X9kpw4u8v46mmek-9Xl98uPlXXNx8_X3y4rgYu2VKxthGDctLpUTndCsdGzkTbs1bqTg9cKwvKMYEKdM9cUxoOwmkn-142ouf8nLw9_lv6_VzLTWbyecDdzgaMazasY0IAaIAibY_SIcWcEzozJz_Z9Ms0YA7IzdYckZsDcgPCFOTF9uYpYe0nHP-Z_jIugvdHAZae9x6TyYPHwmL0CYfFjNH_P-EP2qqXWw</recordid><startdate>202308</startdate><enddate>202308</enddate><creator>Heidari, Mohsen</creator><creator>Moattar, Mohammad Hossein</creator><creator>Ghaffari, Hamidreza</creator><general>Elsevier Ltd</general><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7X8</scope><orcidid>https://orcid.org/0000-0002-8968-6744</orcidid></search><sort><creationdate>202308</creationdate><title>Forward propagation dropout in deep neural networks using Jensen–Shannon and random forest feature importance ranking</title><author>Heidari, Mohsen ; Moattar, Mohammad Hossein ; Ghaffari, Hamidreza</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c362t-2415c7f6f9d7f945f2d3254b246989c397a07f25e709b2f1006c5f9f6bb615b33</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Convolutional neural networks</topic><topic>Deep neural networks</topic><topic>Dropout</topic><topic>Jensen–Shannon divergence</topic><topic>Overfitting</topic><topic>Random forest</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Heidari, Mohsen</creatorcontrib><creatorcontrib>Moattar, Mohammad Hossein</creatorcontrib><creatorcontrib>Ghaffari, Hamidreza</creatorcontrib><collection>PubMed</collection><collection>CrossRef</collection><collection>MEDLINE - Academic</collection><jtitle>Neural networks</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Heidari, Mohsen</au><au>Moattar, Mohammad Hossein</au><au>Ghaffari, Hamidreza</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Forward propagation dropout in deep neural networks using Jensen–Shannon and random forest feature importance ranking</atitle><jtitle>Neural networks</jtitle><addtitle>Neural Netw</addtitle><date>2023-08</date><risdate>2023</risdate><volume>165</volume><spage>238</spage><epage>247</epage><pages>238-247</pages><issn>0893-6080</issn><eissn>1879-2782</eissn><abstract>Dropout is a mechanism to prevent deep neural networks from overfitting and improving their generalization. Random dropout is the simplest method, where nodes are randomly terminated at each step of the training phase, which may lead to network accuracy reduction. In dynamic dropout, the importance of each node and its impact on the network performance is calculated, and the important nodes do not participate in the dropout. But the problem is that the importance of the nodes is not calculated consistently. A node may be considered less important and be dropped in one training epoch and on a batch of data before entering the next epoch, in which it may be an important node. On the other hand, calculating the importance of each unit in every training step is costly. In the proposed method, using random forest and Jensen–Shannon divergence, the importance of each node is calculated once. Then, in the forward propagation steps, the importance of the nodes is propagated and used in the dropout mechanism. This method is evaluated and compared with some previously proposed dropout approaches using two different deep neural network architectures on the MNIST, NorB, CIFAR10, CIFAR100, SVHN, and ImageNet datasets. The results suggest that the proposed method has better accuracy with fewer nodes and better generalizability. Also, the evaluations show that the approach has comparable complexity with other approaches and its convergence time is low as compared with state-of-the-art methods.</abstract><cop>United States</cop><pub>Elsevier Ltd</pub><pmid>37307667</pmid><doi>10.1016/j.neunet.2023.05.044</doi><tpages>10</tpages><orcidid>https://orcid.org/0000-0002-8968-6744</orcidid></addata></record>
fulltext fulltext
identifier ISSN: 0893-6080
ispartof Neural networks, 2023-08, Vol.165, p.238-247
issn 0893-6080
1879-2782
language eng
recordid cdi_proquest_miscellaneous_2825500900
source Elsevier ScienceDirect Journals
subjects Convolutional neural networks
Deep neural networks
Dropout
Jensen–Shannon divergence
Overfitting
Random forest
title Forward propagation dropout in deep neural networks using Jensen–Shannon and random forest feature importance ranking
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-09T04%3A47%3A12IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Forward%20propagation%20dropout%20in%20deep%20neural%20networks%20using%20Jensen%E2%80%93Shannon%20and%20random%20forest%20feature%20importance%20ranking&rft.jtitle=Neural%20networks&rft.au=Heidari,%20Mohsen&rft.date=2023-08&rft.volume=165&rft.spage=238&rft.epage=247&rft.pages=238-247&rft.issn=0893-6080&rft.eissn=1879-2782&rft_id=info:doi/10.1016/j.neunet.2023.05.044&rft_dat=%3Cproquest_cross%3E2825500900%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2825500900&rft_id=info:pmid/37307667&rft_els_id=S0893608023002885&rfr_iscdi=true