Forward propagation dropout in deep neural networks using Jensen–Shannon and random forest feature importance ranking
Dropout is a mechanism to prevent deep neural networks from overfitting and improving their generalization. Random dropout is the simplest method, where nodes are randomly terminated at each step of the training phase, which may lead to network accuracy reduction. In dynamic dropout, the importance...
Gespeichert in:
Veröffentlicht in: | Neural networks 2023-08, Vol.165, p.238-247 |
---|---|
Hauptverfasser: | , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 247 |
---|---|
container_issue | |
container_start_page | 238 |
container_title | Neural networks |
container_volume | 165 |
creator | Heidari, Mohsen Moattar, Mohammad Hossein Ghaffari, Hamidreza |
description | Dropout is a mechanism to prevent deep neural networks from overfitting and improving their generalization. Random dropout is the simplest method, where nodes are randomly terminated at each step of the training phase, which may lead to network accuracy reduction. In dynamic dropout, the importance of each node and its impact on the network performance is calculated, and the important nodes do not participate in the dropout. But the problem is that the importance of the nodes is not calculated consistently. A node may be considered less important and be dropped in one training epoch and on a batch of data before entering the next epoch, in which it may be an important node. On the other hand, calculating the importance of each unit in every training step is costly. In the proposed method, using random forest and Jensen–Shannon divergence, the importance of each node is calculated once. Then, in the forward propagation steps, the importance of the nodes is propagated and used in the dropout mechanism. This method is evaluated and compared with some previously proposed dropout approaches using two different deep neural network architectures on the MNIST, NorB, CIFAR10, CIFAR100, SVHN, and ImageNet datasets. The results suggest that the proposed method has better accuracy with fewer nodes and better generalizability. Also, the evaluations show that the approach has comparable complexity with other approaches and its convergence time is low as compared with state-of-the-art methods. |
doi_str_mv | 10.1016/j.neunet.2023.05.044 |
format | Article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_miscellaneous_2825500900</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0893608023002885</els_id><sourcerecordid>2825500900</sourcerecordid><originalsourceid>FETCH-LOGICAL-c362t-2415c7f6f9d7f945f2d3254b246989c397a07f25e709b2f1006c5f9f6bb615b33</originalsourceid><addsrcrecordid>eNp9kM1u1DAUhS1ERaeFN0DISzYJN3ZsxxskVLX8qFIXwNpykuvi6cQOdtIRO96hb9gnwaMpLNn43sU5Pvd8hLxuoG6gke-2dcA14FIzYLwGUUPbPiObplO6Yqpjz8kGOs0rCR2ckrOctwAgu5a_IKdccVBSqg3ZX8W0t2mkc4qzvbWLj4GOZY_rQn1ZEWdagpLdlbHsY7rLdM0-3NIvGDKGx98PX3_YEIrNhpGm8sSJupgwL9ShXdaE1E9zTIsNAx4Ed8X9kpw4u8v46mmek-9Xl98uPlXXNx8_X3y4rgYu2VKxthGDctLpUTndCsdGzkTbs1bqTg9cKwvKMYEKdM9cUxoOwmkn-142ouf8nLw9_lv6_VzLTWbyecDdzgaMazasY0IAaIAibY_SIcWcEzozJz_Z9Ms0YA7IzdYckZsDcgPCFOTF9uYpYe0nHP-Z_jIugvdHAZae9x6TyYPHwmL0CYfFjNH_P-EP2qqXWw</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2825500900</pqid></control><display><type>article</type><title>Forward propagation dropout in deep neural networks using Jensen–Shannon and random forest feature importance ranking</title><source>Elsevier ScienceDirect Journals</source><creator>Heidari, Mohsen ; Moattar, Mohammad Hossein ; Ghaffari, Hamidreza</creator><creatorcontrib>Heidari, Mohsen ; Moattar, Mohammad Hossein ; Ghaffari, Hamidreza</creatorcontrib><description>Dropout is a mechanism to prevent deep neural networks from overfitting and improving their generalization. Random dropout is the simplest method, where nodes are randomly terminated at each step of the training phase, which may lead to network accuracy reduction. In dynamic dropout, the importance of each node and its impact on the network performance is calculated, and the important nodes do not participate in the dropout. But the problem is that the importance of the nodes is not calculated consistently. A node may be considered less important and be dropped in one training epoch and on a batch of data before entering the next epoch, in which it may be an important node. On the other hand, calculating the importance of each unit in every training step is costly. In the proposed method, using random forest and Jensen–Shannon divergence, the importance of each node is calculated once. Then, in the forward propagation steps, the importance of the nodes is propagated and used in the dropout mechanism. This method is evaluated and compared with some previously proposed dropout approaches using two different deep neural network architectures on the MNIST, NorB, CIFAR10, CIFAR100, SVHN, and ImageNet datasets. The results suggest that the proposed method has better accuracy with fewer nodes and better generalizability. Also, the evaluations show that the approach has comparable complexity with other approaches and its convergence time is low as compared with state-of-the-art methods.</description><identifier>ISSN: 0893-6080</identifier><identifier>EISSN: 1879-2782</identifier><identifier>DOI: 10.1016/j.neunet.2023.05.044</identifier><identifier>PMID: 37307667</identifier><language>eng</language><publisher>United States: Elsevier Ltd</publisher><subject>Convolutional neural networks ; Deep neural networks ; Dropout ; Jensen–Shannon divergence ; Overfitting ; Random forest</subject><ispartof>Neural networks, 2023-08, Vol.165, p.238-247</ispartof><rights>2023 Elsevier Ltd</rights><rights>Copyright © 2023 Elsevier Ltd. All rights reserved.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c362t-2415c7f6f9d7f945f2d3254b246989c397a07f25e709b2f1006c5f9f6bb615b33</citedby><cites>FETCH-LOGICAL-c362t-2415c7f6f9d7f945f2d3254b246989c397a07f25e709b2f1006c5f9f6bb615b33</cites><orcidid>0000-0002-8968-6744</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.sciencedirect.com/science/article/pii/S0893608023002885$$EHTML$$P50$$Gelsevier$$H</linktohtml><link.rule.ids>314,776,780,3537,27901,27902,65306</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/37307667$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Heidari, Mohsen</creatorcontrib><creatorcontrib>Moattar, Mohammad Hossein</creatorcontrib><creatorcontrib>Ghaffari, Hamidreza</creatorcontrib><title>Forward propagation dropout in deep neural networks using Jensen–Shannon and random forest feature importance ranking</title><title>Neural networks</title><addtitle>Neural Netw</addtitle><description>Dropout is a mechanism to prevent deep neural networks from overfitting and improving their generalization. Random dropout is the simplest method, where nodes are randomly terminated at each step of the training phase, which may lead to network accuracy reduction. In dynamic dropout, the importance of each node and its impact on the network performance is calculated, and the important nodes do not participate in the dropout. But the problem is that the importance of the nodes is not calculated consistently. A node may be considered less important and be dropped in one training epoch and on a batch of data before entering the next epoch, in which it may be an important node. On the other hand, calculating the importance of each unit in every training step is costly. In the proposed method, using random forest and Jensen–Shannon divergence, the importance of each node is calculated once. Then, in the forward propagation steps, the importance of the nodes is propagated and used in the dropout mechanism. This method is evaluated and compared with some previously proposed dropout approaches using two different deep neural network architectures on the MNIST, NorB, CIFAR10, CIFAR100, SVHN, and ImageNet datasets. The results suggest that the proposed method has better accuracy with fewer nodes and better generalizability. Also, the evaluations show that the approach has comparable complexity with other approaches and its convergence time is low as compared with state-of-the-art methods.</description><subject>Convolutional neural networks</subject><subject>Deep neural networks</subject><subject>Dropout</subject><subject>Jensen–Shannon divergence</subject><subject>Overfitting</subject><subject>Random forest</subject><issn>0893-6080</issn><issn>1879-2782</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><recordid>eNp9kM1u1DAUhS1ERaeFN0DISzYJN3ZsxxskVLX8qFIXwNpykuvi6cQOdtIRO96hb9gnwaMpLNn43sU5Pvd8hLxuoG6gke-2dcA14FIzYLwGUUPbPiObplO6Yqpjz8kGOs0rCR2ckrOctwAgu5a_IKdccVBSqg3ZX8W0t2mkc4qzvbWLj4GOZY_rQn1ZEWdagpLdlbHsY7rLdM0-3NIvGDKGx98PX3_YEIrNhpGm8sSJupgwL9ShXdaE1E9zTIsNAx4Ed8X9kpw4u8v46mmek-9Xl98uPlXXNx8_X3y4rgYu2VKxthGDctLpUTndCsdGzkTbs1bqTg9cKwvKMYEKdM9cUxoOwmkn-142ouf8nLw9_lv6_VzLTWbyecDdzgaMazasY0IAaIAibY_SIcWcEzozJz_Z9Ms0YA7IzdYckZsDcgPCFOTF9uYpYe0nHP-Z_jIugvdHAZae9x6TyYPHwmL0CYfFjNH_P-EP2qqXWw</recordid><startdate>202308</startdate><enddate>202308</enddate><creator>Heidari, Mohsen</creator><creator>Moattar, Mohammad Hossein</creator><creator>Ghaffari, Hamidreza</creator><general>Elsevier Ltd</general><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7X8</scope><orcidid>https://orcid.org/0000-0002-8968-6744</orcidid></search><sort><creationdate>202308</creationdate><title>Forward propagation dropout in deep neural networks using Jensen–Shannon and random forest feature importance ranking</title><author>Heidari, Mohsen ; Moattar, Mohammad Hossein ; Ghaffari, Hamidreza</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c362t-2415c7f6f9d7f945f2d3254b246989c397a07f25e709b2f1006c5f9f6bb615b33</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Convolutional neural networks</topic><topic>Deep neural networks</topic><topic>Dropout</topic><topic>Jensen–Shannon divergence</topic><topic>Overfitting</topic><topic>Random forest</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Heidari, Mohsen</creatorcontrib><creatorcontrib>Moattar, Mohammad Hossein</creatorcontrib><creatorcontrib>Ghaffari, Hamidreza</creatorcontrib><collection>PubMed</collection><collection>CrossRef</collection><collection>MEDLINE - Academic</collection><jtitle>Neural networks</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Heidari, Mohsen</au><au>Moattar, Mohammad Hossein</au><au>Ghaffari, Hamidreza</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Forward propagation dropout in deep neural networks using Jensen–Shannon and random forest feature importance ranking</atitle><jtitle>Neural networks</jtitle><addtitle>Neural Netw</addtitle><date>2023-08</date><risdate>2023</risdate><volume>165</volume><spage>238</spage><epage>247</epage><pages>238-247</pages><issn>0893-6080</issn><eissn>1879-2782</eissn><abstract>Dropout is a mechanism to prevent deep neural networks from overfitting and improving their generalization. Random dropout is the simplest method, where nodes are randomly terminated at each step of the training phase, which may lead to network accuracy reduction. In dynamic dropout, the importance of each node and its impact on the network performance is calculated, and the important nodes do not participate in the dropout. But the problem is that the importance of the nodes is not calculated consistently. A node may be considered less important and be dropped in one training epoch and on a batch of data before entering the next epoch, in which it may be an important node. On the other hand, calculating the importance of each unit in every training step is costly. In the proposed method, using random forest and Jensen–Shannon divergence, the importance of each node is calculated once. Then, in the forward propagation steps, the importance of the nodes is propagated and used in the dropout mechanism. This method is evaluated and compared with some previously proposed dropout approaches using two different deep neural network architectures on the MNIST, NorB, CIFAR10, CIFAR100, SVHN, and ImageNet datasets. The results suggest that the proposed method has better accuracy with fewer nodes and better generalizability. Also, the evaluations show that the approach has comparable complexity with other approaches and its convergence time is low as compared with state-of-the-art methods.</abstract><cop>United States</cop><pub>Elsevier Ltd</pub><pmid>37307667</pmid><doi>10.1016/j.neunet.2023.05.044</doi><tpages>10</tpages><orcidid>https://orcid.org/0000-0002-8968-6744</orcidid></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0893-6080 |
ispartof | Neural networks, 2023-08, Vol.165, p.238-247 |
issn | 0893-6080 1879-2782 |
language | eng |
recordid | cdi_proquest_miscellaneous_2825500900 |
source | Elsevier ScienceDirect Journals |
subjects | Convolutional neural networks Deep neural networks Dropout Jensen–Shannon divergence Overfitting Random forest |
title | Forward propagation dropout in deep neural networks using Jensen–Shannon and random forest feature importance ranking |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-09T04%3A47%3A12IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Forward%20propagation%20dropout%20in%20deep%20neural%20networks%20using%20Jensen%E2%80%93Shannon%20and%20random%20forest%20feature%20importance%20ranking&rft.jtitle=Neural%20networks&rft.au=Heidari,%20Mohsen&rft.date=2023-08&rft.volume=165&rft.spage=238&rft.epage=247&rft.pages=238-247&rft.issn=0893-6080&rft.eissn=1879-2782&rft_id=info:doi/10.1016/j.neunet.2023.05.044&rft_dat=%3Cproquest_cross%3E2825500900%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2825500900&rft_id=info:pmid/37307667&rft_els_id=S0893608023002885&rfr_iscdi=true |