A fine-tuning deep learning with multi-objective-based feature selection approach for the classification of text
Document classification is becoming increasingly essential for the vast number of documents available in digital libraries, emails, the Internet, etc. Textual records frequently contain non-discriminative (noisy and irrelevant) terms that are also high-dimensional, resulting in higher computing cost...
Gespeichert in:
Veröffentlicht in: | Neural computing & applications 2024-03, Vol.36 (7), p.3525-3553 |
---|---|
Hauptverfasser: | , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 3553 |
---|---|
container_issue | 7 |
container_start_page | 3525 |
container_title | Neural computing & applications |
container_volume | 36 |
creator | Dhal, Pradip Azad, Chandrashekhar |
description | Document classification is becoming increasingly essential for the vast number of documents available in digital libraries, emails, the Internet, etc. Textual records frequently contain non-discriminative (noisy and irrelevant) terms that are also high-dimensional, resulting in higher computing costs and poorer learning performance in Text Classification (TC). Feature selection (FS), which tries to discover discriminate terms or features from the textual data, is one of the most effective tasks for this issue. This paper introduces a novel multi-stage term-weighting scheme-based FS model designed for the single-label TC system to obtain the optimal set of features. We have also developed a hybrid deep learning fine-tuning network based on Bidirectional Long Short-Term Memory (BiLSTM) and Convolutional Neural Network (CNN) for the classification stage. The FS approach is worked on two-stage criteria. The filter model is used in the first stage, and the multi-objective wrapper model, an upgraded version of the Whale Optimization Algorithm (WOA) with Particle Swarm Optimization (PSO), is used in the second stage. The objective function in the above wrapper model is based on a tri-objective principle. It uses the Pareto front technique to discover the optimal set of features. Here in the wrapper model, a novel selection strategy has been introduced to select the whale instead of the random whale. The proposed work is evaluated on four popular benchmark text corpora, of which two are binary class, and two are multi-class. The suggested FS technique is compared against classic Machine Learning (ML) and deep learning classifiers. The results of the experiments reveal that the recommended FS technique is more effective in obtaining better results than the other results. |
doi_str_mv | 10.1007/s00521-023-09225-1 |
format | Article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2924060797</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2924060797</sourcerecordid><originalsourceid>FETCH-LOGICAL-c319t-f22e5957e26930a19a011a5d8f40b07a93154755122f447879b969d4c413d6823</originalsourceid><addsrcrecordid>eNp9kMtOwzAQRS0EEqXwA6wssTaMX0m8rCpeUiU2sLbcZNymSpNgOzz-nrRBYsdqNHPvnRkdQq453HKA_C4CaMEZCMnACKEZPyEzrqRkEnRxSmZg1ChnSp6Tixh3AKCyQs9Iv6C-bpGloa3bDa0Qe9qgC8fus05buh-aVLNuvcMy1R_I1i5iRT26NASkEZvDvGup6_vQuXJLfRdo2iItGxdj7evSHfXO04Rf6ZKceddEvPqtc_L2cP-6fGKrl8fn5WLFSslNYl4I1EbnKDIjwXHjgHOnq8IrWEPujORa5VpzIbxSeZGbtclMpUrFZZUVQs7JzbR3_Op9wJjsrhtCO560wggFGeQmH11icpWhizGgt32o9y58Ww72QNZOZO1I1h7JWj6G5BSKo7ndYPhb_U_qB8hIe14</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2924060797</pqid></control><display><type>article</type><title>A fine-tuning deep learning with multi-objective-based feature selection approach for the classification of text</title><source>Springer Nature - Complete Springer Journals</source><creator>Dhal, Pradip ; Azad, Chandrashekhar</creator><creatorcontrib>Dhal, Pradip ; Azad, Chandrashekhar</creatorcontrib><description>Document classification is becoming increasingly essential for the vast number of documents available in digital libraries, emails, the Internet, etc. Textual records frequently contain non-discriminative (noisy and irrelevant) terms that are also high-dimensional, resulting in higher computing costs and poorer learning performance in Text Classification (TC). Feature selection (FS), which tries to discover discriminate terms or features from the textual data, is one of the most effective tasks for this issue. This paper introduces a novel multi-stage term-weighting scheme-based FS model designed for the single-label TC system to obtain the optimal set of features. We have also developed a hybrid deep learning fine-tuning network based on Bidirectional Long Short-Term Memory (BiLSTM) and Convolutional Neural Network (CNN) for the classification stage. The FS approach is worked on two-stage criteria. The filter model is used in the first stage, and the multi-objective wrapper model, an upgraded version of the Whale Optimization Algorithm (WOA) with Particle Swarm Optimization (PSO), is used in the second stage. The objective function in the above wrapper model is based on a tri-objective principle. It uses the Pareto front technique to discover the optimal set of features. Here in the wrapper model, a novel selection strategy has been introduced to select the whale instead of the random whale. The proposed work is evaluated on four popular benchmark text corpora, of which two are binary class, and two are multi-class. The suggested FS technique is compared against classic Machine Learning (ML) and deep learning classifiers. The results of the experiments reveal that the recommended FS technique is more effective in obtaining better results than the other results.</description><identifier>ISSN: 0941-0643</identifier><identifier>EISSN: 1433-3058</identifier><identifier>DOI: 10.1007/s00521-023-09225-1</identifier><language>eng</language><publisher>London: Springer London</publisher><subject>Algorithms ; Artificial Intelligence ; Artificial neural networks ; Classification ; Computational Biology/Bioinformatics ; Computational Science and Engineering ; Computer Science ; Data Mining and Knowledge Discovery ; Deep learning ; Documents ; Feature selection ; Image Processing and Computer Vision ; Machine learning ; Multiple objective analysis ; Original Article ; Particle swarm optimization ; Probability and Statistics in Computer Science</subject><ispartof>Neural computing & applications, 2024-03, Vol.36 (7), p.3525-3553</ispartof><rights>The Author(s), under exclusive licence to Springer-Verlag London Ltd., part of Springer Nature 2023. Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c319t-f22e5957e26930a19a011a5d8f40b07a93154755122f447879b969d4c413d6823</citedby><cites>FETCH-LOGICAL-c319t-f22e5957e26930a19a011a5d8f40b07a93154755122f447879b969d4c413d6823</cites><orcidid>0000-0003-2097-6601</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://link.springer.com/content/pdf/10.1007/s00521-023-09225-1$$EPDF$$P50$$Gspringer$$H</linktopdf><linktohtml>$$Uhttps://link.springer.com/10.1007/s00521-023-09225-1$$EHTML$$P50$$Gspringer$$H</linktohtml><link.rule.ids>314,776,780,27903,27904,41467,42536,51297</link.rule.ids></links><search><creatorcontrib>Dhal, Pradip</creatorcontrib><creatorcontrib>Azad, Chandrashekhar</creatorcontrib><title>A fine-tuning deep learning with multi-objective-based feature selection approach for the classification of text</title><title>Neural computing & applications</title><addtitle>Neural Comput & Applic</addtitle><description>Document classification is becoming increasingly essential for the vast number of documents available in digital libraries, emails, the Internet, etc. Textual records frequently contain non-discriminative (noisy and irrelevant) terms that are also high-dimensional, resulting in higher computing costs and poorer learning performance in Text Classification (TC). Feature selection (FS), which tries to discover discriminate terms or features from the textual data, is one of the most effective tasks for this issue. This paper introduces a novel multi-stage term-weighting scheme-based FS model designed for the single-label TC system to obtain the optimal set of features. We have also developed a hybrid deep learning fine-tuning network based on Bidirectional Long Short-Term Memory (BiLSTM) and Convolutional Neural Network (CNN) for the classification stage. The FS approach is worked on two-stage criteria. The filter model is used in the first stage, and the multi-objective wrapper model, an upgraded version of the Whale Optimization Algorithm (WOA) with Particle Swarm Optimization (PSO), is used in the second stage. The objective function in the above wrapper model is based on a tri-objective principle. It uses the Pareto front technique to discover the optimal set of features. Here in the wrapper model, a novel selection strategy has been introduced to select the whale instead of the random whale. The proposed work is evaluated on four popular benchmark text corpora, of which two are binary class, and two are multi-class. The suggested FS technique is compared against classic Machine Learning (ML) and deep learning classifiers. The results of the experiments reveal that the recommended FS technique is more effective in obtaining better results than the other results.</description><subject>Algorithms</subject><subject>Artificial Intelligence</subject><subject>Artificial neural networks</subject><subject>Classification</subject><subject>Computational Biology/Bioinformatics</subject><subject>Computational Science and Engineering</subject><subject>Computer Science</subject><subject>Data Mining and Knowledge Discovery</subject><subject>Deep learning</subject><subject>Documents</subject><subject>Feature selection</subject><subject>Image Processing and Computer Vision</subject><subject>Machine learning</subject><subject>Multiple objective analysis</subject><subject>Original Article</subject><subject>Particle swarm optimization</subject><subject>Probability and Statistics in Computer Science</subject><issn>0941-0643</issn><issn>1433-3058</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><recordid>eNp9kMtOwzAQRS0EEqXwA6wssTaMX0m8rCpeUiU2sLbcZNymSpNgOzz-nrRBYsdqNHPvnRkdQq453HKA_C4CaMEZCMnACKEZPyEzrqRkEnRxSmZg1ChnSp6Tixh3AKCyQs9Iv6C-bpGloa3bDa0Qe9qgC8fus05buh-aVLNuvcMy1R_I1i5iRT26NASkEZvDvGup6_vQuXJLfRdo2iItGxdj7evSHfXO04Rf6ZKceddEvPqtc_L2cP-6fGKrl8fn5WLFSslNYl4I1EbnKDIjwXHjgHOnq8IrWEPujORa5VpzIbxSeZGbtclMpUrFZZUVQs7JzbR3_Op9wJjsrhtCO560wggFGeQmH11icpWhizGgt32o9y58Ww72QNZOZO1I1h7JWj6G5BSKo7ndYPhb_U_qB8hIe14</recordid><startdate>20240301</startdate><enddate>20240301</enddate><creator>Dhal, Pradip</creator><creator>Azad, Chandrashekhar</creator><general>Springer London</general><general>Springer Nature B.V</general><scope>AAYXX</scope><scope>CITATION</scope><orcidid>https://orcid.org/0000-0003-2097-6601</orcidid></search><sort><creationdate>20240301</creationdate><title>A fine-tuning deep learning with multi-objective-based feature selection approach for the classification of text</title><author>Dhal, Pradip ; Azad, Chandrashekhar</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c319t-f22e5957e26930a19a011a5d8f40b07a93154755122f447879b969d4c413d6823</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Algorithms</topic><topic>Artificial Intelligence</topic><topic>Artificial neural networks</topic><topic>Classification</topic><topic>Computational Biology/Bioinformatics</topic><topic>Computational Science and Engineering</topic><topic>Computer Science</topic><topic>Data Mining and Knowledge Discovery</topic><topic>Deep learning</topic><topic>Documents</topic><topic>Feature selection</topic><topic>Image Processing and Computer Vision</topic><topic>Machine learning</topic><topic>Multiple objective analysis</topic><topic>Original Article</topic><topic>Particle swarm optimization</topic><topic>Probability and Statistics in Computer Science</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Dhal, Pradip</creatorcontrib><creatorcontrib>Azad, Chandrashekhar</creatorcontrib><collection>CrossRef</collection><jtitle>Neural computing & applications</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Dhal, Pradip</au><au>Azad, Chandrashekhar</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>A fine-tuning deep learning with multi-objective-based feature selection approach for the classification of text</atitle><jtitle>Neural computing & applications</jtitle><stitle>Neural Comput & Applic</stitle><date>2024-03-01</date><risdate>2024</risdate><volume>36</volume><issue>7</issue><spage>3525</spage><epage>3553</epage><pages>3525-3553</pages><issn>0941-0643</issn><eissn>1433-3058</eissn><abstract>Document classification is becoming increasingly essential for the vast number of documents available in digital libraries, emails, the Internet, etc. Textual records frequently contain non-discriminative (noisy and irrelevant) terms that are also high-dimensional, resulting in higher computing costs and poorer learning performance in Text Classification (TC). Feature selection (FS), which tries to discover discriminate terms or features from the textual data, is one of the most effective tasks for this issue. This paper introduces a novel multi-stage term-weighting scheme-based FS model designed for the single-label TC system to obtain the optimal set of features. We have also developed a hybrid deep learning fine-tuning network based on Bidirectional Long Short-Term Memory (BiLSTM) and Convolutional Neural Network (CNN) for the classification stage. The FS approach is worked on two-stage criteria. The filter model is used in the first stage, and the multi-objective wrapper model, an upgraded version of the Whale Optimization Algorithm (WOA) with Particle Swarm Optimization (PSO), is used in the second stage. The objective function in the above wrapper model is based on a tri-objective principle. It uses the Pareto front technique to discover the optimal set of features. Here in the wrapper model, a novel selection strategy has been introduced to select the whale instead of the random whale. The proposed work is evaluated on four popular benchmark text corpora, of which two are binary class, and two are multi-class. The suggested FS technique is compared against classic Machine Learning (ML) and deep learning classifiers. The results of the experiments reveal that the recommended FS technique is more effective in obtaining better results than the other results.</abstract><cop>London</cop><pub>Springer London</pub><doi>10.1007/s00521-023-09225-1</doi><tpages>29</tpages><orcidid>https://orcid.org/0000-0003-2097-6601</orcidid></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0941-0643 |
ispartof | Neural computing & applications, 2024-03, Vol.36 (7), p.3525-3553 |
issn | 0941-0643 1433-3058 |
language | eng |
recordid | cdi_proquest_journals_2924060797 |
source | Springer Nature - Complete Springer Journals |
subjects | Algorithms Artificial Intelligence Artificial neural networks Classification Computational Biology/Bioinformatics Computational Science and Engineering Computer Science Data Mining and Knowledge Discovery Deep learning Documents Feature selection Image Processing and Computer Vision Machine learning Multiple objective analysis Original Article Particle swarm optimization Probability and Statistics in Computer Science |
title | A fine-tuning deep learning with multi-objective-based feature selection approach for the classification of text |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-27T02%3A50%3A29IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=A%20fine-tuning%20deep%20learning%20with%20multi-objective-based%20feature%20selection%20approach%20for%20the%20classification%20of%20text&rft.jtitle=Neural%20computing%20&%20applications&rft.au=Dhal,%20Pradip&rft.date=2024-03-01&rft.volume=36&rft.issue=7&rft.spage=3525&rft.epage=3553&rft.pages=3525-3553&rft.issn=0941-0643&rft.eissn=1433-3058&rft_id=info:doi/10.1007/s00521-023-09225-1&rft_dat=%3Cproquest_cross%3E2924060797%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2924060797&rft_id=info:pmid/&rfr_iscdi=true |