A fine-tuning deep learning with multi-objective-based feature selection approach for the classification of text

Document classification is becoming increasingly essential for the vast number of documents available in digital libraries, emails, the Internet, etc. Textual records frequently contain non-discriminative (noisy and irrelevant) terms that are also high-dimensional, resulting in higher computing cost...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Neural computing & applications 2024-03, Vol.36 (7), p.3525-3553
Hauptverfasser: Dhal, Pradip, Azad, Chandrashekhar
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 3553
container_issue 7
container_start_page 3525
container_title Neural computing & applications
container_volume 36
creator Dhal, Pradip
Azad, Chandrashekhar
description Document classification is becoming increasingly essential for the vast number of documents available in digital libraries, emails, the Internet, etc. Textual records frequently contain non-discriminative (noisy and irrelevant) terms that are also high-dimensional, resulting in higher computing costs and poorer learning performance in Text Classification (TC). Feature selection (FS), which tries to discover discriminate terms or features from the textual data, is one of the most effective tasks for this issue. This paper introduces a novel multi-stage term-weighting scheme-based FS model designed for the single-label TC system to obtain the optimal set of features. We have also developed a hybrid deep learning fine-tuning network based on Bidirectional Long Short-Term Memory (BiLSTM) and Convolutional Neural Network (CNN) for the classification stage. The FS approach is worked on two-stage criteria. The filter model is used in the first stage, and the multi-objective wrapper model, an upgraded version of the Whale Optimization Algorithm (WOA) with Particle Swarm Optimization (PSO), is used in the second stage. The objective function in the above wrapper model is based on a tri-objective principle. It uses the Pareto front technique to discover the optimal set of features. Here in the wrapper model, a novel selection strategy has been introduced to select the whale instead of the random whale. The proposed work is evaluated on four popular benchmark text corpora, of which two are binary class, and two are multi-class. The suggested FS technique is compared against classic Machine Learning (ML) and deep learning classifiers. The results of the experiments reveal that the recommended FS technique is more effective in obtaining better results than the other results.
doi_str_mv 10.1007/s00521-023-09225-1
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2924060797</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2924060797</sourcerecordid><originalsourceid>FETCH-LOGICAL-c319t-f22e5957e26930a19a011a5d8f40b07a93154755122f447879b969d4c413d6823</originalsourceid><addsrcrecordid>eNp9kMtOwzAQRS0EEqXwA6wssTaMX0m8rCpeUiU2sLbcZNymSpNgOzz-nrRBYsdqNHPvnRkdQq453HKA_C4CaMEZCMnACKEZPyEzrqRkEnRxSmZg1ChnSp6Tixh3AKCyQs9Iv6C-bpGloa3bDa0Qe9qgC8fus05buh-aVLNuvcMy1R_I1i5iRT26NASkEZvDvGup6_vQuXJLfRdo2iItGxdj7evSHfXO04Rf6ZKceddEvPqtc_L2cP-6fGKrl8fn5WLFSslNYl4I1EbnKDIjwXHjgHOnq8IrWEPujORa5VpzIbxSeZGbtclMpUrFZZUVQs7JzbR3_Op9wJjsrhtCO560wggFGeQmH11icpWhizGgt32o9y58Ww72QNZOZO1I1h7JWj6G5BSKo7ndYPhb_U_qB8hIe14</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2924060797</pqid></control><display><type>article</type><title>A fine-tuning deep learning with multi-objective-based feature selection approach for the classification of text</title><source>Springer Nature - Complete Springer Journals</source><creator>Dhal, Pradip ; Azad, Chandrashekhar</creator><creatorcontrib>Dhal, Pradip ; Azad, Chandrashekhar</creatorcontrib><description>Document classification is becoming increasingly essential for the vast number of documents available in digital libraries, emails, the Internet, etc. Textual records frequently contain non-discriminative (noisy and irrelevant) terms that are also high-dimensional, resulting in higher computing costs and poorer learning performance in Text Classification (TC). Feature selection (FS), which tries to discover discriminate terms or features from the textual data, is one of the most effective tasks for this issue. This paper introduces a novel multi-stage term-weighting scheme-based FS model designed for the single-label TC system to obtain the optimal set of features. We have also developed a hybrid deep learning fine-tuning network based on Bidirectional Long Short-Term Memory (BiLSTM) and Convolutional Neural Network (CNN) for the classification stage. The FS approach is worked on two-stage criteria. The filter model is used in the first stage, and the multi-objective wrapper model, an upgraded version of the Whale Optimization Algorithm (WOA) with Particle Swarm Optimization (PSO), is used in the second stage. The objective function in the above wrapper model is based on a tri-objective principle. It uses the Pareto front technique to discover the optimal set of features. Here in the wrapper model, a novel selection strategy has been introduced to select the whale instead of the random whale. The proposed work is evaluated on four popular benchmark text corpora, of which two are binary class, and two are multi-class. The suggested FS technique is compared against classic Machine Learning (ML) and deep learning classifiers. The results of the experiments reveal that the recommended FS technique is more effective in obtaining better results than the other results.</description><identifier>ISSN: 0941-0643</identifier><identifier>EISSN: 1433-3058</identifier><identifier>DOI: 10.1007/s00521-023-09225-1</identifier><language>eng</language><publisher>London: Springer London</publisher><subject>Algorithms ; Artificial Intelligence ; Artificial neural networks ; Classification ; Computational Biology/Bioinformatics ; Computational Science and Engineering ; Computer Science ; Data Mining and Knowledge Discovery ; Deep learning ; Documents ; Feature selection ; Image Processing and Computer Vision ; Machine learning ; Multiple objective analysis ; Original Article ; Particle swarm optimization ; Probability and Statistics in Computer Science</subject><ispartof>Neural computing &amp; applications, 2024-03, Vol.36 (7), p.3525-3553</ispartof><rights>The Author(s), under exclusive licence to Springer-Verlag London Ltd., part of Springer Nature 2023. Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c319t-f22e5957e26930a19a011a5d8f40b07a93154755122f447879b969d4c413d6823</citedby><cites>FETCH-LOGICAL-c319t-f22e5957e26930a19a011a5d8f40b07a93154755122f447879b969d4c413d6823</cites><orcidid>0000-0003-2097-6601</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://link.springer.com/content/pdf/10.1007/s00521-023-09225-1$$EPDF$$P50$$Gspringer$$H</linktopdf><linktohtml>$$Uhttps://link.springer.com/10.1007/s00521-023-09225-1$$EHTML$$P50$$Gspringer$$H</linktohtml><link.rule.ids>314,776,780,27903,27904,41467,42536,51297</link.rule.ids></links><search><creatorcontrib>Dhal, Pradip</creatorcontrib><creatorcontrib>Azad, Chandrashekhar</creatorcontrib><title>A fine-tuning deep learning with multi-objective-based feature selection approach for the classification of text</title><title>Neural computing &amp; applications</title><addtitle>Neural Comput &amp; Applic</addtitle><description>Document classification is becoming increasingly essential for the vast number of documents available in digital libraries, emails, the Internet, etc. Textual records frequently contain non-discriminative (noisy and irrelevant) terms that are also high-dimensional, resulting in higher computing costs and poorer learning performance in Text Classification (TC). Feature selection (FS), which tries to discover discriminate terms or features from the textual data, is one of the most effective tasks for this issue. This paper introduces a novel multi-stage term-weighting scheme-based FS model designed for the single-label TC system to obtain the optimal set of features. We have also developed a hybrid deep learning fine-tuning network based on Bidirectional Long Short-Term Memory (BiLSTM) and Convolutional Neural Network (CNN) for the classification stage. The FS approach is worked on two-stage criteria. The filter model is used in the first stage, and the multi-objective wrapper model, an upgraded version of the Whale Optimization Algorithm (WOA) with Particle Swarm Optimization (PSO), is used in the second stage. The objective function in the above wrapper model is based on a tri-objective principle. It uses the Pareto front technique to discover the optimal set of features. Here in the wrapper model, a novel selection strategy has been introduced to select the whale instead of the random whale. The proposed work is evaluated on four popular benchmark text corpora, of which two are binary class, and two are multi-class. The suggested FS technique is compared against classic Machine Learning (ML) and deep learning classifiers. The results of the experiments reveal that the recommended FS technique is more effective in obtaining better results than the other results.</description><subject>Algorithms</subject><subject>Artificial Intelligence</subject><subject>Artificial neural networks</subject><subject>Classification</subject><subject>Computational Biology/Bioinformatics</subject><subject>Computational Science and Engineering</subject><subject>Computer Science</subject><subject>Data Mining and Knowledge Discovery</subject><subject>Deep learning</subject><subject>Documents</subject><subject>Feature selection</subject><subject>Image Processing and Computer Vision</subject><subject>Machine learning</subject><subject>Multiple objective analysis</subject><subject>Original Article</subject><subject>Particle swarm optimization</subject><subject>Probability and Statistics in Computer Science</subject><issn>0941-0643</issn><issn>1433-3058</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><recordid>eNp9kMtOwzAQRS0EEqXwA6wssTaMX0m8rCpeUiU2sLbcZNymSpNgOzz-nrRBYsdqNHPvnRkdQq453HKA_C4CaMEZCMnACKEZPyEzrqRkEnRxSmZg1ChnSp6Tixh3AKCyQs9Iv6C-bpGloa3bDa0Qe9qgC8fus05buh-aVLNuvcMy1R_I1i5iRT26NASkEZvDvGup6_vQuXJLfRdo2iItGxdj7evSHfXO04Rf6ZKceddEvPqtc_L2cP-6fGKrl8fn5WLFSslNYl4I1EbnKDIjwXHjgHOnq8IrWEPujORa5VpzIbxSeZGbtclMpUrFZZUVQs7JzbR3_Op9wJjsrhtCO560wggFGeQmH11icpWhizGgt32o9y58Ww72QNZOZO1I1h7JWj6G5BSKo7ndYPhb_U_qB8hIe14</recordid><startdate>20240301</startdate><enddate>20240301</enddate><creator>Dhal, Pradip</creator><creator>Azad, Chandrashekhar</creator><general>Springer London</general><general>Springer Nature B.V</general><scope>AAYXX</scope><scope>CITATION</scope><orcidid>https://orcid.org/0000-0003-2097-6601</orcidid></search><sort><creationdate>20240301</creationdate><title>A fine-tuning deep learning with multi-objective-based feature selection approach for the classification of text</title><author>Dhal, Pradip ; Azad, Chandrashekhar</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c319t-f22e5957e26930a19a011a5d8f40b07a93154755122f447879b969d4c413d6823</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Algorithms</topic><topic>Artificial Intelligence</topic><topic>Artificial neural networks</topic><topic>Classification</topic><topic>Computational Biology/Bioinformatics</topic><topic>Computational Science and Engineering</topic><topic>Computer Science</topic><topic>Data Mining and Knowledge Discovery</topic><topic>Deep learning</topic><topic>Documents</topic><topic>Feature selection</topic><topic>Image Processing and Computer Vision</topic><topic>Machine learning</topic><topic>Multiple objective analysis</topic><topic>Original Article</topic><topic>Particle swarm optimization</topic><topic>Probability and Statistics in Computer Science</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Dhal, Pradip</creatorcontrib><creatorcontrib>Azad, Chandrashekhar</creatorcontrib><collection>CrossRef</collection><jtitle>Neural computing &amp; applications</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Dhal, Pradip</au><au>Azad, Chandrashekhar</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>A fine-tuning deep learning with multi-objective-based feature selection approach for the classification of text</atitle><jtitle>Neural computing &amp; applications</jtitle><stitle>Neural Comput &amp; Applic</stitle><date>2024-03-01</date><risdate>2024</risdate><volume>36</volume><issue>7</issue><spage>3525</spage><epage>3553</epage><pages>3525-3553</pages><issn>0941-0643</issn><eissn>1433-3058</eissn><abstract>Document classification is becoming increasingly essential for the vast number of documents available in digital libraries, emails, the Internet, etc. Textual records frequently contain non-discriminative (noisy and irrelevant) terms that are also high-dimensional, resulting in higher computing costs and poorer learning performance in Text Classification (TC). Feature selection (FS), which tries to discover discriminate terms or features from the textual data, is one of the most effective tasks for this issue. This paper introduces a novel multi-stage term-weighting scheme-based FS model designed for the single-label TC system to obtain the optimal set of features. We have also developed a hybrid deep learning fine-tuning network based on Bidirectional Long Short-Term Memory (BiLSTM) and Convolutional Neural Network (CNN) for the classification stage. The FS approach is worked on two-stage criteria. The filter model is used in the first stage, and the multi-objective wrapper model, an upgraded version of the Whale Optimization Algorithm (WOA) with Particle Swarm Optimization (PSO), is used in the second stage. The objective function in the above wrapper model is based on a tri-objective principle. It uses the Pareto front technique to discover the optimal set of features. Here in the wrapper model, a novel selection strategy has been introduced to select the whale instead of the random whale. The proposed work is evaluated on four popular benchmark text corpora, of which two are binary class, and two are multi-class. The suggested FS technique is compared against classic Machine Learning (ML) and deep learning classifiers. The results of the experiments reveal that the recommended FS technique is more effective in obtaining better results than the other results.</abstract><cop>London</cop><pub>Springer London</pub><doi>10.1007/s00521-023-09225-1</doi><tpages>29</tpages><orcidid>https://orcid.org/0000-0003-2097-6601</orcidid></addata></record>
fulltext fulltext
identifier ISSN: 0941-0643
ispartof Neural computing & applications, 2024-03, Vol.36 (7), p.3525-3553
issn 0941-0643
1433-3058
language eng
recordid cdi_proquest_journals_2924060797
source Springer Nature - Complete Springer Journals
subjects Algorithms
Artificial Intelligence
Artificial neural networks
Classification
Computational Biology/Bioinformatics
Computational Science and Engineering
Computer Science
Data Mining and Knowledge Discovery
Deep learning
Documents
Feature selection
Image Processing and Computer Vision
Machine learning
Multiple objective analysis
Original Article
Particle swarm optimization
Probability and Statistics in Computer Science
title A fine-tuning deep learning with multi-objective-based feature selection approach for the classification of text
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-27T02%3A50%3A29IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=A%20fine-tuning%20deep%20learning%20with%20multi-objective-based%20feature%20selection%20approach%20for%20the%20classification%20of%20text&rft.jtitle=Neural%20computing%20&%20applications&rft.au=Dhal,%20Pradip&rft.date=2024-03-01&rft.volume=36&rft.issue=7&rft.spage=3525&rft.epage=3553&rft.pages=3525-3553&rft.issn=0941-0643&rft.eissn=1433-3058&rft_id=info:doi/10.1007/s00521-023-09225-1&rft_dat=%3Cproquest_cross%3E2924060797%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2924060797&rft_id=info:pmid/&rfr_iscdi=true