Fine-Tuned T5 Transformer with LSTM and Spider Monkey Optimizer for Redundancy Reduction in Automatic Question Generation

The significance of Automatic Question Generation (AQG) lies in its potential to support educators and streamline assessment processes. Notable improvements in AQG are seen with the use of language models, ranging from LSTMs to Transformers. However, there is a need for improvement in the probabilis...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:SN computer science 2024-06, Vol.5 (5), p.475, Article 475
Hauptverfasser: Tharaniya sairaj, R., Balasundaram, S. R.
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue 5
container_start_page 475
container_title SN computer science
container_volume 5
creator Tharaniya sairaj, R.
Balasundaram, S. R.
description The significance of Automatic Question Generation (AQG) lies in its potential to support educators and streamline assessment processes. Notable improvements in AQG are seen with the use of language models, ranging from LSTMs to Transformers. However, there is a need for improvement in the probabilistic scoring technique employed for next word generation in the target question. In this regard, it is noted that template-based methods offer potential for enhancement, although they may result in the generation of fewer or redundant questions due to the utilization of fixed templates. This research aims to address this gap by proposing a hybrid model that combines the the advantages of template-based and Transformer-based AQG approaches. The template-based LSTM approach is explored to learn adaptable question templates. On the other side, the Transformer model is explored to reduce redundancy in the auto-generated questions. The proposed work finetunes the pipelined T5 Transformer model using the Spider Monkey Optimizer over the LSTM-generated templates. The choice of Spider Monkey Optimizer enhances the selection of the named entity in question tail (tail entity) through dynamic sub-search space division for efficient exploration and exploitation, and self-organization based on local and global scoring. This ensures that the named entity in the question tail is non-redundant (diverse) and regards both structural and contextual coherence to the auto-generated question. Experimental findings highlight improvements in how well the diversely selected named entities are relevant to the generated questions through higher precision, recall and f1-scores in the pipelining phase. Moreover, the study shows that the Spider Monkey Optimizer performs better in selecting tail entities, and it consistently outperforms other algorithms in F1-score and convergence time across all datasets, with its time complexity increasing linearly as dataset size grows. The finetuned pipelined T5 model (proposed model) exhibits improved ROUGE scores over baselines with reduced computational overhead and shorter inference time in the generative phase across datasets in linear convergence time.
doi_str_mv 10.1007/s42979-024-02826-0
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_3048639198</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>3048639198</sourcerecordid><originalsourceid>FETCH-LOGICAL-c1850-fe054e2ad782a4317b44131bfd8474a1dd1ce5507395939a1710b8024c983203</originalsourceid><addsrcrecordid>eNp9UE1LAzEUDKJgqf0DngKeV18-9utYiq1CS9HuPaSbrKba7JrsIuuvN9sV9OTh8YZhZt5jELomcEsA0jvPaZ7mEVAeJqNJBGdoQpOERFkO6fkffIlm3h8AgMbAeRJPUL80VkdFZ7XCRYwLJ62vanfUDn-a9hWvd8UGS6vwrjEqkJvavukeb5vWHM1XIIIYP2vVWSVt2Z9g2ZraYmPxvGvro2xNiZ867U_sSlvt5ACv0EUl372e_ewpKpb3xeIhWm9Xj4v5OipJFkNUaYi5plKlGZWckXTPOWFkX6mMp1wSpUip4xhSlsc5yyVJCeyz0EWZZ4wCm6KbMbZx9cfwhTjUnbPhomDAs4TlJAiniI6q0tXeO12JxpmjdL0gIIaSxViyCMHiVLIYotlo8kFsX7T7jf7H9Q2e_n6Z</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3048639198</pqid></control><display><type>article</type><title>Fine-Tuned T5 Transformer with LSTM and Spider Monkey Optimizer for Redundancy Reduction in Automatic Question Generation</title><source>SpringerLink Journals - AutoHoldings</source><creator>Tharaniya sairaj, R. ; Balasundaram, S. R.</creator><creatorcontrib>Tharaniya sairaj, R. ; Balasundaram, S. R.</creatorcontrib><description>The significance of Automatic Question Generation (AQG) lies in its potential to support educators and streamline assessment processes. Notable improvements in AQG are seen with the use of language models, ranging from LSTMs to Transformers. However, there is a need for improvement in the probabilistic scoring technique employed for next word generation in the target question. In this regard, it is noted that template-based methods offer potential for enhancement, although they may result in the generation of fewer or redundant questions due to the utilization of fixed templates. This research aims to address this gap by proposing a hybrid model that combines the the advantages of template-based and Transformer-based AQG approaches. The template-based LSTM approach is explored to learn adaptable question templates. On the other side, the Transformer model is explored to reduce redundancy in the auto-generated questions. The proposed work finetunes the pipelined T5 Transformer model using the Spider Monkey Optimizer over the LSTM-generated templates. The choice of Spider Monkey Optimizer enhances the selection of the named entity in question tail (tail entity) through dynamic sub-search space division for efficient exploration and exploitation, and self-organization based on local and global scoring. This ensures that the named entity in the question tail is non-redundant (diverse) and regards both structural and contextual coherence to the auto-generated question. Experimental findings highlight improvements in how well the diversely selected named entities are relevant to the generated questions through higher precision, recall and f1-scores in the pipelining phase. Moreover, the study shows that the Spider Monkey Optimizer performs better in selecting tail entities, and it consistently outperforms other algorithms in F1-score and convergence time across all datasets, with its time complexity increasing linearly as dataset size grows. The finetuned pipelined T5 model (proposed model) exhibits improved ROUGE scores over baselines with reduced computational overhead and shorter inference time in the generative phase across datasets in linear convergence time.</description><identifier>ISSN: 2661-8907</identifier><identifier>ISSN: 2662-995X</identifier><identifier>EISSN: 2661-8907</identifier><identifier>DOI: 10.1007/s42979-024-02826-0</identifier><language>eng</language><publisher>Singapore: Springer Nature Singapore</publisher><subject>Algorithms ; Automation ; Computer Imaging ; Computer Science ; Computer Systems Organization and Communication Networks ; Convergence ; Data base management systems ; Data integrity ; Data Structures and Information Theory ; Datasets ; Deep learning ; Emerging Applications of Data Science for Real-World Problems ; Information Systems and Communication Service ; Language ; Monkeys ; Neural networks ; Original Research ; Pattern Recognition and Graphics ; Questions ; Redundancy ; Software Engineering/Programming and Operating Systems ; Transformers ; Vision</subject><ispartof>SN computer science, 2024-06, Vol.5 (5), p.475, Article 475</ispartof><rights>The Author(s), under exclusive licence to Springer Nature Singapore Pte Ltd. 2024. Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c1850-fe054e2ad782a4317b44131bfd8474a1dd1ce5507395939a1710b8024c983203</cites><orcidid>0000-0002-7015-4414</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://link.springer.com/content/pdf/10.1007/s42979-024-02826-0$$EPDF$$P50$$Gspringer$$H</linktopdf><linktohtml>$$Uhttps://link.springer.com/10.1007/s42979-024-02826-0$$EHTML$$P50$$Gspringer$$H</linktohtml><link.rule.ids>314,776,780,27901,27902,41464,42533,51294</link.rule.ids></links><search><creatorcontrib>Tharaniya sairaj, R.</creatorcontrib><creatorcontrib>Balasundaram, S. R.</creatorcontrib><title>Fine-Tuned T5 Transformer with LSTM and Spider Monkey Optimizer for Redundancy Reduction in Automatic Question Generation</title><title>SN computer science</title><addtitle>SN COMPUT. SCI</addtitle><description>The significance of Automatic Question Generation (AQG) lies in its potential to support educators and streamline assessment processes. Notable improvements in AQG are seen with the use of language models, ranging from LSTMs to Transformers. However, there is a need for improvement in the probabilistic scoring technique employed for next word generation in the target question. In this regard, it is noted that template-based methods offer potential for enhancement, although they may result in the generation of fewer or redundant questions due to the utilization of fixed templates. This research aims to address this gap by proposing a hybrid model that combines the the advantages of template-based and Transformer-based AQG approaches. The template-based LSTM approach is explored to learn adaptable question templates. On the other side, the Transformer model is explored to reduce redundancy in the auto-generated questions. The proposed work finetunes the pipelined T5 Transformer model using the Spider Monkey Optimizer over the LSTM-generated templates. The choice of Spider Monkey Optimizer enhances the selection of the named entity in question tail (tail entity) through dynamic sub-search space division for efficient exploration and exploitation, and self-organization based on local and global scoring. This ensures that the named entity in the question tail is non-redundant (diverse) and regards both structural and contextual coherence to the auto-generated question. Experimental findings highlight improvements in how well the diversely selected named entities are relevant to the generated questions through higher precision, recall and f1-scores in the pipelining phase. Moreover, the study shows that the Spider Monkey Optimizer performs better in selecting tail entities, and it consistently outperforms other algorithms in F1-score and convergence time across all datasets, with its time complexity increasing linearly as dataset size grows. The finetuned pipelined T5 model (proposed model) exhibits improved ROUGE scores over baselines with reduced computational overhead and shorter inference time in the generative phase across datasets in linear convergence time.</description><subject>Algorithms</subject><subject>Automation</subject><subject>Computer Imaging</subject><subject>Computer Science</subject><subject>Computer Systems Organization and Communication Networks</subject><subject>Convergence</subject><subject>Data base management systems</subject><subject>Data integrity</subject><subject>Data Structures and Information Theory</subject><subject>Datasets</subject><subject>Deep learning</subject><subject>Emerging Applications of Data Science for Real-World Problems</subject><subject>Information Systems and Communication Service</subject><subject>Language</subject><subject>Monkeys</subject><subject>Neural networks</subject><subject>Original Research</subject><subject>Pattern Recognition and Graphics</subject><subject>Questions</subject><subject>Redundancy</subject><subject>Software Engineering/Programming and Operating Systems</subject><subject>Transformers</subject><subject>Vision</subject><issn>2661-8907</issn><issn>2662-995X</issn><issn>2661-8907</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><recordid>eNp9UE1LAzEUDKJgqf0DngKeV18-9utYiq1CS9HuPaSbrKba7JrsIuuvN9sV9OTh8YZhZt5jELomcEsA0jvPaZ7mEVAeJqNJBGdoQpOERFkO6fkffIlm3h8AgMbAeRJPUL80VkdFZ7XCRYwLJ62vanfUDn-a9hWvd8UGS6vwrjEqkJvavukeb5vWHM1XIIIYP2vVWSVt2Z9g2ZraYmPxvGvro2xNiZ867U_sSlvt5ACv0EUl372e_ewpKpb3xeIhWm9Xj4v5OipJFkNUaYi5plKlGZWckXTPOWFkX6mMp1wSpUip4xhSlsc5yyVJCeyz0EWZZ4wCm6KbMbZx9cfwhTjUnbPhomDAs4TlJAiniI6q0tXeO12JxpmjdL0gIIaSxViyCMHiVLIYotlo8kFsX7T7jf7H9Q2e_n6Z</recordid><startdate>20240601</startdate><enddate>20240601</enddate><creator>Tharaniya sairaj, R.</creator><creator>Balasundaram, S. R.</creator><general>Springer Nature Singapore</general><general>Springer Nature B.V</general><scope>AAYXX</scope><scope>CITATION</scope><scope>JQ2</scope><orcidid>https://orcid.org/0000-0002-7015-4414</orcidid></search><sort><creationdate>20240601</creationdate><title>Fine-Tuned T5 Transformer with LSTM and Spider Monkey Optimizer for Redundancy Reduction in Automatic Question Generation</title><author>Tharaniya sairaj, R. ; Balasundaram, S. R.</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c1850-fe054e2ad782a4317b44131bfd8474a1dd1ce5507395939a1710b8024c983203</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Algorithms</topic><topic>Automation</topic><topic>Computer Imaging</topic><topic>Computer Science</topic><topic>Computer Systems Organization and Communication Networks</topic><topic>Convergence</topic><topic>Data base management systems</topic><topic>Data integrity</topic><topic>Data Structures and Information Theory</topic><topic>Datasets</topic><topic>Deep learning</topic><topic>Emerging Applications of Data Science for Real-World Problems</topic><topic>Information Systems and Communication Service</topic><topic>Language</topic><topic>Monkeys</topic><topic>Neural networks</topic><topic>Original Research</topic><topic>Pattern Recognition and Graphics</topic><topic>Questions</topic><topic>Redundancy</topic><topic>Software Engineering/Programming and Operating Systems</topic><topic>Transformers</topic><topic>Vision</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Tharaniya sairaj, R.</creatorcontrib><creatorcontrib>Balasundaram, S. R.</creatorcontrib><collection>CrossRef</collection><collection>ProQuest Computer Science Collection</collection><jtitle>SN computer science</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Tharaniya sairaj, R.</au><au>Balasundaram, S. R.</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Fine-Tuned T5 Transformer with LSTM and Spider Monkey Optimizer for Redundancy Reduction in Automatic Question Generation</atitle><jtitle>SN computer science</jtitle><stitle>SN COMPUT. SCI</stitle><date>2024-06-01</date><risdate>2024</risdate><volume>5</volume><issue>5</issue><spage>475</spage><pages>475-</pages><artnum>475</artnum><issn>2661-8907</issn><issn>2662-995X</issn><eissn>2661-8907</eissn><abstract>The significance of Automatic Question Generation (AQG) lies in its potential to support educators and streamline assessment processes. Notable improvements in AQG are seen with the use of language models, ranging from LSTMs to Transformers. However, there is a need for improvement in the probabilistic scoring technique employed for next word generation in the target question. In this regard, it is noted that template-based methods offer potential for enhancement, although they may result in the generation of fewer or redundant questions due to the utilization of fixed templates. This research aims to address this gap by proposing a hybrid model that combines the the advantages of template-based and Transformer-based AQG approaches. The template-based LSTM approach is explored to learn adaptable question templates. On the other side, the Transformer model is explored to reduce redundancy in the auto-generated questions. The proposed work finetunes the pipelined T5 Transformer model using the Spider Monkey Optimizer over the LSTM-generated templates. The choice of Spider Monkey Optimizer enhances the selection of the named entity in question tail (tail entity) through dynamic sub-search space division for efficient exploration and exploitation, and self-organization based on local and global scoring. This ensures that the named entity in the question tail is non-redundant (diverse) and regards both structural and contextual coherence to the auto-generated question. Experimental findings highlight improvements in how well the diversely selected named entities are relevant to the generated questions through higher precision, recall and f1-scores in the pipelining phase. Moreover, the study shows that the Spider Monkey Optimizer performs better in selecting tail entities, and it consistently outperforms other algorithms in F1-score and convergence time across all datasets, with its time complexity increasing linearly as dataset size grows. The finetuned pipelined T5 model (proposed model) exhibits improved ROUGE scores over baselines with reduced computational overhead and shorter inference time in the generative phase across datasets in linear convergence time.</abstract><cop>Singapore</cop><pub>Springer Nature Singapore</pub><doi>10.1007/s42979-024-02826-0</doi><orcidid>https://orcid.org/0000-0002-7015-4414</orcidid></addata></record>
fulltext fulltext
identifier ISSN: 2661-8907
ispartof SN computer science, 2024-06, Vol.5 (5), p.475, Article 475
issn 2661-8907
2662-995X
2661-8907
language eng
recordid cdi_proquest_journals_3048639198
source SpringerLink Journals - AutoHoldings
subjects Algorithms
Automation
Computer Imaging
Computer Science
Computer Systems Organization and Communication Networks
Convergence
Data base management systems
Data integrity
Data Structures and Information Theory
Datasets
Deep learning
Emerging Applications of Data Science for Real-World Problems
Information Systems and Communication Service
Language
Monkeys
Neural networks
Original Research
Pattern Recognition and Graphics
Questions
Redundancy
Software Engineering/Programming and Operating Systems
Transformers
Vision
title Fine-Tuned T5 Transformer with LSTM and Spider Monkey Optimizer for Redundancy Reduction in Automatic Question Generation
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-29T16%3A23%3A33IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Fine-Tuned%20T5%20Transformer%20with%20LSTM%20and%20Spider%20Monkey%20Optimizer%20for%20Redundancy%20Reduction%20in%20Automatic%20Question%20Generation&rft.jtitle=SN%20computer%20science&rft.au=Tharaniya%20sairaj,%20R.&rft.date=2024-06-01&rft.volume=5&rft.issue=5&rft.spage=475&rft.pages=475-&rft.artnum=475&rft.issn=2661-8907&rft.eissn=2661-8907&rft_id=info:doi/10.1007/s42979-024-02826-0&rft_dat=%3Cproquest_cross%3E3048639198%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=3048639198&rft_id=info:pmid/&rfr_iscdi=true