Generative Software Engineering

The rapid development of deep learning techniques, improved computational power, and the availability of vast training data have led to significant advancements in pre-trained models and large language models (LLMs). Pre-trained models based on architectures such as BERT and Transformer, as well as...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Huang, Yuan, Chen, Yinan, Chen, Xiangping, Chen, Junqi, Peng, Rui, Tang, Zhicao, Huang, Jinbo, Xu, Furen, Zheng, Zibin
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Huang, Yuan
Chen, Yinan
Chen, Xiangping
Chen, Junqi
Peng, Rui
Tang, Zhicao
Huang, Jinbo
Xu, Furen
Zheng, Zibin
description The rapid development of deep learning techniques, improved computational power, and the availability of vast training data have led to significant advancements in pre-trained models and large language models (LLMs). Pre-trained models based on architectures such as BERT and Transformer, as well as LLMs like ChatGPT, have demonstrated remarkable language capabilities and found applications in Software engineering. Software engineering tasks can be divided into many categories, among which generative tasks are the most concern by researchers, where pre-trained models and LLMs possess powerful language representation and contextual awareness capabilities, enabling them to leverage diverse training data and adapt to generative tasks through fine-tuning, transfer learning, and prompt engineering. These advantages make them effective tools in generative tasks and have demonstrated excellent performance. In this paper, we present a comprehensive literature review of generative tasks in SE using pre-trained models and LLMs. We accurately categorize SE generative tasks based on software engineering methodologies and summarize the advanced pre-trained models and LLMs involved, as well as the datasets and evaluation metrics used. Additionally, we identify key strengths, weaknesses, and gaps in existing approaches, and propose potential research directions. This review aims to provide researchers and practitioners with an in-depth analysis and guidance on the application of pre-trained models and LLMs in generative tasks within SE.
doi_str_mv 10.48550/arxiv.2403.02583
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2403_02583</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2403_02583</sourcerecordid><originalsourceid>FETCH-LOGICAL-a673-e25454198dcf54250d416f34bb41af16197bdd11545f1f5dbd5a5350d1d205aa3</originalsourceid><addsrcrecordid>eNotzr0KwjAUhuEsDqJegFO9gdacJKeto0itguBg93JikhLQKFH8uXt_p295-XgYGwPPVInIpxQf_pYJxWXGBZayz5LaBhvp6m92sju5652inVSh88Ha6EM3ZD1Hh4sd_XfAmmXVLFbpZluvF_NNSnkhUytQoYJZafYOlUBuFOROKq0VkIMcZoU2BuBdOXBotEFC-c7ACI5EcsCS3-1X2J6jP1J8th9p-5XKF1GHN1A</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Generative Software Engineering</title><source>arXiv.org</source><creator>Huang, Yuan ; Chen, Yinan ; Chen, Xiangping ; Chen, Junqi ; Peng, Rui ; Tang, Zhicao ; Huang, Jinbo ; Xu, Furen ; Zheng, Zibin</creator><creatorcontrib>Huang, Yuan ; Chen, Yinan ; Chen, Xiangping ; Chen, Junqi ; Peng, Rui ; Tang, Zhicao ; Huang, Jinbo ; Xu, Furen ; Zheng, Zibin</creatorcontrib><description>The rapid development of deep learning techniques, improved computational power, and the availability of vast training data have led to significant advancements in pre-trained models and large language models (LLMs). Pre-trained models based on architectures such as BERT and Transformer, as well as LLMs like ChatGPT, have demonstrated remarkable language capabilities and found applications in Software engineering. Software engineering tasks can be divided into many categories, among which generative tasks are the most concern by researchers, where pre-trained models and LLMs possess powerful language representation and contextual awareness capabilities, enabling them to leverage diverse training data and adapt to generative tasks through fine-tuning, transfer learning, and prompt engineering. These advantages make them effective tools in generative tasks and have demonstrated excellent performance. In this paper, we present a comprehensive literature review of generative tasks in SE using pre-trained models and LLMs. We accurately categorize SE generative tasks based on software engineering methodologies and summarize the advanced pre-trained models and LLMs involved, as well as the datasets and evaluation metrics used. Additionally, we identify key strengths, weaknesses, and gaps in existing approaches, and propose potential research directions. This review aims to provide researchers and practitioners with an in-depth analysis and guidance on the application of pre-trained models and LLMs in generative tasks within SE.</description><identifier>DOI: 10.48550/arxiv.2403.02583</identifier><language>eng</language><subject>Computer Science - Software Engineering</subject><creationdate>2024-03</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2403.02583$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2403.02583$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Huang, Yuan</creatorcontrib><creatorcontrib>Chen, Yinan</creatorcontrib><creatorcontrib>Chen, Xiangping</creatorcontrib><creatorcontrib>Chen, Junqi</creatorcontrib><creatorcontrib>Peng, Rui</creatorcontrib><creatorcontrib>Tang, Zhicao</creatorcontrib><creatorcontrib>Huang, Jinbo</creatorcontrib><creatorcontrib>Xu, Furen</creatorcontrib><creatorcontrib>Zheng, Zibin</creatorcontrib><title>Generative Software Engineering</title><description>The rapid development of deep learning techniques, improved computational power, and the availability of vast training data have led to significant advancements in pre-trained models and large language models (LLMs). Pre-trained models based on architectures such as BERT and Transformer, as well as LLMs like ChatGPT, have demonstrated remarkable language capabilities and found applications in Software engineering. Software engineering tasks can be divided into many categories, among which generative tasks are the most concern by researchers, where pre-trained models and LLMs possess powerful language representation and contextual awareness capabilities, enabling them to leverage diverse training data and adapt to generative tasks through fine-tuning, transfer learning, and prompt engineering. These advantages make them effective tools in generative tasks and have demonstrated excellent performance. In this paper, we present a comprehensive literature review of generative tasks in SE using pre-trained models and LLMs. We accurately categorize SE generative tasks based on software engineering methodologies and summarize the advanced pre-trained models and LLMs involved, as well as the datasets and evaluation metrics used. Additionally, we identify key strengths, weaknesses, and gaps in existing approaches, and propose potential research directions. This review aims to provide researchers and practitioners with an in-depth analysis and guidance on the application of pre-trained models and LLMs in generative tasks within SE.</description><subject>Computer Science - Software Engineering</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotzr0KwjAUhuEsDqJegFO9gdacJKeto0itguBg93JikhLQKFH8uXt_p295-XgYGwPPVInIpxQf_pYJxWXGBZayz5LaBhvp6m92sju5652inVSh88Ha6EM3ZD1Hh4sd_XfAmmXVLFbpZluvF_NNSnkhUytQoYJZafYOlUBuFOROKq0VkIMcZoU2BuBdOXBotEFC-c7ACI5EcsCS3-1X2J6jP1J8th9p-5XKF1GHN1A</recordid><startdate>20240304</startdate><enddate>20240304</enddate><creator>Huang, Yuan</creator><creator>Chen, Yinan</creator><creator>Chen, Xiangping</creator><creator>Chen, Junqi</creator><creator>Peng, Rui</creator><creator>Tang, Zhicao</creator><creator>Huang, Jinbo</creator><creator>Xu, Furen</creator><creator>Zheng, Zibin</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20240304</creationdate><title>Generative Software Engineering</title><author>Huang, Yuan ; Chen, Yinan ; Chen, Xiangping ; Chen, Junqi ; Peng, Rui ; Tang, Zhicao ; Huang, Jinbo ; Xu, Furen ; Zheng, Zibin</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a673-e25454198dcf54250d416f34bb41af16197bdd11545f1f5dbd5a5350d1d205aa3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computer Science - Software Engineering</topic><toplevel>online_resources</toplevel><creatorcontrib>Huang, Yuan</creatorcontrib><creatorcontrib>Chen, Yinan</creatorcontrib><creatorcontrib>Chen, Xiangping</creatorcontrib><creatorcontrib>Chen, Junqi</creatorcontrib><creatorcontrib>Peng, Rui</creatorcontrib><creatorcontrib>Tang, Zhicao</creatorcontrib><creatorcontrib>Huang, Jinbo</creatorcontrib><creatorcontrib>Xu, Furen</creatorcontrib><creatorcontrib>Zheng, Zibin</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Huang, Yuan</au><au>Chen, Yinan</au><au>Chen, Xiangping</au><au>Chen, Junqi</au><au>Peng, Rui</au><au>Tang, Zhicao</au><au>Huang, Jinbo</au><au>Xu, Furen</au><au>Zheng, Zibin</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Generative Software Engineering</atitle><date>2024-03-04</date><risdate>2024</risdate><abstract>The rapid development of deep learning techniques, improved computational power, and the availability of vast training data have led to significant advancements in pre-trained models and large language models (LLMs). Pre-trained models based on architectures such as BERT and Transformer, as well as LLMs like ChatGPT, have demonstrated remarkable language capabilities and found applications in Software engineering. Software engineering tasks can be divided into many categories, among which generative tasks are the most concern by researchers, where pre-trained models and LLMs possess powerful language representation and contextual awareness capabilities, enabling them to leverage diverse training data and adapt to generative tasks through fine-tuning, transfer learning, and prompt engineering. These advantages make them effective tools in generative tasks and have demonstrated excellent performance. In this paper, we present a comprehensive literature review of generative tasks in SE using pre-trained models and LLMs. We accurately categorize SE generative tasks based on software engineering methodologies and summarize the advanced pre-trained models and LLMs involved, as well as the datasets and evaluation metrics used. Additionally, we identify key strengths, weaknesses, and gaps in existing approaches, and propose potential research directions. This review aims to provide researchers and practitioners with an in-depth analysis and guidance on the application of pre-trained models and LLMs in generative tasks within SE.</abstract><doi>10.48550/arxiv.2403.02583</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2403.02583
ispartof
issn
language eng
recordid cdi_arxiv_primary_2403_02583
source arXiv.org
subjects Computer Science - Software Engineering
title Generative Software Engineering
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-22T09%3A29%3A28IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Generative%20Software%20Engineering&rft.au=Huang,%20Yuan&rft.date=2024-03-04&rft_id=info:doi/10.48550/arxiv.2403.02583&rft_dat=%3Carxiv_GOX%3E2403_02583%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true