The Performance of the LSTM-based Code Generated by Large Language Models (LLMs) in Forecasting Time Series Data

As an intriguing case is the goodness of the machine and deep learning models generated by these LLMs in conducting automated scientific data analysis, where a data analyst may not have enough expertise in manually coding and optimizing complex deep learning models and codes and thus may opt to leve...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:arXiv.org 2024-11
Hauptverfasser: Gopali, Saroj, Siami-Namini, Sima, Abri, Faranak, Namin, Akbar Siami
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Gopali, Saroj
Siami-Namini, Sima
Abri, Faranak
Namin, Akbar Siami
description As an intriguing case is the goodness of the machine and deep learning models generated by these LLMs in conducting automated scientific data analysis, where a data analyst may not have enough expertise in manually coding and optimizing complex deep learning models and codes and thus may opt to leverage LLMs to generate the required models. This paper investigates and compares the performance of the mainstream LLMs, such as ChatGPT, PaLM, LLama, and Falcon, in generating deep learning models for analyzing time series data, an important and popular data type with its prevalent applications in many application domains including financial and stock market. This research conducts a set of controlled experiments where the prompts for generating deep learning-based models are controlled with respect to sensitivity levels of four criteria including 1) Clarify and Specificity, 2) Objective and Intent, 3) Contextual Information, and 4) Format and Style. While the results are relatively mix, we observe some distinct patterns. We notice that using LLMs, we are able to generate deep learning-based models with executable codes for each dataset seperatly whose performance are comparable with the manually crafted and optimized LSTM models for predicting the whole time series dataset. We also noticed that ChatGPT outperforms the other LLMs in generating more accurate models. Furthermore, we observed that the goodness of the generated models vary with respect to the ``temperature'' parameter used in configuring LLMS. The results can be beneficial for data analysts and practitioners who would like to leverage generative AIs to produce good prediction models with acceptable goodness.
format Article
fullrecord <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_3134987124</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>3134987124</sourcerecordid><originalsourceid>FETCH-proquest_journals_31349871243</originalsourceid><addsrcrecordid>eNqNir0KwkAQhA9BMKjvsGCjRSC5ixprf4sEhKSXVTcxktzp7aXw7b3CB7CamW--gQikUnGYJlKOxJT5GUWRXK3lcqkC8SofBGeylbEd6huBqcB5lBVlHl6R6Q5bcyc4kiaLzs_rBzK0tVdQ1z36knuhZZhnWc4LaDQcjKUbsmt0DWXTERRkG2LYocOJGFbYMk1_ORazw77cnsKXNe-e2F2eprfaXxcVq2STrmOZqP-sL0ipSB8</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3134987124</pqid></control><display><type>article</type><title>The Performance of the LSTM-based Code Generated by Large Language Models (LLMs) in Forecasting Time Series Data</title><source>Free E- Journals</source><creator>Gopali, Saroj ; Siami-Namini, Sima ; Abri, Faranak ; Namin, Akbar Siami</creator><creatorcontrib>Gopali, Saroj ; Siami-Namini, Sima ; Abri, Faranak ; Namin, Akbar Siami</creatorcontrib><description>As an intriguing case is the goodness of the machine and deep learning models generated by these LLMs in conducting automated scientific data analysis, where a data analyst may not have enough expertise in manually coding and optimizing complex deep learning models and codes and thus may opt to leverage LLMs to generate the required models. This paper investigates and compares the performance of the mainstream LLMs, such as ChatGPT, PaLM, LLama, and Falcon, in generating deep learning models for analyzing time series data, an important and popular data type with its prevalent applications in many application domains including financial and stock market. This research conducts a set of controlled experiments where the prompts for generating deep learning-based models are controlled with respect to sensitivity levels of four criteria including 1) Clarify and Specificity, 2) Objective and Intent, 3) Contextual Information, and 4) Format and Style. While the results are relatively mix, we observe some distinct patterns. We notice that using LLMs, we are able to generate deep learning-based models with executable codes for each dataset seperatly whose performance are comparable with the manually crafted and optimized LSTM models for predicting the whole time series dataset. We also noticed that ChatGPT outperforms the other LLMs in generating more accurate models. Furthermore, we observed that the goodness of the generated models vary with respect to the ``temperature'' parameter used in configuring LLMS. The results can be beneficial for data analysts and practitioners who would like to leverage generative AIs to produce good prediction models with acceptable goodness.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Chatbots ; Data analysis ; Datasets ; Deep learning ; Generative artificial intelligence ; Large language models ; Machine learning ; Prediction models ; Predictions ; Time series</subject><ispartof>arXiv.org, 2024-11</ispartof><rights>2024. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>780,784</link.rule.ids></links><search><creatorcontrib>Gopali, Saroj</creatorcontrib><creatorcontrib>Siami-Namini, Sima</creatorcontrib><creatorcontrib>Abri, Faranak</creatorcontrib><creatorcontrib>Namin, Akbar Siami</creatorcontrib><title>The Performance of the LSTM-based Code Generated by Large Language Models (LLMs) in Forecasting Time Series Data</title><title>arXiv.org</title><description>As an intriguing case is the goodness of the machine and deep learning models generated by these LLMs in conducting automated scientific data analysis, where a data analyst may not have enough expertise in manually coding and optimizing complex deep learning models and codes and thus may opt to leverage LLMs to generate the required models. This paper investigates and compares the performance of the mainstream LLMs, such as ChatGPT, PaLM, LLama, and Falcon, in generating deep learning models for analyzing time series data, an important and popular data type with its prevalent applications in many application domains including financial and stock market. This research conducts a set of controlled experiments where the prompts for generating deep learning-based models are controlled with respect to sensitivity levels of four criteria including 1) Clarify and Specificity, 2) Objective and Intent, 3) Contextual Information, and 4) Format and Style. While the results are relatively mix, we observe some distinct patterns. We notice that using LLMs, we are able to generate deep learning-based models with executable codes for each dataset seperatly whose performance are comparable with the manually crafted and optimized LSTM models for predicting the whole time series dataset. We also noticed that ChatGPT outperforms the other LLMs in generating more accurate models. Furthermore, we observed that the goodness of the generated models vary with respect to the ``temperature'' parameter used in configuring LLMS. The results can be beneficial for data analysts and practitioners who would like to leverage generative AIs to produce good prediction models with acceptable goodness.</description><subject>Chatbots</subject><subject>Data analysis</subject><subject>Datasets</subject><subject>Deep learning</subject><subject>Generative artificial intelligence</subject><subject>Large language models</subject><subject>Machine learning</subject><subject>Prediction models</subject><subject>Predictions</subject><subject>Time series</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><recordid>eNqNir0KwkAQhA9BMKjvsGCjRSC5ixprf4sEhKSXVTcxktzp7aXw7b3CB7CamW--gQikUnGYJlKOxJT5GUWRXK3lcqkC8SofBGeylbEd6huBqcB5lBVlHl6R6Q5bcyc4kiaLzs_rBzK0tVdQ1z36knuhZZhnWc4LaDQcjKUbsmt0DWXTERRkG2LYocOJGFbYMk1_ORazw77cnsKXNe-e2F2eprfaXxcVq2STrmOZqP-sL0ipSB8</recordid><startdate>20241127</startdate><enddate>20241127</enddate><creator>Gopali, Saroj</creator><creator>Siami-Namini, Sima</creator><creator>Abri, Faranak</creator><creator>Namin, Akbar Siami</creator><general>Cornell University Library, arXiv.org</general><scope>7X5</scope><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BEZIV</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>K6~</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20241127</creationdate><title>The Performance of the LSTM-based Code Generated by Large Language Models (LLMs) in Forecasting Time Series Data</title><author>Gopali, Saroj ; Siami-Namini, Sima ; Abri, Faranak ; Namin, Akbar Siami</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_31349871243</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Chatbots</topic><topic>Data analysis</topic><topic>Datasets</topic><topic>Deep learning</topic><topic>Generative artificial intelligence</topic><topic>Large language models</topic><topic>Machine learning</topic><topic>Prediction models</topic><topic>Predictions</topic><topic>Time series</topic><toplevel>online_resources</toplevel><creatorcontrib>Gopali, Saroj</creatorcontrib><creatorcontrib>Siami-Namini, Sima</creatorcontrib><creatorcontrib>Abri, Faranak</creatorcontrib><creatorcontrib>Namin, Akbar Siami</creatorcontrib><collection>Entrepreneurship Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Business Premium Collection</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Business Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Access via ProQuest (Open Access)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Gopali, Saroj</au><au>Siami-Namini, Sima</au><au>Abri, Faranak</au><au>Namin, Akbar Siami</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>The Performance of the LSTM-based Code Generated by Large Language Models (LLMs) in Forecasting Time Series Data</atitle><jtitle>arXiv.org</jtitle><date>2024-11-27</date><risdate>2024</risdate><eissn>2331-8422</eissn><abstract>As an intriguing case is the goodness of the machine and deep learning models generated by these LLMs in conducting automated scientific data analysis, where a data analyst may not have enough expertise in manually coding and optimizing complex deep learning models and codes and thus may opt to leverage LLMs to generate the required models. This paper investigates and compares the performance of the mainstream LLMs, such as ChatGPT, PaLM, LLama, and Falcon, in generating deep learning models for analyzing time series data, an important and popular data type with its prevalent applications in many application domains including financial and stock market. This research conducts a set of controlled experiments where the prompts for generating deep learning-based models are controlled with respect to sensitivity levels of four criteria including 1) Clarify and Specificity, 2) Objective and Intent, 3) Contextual Information, and 4) Format and Style. While the results are relatively mix, we observe some distinct patterns. We notice that using LLMs, we are able to generate deep learning-based models with executable codes for each dataset seperatly whose performance are comparable with the manually crafted and optimized LSTM models for predicting the whole time series dataset. We also noticed that ChatGPT outperforms the other LLMs in generating more accurate models. Furthermore, we observed that the goodness of the generated models vary with respect to the ``temperature'' parameter used in configuring LLMS. The results can be beneficial for data analysts and practitioners who would like to leverage generative AIs to produce good prediction models with acceptable goodness.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2024-11
issn 2331-8422
language eng
recordid cdi_proquest_journals_3134987124
source Free E- Journals
subjects Chatbots
Data analysis
Datasets
Deep learning
Generative artificial intelligence
Large language models
Machine learning
Prediction models
Predictions
Time series
title The Performance of the LSTM-based Code Generated by Large Language Models (LLMs) in Forecasting Time Series Data
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-28T02%3A48%3A37IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=The%20Performance%20of%20the%20LSTM-based%20Code%20Generated%20by%20Large%20Language%20Models%20(LLMs)%20in%20Forecasting%20Time%20Series%20Data&rft.jtitle=arXiv.org&rft.au=Gopali,%20Saroj&rft.date=2024-11-27&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E3134987124%3C/proquest%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=3134987124&rft_id=info:pmid/&rfr_iscdi=true