Examining User-Friendly and Open-Sourced Large GPT Models: A Survey on Language, Multimodal, and Scientific GPT Models

Generative pre-trained transformer (GPT) models have revolutionized the field of natural language processing (NLP) with remarkable performance in various tasks and also extend their power to multimodal domains. Despite their success, large GPT models like GPT-4 face inherent limitations such as cons...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Gao, Kaiyuan, He, Sunan, He, Zhenyu, Lin, Jiacheng, Pei, QiZhi, Shao, Jie, Zhang, Wei
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Gao, Kaiyuan
He, Sunan
He, Zhenyu
Lin, Jiacheng
Pei, QiZhi
Shao, Jie
Zhang, Wei
description Generative pre-trained transformer (GPT) models have revolutionized the field of natural language processing (NLP) with remarkable performance in various tasks and also extend their power to multimodal domains. Despite their success, large GPT models like GPT-4 face inherent limitations such as considerable size, high computational requirements, complex deployment processes, and closed development loops. These constraints restrict their widespread adoption and raise concerns regarding their responsible development and usage. The need for user-friendly, relatively small, and open-sourced alternative GPT models arises from the desire to overcome these limitations while retaining high performance. In this survey paper, we provide an examination of alternative open-sourced models of large GPTs, focusing on user-friendly and relatively small models that facilitate easier deployment and accessibility. Through this extensive survey, we aim to equip researchers, practitioners, and enthusiasts with a thorough understanding of user-friendly and relatively small open-sourced models of large GPTs, their current state, challenges, and future research directions, inspiring the development of more efficient, accessible, and versatile GPT models that cater to the broader scientific community and advance the field of general artificial intelligence. The source contents are continuously updating in https://github.com/GPT-Alternatives/gpt_alternatives.
doi_str_mv 10.48550/arxiv.2308.14149
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2308_14149</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2308_14149</sourcerecordid><originalsourceid>FETCH-LOGICAL-a679-b90e6aa1dfefb7c05a7bf7f25e31ba0354856a199f3d31ee0df03f1198cc2c0b3</originalsourceid><addsrcrecordid>eNpNkL1ugzAUhVk6VGkfoFP9AIHamD93i6IkrUSUStAZXexrZAlMZAIKb1-aZuh0hqPzSd_xvBdGgyiLY_oG7mqmIOQ0C1jEIvHoTbsrdMYa25DvAZ2_dwatamcCVpHTGa1f9KOTqEgOrkFy-CrJsVfYDu9kQ4rRTTiT3i6tbUZocE2OY3sxXa-gXd8ghVyIF6ON_Dd-8h40tAM-33Pllftduf3w89Phc7vJfUhS4deCYgLAlEZdp5LGkNY61WGMnNVAebxYJcCE0FxxhkiVplwzJjIpQ0lrvvJe_7A38ersTAdurn4PqG4H8B8K9lZn</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Examining User-Friendly and Open-Sourced Large GPT Models: A Survey on Language, Multimodal, and Scientific GPT Models</title><source>arXiv.org</source><creator>Gao, Kaiyuan ; He, Sunan ; He, Zhenyu ; Lin, Jiacheng ; Pei, QiZhi ; Shao, Jie ; Zhang, Wei</creator><creatorcontrib>Gao, Kaiyuan ; He, Sunan ; He, Zhenyu ; Lin, Jiacheng ; Pei, QiZhi ; Shao, Jie ; Zhang, Wei</creatorcontrib><description>Generative pre-trained transformer (GPT) models have revolutionized the field of natural language processing (NLP) with remarkable performance in various tasks and also extend their power to multimodal domains. Despite their success, large GPT models like GPT-4 face inherent limitations such as considerable size, high computational requirements, complex deployment processes, and closed development loops. These constraints restrict their widespread adoption and raise concerns regarding their responsible development and usage. The need for user-friendly, relatively small, and open-sourced alternative GPT models arises from the desire to overcome these limitations while retaining high performance. In this survey paper, we provide an examination of alternative open-sourced models of large GPTs, focusing on user-friendly and relatively small models that facilitate easier deployment and accessibility. Through this extensive survey, we aim to equip researchers, practitioners, and enthusiasts with a thorough understanding of user-friendly and relatively small open-sourced models of large GPTs, their current state, challenges, and future research directions, inspiring the development of more efficient, accessible, and versatile GPT models that cater to the broader scientific community and advance the field of general artificial intelligence. The source contents are continuously updating in https://github.com/GPT-Alternatives/gpt_alternatives.</description><identifier>DOI: 10.48550/arxiv.2308.14149</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Computation and Language</subject><creationdate>2023-08</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,778,883</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2308.14149$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2308.14149$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Gao, Kaiyuan</creatorcontrib><creatorcontrib>He, Sunan</creatorcontrib><creatorcontrib>He, Zhenyu</creatorcontrib><creatorcontrib>Lin, Jiacheng</creatorcontrib><creatorcontrib>Pei, QiZhi</creatorcontrib><creatorcontrib>Shao, Jie</creatorcontrib><creatorcontrib>Zhang, Wei</creatorcontrib><title>Examining User-Friendly and Open-Sourced Large GPT Models: A Survey on Language, Multimodal, and Scientific GPT Models</title><description>Generative pre-trained transformer (GPT) models have revolutionized the field of natural language processing (NLP) with remarkable performance in various tasks and also extend their power to multimodal domains. Despite their success, large GPT models like GPT-4 face inherent limitations such as considerable size, high computational requirements, complex deployment processes, and closed development loops. These constraints restrict their widespread adoption and raise concerns regarding their responsible development and usage. The need for user-friendly, relatively small, and open-sourced alternative GPT models arises from the desire to overcome these limitations while retaining high performance. In this survey paper, we provide an examination of alternative open-sourced models of large GPTs, focusing on user-friendly and relatively small models that facilitate easier deployment and accessibility. Through this extensive survey, we aim to equip researchers, practitioners, and enthusiasts with a thorough understanding of user-friendly and relatively small open-sourced models of large GPTs, their current state, challenges, and future research directions, inspiring the development of more efficient, accessible, and versatile GPT models that cater to the broader scientific community and advance the field of general artificial intelligence. The source contents are continuously updating in https://github.com/GPT-Alternatives/gpt_alternatives.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Computation and Language</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNpNkL1ugzAUhVk6VGkfoFP9AIHamD93i6IkrUSUStAZXexrZAlMZAIKb1-aZuh0hqPzSd_xvBdGgyiLY_oG7mqmIOQ0C1jEIvHoTbsrdMYa25DvAZ2_dwatamcCVpHTGa1f9KOTqEgOrkFy-CrJsVfYDu9kQ4rRTTiT3i6tbUZocE2OY3sxXa-gXd8ghVyIF6ON_Dd-8h40tAM-33Pllftduf3w89Phc7vJfUhS4deCYgLAlEZdp5LGkNY61WGMnNVAebxYJcCE0FxxhkiVplwzJjIpQ0lrvvJe_7A38ersTAdurn4PqG4H8B8K9lZn</recordid><startdate>20230827</startdate><enddate>20230827</enddate><creator>Gao, Kaiyuan</creator><creator>He, Sunan</creator><creator>He, Zhenyu</creator><creator>Lin, Jiacheng</creator><creator>Pei, QiZhi</creator><creator>Shao, Jie</creator><creator>Zhang, Wei</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20230827</creationdate><title>Examining User-Friendly and Open-Sourced Large GPT Models: A Survey on Language, Multimodal, and Scientific GPT Models</title><author>Gao, Kaiyuan ; He, Sunan ; He, Zhenyu ; Lin, Jiacheng ; Pei, QiZhi ; Shao, Jie ; Zhang, Wei</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a679-b90e6aa1dfefb7c05a7bf7f25e31ba0354856a199f3d31ee0df03f1198cc2c0b3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Computation and Language</topic><toplevel>online_resources</toplevel><creatorcontrib>Gao, Kaiyuan</creatorcontrib><creatorcontrib>He, Sunan</creatorcontrib><creatorcontrib>He, Zhenyu</creatorcontrib><creatorcontrib>Lin, Jiacheng</creatorcontrib><creatorcontrib>Pei, QiZhi</creatorcontrib><creatorcontrib>Shao, Jie</creatorcontrib><creatorcontrib>Zhang, Wei</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Gao, Kaiyuan</au><au>He, Sunan</au><au>He, Zhenyu</au><au>Lin, Jiacheng</au><au>Pei, QiZhi</au><au>Shao, Jie</au><au>Zhang, Wei</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Examining User-Friendly and Open-Sourced Large GPT Models: A Survey on Language, Multimodal, and Scientific GPT Models</atitle><date>2023-08-27</date><risdate>2023</risdate><abstract>Generative pre-trained transformer (GPT) models have revolutionized the field of natural language processing (NLP) with remarkable performance in various tasks and also extend their power to multimodal domains. Despite their success, large GPT models like GPT-4 face inherent limitations such as considerable size, high computational requirements, complex deployment processes, and closed development loops. These constraints restrict their widespread adoption and raise concerns regarding their responsible development and usage. The need for user-friendly, relatively small, and open-sourced alternative GPT models arises from the desire to overcome these limitations while retaining high performance. In this survey paper, we provide an examination of alternative open-sourced models of large GPTs, focusing on user-friendly and relatively small models that facilitate easier deployment and accessibility. Through this extensive survey, we aim to equip researchers, practitioners, and enthusiasts with a thorough understanding of user-friendly and relatively small open-sourced models of large GPTs, their current state, challenges, and future research directions, inspiring the development of more efficient, accessible, and versatile GPT models that cater to the broader scientific community and advance the field of general artificial intelligence. The source contents are continuously updating in https://github.com/GPT-Alternatives/gpt_alternatives.</abstract><doi>10.48550/arxiv.2308.14149</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2308.14149
ispartof
issn
language eng
recordid cdi_arxiv_primary_2308_14149
source arXiv.org
subjects Computer Science - Artificial Intelligence
Computer Science - Computation and Language
title Examining User-Friendly and Open-Sourced Large GPT Models: A Survey on Language, Multimodal, and Scientific GPT Models
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-15T12%3A48%3A44IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Examining%20User-Friendly%20and%20Open-Sourced%20Large%20GPT%20Models:%20A%20Survey%20on%20Language,%20Multimodal,%20and%20Scientific%20GPT%20Models&rft.au=Gao,%20Kaiyuan&rft.date=2023-08-27&rft_id=info:doi/10.48550/arxiv.2308.14149&rft_dat=%3Carxiv_GOX%3E2308_14149%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true