Towards Publicly Accountable Frontier LLMs: Building an External Scrutiny Ecosystem under the ASPIRE Framework

With the increasing integration of frontier large language models (LLMs) into society and the economy, decisions related to their training, deployment, and use have far-reaching implications. These decisions should not be left solely in the hands of frontier LLM developers. LLM users, civil society...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Anderljung, Markus, Smith, Everett Thornton, O'Brien, Joe, Soder, Lisa, Bucknall, Benjamin, Bluemke, Emma, Schuett, Jonas, Trager, Robert, Strahm, Lacey, Chowdhury, Rumman
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Anderljung, Markus
Smith, Everett Thornton
O'Brien, Joe
Soder, Lisa
Bucknall, Benjamin
Bluemke, Emma
Schuett, Jonas
Trager, Robert
Strahm, Lacey
Chowdhury, Rumman
description With the increasing integration of frontier large language models (LLMs) into society and the economy, decisions related to their training, deployment, and use have far-reaching implications. These decisions should not be left solely in the hands of frontier LLM developers. LLM users, civil society and policymakers need trustworthy sources of information to steer such decisions for the better. Involving outside actors in the evaluation of these systems - what we term 'external scrutiny' - via red-teaming, auditing, and external researcher access, offers a solution. Though there are encouraging signs of increasing external scrutiny of frontier LLMs, its success is not assured. In this paper, we survey six requirements for effective external scrutiny of frontier AI systems and organize them under the ASPIRE framework: Access, Searching attitude, Proportionality to the risks, Independence, Resources, and Expertise. We then illustrate how external scrutiny might function throughout the AI lifecycle and offer recommendations to policymakers.
doi_str_mv 10.48550/arxiv.2311.14711
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2311_14711</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2311_14711</sourcerecordid><originalsourceid>FETCH-LOGICAL-a671-d61a7f0bafd4f9586af53aeec5c1efe95c3fd9385c5370ee4de6a73f7c39f1703</originalsourceid><addsrcrecordid>eNotz0FPgzAYxnEuHsz0A3iyXwCkllLwhgvTJRgXx528tG-1sRRTihvf3jk9Pafnn_yi6IamSVZwnt6BP5rv5J5RmtBMUHoZuXY8gFcT2c29NdIupJJynF2A3iLZ-NEFg540zcv0QB5nY5Vx7wQcqY8BvQNL9tLPwbiF1HKclingQGanTp_wgaTa77Zv9akDAx5G_3kVXWiwE17_7ypqN3W7fo6b16ftumpiyAWNVU5B6LQHrTJd8iIHzRkgSi4paiy5ZFqVrOCSM5EiZgpzEEwLyUpNRcpW0e1f9gzuvrwZwC_dL7w7w9kPMkpVSQ</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Towards Publicly Accountable Frontier LLMs: Building an External Scrutiny Ecosystem under the ASPIRE Framework</title><source>arXiv.org</source><creator>Anderljung, Markus ; Smith, Everett Thornton ; O'Brien, Joe ; Soder, Lisa ; Bucknall, Benjamin ; Bluemke, Emma ; Schuett, Jonas ; Trager, Robert ; Strahm, Lacey ; Chowdhury, Rumman</creator><creatorcontrib>Anderljung, Markus ; Smith, Everett Thornton ; O'Brien, Joe ; Soder, Lisa ; Bucknall, Benjamin ; Bluemke, Emma ; Schuett, Jonas ; Trager, Robert ; Strahm, Lacey ; Chowdhury, Rumman</creatorcontrib><description>With the increasing integration of frontier large language models (LLMs) into society and the economy, decisions related to their training, deployment, and use have far-reaching implications. These decisions should not be left solely in the hands of frontier LLM developers. LLM users, civil society and policymakers need trustworthy sources of information to steer such decisions for the better. Involving outside actors in the evaluation of these systems - what we term 'external scrutiny' - via red-teaming, auditing, and external researcher access, offers a solution. Though there are encouraging signs of increasing external scrutiny of frontier LLMs, its success is not assured. In this paper, we survey six requirements for effective external scrutiny of frontier AI systems and organize them under the ASPIRE framework: Access, Searching attitude, Proportionality to the risks, Independence, Resources, and Expertise. We then illustrate how external scrutiny might function throughout the AI lifecycle and offer recommendations to policymakers.</description><identifier>DOI: 10.48550/arxiv.2311.14711</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Computers and Society</subject><creationdate>2023-11</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2311.14711$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2311.14711$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Anderljung, Markus</creatorcontrib><creatorcontrib>Smith, Everett Thornton</creatorcontrib><creatorcontrib>O'Brien, Joe</creatorcontrib><creatorcontrib>Soder, Lisa</creatorcontrib><creatorcontrib>Bucknall, Benjamin</creatorcontrib><creatorcontrib>Bluemke, Emma</creatorcontrib><creatorcontrib>Schuett, Jonas</creatorcontrib><creatorcontrib>Trager, Robert</creatorcontrib><creatorcontrib>Strahm, Lacey</creatorcontrib><creatorcontrib>Chowdhury, Rumman</creatorcontrib><title>Towards Publicly Accountable Frontier LLMs: Building an External Scrutiny Ecosystem under the ASPIRE Framework</title><description>With the increasing integration of frontier large language models (LLMs) into society and the economy, decisions related to their training, deployment, and use have far-reaching implications. These decisions should not be left solely in the hands of frontier LLM developers. LLM users, civil society and policymakers need trustworthy sources of information to steer such decisions for the better. Involving outside actors in the evaluation of these systems - what we term 'external scrutiny' - via red-teaming, auditing, and external researcher access, offers a solution. Though there are encouraging signs of increasing external scrutiny of frontier LLMs, its success is not assured. In this paper, we survey six requirements for effective external scrutiny of frontier AI systems and organize them under the ASPIRE framework: Access, Searching attitude, Proportionality to the risks, Independence, Resources, and Expertise. We then illustrate how external scrutiny might function throughout the AI lifecycle and offer recommendations to policymakers.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Computers and Society</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotz0FPgzAYxnEuHsz0A3iyXwCkllLwhgvTJRgXx528tG-1sRRTihvf3jk9Pafnn_yi6IamSVZwnt6BP5rv5J5RmtBMUHoZuXY8gFcT2c29NdIupJJynF2A3iLZ-NEFg540zcv0QB5nY5Vx7wQcqY8BvQNL9tLPwbiF1HKclingQGanTp_wgaTa77Zv9akDAx5G_3kVXWiwE17_7ypqN3W7fo6b16ftumpiyAWNVU5B6LQHrTJd8iIHzRkgSi4paiy5ZFqVrOCSM5EiZgpzEEwLyUpNRcpW0e1f9gzuvrwZwC_dL7w7w9kPMkpVSQ</recordid><startdate>20231115</startdate><enddate>20231115</enddate><creator>Anderljung, Markus</creator><creator>Smith, Everett Thornton</creator><creator>O'Brien, Joe</creator><creator>Soder, Lisa</creator><creator>Bucknall, Benjamin</creator><creator>Bluemke, Emma</creator><creator>Schuett, Jonas</creator><creator>Trager, Robert</creator><creator>Strahm, Lacey</creator><creator>Chowdhury, Rumman</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20231115</creationdate><title>Towards Publicly Accountable Frontier LLMs: Building an External Scrutiny Ecosystem under the ASPIRE Framework</title><author>Anderljung, Markus ; Smith, Everett Thornton ; O'Brien, Joe ; Soder, Lisa ; Bucknall, Benjamin ; Bluemke, Emma ; Schuett, Jonas ; Trager, Robert ; Strahm, Lacey ; Chowdhury, Rumman</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a671-d61a7f0bafd4f9586af53aeec5c1efe95c3fd9385c5370ee4de6a73f7c39f1703</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Computers and Society</topic><toplevel>online_resources</toplevel><creatorcontrib>Anderljung, Markus</creatorcontrib><creatorcontrib>Smith, Everett Thornton</creatorcontrib><creatorcontrib>O'Brien, Joe</creatorcontrib><creatorcontrib>Soder, Lisa</creatorcontrib><creatorcontrib>Bucknall, Benjamin</creatorcontrib><creatorcontrib>Bluemke, Emma</creatorcontrib><creatorcontrib>Schuett, Jonas</creatorcontrib><creatorcontrib>Trager, Robert</creatorcontrib><creatorcontrib>Strahm, Lacey</creatorcontrib><creatorcontrib>Chowdhury, Rumman</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Anderljung, Markus</au><au>Smith, Everett Thornton</au><au>O'Brien, Joe</au><au>Soder, Lisa</au><au>Bucknall, Benjamin</au><au>Bluemke, Emma</au><au>Schuett, Jonas</au><au>Trager, Robert</au><au>Strahm, Lacey</au><au>Chowdhury, Rumman</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Towards Publicly Accountable Frontier LLMs: Building an External Scrutiny Ecosystem under the ASPIRE Framework</atitle><date>2023-11-15</date><risdate>2023</risdate><abstract>With the increasing integration of frontier large language models (LLMs) into society and the economy, decisions related to their training, deployment, and use have far-reaching implications. These decisions should not be left solely in the hands of frontier LLM developers. LLM users, civil society and policymakers need trustworthy sources of information to steer such decisions for the better. Involving outside actors in the evaluation of these systems - what we term 'external scrutiny' - via red-teaming, auditing, and external researcher access, offers a solution. Though there are encouraging signs of increasing external scrutiny of frontier LLMs, its success is not assured. In this paper, we survey six requirements for effective external scrutiny of frontier AI systems and organize them under the ASPIRE framework: Access, Searching attitude, Proportionality to the risks, Independence, Resources, and Expertise. We then illustrate how external scrutiny might function throughout the AI lifecycle and offer recommendations to policymakers.</abstract><doi>10.48550/arxiv.2311.14711</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2311.14711
ispartof
issn
language eng
recordid cdi_arxiv_primary_2311_14711
source arXiv.org
subjects Computer Science - Artificial Intelligence
Computer Science - Computers and Society
title Towards Publicly Accountable Frontier LLMs: Building an External Scrutiny Ecosystem under the ASPIRE Framework
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-28T09%3A55%3A48IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Towards%20Publicly%20Accountable%20Frontier%20LLMs:%20Building%20an%20External%20Scrutiny%20Ecosystem%20under%20the%20ASPIRE%20Framework&rft.au=Anderljung,%20Markus&rft.date=2023-11-15&rft_id=info:doi/10.48550/arxiv.2311.14711&rft_dat=%3Carxiv_GOX%3E2311_14711%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true