Towards Publicly Accountable Frontier LLMs: Building an External Scrutiny Ecosystem under the ASPIRE Framework
With the increasing integration of frontier large language models (LLMs) into society and the economy, decisions related to their training, deployment, and use have far-reaching implications. These decisions should not be left solely in the hands of frontier LLM developers. LLM users, civil society...
Gespeichert in:
Hauptverfasser: | , , , , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Anderljung, Markus Smith, Everett Thornton O'Brien, Joe Soder, Lisa Bucknall, Benjamin Bluemke, Emma Schuett, Jonas Trager, Robert Strahm, Lacey Chowdhury, Rumman |
description | With the increasing integration of frontier large language models (LLMs) into
society and the economy, decisions related to their training, deployment, and
use have far-reaching implications. These decisions should not be left solely
in the hands of frontier LLM developers. LLM users, civil society and
policymakers need trustworthy sources of information to steer such decisions
for the better. Involving outside actors in the evaluation of these systems -
what we term 'external scrutiny' - via red-teaming, auditing, and external
researcher access, offers a solution. Though there are encouraging signs of
increasing external scrutiny of frontier LLMs, its success is not assured. In
this paper, we survey six requirements for effective external scrutiny of
frontier AI systems and organize them under the ASPIRE framework: Access,
Searching attitude, Proportionality to the risks, Independence, Resources, and
Expertise. We then illustrate how external scrutiny might function throughout
the AI lifecycle and offer recommendations to policymakers. |
doi_str_mv | 10.48550/arxiv.2311.14711 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2311_14711</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2311_14711</sourcerecordid><originalsourceid>FETCH-LOGICAL-a671-d61a7f0bafd4f9586af53aeec5c1efe95c3fd9385c5370ee4de6a73f7c39f1703</originalsourceid><addsrcrecordid>eNotz0FPgzAYxnEuHsz0A3iyXwCkllLwhgvTJRgXx528tG-1sRRTihvf3jk9Pafnn_yi6IamSVZwnt6BP5rv5J5RmtBMUHoZuXY8gFcT2c29NdIupJJynF2A3iLZ-NEFg540zcv0QB5nY5Vx7wQcqY8BvQNL9tLPwbiF1HKclingQGanTp_wgaTa77Zv9akDAx5G_3kVXWiwE17_7ypqN3W7fo6b16ftumpiyAWNVU5B6LQHrTJd8iIHzRkgSi4paiy5ZFqVrOCSM5EiZgpzEEwLyUpNRcpW0e1f9gzuvrwZwC_dL7w7w9kPMkpVSQ</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Towards Publicly Accountable Frontier LLMs: Building an External Scrutiny Ecosystem under the ASPIRE Framework</title><source>arXiv.org</source><creator>Anderljung, Markus ; Smith, Everett Thornton ; O'Brien, Joe ; Soder, Lisa ; Bucknall, Benjamin ; Bluemke, Emma ; Schuett, Jonas ; Trager, Robert ; Strahm, Lacey ; Chowdhury, Rumman</creator><creatorcontrib>Anderljung, Markus ; Smith, Everett Thornton ; O'Brien, Joe ; Soder, Lisa ; Bucknall, Benjamin ; Bluemke, Emma ; Schuett, Jonas ; Trager, Robert ; Strahm, Lacey ; Chowdhury, Rumman</creatorcontrib><description>With the increasing integration of frontier large language models (LLMs) into
society and the economy, decisions related to their training, deployment, and
use have far-reaching implications. These decisions should not be left solely
in the hands of frontier LLM developers. LLM users, civil society and
policymakers need trustworthy sources of information to steer such decisions
for the better. Involving outside actors in the evaluation of these systems -
what we term 'external scrutiny' - via red-teaming, auditing, and external
researcher access, offers a solution. Though there are encouraging signs of
increasing external scrutiny of frontier LLMs, its success is not assured. In
this paper, we survey six requirements for effective external scrutiny of
frontier AI systems and organize them under the ASPIRE framework: Access,
Searching attitude, Proportionality to the risks, Independence, Resources, and
Expertise. We then illustrate how external scrutiny might function throughout
the AI lifecycle and offer recommendations to policymakers.</description><identifier>DOI: 10.48550/arxiv.2311.14711</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Computers and Society</subject><creationdate>2023-11</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2311.14711$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2311.14711$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Anderljung, Markus</creatorcontrib><creatorcontrib>Smith, Everett Thornton</creatorcontrib><creatorcontrib>O'Brien, Joe</creatorcontrib><creatorcontrib>Soder, Lisa</creatorcontrib><creatorcontrib>Bucknall, Benjamin</creatorcontrib><creatorcontrib>Bluemke, Emma</creatorcontrib><creatorcontrib>Schuett, Jonas</creatorcontrib><creatorcontrib>Trager, Robert</creatorcontrib><creatorcontrib>Strahm, Lacey</creatorcontrib><creatorcontrib>Chowdhury, Rumman</creatorcontrib><title>Towards Publicly Accountable Frontier LLMs: Building an External Scrutiny Ecosystem under the ASPIRE Framework</title><description>With the increasing integration of frontier large language models (LLMs) into
society and the economy, decisions related to their training, deployment, and
use have far-reaching implications. These decisions should not be left solely
in the hands of frontier LLM developers. LLM users, civil society and
policymakers need trustworthy sources of information to steer such decisions
for the better. Involving outside actors in the evaluation of these systems -
what we term 'external scrutiny' - via red-teaming, auditing, and external
researcher access, offers a solution. Though there are encouraging signs of
increasing external scrutiny of frontier LLMs, its success is not assured. In
this paper, we survey six requirements for effective external scrutiny of
frontier AI systems and organize them under the ASPIRE framework: Access,
Searching attitude, Proportionality to the risks, Independence, Resources, and
Expertise. We then illustrate how external scrutiny might function throughout
the AI lifecycle and offer recommendations to policymakers.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Computers and Society</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotz0FPgzAYxnEuHsz0A3iyXwCkllLwhgvTJRgXx528tG-1sRRTihvf3jk9Pafnn_yi6IamSVZwnt6BP5rv5J5RmtBMUHoZuXY8gFcT2c29NdIupJJynF2A3iLZ-NEFg540zcv0QB5nY5Vx7wQcqY8BvQNL9tLPwbiF1HKclingQGanTp_wgaTa77Zv9akDAx5G_3kVXWiwE17_7ypqN3W7fo6b16ftumpiyAWNVU5B6LQHrTJd8iIHzRkgSi4paiy5ZFqVrOCSM5EiZgpzEEwLyUpNRcpW0e1f9gzuvrwZwC_dL7w7w9kPMkpVSQ</recordid><startdate>20231115</startdate><enddate>20231115</enddate><creator>Anderljung, Markus</creator><creator>Smith, Everett Thornton</creator><creator>O'Brien, Joe</creator><creator>Soder, Lisa</creator><creator>Bucknall, Benjamin</creator><creator>Bluemke, Emma</creator><creator>Schuett, Jonas</creator><creator>Trager, Robert</creator><creator>Strahm, Lacey</creator><creator>Chowdhury, Rumman</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20231115</creationdate><title>Towards Publicly Accountable Frontier LLMs: Building an External Scrutiny Ecosystem under the ASPIRE Framework</title><author>Anderljung, Markus ; Smith, Everett Thornton ; O'Brien, Joe ; Soder, Lisa ; Bucknall, Benjamin ; Bluemke, Emma ; Schuett, Jonas ; Trager, Robert ; Strahm, Lacey ; Chowdhury, Rumman</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a671-d61a7f0bafd4f9586af53aeec5c1efe95c3fd9385c5370ee4de6a73f7c39f1703</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Computers and Society</topic><toplevel>online_resources</toplevel><creatorcontrib>Anderljung, Markus</creatorcontrib><creatorcontrib>Smith, Everett Thornton</creatorcontrib><creatorcontrib>O'Brien, Joe</creatorcontrib><creatorcontrib>Soder, Lisa</creatorcontrib><creatorcontrib>Bucknall, Benjamin</creatorcontrib><creatorcontrib>Bluemke, Emma</creatorcontrib><creatorcontrib>Schuett, Jonas</creatorcontrib><creatorcontrib>Trager, Robert</creatorcontrib><creatorcontrib>Strahm, Lacey</creatorcontrib><creatorcontrib>Chowdhury, Rumman</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Anderljung, Markus</au><au>Smith, Everett Thornton</au><au>O'Brien, Joe</au><au>Soder, Lisa</au><au>Bucknall, Benjamin</au><au>Bluemke, Emma</au><au>Schuett, Jonas</au><au>Trager, Robert</au><au>Strahm, Lacey</au><au>Chowdhury, Rumman</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Towards Publicly Accountable Frontier LLMs: Building an External Scrutiny Ecosystem under the ASPIRE Framework</atitle><date>2023-11-15</date><risdate>2023</risdate><abstract>With the increasing integration of frontier large language models (LLMs) into
society and the economy, decisions related to their training, deployment, and
use have far-reaching implications. These decisions should not be left solely
in the hands of frontier LLM developers. LLM users, civil society and
policymakers need trustworthy sources of information to steer such decisions
for the better. Involving outside actors in the evaluation of these systems -
what we term 'external scrutiny' - via red-teaming, auditing, and external
researcher access, offers a solution. Though there are encouraging signs of
increasing external scrutiny of frontier LLMs, its success is not assured. In
this paper, we survey six requirements for effective external scrutiny of
frontier AI systems and organize them under the ASPIRE framework: Access,
Searching attitude, Proportionality to the risks, Independence, Resources, and
Expertise. We then illustrate how external scrutiny might function throughout
the AI lifecycle and offer recommendations to policymakers.</abstract><doi>10.48550/arxiv.2311.14711</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2311.14711 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2311_14711 |
source | arXiv.org |
subjects | Computer Science - Artificial Intelligence Computer Science - Computers and Society |
title | Towards Publicly Accountable Frontier LLMs: Building an External Scrutiny Ecosystem under the ASPIRE Framework |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-28T09%3A55%3A48IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Towards%20Publicly%20Accountable%20Frontier%20LLMs:%20Building%20an%20External%20Scrutiny%20Ecosystem%20under%20the%20ASPIRE%20Framework&rft.au=Anderljung,%20Markus&rft.date=2023-11-15&rft_id=info:doi/10.48550/arxiv.2311.14711&rft_dat=%3Carxiv_GOX%3E2311_14711%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |