Ethics and Governance of Artificial Intelligence: Evidence from a Survey of Machine Learning Researchers

Machine learning (ML) and artificial intelligence (AI) researchers play an important role in the ethics and governance of AI, including through their work, advocacy, and choice of employment. Nevertheless, this influential group's attitudes are not well understood, undermining our ability to di...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:The Journal of artificial intelligence research 2021, Vol.71, p.591
Hauptverfasser: Zhang, Baobao, Anderljung, Markus, Kahn, Lauren, Dreksler, Noemi, Horowitz, Michael C., Dafoe, Allan
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page 591
container_title The Journal of artificial intelligence research
container_volume 71
creator Zhang, Baobao
Anderljung, Markus
Kahn, Lauren
Dreksler, Noemi
Horowitz, Michael C.
Dafoe, Allan
description Machine learning (ML) and artificial intelligence (AI) researchers play an important role in the ethics and governance of AI, including through their work, advocacy, and choice of employment. Nevertheless, this influential group's attitudes are not well understood, undermining our ability to discern consensuses or disagreements between AI/ML researchers. To examine these researchers' views, we conducted a survey of those who published in two top AI/ML conferences (N = 524). We compare these results with those from a 2016 survey of AI/ML researchers (Grace et al., 2018) and a 2018 survey of the US public (Zhang & Dafoe, 2020). We find that AI/ML researchers place high levels of trust in international organizations and scientific organizations to shape the development and use of AI in the public interest; moderate trust in most Western tech companies; and low trust in national militaries, Chinese tech companies, and Facebook. While the respondents were overwhelmingly opposed to AI/ML researchers working on lethal autonomous weapons, they are less opposed to researchers working on other military applications of AI, particularly logistics algorithms. A strong majority of respondents think that AI safety research should be prioritized and that ML institutions should conduct pre-publication review to assess potential harms. Being closer to the technology itself, AI/ML researchers are well placed to highlight new risks and develop technical solutions, so this novel attempt to measure their attitudes has broad relevance. The findings should help to improve how researchers, private sector executives, and policymakers think about regulations, governance frameworks, guiding principles, and national and international governance strategies for AI. This article appears in the special track on AI & Society.
doi_str_mv 10.1613/jair.1.12895
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2567811157</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2567811157</sourcerecordid><originalsourceid>FETCH-LOGICAL-c178t-d142b291f8dc3bd8a1f3a1e289da47cbdad120c77415d444eab4b315a3cfd9463</originalsourceid><addsrcrecordid>eNpNkE1Lw0AQhhdRsFZv_oAFr6Zmkk028VZKWwsVwY_zMtmPZku6qbtpof_exHrwNA_MOwPvQ8g9xBPIIX3aovUTmEBSlNkFGUHM86jkGb_8x9fkJoRtHEPJkmJE6nlXWxkoOkWX7VF7h05q2ho69Z01Vlps6Mp1umnsRverZzo_WjUQNb7dUaQfB3_Up-HkFWVtnaZrjd5Zt6HvOvQoa-3DLbky2AR99zfH5Gsx_5y9ROu35Wo2XUcSeNFFClhSJSWYQsm0UgWCSRF030gh47JSqCCJJecMMsUY01ixKoUMU2lUyfJ0TB7Of_e-_T7o0Ilte-hbNUEkWc4LAMh4n3o8p6RvQ_DaiL23O_QnAbEYXIrBpQDx6zL9AWooaKU</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2567811157</pqid></control><display><type>article</type><title>Ethics and Governance of Artificial Intelligence: Evidence from a Survey of Machine Learning Researchers</title><source>DOAJ Directory of Open Access Journals</source><source>Elektronische Zeitschriftenbibliothek - Frei zugängliche E-Journals</source><source>Free E- Journals</source><creator>Zhang, Baobao ; Anderljung, Markus ; Kahn, Lauren ; Dreksler, Noemi ; Horowitz, Michael C. ; Dafoe, Allan</creator><creatorcontrib>Zhang, Baobao ; Anderljung, Markus ; Kahn, Lauren ; Dreksler, Noemi ; Horowitz, Michael C. ; Dafoe, Allan</creatorcontrib><description>Machine learning (ML) and artificial intelligence (AI) researchers play an important role in the ethics and governance of AI, including through their work, advocacy, and choice of employment. Nevertheless, this influential group's attitudes are not well understood, undermining our ability to discern consensuses or disagreements between AI/ML researchers. To examine these researchers' views, we conducted a survey of those who published in two top AI/ML conferences (N = 524). We compare these results with those from a 2016 survey of AI/ML researchers (Grace et al., 2018) and a 2018 survey of the US public (Zhang &amp; Dafoe, 2020). We find that AI/ML researchers place high levels of trust in international organizations and scientific organizations to shape the development and use of AI in the public interest; moderate trust in most Western tech companies; and low trust in national militaries, Chinese tech companies, and Facebook. While the respondents were overwhelmingly opposed to AI/ML researchers working on lethal autonomous weapons, they are less opposed to researchers working on other military applications of AI, particularly logistics algorithms. A strong majority of respondents think that AI safety research should be prioritized and that ML institutions should conduct pre-publication review to assess potential harms. Being closer to the technology itself, AI/ML researchers are well placed to highlight new risks and develop technical solutions, so this novel attempt to measure their attitudes has broad relevance. The findings should help to improve how researchers, private sector executives, and policymakers think about regulations, governance frameworks, guiding principles, and national and international governance strategies for AI. This article appears in the special track on AI &amp; Society.</description><identifier>ISSN: 1076-9757</identifier><identifier>EISSN: 1076-9757</identifier><identifier>EISSN: 1943-5037</identifier><identifier>DOI: 10.1613/jair.1.12895</identifier><language>eng</language><publisher>San Francisco: AI Access Foundation</publisher><subject>Algorithms ; Artificial intelligence ; Attitudes ; Ethics ; International organizations ; Logistics ; Machine learning ; Military applications ; Researchers</subject><ispartof>The Journal of artificial intelligence research, 2021, Vol.71, p.591</ispartof><rights>2021. Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the associated terms available at https://www.jair.org/index.php/jair/about</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c178t-d142b291f8dc3bd8a1f3a1e289da47cbdad120c77415d444eab4b315a3cfd9463</citedby></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,780,784,864,4024,27923,27924,27925</link.rule.ids></links><search><creatorcontrib>Zhang, Baobao</creatorcontrib><creatorcontrib>Anderljung, Markus</creatorcontrib><creatorcontrib>Kahn, Lauren</creatorcontrib><creatorcontrib>Dreksler, Noemi</creatorcontrib><creatorcontrib>Horowitz, Michael C.</creatorcontrib><creatorcontrib>Dafoe, Allan</creatorcontrib><title>Ethics and Governance of Artificial Intelligence: Evidence from a Survey of Machine Learning Researchers</title><title>The Journal of artificial intelligence research</title><description>Machine learning (ML) and artificial intelligence (AI) researchers play an important role in the ethics and governance of AI, including through their work, advocacy, and choice of employment. Nevertheless, this influential group's attitudes are not well understood, undermining our ability to discern consensuses or disagreements between AI/ML researchers. To examine these researchers' views, we conducted a survey of those who published in two top AI/ML conferences (N = 524). We compare these results with those from a 2016 survey of AI/ML researchers (Grace et al., 2018) and a 2018 survey of the US public (Zhang &amp; Dafoe, 2020). We find that AI/ML researchers place high levels of trust in international organizations and scientific organizations to shape the development and use of AI in the public interest; moderate trust in most Western tech companies; and low trust in national militaries, Chinese tech companies, and Facebook. While the respondents were overwhelmingly opposed to AI/ML researchers working on lethal autonomous weapons, they are less opposed to researchers working on other military applications of AI, particularly logistics algorithms. A strong majority of respondents think that AI safety research should be prioritized and that ML institutions should conduct pre-publication review to assess potential harms. Being closer to the technology itself, AI/ML researchers are well placed to highlight new risks and develop technical solutions, so this novel attempt to measure their attitudes has broad relevance. The findings should help to improve how researchers, private sector executives, and policymakers think about regulations, governance frameworks, guiding principles, and national and international governance strategies for AI. This article appears in the special track on AI &amp; Society.</description><subject>Algorithms</subject><subject>Artificial intelligence</subject><subject>Attitudes</subject><subject>Ethics</subject><subject>International organizations</subject><subject>Logistics</subject><subject>Machine learning</subject><subject>Military applications</subject><subject>Researchers</subject><issn>1076-9757</issn><issn>1076-9757</issn><issn>1943-5037</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><sourceid>GNUQQ</sourceid><recordid>eNpNkE1Lw0AQhhdRsFZv_oAFr6Zmkk028VZKWwsVwY_zMtmPZku6qbtpof_exHrwNA_MOwPvQ8g9xBPIIX3aovUTmEBSlNkFGUHM86jkGb_8x9fkJoRtHEPJkmJE6nlXWxkoOkWX7VF7h05q2ho69Z01Vlps6Mp1umnsRverZzo_WjUQNb7dUaQfB3_Up-HkFWVtnaZrjd5Zt6HvOvQoa-3DLbky2AR99zfH5Gsx_5y9ROu35Wo2XUcSeNFFClhSJSWYQsm0UgWCSRF030gh47JSqCCJJecMMsUY01ixKoUMU2lUyfJ0TB7Of_e-_T7o0Ilte-hbNUEkWc4LAMh4n3o8p6RvQ_DaiL23O_QnAbEYXIrBpQDx6zL9AWooaKU</recordid><startdate>2021</startdate><enddate>2021</enddate><creator>Zhang, Baobao</creator><creator>Anderljung, Markus</creator><creator>Kahn, Lauren</creator><creator>Dreksler, Noemi</creator><creator>Horowitz, Michael C.</creator><creator>Dafoe, Allan</creator><general>AI Access Foundation</general><scope>AAYXX</scope><scope>CITATION</scope><scope>8FE</scope><scope>8FG</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>ARAPS</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>COVID</scope><scope>DWQXO</scope><scope>GNUQQ</scope><scope>HCIFZ</scope><scope>JQ2</scope><scope>K7-</scope><scope>P62</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope></search><sort><creationdate>2021</creationdate><title>Ethics and Governance of Artificial Intelligence: Evidence from a Survey of Machine Learning Researchers</title><author>Zhang, Baobao ; Anderljung, Markus ; Kahn, Lauren ; Dreksler, Noemi ; Horowitz, Michael C. ; Dafoe, Allan</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c178t-d142b291f8dc3bd8a1f3a1e289da47cbdad120c77415d444eab4b315a3cfd9463</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Algorithms</topic><topic>Artificial intelligence</topic><topic>Attitudes</topic><topic>Ethics</topic><topic>International organizations</topic><topic>Logistics</topic><topic>Machine learning</topic><topic>Military applications</topic><topic>Researchers</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Zhang, Baobao</creatorcontrib><creatorcontrib>Anderljung, Markus</creatorcontrib><creatorcontrib>Kahn, Lauren</creatorcontrib><creatorcontrib>Dreksler, Noemi</creatorcontrib><creatorcontrib>Horowitz, Michael C.</creatorcontrib><creatorcontrib>Dafoe, Allan</creatorcontrib><collection>CrossRef</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>Advanced Technologies &amp; Aerospace Collection</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>Coronavirus Research Database</collection><collection>ProQuest Central Korea</collection><collection>ProQuest Central Student</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Computer Science Collection</collection><collection>Computer Science Database</collection><collection>ProQuest Advanced Technologies &amp; Aerospace Collection</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><jtitle>The Journal of artificial intelligence research</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Zhang, Baobao</au><au>Anderljung, Markus</au><au>Kahn, Lauren</au><au>Dreksler, Noemi</au><au>Horowitz, Michael C.</au><au>Dafoe, Allan</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Ethics and Governance of Artificial Intelligence: Evidence from a Survey of Machine Learning Researchers</atitle><jtitle>The Journal of artificial intelligence research</jtitle><date>2021</date><risdate>2021</risdate><volume>71</volume><spage>591</spage><pages>591-</pages><issn>1076-9757</issn><eissn>1076-9757</eissn><eissn>1943-5037</eissn><abstract>Machine learning (ML) and artificial intelligence (AI) researchers play an important role in the ethics and governance of AI, including through their work, advocacy, and choice of employment. Nevertheless, this influential group's attitudes are not well understood, undermining our ability to discern consensuses or disagreements between AI/ML researchers. To examine these researchers' views, we conducted a survey of those who published in two top AI/ML conferences (N = 524). We compare these results with those from a 2016 survey of AI/ML researchers (Grace et al., 2018) and a 2018 survey of the US public (Zhang &amp; Dafoe, 2020). We find that AI/ML researchers place high levels of trust in international organizations and scientific organizations to shape the development and use of AI in the public interest; moderate trust in most Western tech companies; and low trust in national militaries, Chinese tech companies, and Facebook. While the respondents were overwhelmingly opposed to AI/ML researchers working on lethal autonomous weapons, they are less opposed to researchers working on other military applications of AI, particularly logistics algorithms. A strong majority of respondents think that AI safety research should be prioritized and that ML institutions should conduct pre-publication review to assess potential harms. Being closer to the technology itself, AI/ML researchers are well placed to highlight new risks and develop technical solutions, so this novel attempt to measure their attitudes has broad relevance. The findings should help to improve how researchers, private sector executives, and policymakers think about regulations, governance frameworks, guiding principles, and national and international governance strategies for AI. This article appears in the special track on AI &amp; Society.</abstract><cop>San Francisco</cop><pub>AI Access Foundation</pub><doi>10.1613/jair.1.12895</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 1076-9757
ispartof The Journal of artificial intelligence research, 2021, Vol.71, p.591
issn 1076-9757
1076-9757
1943-5037
language eng
recordid cdi_proquest_journals_2567811157
source DOAJ Directory of Open Access Journals; Elektronische Zeitschriftenbibliothek - Frei zugängliche E-Journals; Free E- Journals
subjects Algorithms
Artificial intelligence
Attitudes
Ethics
International organizations
Logistics
Machine learning
Military applications
Researchers
title Ethics and Governance of Artificial Intelligence: Evidence from a Survey of Machine Learning Researchers
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-06T03%3A42%3A03IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Ethics%20and%20Governance%20of%20Artificial%20Intelligence:%20Evidence%20from%20a%20Survey%20of%20Machine%20Learning%20Researchers&rft.jtitle=The%20Journal%20of%20artificial%20intelligence%20research&rft.au=Zhang,%20Baobao&rft.date=2021&rft.volume=71&rft.spage=591&rft.pages=591-&rft.issn=1076-9757&rft.eissn=1076-9757&rft_id=info:doi/10.1613/jair.1.12895&rft_dat=%3Cproquest_cross%3E2567811157%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2567811157&rft_id=info:pmid/&rfr_iscdi=true