Dense reinforcement learning for safety validation of autonomous vehicles

One critical bottleneck that impedes the development and deployment of autonomous vehicles is the prohibitively high economic and time costs required to validate their safety in a naturalistic driving environment, owing to the rarity of safety-critical events 1 . Here we report the development of an...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Nature (London) 2023-03, Vol.615 (7953), p.620-627
Hauptverfasser: Feng, Shuo, Sun, Haowei, Yan, Xintao, Zhu, Haojie, Zou, Zhengxia, Shen, Shengyin, Liu, Henry X.
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 627
container_issue 7953
container_start_page 620
container_title Nature (London)
container_volume 615
creator Feng, Shuo
Sun, Haowei
Yan, Xintao
Zhu, Haojie
Zou, Zhengxia
Shen, Shengyin
Liu, Henry X.
description One critical bottleneck that impedes the development and deployment of autonomous vehicles is the prohibitively high economic and time costs required to validate their safety in a naturalistic driving environment, owing to the rarity of safety-critical events 1 . Here we report the development of an intelligent testing environment, where artificial-intelligence-based background agents are trained to validate the safety performances of autonomous vehicles in an accelerated mode, without loss of unbiasedness. From naturalistic driving data, the background agents learn what adversarial manoeuvre to execute through a dense deep-reinforcement-learning (D2RL) approach, in which Markov decision processes are edited by removing non-safety-critical states and reconnecting critical ones so that the information in the training data is densified. D2RL enables neural networks to learn from densified information with safety-critical events and achieves tasks that are intractable for traditional deep-reinforcement-learning approaches. We demonstrate the effectiveness of our approach by testing a highly automated vehicle in both highway and urban test tracks with an augmented-reality environment, combining simulated background vehicles with physical road infrastructure and a real autonomous test vehicle. Our results show that the D2RL-trained agents can accelerate the evaluation process by multiple orders of magnitude (10 3 to 10 5 times faster). In addition, D2RL will enable accelerated testing and training with other safety-critical autonomous systems. An intelligent environment has been developed for testing the safety performance of autonomous vehicles and its effectiveness has been demonstrated for highway and urban test tracks in an augmented-reality environment.
doi_str_mv 10.1038/s41586-023-05732-2
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_miscellaneous_2790051992</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2790480668</sourcerecordid><originalsourceid>FETCH-LOGICAL-c375t-916d378dbd15249a357d0f0cd22dc65971338583781d20668606f712c9e952a93</originalsourceid><addsrcrecordid>eNp9kElLxDAYQIMozrj8AQ9S8OKlmqXZjuI6IHjRc4jNV-3QJmPSCvPvzSwqePAUSF5evjyETgi-IJipy1QRrkSJKSsxl4yWdAdNSSVFWQkld9EUY6pKrJiYoIOU5hhjTmS1jyZM6EozJqdodgM-QRGh9U2INfTgh6IDG33r34q8VSTbwLAsPm3XOju0wRehKew4BB_6MKbiE97buoN0hPYa2yU43q6H6OXu9vn6oXx8up9dXz2WNZN8KDURjknlXh3htNKWcelwg2tHqasF15IwprjKCHEUC6EEFo0ktNagObWaHaLzjXcRw8cIaTB9m2roOushz2Oo1Kt_ak0zevYHnYcx-jzdmqrUyp8puqHqGFKK0JhFbHsbl4ZgswptNqFNDm3Woc1KfbpVj689uJ8r32UzwDZAykf-DeLv2_9ovwBxp4cq</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2790480668</pqid></control><display><type>article</type><title>Dense reinforcement learning for safety validation of autonomous vehicles</title><source>MEDLINE</source><source>Nature Journals Online</source><source>SpringerLink Journals - AutoHoldings</source><creator>Feng, Shuo ; Sun, Haowei ; Yan, Xintao ; Zhu, Haojie ; Zou, Zhengxia ; Shen, Shengyin ; Liu, Henry X.</creator><creatorcontrib>Feng, Shuo ; Sun, Haowei ; Yan, Xintao ; Zhu, Haojie ; Zou, Zhengxia ; Shen, Shengyin ; Liu, Henry X.</creatorcontrib><description>One critical bottleneck that impedes the development and deployment of autonomous vehicles is the prohibitively high economic and time costs required to validate their safety in a naturalistic driving environment, owing to the rarity of safety-critical events 1 . Here we report the development of an intelligent testing environment, where artificial-intelligence-based background agents are trained to validate the safety performances of autonomous vehicles in an accelerated mode, without loss of unbiasedness. From naturalistic driving data, the background agents learn what adversarial manoeuvre to execute through a dense deep-reinforcement-learning (D2RL) approach, in which Markov decision processes are edited by removing non-safety-critical states and reconnecting critical ones so that the information in the training data is densified. D2RL enables neural networks to learn from densified information with safety-critical events and achieves tasks that are intractable for traditional deep-reinforcement-learning approaches. We demonstrate the effectiveness of our approach by testing a highly automated vehicle in both highway and urban test tracks with an augmented-reality environment, combining simulated background vehicles with physical road infrastructure and a real autonomous test vehicle. Our results show that the D2RL-trained agents can accelerate the evaluation process by multiple orders of magnitude (10 3 to 10 5 times faster). In addition, D2RL will enable accelerated testing and training with other safety-critical autonomous systems. An intelligent environment has been developed for testing the safety performance of autonomous vehicles and its effectiveness has been demonstrated for highway and urban test tracks in an augmented-reality environment.</description><identifier>ISSN: 0028-0836</identifier><identifier>EISSN: 1476-4687</identifier><identifier>DOI: 10.1038/s41586-023-05732-2</identifier><identifier>PMID: 36949337</identifier><language>eng</language><publisher>London: Nature Publishing Group UK</publisher><subject>639/166/986 ; 639/166/988 ; Accelerated tests ; Artificial intelligence ; Augmented reality ; Automation ; Automation - methods ; Automation - standards ; Automobile Driving ; Autonomous vehicles ; Autonomous Vehicles - standards ; Deep Learning ; Driving ability ; Efficiency ; Humanities and Social Sciences ; Humans ; Learning ; Machine learning ; Markov analysis ; Markov processes ; multidisciplinary ; Neural networks ; Reinforcement ; Reproducibility of Results ; Roads &amp; highways ; Safety ; Safety critical ; Science ; Science (multidisciplinary) ; Simulation ; Test vehicles ; Training</subject><ispartof>Nature (London), 2023-03, Vol.615 (7953), p.620-627</ispartof><rights>The Author(s), under exclusive licence to Springer Nature Limited 2023. Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.</rights><rights>2023. The Author(s), under exclusive licence to Springer Nature Limited.</rights><rights>Copyright Nature Publishing Group Mar 23, 2023</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c375t-916d378dbd15249a357d0f0cd22dc65971338583781d20668606f712c9e952a93</citedby><cites>FETCH-LOGICAL-c375t-916d378dbd15249a357d0f0cd22dc65971338583781d20668606f712c9e952a93</cites><orcidid>0000-0002-0569-5628 ; 0000-0002-3685-9920 ; 0000-0002-2117-4427</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://link.springer.com/content/pdf/10.1038/s41586-023-05732-2$$EPDF$$P50$$Gspringer$$H</linktopdf><linktohtml>$$Uhttps://link.springer.com/10.1038/s41586-023-05732-2$$EHTML$$P50$$Gspringer$$H</linktohtml><link.rule.ids>314,776,780,27903,27904,41467,42536,51297</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/36949337$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Feng, Shuo</creatorcontrib><creatorcontrib>Sun, Haowei</creatorcontrib><creatorcontrib>Yan, Xintao</creatorcontrib><creatorcontrib>Zhu, Haojie</creatorcontrib><creatorcontrib>Zou, Zhengxia</creatorcontrib><creatorcontrib>Shen, Shengyin</creatorcontrib><creatorcontrib>Liu, Henry X.</creatorcontrib><title>Dense reinforcement learning for safety validation of autonomous vehicles</title><title>Nature (London)</title><addtitle>Nature</addtitle><addtitle>Nature</addtitle><description>One critical bottleneck that impedes the development and deployment of autonomous vehicles is the prohibitively high economic and time costs required to validate their safety in a naturalistic driving environment, owing to the rarity of safety-critical events 1 . Here we report the development of an intelligent testing environment, where artificial-intelligence-based background agents are trained to validate the safety performances of autonomous vehicles in an accelerated mode, without loss of unbiasedness. From naturalistic driving data, the background agents learn what adversarial manoeuvre to execute through a dense deep-reinforcement-learning (D2RL) approach, in which Markov decision processes are edited by removing non-safety-critical states and reconnecting critical ones so that the information in the training data is densified. D2RL enables neural networks to learn from densified information with safety-critical events and achieves tasks that are intractable for traditional deep-reinforcement-learning approaches. We demonstrate the effectiveness of our approach by testing a highly automated vehicle in both highway and urban test tracks with an augmented-reality environment, combining simulated background vehicles with physical road infrastructure and a real autonomous test vehicle. Our results show that the D2RL-trained agents can accelerate the evaluation process by multiple orders of magnitude (10 3 to 10 5 times faster). In addition, D2RL will enable accelerated testing and training with other safety-critical autonomous systems. An intelligent environment has been developed for testing the safety performance of autonomous vehicles and its effectiveness has been demonstrated for highway and urban test tracks in an augmented-reality environment.</description><subject>639/166/986</subject><subject>639/166/988</subject><subject>Accelerated tests</subject><subject>Artificial intelligence</subject><subject>Augmented reality</subject><subject>Automation</subject><subject>Automation - methods</subject><subject>Automation - standards</subject><subject>Automobile Driving</subject><subject>Autonomous vehicles</subject><subject>Autonomous Vehicles - standards</subject><subject>Deep Learning</subject><subject>Driving ability</subject><subject>Efficiency</subject><subject>Humanities and Social Sciences</subject><subject>Humans</subject><subject>Learning</subject><subject>Machine learning</subject><subject>Markov analysis</subject><subject>Markov processes</subject><subject>multidisciplinary</subject><subject>Neural networks</subject><subject>Reinforcement</subject><subject>Reproducibility of Results</subject><subject>Roads &amp; highways</subject><subject>Safety</subject><subject>Safety critical</subject><subject>Science</subject><subject>Science (multidisciplinary)</subject><subject>Simulation</subject><subject>Test vehicles</subject><subject>Training</subject><issn>0028-0836</issn><issn>1476-4687</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>EIF</sourceid><sourceid>8G5</sourceid><sourceid>BEC</sourceid><sourceid>BENPR</sourceid><sourceid>GUQSH</sourceid><sourceid>M2O</sourceid><recordid>eNp9kElLxDAYQIMozrj8AQ9S8OKlmqXZjuI6IHjRc4jNV-3QJmPSCvPvzSwqePAUSF5evjyETgi-IJipy1QRrkSJKSsxl4yWdAdNSSVFWQkld9EUY6pKrJiYoIOU5hhjTmS1jyZM6EozJqdodgM-QRGh9U2INfTgh6IDG33r34q8VSTbwLAsPm3XOju0wRehKew4BB_6MKbiE97buoN0hPYa2yU43q6H6OXu9vn6oXx8up9dXz2WNZN8KDURjknlXh3htNKWcelwg2tHqasF15IwprjKCHEUC6EEFo0ktNagObWaHaLzjXcRw8cIaTB9m2roOushz2Oo1Kt_ak0zevYHnYcx-jzdmqrUyp8puqHqGFKK0JhFbHsbl4ZgswptNqFNDm3Woc1KfbpVj689uJ8r32UzwDZAykf-DeLv2_9ovwBxp4cq</recordid><startdate>20230323</startdate><enddate>20230323</enddate><creator>Feng, Shuo</creator><creator>Sun, Haowei</creator><creator>Yan, Xintao</creator><creator>Zhu, Haojie</creator><creator>Zou, Zhengxia</creator><creator>Shen, Shengyin</creator><creator>Liu, Henry X.</creator><general>Nature Publishing Group UK</general><general>Nature Publishing Group</general><scope>CGR</scope><scope>CUY</scope><scope>CVF</scope><scope>ECM</scope><scope>EIF</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>3V.</scope><scope>7QG</scope><scope>7QL</scope><scope>7QP</scope><scope>7QR</scope><scope>7RV</scope><scope>7SN</scope><scope>7SS</scope><scope>7ST</scope><scope>7T5</scope><scope>7TG</scope><scope>7TK</scope><scope>7TM</scope><scope>7TO</scope><scope>7U9</scope><scope>7X2</scope><scope>7X7</scope><scope>7XB</scope><scope>88A</scope><scope>88E</scope><scope>88G</scope><scope>88I</scope><scope>8AF</scope><scope>8AO</scope><scope>8C1</scope><scope>8FD</scope><scope>8FE</scope><scope>8FG</scope><scope>8FH</scope><scope>8FI</scope><scope>8FJ</scope><scope>8FK</scope><scope>8G5</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AEUYN</scope><scope>AFKRA</scope><scope>ARAPS</scope><scope>ATCPS</scope><scope>AZQEC</scope><scope>BBNVY</scope><scope>BEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>BHPHI</scope><scope>BKSAR</scope><scope>C1K</scope><scope>CCPQU</scope><scope>D1I</scope><scope>DWQXO</scope><scope>FR3</scope><scope>FYUFA</scope><scope>GHDGH</scope><scope>GNUQQ</scope><scope>GUQSH</scope><scope>H94</scope><scope>HCIFZ</scope><scope>K9.</scope><scope>KB.</scope><scope>KB0</scope><scope>KL.</scope><scope>L6V</scope><scope>LK8</scope><scope>M0K</scope><scope>M0S</scope><scope>M1P</scope><scope>M2M</scope><scope>M2O</scope><scope>M2P</scope><scope>M7N</scope><scope>M7P</scope><scope>M7S</scope><scope>MBDVC</scope><scope>NAPCQ</scope><scope>P5Z</scope><scope>P62</scope><scope>P64</scope><scope>PATMY</scope><scope>PCBAR</scope><scope>PDBOC</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PSYQQ</scope><scope>PTHSS</scope><scope>PYCSY</scope><scope>Q9U</scope><scope>R05</scope><scope>RC3</scope><scope>S0X</scope><scope>SOI</scope><scope>7X8</scope><orcidid>https://orcid.org/0000-0002-0569-5628</orcidid><orcidid>https://orcid.org/0000-0002-3685-9920</orcidid><orcidid>https://orcid.org/0000-0002-2117-4427</orcidid></search><sort><creationdate>20230323</creationdate><title>Dense reinforcement learning for safety validation of autonomous vehicles</title><author>Feng, Shuo ; Sun, Haowei ; Yan, Xintao ; Zhu, Haojie ; Zou, Zhengxia ; Shen, Shengyin ; Liu, Henry X.</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c375t-916d378dbd15249a357d0f0cd22dc65971338583781d20668606f712c9e952a93</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>639/166/986</topic><topic>639/166/988</topic><topic>Accelerated tests</topic><topic>Artificial intelligence</topic><topic>Augmented reality</topic><topic>Automation</topic><topic>Automation - methods</topic><topic>Automation - standards</topic><topic>Automobile Driving</topic><topic>Autonomous vehicles</topic><topic>Autonomous Vehicles - standards</topic><topic>Deep Learning</topic><topic>Driving ability</topic><topic>Efficiency</topic><topic>Humanities and Social Sciences</topic><topic>Humans</topic><topic>Learning</topic><topic>Machine learning</topic><topic>Markov analysis</topic><topic>Markov processes</topic><topic>multidisciplinary</topic><topic>Neural networks</topic><topic>Reinforcement</topic><topic>Reproducibility of Results</topic><topic>Roads &amp; highways</topic><topic>Safety</topic><topic>Safety critical</topic><topic>Science</topic><topic>Science (multidisciplinary)</topic><topic>Simulation</topic><topic>Test vehicles</topic><topic>Training</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Feng, Shuo</creatorcontrib><creatorcontrib>Sun, Haowei</creatorcontrib><creatorcontrib>Yan, Xintao</creatorcontrib><creatorcontrib>Zhu, Haojie</creatorcontrib><creatorcontrib>Zou, Zhengxia</creatorcontrib><creatorcontrib>Shen, Shengyin</creatorcontrib><creatorcontrib>Liu, Henry X.</creatorcontrib><collection>Medline</collection><collection>MEDLINE</collection><collection>MEDLINE (Ovid)</collection><collection>MEDLINE</collection><collection>MEDLINE</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>ProQuest Central (Corporate)</collection><collection>Animal Behavior Abstracts</collection><collection>Bacteriology Abstracts (Microbiology B)</collection><collection>Calcium &amp; Calcified Tissue Abstracts</collection><collection>Chemoreception Abstracts</collection><collection>Nursing &amp; Allied Health Database</collection><collection>Ecology Abstracts</collection><collection>Entomology Abstracts (Full archive)</collection><collection>Environment Abstracts</collection><collection>Immunology Abstracts</collection><collection>Meteorological &amp; Geoastrophysical Abstracts</collection><collection>Neurosciences Abstracts</collection><collection>Nucleic Acids Abstracts</collection><collection>Oncogenes and Growth Factors Abstracts</collection><collection>Virology and AIDS Abstracts</collection><collection>Agricultural Science Collection</collection><collection>Health &amp; Medical Collection</collection><collection>ProQuest Central (purchase pre-March 2016)</collection><collection>Biology Database (Alumni Edition)</collection><collection>Medical Database (Alumni Edition)</collection><collection>Psychology Database (Alumni)</collection><collection>Science Database (Alumni Edition)</collection><collection>STEM Database</collection><collection>ProQuest Pharma Collection</collection><collection>Public Health Database</collection><collection>Technology Research Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>ProQuest Natural Science Collection</collection><collection>Hospital Premium Collection</collection><collection>Hospital Premium Collection (Alumni Edition)</collection><collection>ProQuest Central (Alumni) (purchase pre-March 2016)</collection><collection>Research Library (Alumni Edition)</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest One Sustainability</collection><collection>ProQuest Central UK/Ireland</collection><collection>Advanced Technologies &amp; Aerospace Collection</collection><collection>Agricultural &amp; Environmental Science Collection</collection><collection>ProQuest Central Essentials</collection><collection>Biological Science Collection</collection><collection>eLibrary</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>Natural Science Collection</collection><collection>Earth, Atmospheric &amp; Aquatic Science Collection</collection><collection>Environmental Sciences and Pollution Management</collection><collection>ProQuest One Community College</collection><collection>ProQuest Materials Science Collection</collection><collection>ProQuest Central Korea</collection><collection>Engineering Research Database</collection><collection>Health Research Premium Collection</collection><collection>Health Research Premium Collection (Alumni)</collection><collection>ProQuest Central Student</collection><collection>Research Library Prep</collection><collection>AIDS and Cancer Research Abstracts</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Health &amp; Medical Complete (Alumni)</collection><collection>Materials Science Database</collection><collection>Nursing &amp; Allied Health Database (Alumni Edition)</collection><collection>Meteorological &amp; Geoastrophysical Abstracts - Academic</collection><collection>ProQuest Engineering Collection</collection><collection>ProQuest Biological Science Collection</collection><collection>Agricultural Science Database</collection><collection>Health &amp; Medical Collection (Alumni Edition)</collection><collection>Medical Database</collection><collection>ProQuest Psychology</collection><collection>Research Library</collection><collection>Science Database</collection><collection>Algology Mycology and Protozoology Abstracts (Microbiology C)</collection><collection>Biological Science Database</collection><collection>Engineering Database</collection><collection>Research Library (Corporate)</collection><collection>Nursing &amp; Allied Health Premium</collection><collection>Advanced Technologies &amp; Aerospace Database</collection><collection>ProQuest Advanced Technologies &amp; Aerospace Collection</collection><collection>Biotechnology and BioEngineering Abstracts</collection><collection>Environmental Science Database</collection><collection>Earth, Atmospheric &amp; Aquatic Science Database</collection><collection>Materials Science Collection</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest One Psychology</collection><collection>Engineering Collection</collection><collection>Environmental Science Collection</collection><collection>ProQuest Central Basic</collection><collection>University of Michigan</collection><collection>Genetics Abstracts</collection><collection>SIRS Editorial</collection><collection>Environment Abstracts</collection><collection>MEDLINE - Academic</collection><jtitle>Nature (London)</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Feng, Shuo</au><au>Sun, Haowei</au><au>Yan, Xintao</au><au>Zhu, Haojie</au><au>Zou, Zhengxia</au><au>Shen, Shengyin</au><au>Liu, Henry X.</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Dense reinforcement learning for safety validation of autonomous vehicles</atitle><jtitle>Nature (London)</jtitle><stitle>Nature</stitle><addtitle>Nature</addtitle><date>2023-03-23</date><risdate>2023</risdate><volume>615</volume><issue>7953</issue><spage>620</spage><epage>627</epage><pages>620-627</pages><issn>0028-0836</issn><eissn>1476-4687</eissn><abstract>One critical bottleneck that impedes the development and deployment of autonomous vehicles is the prohibitively high economic and time costs required to validate their safety in a naturalistic driving environment, owing to the rarity of safety-critical events 1 . Here we report the development of an intelligent testing environment, where artificial-intelligence-based background agents are trained to validate the safety performances of autonomous vehicles in an accelerated mode, without loss of unbiasedness. From naturalistic driving data, the background agents learn what adversarial manoeuvre to execute through a dense deep-reinforcement-learning (D2RL) approach, in which Markov decision processes are edited by removing non-safety-critical states and reconnecting critical ones so that the information in the training data is densified. D2RL enables neural networks to learn from densified information with safety-critical events and achieves tasks that are intractable for traditional deep-reinforcement-learning approaches. We demonstrate the effectiveness of our approach by testing a highly automated vehicle in both highway and urban test tracks with an augmented-reality environment, combining simulated background vehicles with physical road infrastructure and a real autonomous test vehicle. Our results show that the D2RL-trained agents can accelerate the evaluation process by multiple orders of magnitude (10 3 to 10 5 times faster). In addition, D2RL will enable accelerated testing and training with other safety-critical autonomous systems. An intelligent environment has been developed for testing the safety performance of autonomous vehicles and its effectiveness has been demonstrated for highway and urban test tracks in an augmented-reality environment.</abstract><cop>London</cop><pub>Nature Publishing Group UK</pub><pmid>36949337</pmid><doi>10.1038/s41586-023-05732-2</doi><tpages>8</tpages><orcidid>https://orcid.org/0000-0002-0569-5628</orcidid><orcidid>https://orcid.org/0000-0002-3685-9920</orcidid><orcidid>https://orcid.org/0000-0002-2117-4427</orcidid></addata></record>
fulltext fulltext
identifier ISSN: 0028-0836
ispartof Nature (London), 2023-03, Vol.615 (7953), p.620-627
issn 0028-0836
1476-4687
language eng
recordid cdi_proquest_miscellaneous_2790051992
source MEDLINE; Nature Journals Online; SpringerLink Journals - AutoHoldings
subjects 639/166/986
639/166/988
Accelerated tests
Artificial intelligence
Augmented reality
Automation
Automation - methods
Automation - standards
Automobile Driving
Autonomous vehicles
Autonomous Vehicles - standards
Deep Learning
Driving ability
Efficiency
Humanities and Social Sciences
Humans
Learning
Machine learning
Markov analysis
Markov processes
multidisciplinary
Neural networks
Reinforcement
Reproducibility of Results
Roads & highways
Safety
Safety critical
Science
Science (multidisciplinary)
Simulation
Test vehicles
Training
title Dense reinforcement learning for safety validation of autonomous vehicles
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-27T19%3A45%3A01IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Dense%20reinforcement%20learning%20for%20safety%20validation%20of%20autonomous%20vehicles&rft.jtitle=Nature%20(London)&rft.au=Feng,%20Shuo&rft.date=2023-03-23&rft.volume=615&rft.issue=7953&rft.spage=620&rft.epage=627&rft.pages=620-627&rft.issn=0028-0836&rft.eissn=1476-4687&rft_id=info:doi/10.1038/s41586-023-05732-2&rft_dat=%3Cproquest_cross%3E2790480668%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2790480668&rft_id=info:pmid/36949337&rfr_iscdi=true