An Autonomous Driving Model Integrated with BEV-V2X Perception, Fusion Prediction of Motion and Occupancy, and Driving Planning, in Complex Traffic Intersections

The comprehensiveness of vehicle-to-everything (V2X) recognition enriches and holistically shapes the global Birds-Eye-View (BEV) perception, incorporating rich semantics and integrating driving scene information, thereby serving features of vehicle state prediction, decision-making and driving plan...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Li, Fukang, Ou, Wenlin, Gao, Kunpeng, Pang, Yuwen, Li, Yifei, Fan, Henry
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Li, Fukang
Ou, Wenlin
Gao, Kunpeng
Pang, Yuwen
Li, Yifei
Fan, Henry
description The comprehensiveness of vehicle-to-everything (V2X) recognition enriches and holistically shapes the global Birds-Eye-View (BEV) perception, incorporating rich semantics and integrating driving scene information, thereby serving features of vehicle state prediction, decision-making and driving planning. Utilizing V2X message sets to form BEV map proves to be an effective perception method for connected and automated vehicles (CAVs). Specifically, Map Msg. (MAP), Signal Phase And Timing (SPAT) and Roadside Information (RSI) contributes to the achievement of road connectivity, synchronized traffic signal navigation and obstacle warning. Moreover, harnessing time-sequential Basic Safety Msg. (BSM) data from multiple vehicles allows for the real-time perception and future state prediction. Therefore, this paper develops a comprehensive autonomous driving model that relies on BEV-V2X perception, Interacting Multiple model Unscented Kalman Filter (IMM-UKF)-based fusion prediction, and deep reinforcement learning (DRL)-based decision making and planning. We integrated them into a DRL environment to develop an optimal set of unified driving behaviors that encompass obstacle avoidance, lane changes, overtaking, turning maneuver, and synchronized traffic signal navigation. Consequently, a complex traffic intersection scenario was simulated, and the well-trained model was applied for driving planning. The observed driving behavior closely resembled that of an experienced driver, exhibiting anticipatory actions and revealing notable operational highlights of driving policy.
doi_str_mv 10.48550/arxiv.2312.05104
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2312_05104</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2312_05104</sourcerecordid><originalsourceid>FETCH-arxiv_primary_2312_051043</originalsourceid><addsrcrecordid>eNqFj0FuwkAMRWfTBWp7AFb1AUJIApG6pRREF4gsItRdNJo41FLiiTwTCsfpTRtG7ZrV__6y_fWUmqZJvHzN82Su5ULnOFukWZzkabKcqJ8Vw2rwlm1nBwfvQmfiE-xtjS18sMeTaI81fJP_grfNcXbMPqFAMdh7shzBdnCjQiFYk7lFYJvxPDjNNRyMGXrN5hqF8b-gaDXzaCIghrXt-hYvUIpuGjKhVxyGd-5JPTS6dfj8p4_qZbsp17tZgKl6oU7LtbpBVQFqcX_jFzgmWT8</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>An Autonomous Driving Model Integrated with BEV-V2X Perception, Fusion Prediction of Motion and Occupancy, and Driving Planning, in Complex Traffic Intersections</title><source>arXiv.org</source><creator>Li, Fukang ; Ou, Wenlin ; Gao, Kunpeng ; Pang, Yuwen ; Li, Yifei ; Fan, Henry</creator><creatorcontrib>Li, Fukang ; Ou, Wenlin ; Gao, Kunpeng ; Pang, Yuwen ; Li, Yifei ; Fan, Henry</creatorcontrib><description>The comprehensiveness of vehicle-to-everything (V2X) recognition enriches and holistically shapes the global Birds-Eye-View (BEV) perception, incorporating rich semantics and integrating driving scene information, thereby serving features of vehicle state prediction, decision-making and driving planning. Utilizing V2X message sets to form BEV map proves to be an effective perception method for connected and automated vehicles (CAVs). Specifically, Map Msg. (MAP), Signal Phase And Timing (SPAT) and Roadside Information (RSI) contributes to the achievement of road connectivity, synchronized traffic signal navigation and obstacle warning. Moreover, harnessing time-sequential Basic Safety Msg. (BSM) data from multiple vehicles allows for the real-time perception and future state prediction. Therefore, this paper develops a comprehensive autonomous driving model that relies on BEV-V2X perception, Interacting Multiple model Unscented Kalman Filter (IMM-UKF)-based fusion prediction, and deep reinforcement learning (DRL)-based decision making and planning. We integrated them into a DRL environment to develop an optimal set of unified driving behaviors that encompass obstacle avoidance, lane changes, overtaking, turning maneuver, and synchronized traffic signal navigation. Consequently, a complex traffic intersection scenario was simulated, and the well-trained model was applied for driving planning. The observed driving behavior closely resembled that of an experienced driver, exhibiting anticipatory actions and revealing notable operational highlights of driving policy.</description><identifier>DOI: 10.48550/arxiv.2312.05104</identifier><language>eng</language><subject>Computer Science - Robotics</subject><creationdate>2023-12</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2312.05104$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2312.05104$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Li, Fukang</creatorcontrib><creatorcontrib>Ou, Wenlin</creatorcontrib><creatorcontrib>Gao, Kunpeng</creatorcontrib><creatorcontrib>Pang, Yuwen</creatorcontrib><creatorcontrib>Li, Yifei</creatorcontrib><creatorcontrib>Fan, Henry</creatorcontrib><title>An Autonomous Driving Model Integrated with BEV-V2X Perception, Fusion Prediction of Motion and Occupancy, and Driving Planning, in Complex Traffic Intersections</title><description>The comprehensiveness of vehicle-to-everything (V2X) recognition enriches and holistically shapes the global Birds-Eye-View (BEV) perception, incorporating rich semantics and integrating driving scene information, thereby serving features of vehicle state prediction, decision-making and driving planning. Utilizing V2X message sets to form BEV map proves to be an effective perception method for connected and automated vehicles (CAVs). Specifically, Map Msg. (MAP), Signal Phase And Timing (SPAT) and Roadside Information (RSI) contributes to the achievement of road connectivity, synchronized traffic signal navigation and obstacle warning. Moreover, harnessing time-sequential Basic Safety Msg. (BSM) data from multiple vehicles allows for the real-time perception and future state prediction. Therefore, this paper develops a comprehensive autonomous driving model that relies on BEV-V2X perception, Interacting Multiple model Unscented Kalman Filter (IMM-UKF)-based fusion prediction, and deep reinforcement learning (DRL)-based decision making and planning. We integrated them into a DRL environment to develop an optimal set of unified driving behaviors that encompass obstacle avoidance, lane changes, overtaking, turning maneuver, and synchronized traffic signal navigation. Consequently, a complex traffic intersection scenario was simulated, and the well-trained model was applied for driving planning. The observed driving behavior closely resembled that of an experienced driver, exhibiting anticipatory actions and revealing notable operational highlights of driving policy.</description><subject>Computer Science - Robotics</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNqFj0FuwkAMRWfTBWp7AFb1AUJIApG6pRREF4gsItRdNJo41FLiiTwTCsfpTRtG7ZrV__6y_fWUmqZJvHzN82Su5ULnOFukWZzkabKcqJ8Vw2rwlm1nBwfvQmfiE-xtjS18sMeTaI81fJP_grfNcXbMPqFAMdh7shzBdnCjQiFYk7lFYJvxPDjNNRyMGXrN5hqF8b-gaDXzaCIghrXt-hYvUIpuGjKhVxyGd-5JPTS6dfj8p4_qZbsp17tZgKl6oU7LtbpBVQFqcX_jFzgmWT8</recordid><startdate>20231208</startdate><enddate>20231208</enddate><creator>Li, Fukang</creator><creator>Ou, Wenlin</creator><creator>Gao, Kunpeng</creator><creator>Pang, Yuwen</creator><creator>Li, Yifei</creator><creator>Fan, Henry</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20231208</creationdate><title>An Autonomous Driving Model Integrated with BEV-V2X Perception, Fusion Prediction of Motion and Occupancy, and Driving Planning, in Complex Traffic Intersections</title><author>Li, Fukang ; Ou, Wenlin ; Gao, Kunpeng ; Pang, Yuwen ; Li, Yifei ; Fan, Henry</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-arxiv_primary_2312_051043</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Computer Science - Robotics</topic><toplevel>online_resources</toplevel><creatorcontrib>Li, Fukang</creatorcontrib><creatorcontrib>Ou, Wenlin</creatorcontrib><creatorcontrib>Gao, Kunpeng</creatorcontrib><creatorcontrib>Pang, Yuwen</creatorcontrib><creatorcontrib>Li, Yifei</creatorcontrib><creatorcontrib>Fan, Henry</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Li, Fukang</au><au>Ou, Wenlin</au><au>Gao, Kunpeng</au><au>Pang, Yuwen</au><au>Li, Yifei</au><au>Fan, Henry</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>An Autonomous Driving Model Integrated with BEV-V2X Perception, Fusion Prediction of Motion and Occupancy, and Driving Planning, in Complex Traffic Intersections</atitle><date>2023-12-08</date><risdate>2023</risdate><abstract>The comprehensiveness of vehicle-to-everything (V2X) recognition enriches and holistically shapes the global Birds-Eye-View (BEV) perception, incorporating rich semantics and integrating driving scene information, thereby serving features of vehicle state prediction, decision-making and driving planning. Utilizing V2X message sets to form BEV map proves to be an effective perception method for connected and automated vehicles (CAVs). Specifically, Map Msg. (MAP), Signal Phase And Timing (SPAT) and Roadside Information (RSI) contributes to the achievement of road connectivity, synchronized traffic signal navigation and obstacle warning. Moreover, harnessing time-sequential Basic Safety Msg. (BSM) data from multiple vehicles allows for the real-time perception and future state prediction. Therefore, this paper develops a comprehensive autonomous driving model that relies on BEV-V2X perception, Interacting Multiple model Unscented Kalman Filter (IMM-UKF)-based fusion prediction, and deep reinforcement learning (DRL)-based decision making and planning. We integrated them into a DRL environment to develop an optimal set of unified driving behaviors that encompass obstacle avoidance, lane changes, overtaking, turning maneuver, and synchronized traffic signal navigation. Consequently, a complex traffic intersection scenario was simulated, and the well-trained model was applied for driving planning. The observed driving behavior closely resembled that of an experienced driver, exhibiting anticipatory actions and revealing notable operational highlights of driving policy.</abstract><doi>10.48550/arxiv.2312.05104</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2312.05104
ispartof
issn
language eng
recordid cdi_arxiv_primary_2312_05104
source arXiv.org
subjects Computer Science - Robotics
title An Autonomous Driving Model Integrated with BEV-V2X Perception, Fusion Prediction of Motion and Occupancy, and Driving Planning, in Complex Traffic Intersections
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-23T13%3A13%3A42IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=An%20Autonomous%20Driving%20Model%20Integrated%20with%20BEV-V2X%20Perception,%20Fusion%20Prediction%20of%20Motion%20and%20Occupancy,%20and%20Driving%20Planning,%20in%20Complex%20Traffic%20Intersections&rft.au=Li,%20Fukang&rft.date=2023-12-08&rft_id=info:doi/10.48550/arxiv.2312.05104&rft_dat=%3Carxiv_GOX%3E2312_05104%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true