Shared Cross-Modal Trajectory Prediction for Autonomous Driving

Predicting future trajectories of traffic agents in highly interactive environments is an essential and challenging problem for the safe operation of autonomous driving systems. On the basis of the fact that self-driving vehicles are equipped with various types of sensors (e.g., LiDAR scanner, RGB c...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Choi, Chiho, Choi, Joon Hee, Li, Jiachen, Malla, Srikanth
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Choi, Chiho
Choi, Joon Hee
Li, Jiachen
Malla, Srikanth
description Predicting future trajectories of traffic agents in highly interactive environments is an essential and challenging problem for the safe operation of autonomous driving systems. On the basis of the fact that self-driving vehicles are equipped with various types of sensors (e.g., LiDAR scanner, RGB camera, radar, etc.), we propose a Cross-Modal Embedding framework that aims to benefit from the use of multiple input modalities. At training time, our model learns to embed a set of complementary features in a shared latent space by jointly optimizing the objective functions across different types of input data. At test time, a single input modality (e.g., LiDAR data) is required to generate predictions from the input perspective (i.e., in the LiDAR space), while taking advantages from the model trained with multiple sensor modalities. An extensive evaluation is conducted to show the efficacy of the proposed framework using two benchmark driving datasets.
doi_str_mv 10.48550/arxiv.2011.08436
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2011_08436</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2011_08436</sourcerecordid><originalsourceid>FETCH-LOGICAL-a676-5c5c68be7dedae8b755b81ffd08b5336dd082dff8db0e51636c76d561db4af1a3</originalsourceid><addsrcrecordid>eNotz81qAjEUBeBsuii2D9BV8wIzTczkTnYi019QFDr74SY30YhOSmaU-va1tqtz4MCBj7EHKcrKaC2eMH_HUzkVUpbCVApu2exzi9kTb3IahmKZCPe8zbjzbkz5zNeXLboxpp6HlPn8OKY-HdJx4M85nmK_uWM3AfeDv__PCWtfX9rmvVis3j6a-aJAqKHQTjsw1tfkCb2xtdbWyBBIGKuVArqUKYVgyAqvJShwNZAGSbbCIFFN2OPf7VXQfeV4wHzufiXdVaJ-AChDRKw</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Shared Cross-Modal Trajectory Prediction for Autonomous Driving</title><source>arXiv.org</source><creator>Choi, Chiho ; Choi, Joon Hee ; Li, Jiachen ; Malla, Srikanth</creator><creatorcontrib>Choi, Chiho ; Choi, Joon Hee ; Li, Jiachen ; Malla, Srikanth</creatorcontrib><description>Predicting future trajectories of traffic agents in highly interactive environments is an essential and challenging problem for the safe operation of autonomous driving systems. On the basis of the fact that self-driving vehicles are equipped with various types of sensors (e.g., LiDAR scanner, RGB camera, radar, etc.), we propose a Cross-Modal Embedding framework that aims to benefit from the use of multiple input modalities. At training time, our model learns to embed a set of complementary features in a shared latent space by jointly optimizing the objective functions across different types of input data. At test time, a single input modality (e.g., LiDAR data) is required to generate predictions from the input perspective (i.e., in the LiDAR space), while taking advantages from the model trained with multiple sensor modalities. An extensive evaluation is conducted to show the efficacy of the proposed framework using two benchmark driving datasets.</description><identifier>DOI: 10.48550/arxiv.2011.08436</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Computer Vision and Pattern Recognition ; Computer Science - Robotics</subject><creationdate>2020-11</creationdate><rights>http://creativecommons.org/licenses/by-nc-nd/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2011.08436$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2011.08436$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Choi, Chiho</creatorcontrib><creatorcontrib>Choi, Joon Hee</creatorcontrib><creatorcontrib>Li, Jiachen</creatorcontrib><creatorcontrib>Malla, Srikanth</creatorcontrib><title>Shared Cross-Modal Trajectory Prediction for Autonomous Driving</title><description>Predicting future trajectories of traffic agents in highly interactive environments is an essential and challenging problem for the safe operation of autonomous driving systems. On the basis of the fact that self-driving vehicles are equipped with various types of sensors (e.g., LiDAR scanner, RGB camera, radar, etc.), we propose a Cross-Modal Embedding framework that aims to benefit from the use of multiple input modalities. At training time, our model learns to embed a set of complementary features in a shared latent space by jointly optimizing the objective functions across different types of input data. At test time, a single input modality (e.g., LiDAR data) is required to generate predictions from the input perspective (i.e., in the LiDAR space), while taking advantages from the model trained with multiple sensor modalities. An extensive evaluation is conducted to show the efficacy of the proposed framework using two benchmark driving datasets.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Computer Vision and Pattern Recognition</subject><subject>Computer Science - Robotics</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotz81qAjEUBeBsuii2D9BV8wIzTczkTnYi019QFDr74SY30YhOSmaU-va1tqtz4MCBj7EHKcrKaC2eMH_HUzkVUpbCVApu2exzi9kTb3IahmKZCPe8zbjzbkz5zNeXLboxpp6HlPn8OKY-HdJx4M85nmK_uWM3AfeDv__PCWtfX9rmvVis3j6a-aJAqKHQTjsw1tfkCb2xtdbWyBBIGKuVArqUKYVgyAqvJShwNZAGSbbCIFFN2OPf7VXQfeV4wHzufiXdVaJ-AChDRKw</recordid><startdate>20201115</startdate><enddate>20201115</enddate><creator>Choi, Chiho</creator><creator>Choi, Joon Hee</creator><creator>Li, Jiachen</creator><creator>Malla, Srikanth</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20201115</creationdate><title>Shared Cross-Modal Trajectory Prediction for Autonomous Driving</title><author>Choi, Chiho ; Choi, Joon Hee ; Li, Jiachen ; Malla, Srikanth</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a676-5c5c68be7dedae8b755b81ffd08b5336dd082dff8db0e51636c76d561db4af1a3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Computer Vision and Pattern Recognition</topic><topic>Computer Science - Robotics</topic><toplevel>online_resources</toplevel><creatorcontrib>Choi, Chiho</creatorcontrib><creatorcontrib>Choi, Joon Hee</creatorcontrib><creatorcontrib>Li, Jiachen</creatorcontrib><creatorcontrib>Malla, Srikanth</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Choi, Chiho</au><au>Choi, Joon Hee</au><au>Li, Jiachen</au><au>Malla, Srikanth</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Shared Cross-Modal Trajectory Prediction for Autonomous Driving</atitle><date>2020-11-15</date><risdate>2020</risdate><abstract>Predicting future trajectories of traffic agents in highly interactive environments is an essential and challenging problem for the safe operation of autonomous driving systems. On the basis of the fact that self-driving vehicles are equipped with various types of sensors (e.g., LiDAR scanner, RGB camera, radar, etc.), we propose a Cross-Modal Embedding framework that aims to benefit from the use of multiple input modalities. At training time, our model learns to embed a set of complementary features in a shared latent space by jointly optimizing the objective functions across different types of input data. At test time, a single input modality (e.g., LiDAR data) is required to generate predictions from the input perspective (i.e., in the LiDAR space), while taking advantages from the model trained with multiple sensor modalities. An extensive evaluation is conducted to show the efficacy of the proposed framework using two benchmark driving datasets.</abstract><doi>10.48550/arxiv.2011.08436</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2011.08436
ispartof
issn
language eng
recordid cdi_arxiv_primary_2011_08436
source arXiv.org
subjects Computer Science - Artificial Intelligence
Computer Science - Computer Vision and Pattern Recognition
Computer Science - Robotics
title Shared Cross-Modal Trajectory Prediction for Autonomous Driving
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-26T08%3A13%3A30IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Shared%20Cross-Modal%20Trajectory%20Prediction%20for%20Autonomous%20Driving&rft.au=Choi,%20Chiho&rft.date=2020-11-15&rft_id=info:doi/10.48550/arxiv.2011.08436&rft_dat=%3Carxiv_GOX%3E2011_08436%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true