Generalized Multi-Objective Reinforcement Learning with Envelope Updates in URLLC-enabled Vehicular Networks
We develop a novel multi-objective reinforcement learning (MORL) framework to jointly optimize wireless network selection and autonomous driving policies in a multi-band vehicular network operating on conventional sub-6GHz spectrum and Terahertz frequencies. The proposed framework is designed to 1....
Gespeichert in:
Hauptverfasser: | , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | We develop a novel multi-objective reinforcement learning (MORL) framework to
jointly optimize wireless network selection and autonomous driving policies in
a multi-band vehicular network operating on conventional sub-6GHz spectrum and
Terahertz frequencies. The proposed framework is designed to 1. maximize the
traffic flow and 2. minimize collisions by controlling the vehicle's motion
dynamics (i.e., speed and acceleration), and enhance the ultra-reliable
low-latency communication (URLLC) while minimizing handoffs (HOs). We cast this
problem as a multi-objective Markov Decision Process (MOMDP) and develop
solutions for both predefined and unknown preferences of the conflicting
objectives. Specifically, deep-Q-network and double deep-Q-network-based
solutions are developed first that consider scalarizing the transportation and
telecommunication rewards using predefined preferences. We then develop a novel
envelope MORL solution which develop policies that address multiple objectives
with unknown preferences to the agent. While this approach reduces reliance on
scalar rewards, policy effectiveness varying with different preferences is a
challenge. To address this, we apply a generalized version of the Bellman
equation and optimize the convex envelope of multi-objective Q values to learn
a unified parametric representation capable of generating optimal policies
across all possible preference configurations. Following an initial learning
phase, our agent can execute optimal policies under any specified preference or
infer preferences from minimal data samples.Numerical results validate the
efficacy of the envelope-based MORL solution and demonstrate interesting
insights related to the inter-dependency of vehicle motion dynamics, HOs, and
the communication data rate. The proposed policies enable autonomous vehicles
to adopt safe driving behaviors with improved connectivity. |
---|---|
DOI: | 10.48550/arxiv.2405.11331 |