RIS-Empowered MEC for URLLC Systems with Digital-Twin-Driven Architecture

This paper investigates a digital twin (DT) and reconfigurable intelligent surface (RIS)-aided mobile edge computing (MEC) system under given constraints on ultra-reliable low latency communication (URLLC). In particular, we focus on the problem of total end-to-end (E2E) latency minimization for the...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE transactions on communications 2024-04, Vol.72 (4), p.1-1
Hauptverfasser: Kurma, Sravani, Katwe, Mayur, Singh, Keshav, Pan, Cunhua, Mumtaz, Shahid, Li, Chih-Peng
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:This paper investigates a digital twin (DT) and reconfigurable intelligent surface (RIS)-aided mobile edge computing (MEC) system under given constraints on ultra-reliable low latency communication (URLLC). In particular, we focus on the problem of total end-to-end (E2E) latency minimization for the considered system under the joint optimization of beamforming design at the RIS, power, bandwidth allocation, processing rates, and task offloading parameters using DT architecture. To tackle the formulated non-convex optimization problem, we first model it as a Markov decision process (MDP). Later, we adopt deep deterministic policy gradient (DDPG) based deep reinforcement learning (DRL) algorithm to solve it effectively. We have compared the DDPG results with proximal policy optimization (PPO), modified PPO (M-PPO), and conventional alternating optimization (AO) algorithms. Simulation results depict that the proposed DT-enabled resource allocation scheme for the RIS-empowered MEC network using DDPG algorithm achieves up to 60% lower transmission delay and 20% lower energy consumption compared to the scheme without an RIS. This confirms the practical advantages of leveraging RIS technology in MEC systems. Results demonstrate that DDPG outperforms M-PPO and PPO in terms of higher reward value and better learning efficiency, while M-PPO and PPO exhibit lower execution time than DDPG and AO due to their advanced policy optimization techniques. Thus, the results validate the effectiveness of the DRL solutions over AO for dynamic resource allocation w.r.t. reduced execution time.
ISSN:0090-6778
1558-0857
DOI:10.1109/TCOMM.2023.3333345