Task offloading method of edge computing in internet of vehicles based on deep reinforcement learning
Compared with the traditional network tasks, the emerging Internet of Vehicles (IoV) technology has higher requirements for network bandwidth and delay. However, due to the limitation of computing resources and battery capacity of existing mobile devices, it is hard to meet the above requirements. H...
Gespeichert in:
Veröffentlicht in: | Cluster computing 2022-04, Vol.25 (2), p.1175-1187 |
---|---|
Hauptverfasser: | , , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Compared with the traditional network tasks, the emerging Internet of Vehicles (IoV) technology has higher requirements for network bandwidth and delay. However, due to the limitation of computing resources and battery capacity of existing mobile devices, it is hard to meet the above requirements. How to complete task offloading and calculation with lower task delay and lower energy consumption is the most important issue. Aiming at the task offloading system of the IoV, this paper considers the situation of multiple MEC servers when modeling, and proposes a dynamic task offloading scheme based on deep reinforcement learning. It improves the traditional Q-Learning algorithm and combines deep learning with reinforcement learning to avoid dimensional disaster in the Q-Learning algorithm. Simulation results show that the proposed algorithm has better performance on delay, energy consumption, and total system overhead under the different number of tasks and wireless channel bandwidth. |
---|---|
ISSN: | 1386-7857 1573-7543 |
DOI: | 10.1007/s10586-021-03532-9 |