Minimization of Energy and Service Latency Computation Offloading using Neural Network in 5G NOMA System

The future Internet of Things (IoT) era is anticipated to support computation-intensive and time-critical applications using edge computing for mobile (MEC), which is regarded as promising technique. However, the transmitting uplink performance will be highly impacted by the hostile wireless channel...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:International Journal of Electronics and Telecommunications 2023-01, Vol.69 (4), p.661-667
Hauptverfasser: Suprith, P G, Riyaz Ahmed, Mohammed
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:The future Internet of Things (IoT) era is anticipated to support computation-intensive and time-critical applications using edge computing for mobile (MEC), which is regarded as promising technique. However, the transmitting uplink performance will be highly impacted by the hostile wireless channel, the low bandwidth, and the low transmission power of IoT devices. Using edge computing for mobile (MEC) to offload tasks becomes a crucial technology to reduce service latency for computation-intensive applications and reduce the computational workloads of mobile devices. Under the restrictions of computation latency and cloud computing capacity, our goal is to reduce the overall energy consumption of all users, including transmission energy and local computation energy. In this article, the Deep Q Network Algorithm (DQNA) to deal with the data rates with respect to the user base in different time slots of 5G NOMA network. The DQNA is optimized by considering more number of cell structures like 2, 4, 6 and 8. Therefore, the DQNA provides the optimal distribution of power among all 3 users in the 5G network, which gives the increased data rates. The existing various power distribution algorithms like frequent pattern (FP), weighted least squares mean error weighted least squares mean error (WLSME), and Random Power and Maximal Power allocation are used to justify the proposed DQNA technique. The proposed technique which gives 81.6% more the data rates when increased the cell structure to 8. Thus 25% more in comparison to other algorithms like FP, WLSME Random Power and Maximal Power allocation.
ISSN:2081-8491
2300-1933
DOI:10.24425/ijet.2023.147685