Back to Search Start Over

Deep Reinforcement Learning techniques for dynamic task offloading in the 5G edge-cloud continuum.

Authors :
Nieto, Gorka
de la Iglesia, Idoia
Lopez-Novoa, Unai
Perfecto, Cristina
Source :
Journal of Cloud Computing (2192-113X); 5/3/2024, Vol. 13 Issue 1, p1-24, 24p
Publication Year :
2024

Abstract

The integration of new Internet of Things (IoT) applications and services heavily relies on task offloading to external devices due to the constrained computing and battery resources of IoT devices. Up to now, Cloud Computing (CC) paradigm has been a good approach for tasks where latency is not critical, but it is not useful when latency matters, so Multi-access Edge Computing (MEC) can be of use. In this work, we propose a distributed Deep Reinforcement Learning (DRL) tool to optimize the binary task offloading decision, this is, the independent decision of where to execute each computing task, depending on many factors. The optimization goal in this work is to maximize the Quality-of-Experience (QoE) when performing tasks, which is defined as a metric related to the battery level of the UE, but subject to satisfying tasks' latency requirements. This distributed DRL approach, specifically an Actor-Critic (AC) algorithm running on each User Equipment (UE), is evaluated through the simulation of two distinct scenarios and outperforms other analyzed baselines in terms of QoE values and/or energy consumption in dynamic environments, also demonstrating that decisions need to be adapted to the environment's evolution. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
2192113X
Volume :
13
Issue :
1
Database :
Complementary Index
Journal :
Journal of Cloud Computing (2192-113X)
Publication Type :
Academic Journal
Accession number :
177044806
Full Text :
https://doi.org/10.1186/s13677-024-00658-0