
Research Article
Double DQN Reinforcement Learning-Based Computational Offloading and Resource Allocation for MEC
@INPROCEEDINGS{10.1007/978-3-031-55471-1_18, author={Chen Zhang and Chunrong Peng and Min Lin and Zhaoyang Du and Celimuge Wu}, title={Double DQN Reinforcement Learning-Based Computational Offloading and Resource Allocation for MEC}, proceedings={Mobile Networks and Management. 13th EAI International Conference, MONAMI 2023, Yingtan, China, October 27-29, 2023, Proceedings}, proceedings_a={MONAMI}, year={2024}, month={3}, keywords={mobile edge computing computational offloading resource allocation single MEC server-multi-user DDQN algorithm}, doi={10.1007/978-3-031-55471-1_18} }
- Chen Zhang
Chunrong Peng
Min Lin
Zhaoyang Du
Celimuge Wu
Year: 2024
Double DQN Reinforcement Learning-Based Computational Offloading and Resource Allocation for MEC
MONAMI
Springer
DOI: 10.1007/978-3-031-55471-1_18
Abstract
In recent years, numerous Deep Reinforcement Learning (DRL) neural network models have been proposed to optimize computational offloading and resource allocation in Mobile Edge Computing (MEC). However, the diversity of computational tasks and the complexity of 5G networks pose significant challenges for current DRL algorithms apply to MEC scenarios. This research focuses on a single MEC server-multi-user scenario and develops a realistic small-scale MEC offloading system. In order to alleviate the problem of overestimation of action value in current Deep Q-learning Network (DQN), we propose a normalized model of Complex network based on Double DQN (DDQN) algorithm to determine the optimal computational offloading and resource allocation strategy. Simulation results demonstrate that DDQN outperforms conventional approaches such as fixed parameter policies and DQN regarding convergence speed, energy consumption and latency. This research showcases the potential of DDQN for achieving efficient optimization in MEC environments.