
Research Article
Deep Reinforcement Learning-Based Joint Task Offloading and Radio Resource Allocation for Platoon-Assisted Vehicular Edge Computing
@INPROCEEDINGS{10.1007/978-3-030-63941-9_6, author={Yi Chen and Xinyu Hu and Haoye Chai and Ke Zhang and Fan Wu and Lisha Gu}, title={Deep Reinforcement Learning-Based Joint Task Offloading and Radio Resource Allocation for Platoon-Assisted Vehicular Edge Computing}, proceedings={6GN for Future Wireless Networks. Third EAI International Conference, 6GN 2020, Tianjin, China, August 15-16, 2020, Proceedings}, proceedings_a={6GN}, year={2021}, month={1}, keywords={Vehicular edge computing Task offloading Resource allocation Platoon}, doi={10.1007/978-3-030-63941-9_6} }
- Yi Chen
Xinyu Hu
Haoye Chai
Ke Zhang
Fan Wu
Lisha Gu
Year: 2021
Deep Reinforcement Learning-Based Joint Task Offloading and Radio Resource Allocation for Platoon-Assisted Vehicular Edge Computing
6GN
Springer
DOI: 10.1007/978-3-030-63941-9_6
Abstract
Platoons, formed by smart vehicles driving in the same patterns, bring potential benefits to road traffic efficiency while providing a promising paradigm to execute computation tasks with onboard computing resources. However, constrained resources of individual vehicles (IV), limited wireless coverage of vehicular communication nodes as well as high mobility of running platoons pose critical challenges on task scheduling and resource management. To address these challenges, we propose a platoon-based vehicular edge computing mechanism, which exploits computation capabilities of both platoons and edge computing enabled Roadside Units (RSUs), and jointly optimizes task offloading target selection and resource allocation. Taking aim at minimize delay cost and energy consumption of the platoon-based task execution, we leverage deep deterministic policy gradient (DDPG) to design a learning algorithm, which efficiently determines target computation servers and obtains optimized resource scheduling strategies. Numerical results demonstrate that our algorithm significantly reduces delay and energy costs in comparing its performance to that of benchmark schemes.