Artificial Intelligence for Communications and Networks. Second EAI International Conference, AICON 2020, Virtual Event, December 19-20, 2020, Proceedings

Research Article

Energy-Efficient Multi-UAV-Enabled Computation Offloading for Industrial Internet of Things via Deep Reinforcement Learning

Download
46 downloads
  • @INPROCEEDINGS{10.1007/978-3-030-69066-3_26,
        author={Shuo Shi and Meng Wang and Xuemai Gu},
        title={Energy-Efficient Multi-UAV-Enabled Computation Offloading for Industrial Internet of Things via Deep Reinforcement Learning},
        proceedings={Artificial Intelligence for Communications and Networks. Second EAI International Conference, AICON 2020, Virtual Event, December 19-20, 2020, Proceedings},
        proceedings_a={AICON},
        year={2021},
        month={7},
        keywords={UAV-enabled IIoT Deep reinforcement learning Computation offloading},
        doi={10.1007/978-3-030-69066-3_26}
    }
    
  • Shuo Shi
    Meng Wang
    Xuemai Gu
    Year: 2021
    Energy-Efficient Multi-UAV-Enabled Computation Offloading for Industrial Internet of Things via Deep Reinforcement Learning
    AICON
    Springer
    DOI: 10.1007/978-3-030-69066-3_26
Shuo Shi1, Meng Wang1, Xuemai Gu1
  • 1: Harbin Institute of Technology

Abstract

Industrial Internet of things (IIoT) has been envisioned as a key technology for Industry 4.0. However, the battery capcity and processing ability of IIoT devices are limited which imposes great challenges when handling tasks with high quality of service (QoS) requirements. Toward this end, in this paper we first use multiple unmanned aerial vehicles (UAVs) equipped with computation resources to offer computation offloading opportunities for IIoT devices due to their high flexibility. Then we formulate the multi-UAV-enabled computation offloading problem as a mixed integer non-linear programming (MINLP) problem and prove its NP-hardness. Furthermore, to obtain the energy-efficient solutions for IIoT devices, we propose an intelligent algorithm called multi-agent deep Q-learning with stochastic prioritized replay (MDSPR). Simulation results show that the proposed MDSPR converges fast and outperforms the normal deep Q-learning (DQN) method and other benchmark algorithms in terms of energy-efficiency and tasks’ successful rate.