IEEE INTERNET OF THINGS JOURNAL, v.12, no.10, pp.14484 - 14497
Abstract
This article studies a new multiagent deep reinforcement learning (MADRL) approach for unmanned aerial vehicle (UAV)-assisted mobile edge computing (MEC) networks, where AAV-mounted servers provide offloading services to mobile users (MUs). We aim to minimize the total energy consumption of MUs by optimizing AAV mobility, AAV-MU association, resource allocation, and task offloading ratios. In the multi-AAV scenario, we model the MEC network as a multiagent partially observable Markov decision process (POMDP), where each AAV agent operates with limited information for decentralized decision-making. Conventional MADRL methods manually design such AAV interaction messages, thereby incurring performance degradation. To address this issue, we propose a new neural network (NN)-based AAV interaction mechanism that generates autonomously task-oriented messages to minimize energy consumption. Such message-generating NNs are developed under the MADRL framework, which allows for joint optimization of AAV interactions and decentralized decisions in an end-to-end manner. Numerical results demonstrate that our approach outperforms traditional MADRL methods and achieves performance close to ideal centralized schemes while maintaining scalability with varying AAV numbers.