This paper addresses the joint design of Unmanned Aerial Vehicles (UAVs) trajectory and radio resource management (RRM) in dynamic wireless environments by leveraging a multi-agent deep reinforcement learning (MADRL) framework. In contrast to prior works that either assume constant synchronization between agents and the controller or overlook the communication cost, we explicitly model the interaction between UAVs and the central controller. We propose an adaptive synchronization strategy that selectively transmits model parameters and experience data based on their relevance, enabling a resource-aware RRM algorithm that optimally balances learning performance and communication overhead. The MADRL agents optimize their trajectories based on rewards that incorporate priorities derived from the RRM layer, which jointly manages both uplink and downlink communications. Simulation results demonstrate that our event-driven synchronization strategy outperforms periodic baselines in both convergence speed and communication overhead
Adaptive Communication for Joint Trajectory and RRM in MADRL-Based UAV Networks / Ferretti, Danila; Spampinato, Leonardo; Testi, Enrico; Buratti, Chiara; Marini, Riccardo. - (2025). (Intervento presentato al convegno 2025 21st International Conference on Wireless and Mobile Computing, Networking and Communications (WiMob) tenutosi a Marrakech, Marocco).
Adaptive Communication for Joint Trajectory and RRM in MADRL-Based UAV Networks
Danila FerrettiPrimo
;
2025
Abstract
This paper addresses the joint design of Unmanned Aerial Vehicles (UAVs) trajectory and radio resource management (RRM) in dynamic wireless environments by leveraging a multi-agent deep reinforcement learning (MADRL) framework. In contrast to prior works that either assume constant synchronization between agents and the controller or overlook the communication cost, we explicitly model the interaction between UAVs and the central controller. We propose an adaptive synchronization strategy that selectively transmits model parameters and experience data based on their relevance, enabling a resource-aware RRM algorithm that optimally balances learning performance and communication overhead. The MADRL agents optimize their trajectories based on rewards that incorporate priorities derived from the RRM layer, which jointly manages both uplink and downlink communications. Simulation results demonstrate that our event-driven synchronization strategy outperforms periodic baselines in both convergence speed and communication overheadI documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.


