Güven, İslamAdam, Evşen Yanmaz2024-01-292024-01-292023-10-30979-840070369-0http://hdl.handle.net/10679/9114https://doi.org/10.1145/3616392.3623414We propose a dynamic path planner that uses a multi-Agent reinforcement learning (MARL) model with novel reward functions for multi-drone search and rescue (SAR) missions. We design a mission environment where a multi-drone team covers an area to detect randomly distributed targets and inform the ground base station (BS) by continuously forming relay chains between the targets and the BS. The training procedure of the agents includes a convolutional neural network (CNN) that uses images which represent trajectory histories and connectivity states of each environment entity such as drones, targets, BS. Agents take actions and get feedback from the environment until the mission is completed. The model is trained with multiple missions with randomized target locations. Our results show that the trained model successfully produces mission plans such that the multi-drone system searches the area efficiently while dynamically forming relay chains. The proposed dynamic method leads up to 45% better total detection and mission times in comparison to a pre-planned optimized path planner.engrestrictedAccessMaintaining connectivity for multi-UAV multi-target search using reinforcement learningconferenceObject10911400112248430001510.1145/3616392.3623414Convolutional neural networksDrone networksMaintaining connectivityMulti-Agent reinforcement learningMulti-UAV path planningReinforcement learningUnmanned aerial vehicles2-s2.0-85178375702