Connected and autonomous vehicles across land, water, and air must often operate in dynamic, unpredictable environments with limited communication, no centralized control, and partial observability. These real-world constraints pose significant challenges for coordination, particularly when vehicles pursue individual objectives. To address this, we propose a decentralized Multi-Agent Reinforcement Learning (MARL) framework that enables vehicles, acting as agents, to communicate selectively based on local goals and observations. This goal-aware communication strategy allows agents to share only relevant information, enhancing collaboration while respecting visibility limitations. We validate our approach in complex multi-agent navigation tasks featuring obstacles and dynamic agent populations. Results show that our method significantly improves task success rates and reduces time-to-goal compared to non-cooperative baselines. Moreover, task performance remains stable as the number of agents increases, demonstrating scalability. These findings highlight the potential of decentralized, goal-driven MARL to support effective coordination in realistic multi-vehicle systems operating across diverse domains.
翻译:陆地、水域及空域中的联网与自主车辆常常需要在动态、不可预测的环境中运行,这些环境具有通信受限、无集中控制及部分可观测性等特点。这些现实约束对协调机制提出了重大挑战,尤其在车辆追求个体目标时更为突出。为此,我们提出了一种去中心化的多智能体强化学习框架,使作为智能体的车辆能够基于局部目标与观测进行选择性通信。这种目标感知的通信策略允许智能体仅共享相关信息,在尊重可见性限制的同时增强了协作能力。我们在包含障碍物和动态智能体群体的复杂多智能体导航任务中验证了该方法。结果表明,与非协作基线相比,我们的方法显著提高了任务成功率并缩短了到达目标的时间。此外,随着智能体数量增加,任务性能保持稳定,体现了良好的可扩展性。这些发现凸显了去中心化、目标驱动的多智能体强化学习在支持跨领域真实多车辆系统实现有效协调方面的潜力。