论文标题
WHO2COM:通过可学习的握手交流的协作感知
Who2com: Collaborative Perception via Learnable Handshake Communication
论文作者
论文摘要
在本文中,我们提出了协作感知的问题,其中机器人可以以可学习的方式将其本地观察结果与邻近代理相结合,以提高感知任务的准确性。与机器人技术和多代理增强学习中的现有工作不同,我们将问题作为一个问题,必须以带宽敏感的方式在一组代理中共享学习的信息,以优化以理解语义细分等场景任务。受网络通信协议的启发,我们提出了一种多阶段握手通信机制,神经网络可以学会压缩每个阶段所需的相关信息。具体来说,具有退化传感器数据的目标代理发送压缩请求,其他代理以匹配分数响应,目标代理确定与谁联系(即接收信息)。我们还基于Airsim模拟器开发了AirSim-CP数据集和指标,其中一组空中机器人感知到多种景观,例如道路,草原,建筑物等。我们表明,对于语义分段任务,我们的手牌沟通方法将准确的沟通准确性远高于分散的基地和替代者,而跨越了20%,则可以将其用于综合的基地和占用范围。
In this paper, we propose the problem of collaborative perception, where robots can combine their local observations with those of neighboring agents in a learnable way to improve accuracy on a perception task. Unlike existing work in robotics and multi-agent reinforcement learning, we formulate the problem as one where learned information must be shared across a set of agents in a bandwidth-sensitive manner to optimize for scene understanding tasks such as semantic segmentation. Inspired by networking communication protocols, we propose a multi-stage handshake communication mechanism where the neural network can learn to compress relevant information needed for each stage. Specifically, a target agent with degraded sensor data sends a compressed request, the other agents respond with matching scores, and the target agent determines who to connect with (i.e., receive information from). We additionally develop the AirSim-CP dataset and metrics based on the AirSim simulator where a group of aerial robots perceive diverse landscapes, such as roads, grasslands, buildings, etc. We show that for the semantic segmentation task, our handshake communication method significantly improves accuracy by approximately 20% over decentralized baselines, and is comparable to centralized ones using a quarter of the bandwidth.