论文标题
通过SIM到sim到sim-to-seg:端到端的越野自动驾驶,没有真实数据
Sim-to-Real via Sim-to-Seg: End-to-end Off-road Autonomous Driving Without Real Data
论文作者
论文摘要
自动驾驶很复杂,需要复杂的3D场景理解,本地化,映射和控制。我们不是通过强化学习(RL)来考虑一种端到端的方法,而不是明确建模和融合这些组件。但是,在现实世界中收集勘探驾驶数据是不切实际和危险的。虽然在模拟和部署视觉SIM到现实技术方面的培训对机器人操纵效果很好,但部署超越受控的工作区观点仍然是一个挑战。在本文中,我们通过提出Sim2Seg来应对这一挑战,Sim2Seg是RCAN的重新想象,它越过了视觉现实差距,无需使用任何真实的数据,就可以越过越野自动驾驶。这是通过学习将随机模拟图像转换为模拟分割和深度图的方法来完成的,随后还可以翻译现实世界的图像。这使我们能够在模拟中训练端到端的RL策略,并直接部署在现实世界中。我们的方法可以在1 GPU的48小时内进行培训,并且可以同等地执行,并且可以进行经典的感知和控制堆栈,在几个月内花费了数千个工程小时。我们希望这项工作激发了未来的端到端自动驾驶研究。
Autonomous driving is complex, requiring sophisticated 3D scene understanding, localization, mapping, and control. Rather than explicitly modelling and fusing each of these components, we instead consider an end-to-end approach via reinforcement learning (RL). However, collecting exploration driving data in the real world is impractical and dangerous. While training in simulation and deploying visual sim-to-real techniques has worked well for robot manipulation, deploying beyond controlled workspace viewpoints remains a challenge. In this paper, we address this challenge by presenting Sim2Seg, a re-imagining of RCAN that crosses the visual reality gap for off-road autonomous driving, without using any real-world data. This is done by learning to translate randomized simulation images into simulated segmentation and depth maps, subsequently enabling real-world images to also be translated. This allows us to train an end-to-end RL policy in simulation, and directly deploy in the real-world. Our approach, which can be trained in 48 hours on 1 GPU, can perform equally as well as a classical perception and control stack that took thousands of engineering hours over several months to build. We hope this work motivates future end-to-end autonomous driving research.