论文标题
可解释的对象引起的自动驾驶汽车的行动决定
Explainable Object-induced Action Decision for Autonomous Vehicles
论文作者
论文摘要
提出了一个新的范式用于自动驾驶。新范式在于端到端和管道方法之间,并受到人类如何解决问题的启发。尽管它依赖场景的理解,但后者仅考虑可能引起危险的物体。这些被称为动作诱导,因为其状态的变化应触发车辆行动。他们还为这些动作定义了一组解释,这些解释应与后者共同生产。提出了BDD100K数据集的扩展,该数据集的注释是针对4种操作和21个解释的注释。然后引入了该问题的新的多任务公式,该公式优化了两个动作命令和解释的准确性。最终提出了CNN架构来解决此问题,通过结合有关诱导对象和全局场景上下文的推理。实验结果表明,解释的需求改善了对动作诱导对象的识别,这又导致了更好的动作预测。
A new paradigm is proposed for autonomous driving. The new paradigm lies between the end-to-end and pipelined approaches, and is inspired by how humans solve the problem. While it relies on scene understanding, the latter only considers objects that could originate hazard. These are denoted as action-inducing, since changes in their state should trigger vehicle actions. They also define a set of explanations for these actions, which should be produced jointly with the latter. An extension of the BDD100K dataset, annotated for a set of 4 actions and 21 explanations, is proposed. A new multi-task formulation of the problem, which optimizes the accuracy of both action commands and explanations, is then introduced. A CNN architecture is finally proposed to solve this problem, by combining reasoning about action inducing objects and global scene context. Experimental results show that the requirement of explanations improves the recognition of action-inducing objects, which in turn leads to better action predictions.