论文标题
学会使用负担能图
Learning to Move with Affordance Maps
论文作者
论文摘要
自主探索和浏览物理空间的能力是几乎任何移动自主剂,从家用机器人真空到自动驾驶汽车的基本要求。传统的基于大满贯的方法进行探索和导航主要集中于利用场景几何形状,但无法建模动态对象(例如其他代理)或语义限制(例如湿地板或门口)。基于学习的RL代理是一种有吸引力的替代方法,因为它们可以同时结合语义和几何信息,但众所周知,它们效率低下,难以推广到新颖的环境,并且难以解释。在本文中,我们将两全其美的最佳方法与模块化方法相结合,该方法学习了一个场景的空间表示,该场景与传统的几何规划师相结合时经过训练可以有效。具体而言,我们设计了一个学者,该代理商可以通过积极的自我监督经验聚集来预测一个空间负担能力图,该空间负担能力图可以阐明场景的哪些部分可以导航。与大多数假定静态世界的模拟环境相反,我们使用包含各种动态参与者和危害的大规模随机生成的地图评估了Vizdoom模拟器中的方法。我们表明,学到的负担能力图可用于增强探索和导航的传统方法,从而提供了显着改善的性能。
The ability to autonomously explore and navigate a physical space is a fundamental requirement for virtually any mobile autonomous agent, from household robotic vacuums to autonomous vehicles. Traditional SLAM-based approaches for exploration and navigation largely focus on leveraging scene geometry, but fail to model dynamic objects (such as other agents) or semantic constraints (such as wet floors or doorways). Learning-based RL agents are an attractive alternative because they can incorporate both semantic and geometric information, but are notoriously sample inefficient, difficult to generalize to novel settings, and are difficult to interpret. In this paper, we combine the best of both worlds with a modular approach that learns a spatial representation of a scene that is trained to be effective when coupled with traditional geometric planners. Specifically, we design an agent that learns to predict a spatial affordance map that elucidates what parts of a scene are navigable through active self-supervised experience gathering. In contrast to most simulation environments that assume a static world, we evaluate our approach in the VizDoom simulator, using large-scale randomly-generated maps containing a variety of dynamic actors and hazards. We show that learned affordance maps can be used to augment traditional approaches for both exploration and navigation, providing significant improvements in performance.