论文标题

可解释的端到端城市自动驾驶,并通过潜在的深度强化学习

Interpretable End-to-end Urban Autonomous Driving with Latent Deep Reinforcement Learning

论文作者

Chen, Jianyu, Li, Shengbo Eben, Tomizuka, Masayoshi

论文摘要

与流行的模块化框架不同,端到端的自主驾驶旨在以综合方式解决感知,决策和控制问题,这可以更适合新场景,并且更容易在大规模上概括。但是,现有的端到端方法通常缺乏解释性,只能处理诸如Lane Keep之类的简单驾驶任务。在本文中,我们提出了一种用于端到端自动驾驶的可解释的深入增强学习方法,该方法能够处理复杂的城市场景。通过增强学习过程共同引入和学习了一个连续的潜在环境模型。借助这种潜在模型,可以生成语义鸟掩模,该面膜被强制在当今的模块化框架中与某个中间属性联系起来,以解释学到的政策的行为。潜在空间还大大降低了增强学习的样本复杂性。与卡拉中的模拟自动驾驶汽车进行比较测试表明,我们方法在城市场景中的性能在拥挤的周围车辆中占主导地位,包括DQN,DDPG,TD3和SAC。此外,通过掩盖的产出,学识渊博的政策能够更好地解释汽车如何理解驾驶环境。这项工作的代码和视频可在我们的GitHub Repo和Project网站上找到。

Unlike popular modularized framework, end-to-end autonomous driving seeks to solve the perception, decision and control problems in an integrated way, which can be more adapting to new scenarios and easier to generalize at scale. However, existing end-to-end approaches are often lack of interpretability, and can only deal with simple driving tasks like lane keeping. In this paper, we propose an interpretable deep reinforcement learning method for end-to-end autonomous driving, which is able to handle complex urban scenarios. A sequential latent environment model is introduced and learned jointly with the reinforcement learning process. With this latent model, a semantic birdeye mask can be generated, which is enforced to connect with a certain intermediate property in today's modularized framework for the purpose of explaining the behaviors of learned policy. The latent space also significantly reduces the sample complexity of reinforcement learning. Comparison tests with a simulated autonomous car in CARLA show that the performance of our method in urban scenarios with crowded surrounding vehicles dominates many baselines including DQN, DDPG, TD3 and SAC. Moreover, through masked outputs, the learned policy is able to provide a better explanation of how the car reasons about the driving environment. The codes and videos of this work are available at our github repo and project website.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源