论文标题

强化学习域的课程学习:框架和调查

Curriculum Learning for Reinforcement Learning Domains: A Framework and Survey

论文作者

Narvekar, Sanmit, Peng, Bei, Leonetti, Matteo, Sinapov, Jivko, Taylor, Matthew E., Stone, Peter

论文摘要

强化学习(RL)是解决顺序决策任务的流行范式,其中代理只有有限的环境反馈。尽管在过去的三十年中取得了许多进步,但在许多领域中学习仍然需要与环境进行大量互动,在现实情况下,这可能会非常昂贵。为了解决这个问题,已将转移学习用于加强学习,以便在开始学习下一个更艰巨的任务时可以利用一项任务中获得的经验。最近,几项研究探索了如何将任务或数据样本本身测序到课程中,以学习一个问题,否则这些问题可能太难从头开始学习。在本文中,我们介绍了一个在强化学习中学习课程学习框架(CL),并利用它根据其假设,能力和目标来调查和对现有的CL方法进行分类。最后,我们使用框架来寻找开放问题,并为未来的RL课程学习研究提出指示。

Reinforcement learning (RL) is a popular paradigm for addressing sequential decision tasks in which the agent has only limited environmental feedback. Despite many advances over the past three decades, learning in many domains still requires a large amount of interaction with the environment, which can be prohibitively expensive in realistic scenarios. To address this problem, transfer learning has been applied to reinforcement learning such that experience gained in one task can be leveraged when starting to learn the next, harder task. More recently, several lines of research have explored how tasks, or data samples themselves, can be sequenced into a curriculum for the purpose of learning a problem that may otherwise be too difficult to learn from scratch. In this article, we present a framework for curriculum learning (CL) in reinforcement learning, and use it to survey and classify existing CL methods in terms of their assumptions, capabilities, and goals. Finally, we use our framework to find open problems and suggest directions for future RL curriculum learning research.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源