论文标题

离线加强学习:关于开放问题的教程,审查和观点

Offline Reinforcement Learning: Tutorial, Review, and Perspectives on Open Problems

论文作者

Levine, Sergey, Kumar, Aviral, Tucker, George, Fu, Justin

论文摘要

在本教程文章中,我们旨在为读者提供开始研究离线增强学习算法所需的概念工具:使用先前收集的数据的强化学习算法,而无需其他在线数据收集。离线增强学习算法具有使大型数据集变成强大决策引擎成为可能的巨大希望。有效的离线增强学习方法将能够从可用数据中提取最大可能的效用的政策,从而可以自动化从医疗保健和教育到机器人技术的广泛决策领域。但是,当前算法的局限性使得这很困难。我们将旨在向读者提供对这些挑战的理解,尤其是在现代深入强化学习方法的背景下,并描述一些潜在的解决方案,这些解决方案已在最近的工作中探讨,以减轻这些挑战,以及最近的应用以及对该领域开放问题的看法的讨论。

In this tutorial article, we aim to provide the reader with the conceptual tools needed to get started on research on offline reinforcement learning algorithms: reinforcement learning algorithms that utilize previously collected data, without additional online data collection. Offline reinforcement learning algorithms hold tremendous promise for making it possible to turn large datasets into powerful decision making engines. Effective offline reinforcement learning methods would be able to extract policies with the maximum possible utility out of the available data, thereby allowing automation of a wide range of decision-making domains, from healthcare and education to robotics. However, the limitations of current algorithms make this difficult. We will aim to provide the reader with an understanding of these challenges, particularly in the context of modern deep reinforcement learning methods, and describe some potential solutions that have been explored in recent work to mitigate these challenges, along with recent applications, and a discussion of perspectives on open problems in the field.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源