论文标题

语言探索语言抽象和预处理的表示

Semantic Exploration from Language Abstractions and Pretrained Representations

论文作者

Tam, Allison C., Rabinowitz, Neil C., Lampinen, Andrew K., Roy, Nicholas A., Chan, Stephanie C. Y., Strouse, DJ, Wang, Jane X., Banino, Andrea, Hill, Felix

论文摘要

有效的探索是增强学习(RL)的挑战。基于新颖的探索方法可能会在高维状态的空间中受到影响,例如连续的部分观察3D环境。我们通过使用语义上有意义的状态抽象来定义新颖性来应对这一挑战,这可以在自然语言塑造的学习表现中找到。特别是,我们评估了在自然图像标题数据集上预估计的视觉语言表示。我们表明,这些预估计的表示形式推动了有意义的,与任务相关的探索,并提高了3D模拟环境的性能。我们还表征了为什么语言以及如何通过考虑使用预告片模型,语言甲骨文和几种消融的表示的影响来为探索提供有用的抽象。我们在两个非常不同的任务领域中展示了我们的方法的好处 - 一个强调对日常物体的识别和操纵,并且需要在广阔的世界中进行导航探索。我们的结果表明,使用语言形式可以改善在具有挑战性的环境中对各种算法和代理的探索。

Effective exploration is a challenge in reinforcement learning (RL). Novelty-based exploration methods can suffer in high-dimensional state spaces, such as continuous partially-observable 3D environments. We address this challenge by defining novelty using semantically meaningful state abstractions, which can be found in learned representations shaped by natural language. In particular, we evaluate vision-language representations, pretrained on natural image captioning datasets. We show that these pretrained representations drive meaningful, task-relevant exploration and improve performance on 3D simulated environments. We also characterize why and how language provides useful abstractions for exploration by considering the impacts of using representations from a pretrained model, a language oracle, and several ablations. We demonstrate the benefits of our approach in two very different task domains -- one that stresses the identification and manipulation of everyday objects, and one that requires navigational exploration in an expansive world. Our results suggest that using language-shaped representations could improve exploration for various algorithms and agents in challenging environments.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源