论文标题

Mira:机器人负担的心理图像

MIRA: Mental Imagery for Robotic Affordances

论文作者

Yen-Chen, Lin, Florence, Pete, Zeng, Andy, Barron, Jonathan T., Du, Yilun, Ma, Wei-Chiu, Simeonov, Anthony, Garcia, Alberto Rodriguez, Isola, Phillip

论文摘要

人类形成3D场景的心理图像,以支持反事实的想象力,计划和运动控制。我们从以前未观察到的观点预测场景的外观和负担能力有助于我们执行操纵任务(例如,6-DOF套件),并以一种轻松的水平来实现现有的机器人学习框架,这是无法触及的。在这项工作中,我们旨在构建可以在想象中的图像之上类似地计划动作的人造系统。为此,我们介绍了用于机器人负担能力的心理图像(MIRA),这是一个动作推理框架,可在循环中使用新颖的视图合成和负担能力预测来优化动作。给定一组2D RGB图像,MIRA构建了一致的3D场景表示形式,通过该表示,我们合成了可与Pixel-Wise的新型拼字图视图,可提供动作优化的预测。我们说明了这种优化过程如何使我们能够概括到有限的示范范围内对6多种机器人操纵任务的平面外旋转,从而铺平了向机器铺平道路,这些机器自主学习了解周围的世界以进行计划行动。

Humans form mental images of 3D scenes to support counterfactual imagination, planning, and motor control. Our abilities to predict the appearance and affordance of the scene from previously unobserved viewpoints aid us in performing manipulation tasks (e.g., 6-DoF kitting) with a level of ease that is currently out of reach for existing robot learning frameworks. In this work, we aim to build artificial systems that can analogously plan actions on top of imagined images. To this end, we introduce Mental Imagery for Robotic Affordances (MIRA), an action reasoning framework that optimizes actions with novel-view synthesis and affordance prediction in the loop. Given a set of 2D RGB images, MIRA builds a consistent 3D scene representation, through which we synthesize novel orthographic views amenable to pixel-wise affordances prediction for action optimization. We illustrate how this optimization process enables us to generalize to unseen out-of-plane rotations for 6-DoF robotic manipulation tasks given a limited number of demonstrations, paving the way toward machines that autonomously learn to understand the world around them for planning actions.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源