论文标题

部分可观测时空混沌系统的无模型预测

Synthesizing Adversarial Visual Scenarios for Model-Based Robotic Control

论文作者

Agarwal, Shubhankar, Chinchali, Sandeep P.

论文摘要

当今的机器人经常与数据驱动的感知和计划模型与经典模型预测控制器(MPC)接口。通常,这种学识渊博的感知/计划模型会在分布(OOD)甚至对抗性视觉输入上产生错误的航路点预测,从而增加了控制成本。但是,当今训练健壮感知模型的方法在很大程度上是任务不合时宜的 - 它们使用随机图像转换或孤立针对视觉模型的对抗性示例来增强数据集。因此,他们经常引入最终良性控制的像素扰动。与先前的工作相反,该工作将对抗性示例用于单步视觉任务,我们的关键贡献是合成针对多步,基于模型的控制量身定制的对抗场景。为此,我们使用可区分的MPC方法来计算基于模型的控制器对状态估计中错误的灵敏度。我们表明,与标准的任务无关数据的增强相比,这些对抗性数据集的重新训练视觉模型将OOD测试方案的控制性能提高了36.2%。我们在机器人导航,机器人中的操纵以及对自动驾驶汽车的控制的示例中演示了我们的方法。

Today's robots often interface with data-driven perception and planning models with classical model-predictive controllers (MPC). Often, such learned perception/planning models produce erroneous waypoint predictions on out-of-distribution (OoD) or even adversarial visual inputs, which increase control costs. However, today's methods to train robust perception models are largely task-agnostic - they augment a dataset using random image transformations or adversarial examples targeted at the vision model in isolation. As such, they often introduce pixel perturbations that are ultimately benign for control. In contrast to prior work that synthesizes adversarial examples for single-step vision tasks, our key contribution is to synthesize adversarial scenarios tailored to multi-step, model-based control. To do so, we use differentiable MPC methods to calculate the sensitivity of a model-based controller to errors in state estimation. We show that re-training vision models on these adversarial datasets improves control performance on OoD test scenarios by up to 36.2% compared to standard task-agnostic data augmentation. We demonstrate our method on examples of robotic navigation, manipulation in RoboSuite, and control of an autonomous air vehicle.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源