论文标题
使用半正式验证评估深钢筋学习模型的安全性
Evaluating the Safety of Deep Reinforcement Learning Models using Semi-Formal Verification
论文作者
论文摘要
深入的加强学习(DRL)在解决实践决策问题方面取得了突破性的成功。特别是机器人技术可能涉及高成本的硬件和人类互动。因此,需要对训练有素的模型进行严格评估,以避免在操作环境中不安全的行为。但是,设计指标来衡量神经网络的安全性是一个开放的问题,因为标准评估参数(例如,全部奖励)的信息不足。在本文中,我们基于间隔分析提出了一种半正式验证方法,用于决策任务,该方法解决了先前验证框架和设计指标的计算要求以衡量模型的安全性。我们的方法相对于正式验证符获得了标准基准测试的可比结果,同时大大减少了计算时间。此外,我们的方法允许在实用应用中有效评估决策模型的安全性能,例如移动机器人无地图导航和操纵器的轨迹生成。
Groundbreaking successes have been achieved by Deep Reinforcement Learning (DRL) in solving practical decision-making problems. Robotics, in particular, can involve high-cost hardware and human interactions. Hence, scrupulous evaluations of trained models are required to avoid unsafe behaviours in the operational environment. However, designing metrics to measure the safety of a neural network is an open problem, since standard evaluation parameters (e.g., total reward) are not informative enough. In this paper, we present a semi-formal verification approach for decision-making tasks, based on interval analysis, that addresses the computational demanding of previous verification frameworks and design metrics to measure the safety of the models. Our method obtains comparable results over standard benchmarks with respect to formal verifiers, while drastically reducing the computation time. Moreover, our approach allows to efficiently evaluate safety properties for decision-making models in practical applications such as mapless navigation for mobile robots and trajectory generation for manipulators.