论文标题

同时从人类姿势和对象线索中学习实时活动识别

Simultaneous Learning from Human Pose and Object Cues for Real-Time Activity Recognition

论文作者

Reily, Brian, Zhu, Qingzhao, Reardon, Christopher, Zhang, Hao

论文摘要

实时人类活动识别在现实世界中以人类为中心的机器人技术(例如辅助生活和人类机器人协作)中起着至关重要的作用。尽管以前基于编码人类姿势的骨骼数据的方法显示了对实时活动识别的有希望的结果,但他们缺乏考虑场景中对象和人类使用中对象提供的上下文的能力,这可以在人类活动类别之间提供进一步的歧视。在本文中,我们通过同时了解人类活动所涉及的人类姿势和对象的观察,提出了一种新颖的人类活动识别方法。我们在统一的数学框架下将人类活动识别作为一个关节优化问题,该框架使用类似回归的损失函数来整合人类姿势和对象提示,并定义结构化的稀疏规范,以识别歧视性的身体关节和对象属性。为了评估我们的方法,我们在两个基准数据集和家庭援助环境中的物理机器人上进行了广泛的实验。实验结果表明,我们的方法的表现优于先前的方法,并以10^4 Hz的处理速度获得了人类活动识别的实时性能。

Real-time human activity recognition plays an essential role in real-world human-centered robotics applications, such as assisted living and human-robot collaboration. Although previous methods based on skeletal data to encode human poses showed promising results on real-time activity recognition, they lacked the capability to consider the context provided by objects within the scene and in use by the humans, which can provide a further discriminant between human activity categories. In this paper, we propose a novel approach to real-time human activity recognition, through simultaneously learning from observations of both human poses and objects involved in the human activity. We formulate human activity recognition as a joint optimization problem under a unified mathematical framework, which uses a regression-like loss function to integrate human pose and object cues and defines structured sparsity-inducing norms to identify discriminative body joints and object attributes. To evaluate our method, we perform extensive experiments on two benchmark datasets and a physical robot in a home assistance setting. Experimental results have shown that our method outperforms previous methods and obtains real-time performance for human activity recognition with a processing speed of 10^4 Hz.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源