论文标题
人们在指导机器人的全景视频中跟踪
People Tracking in Panoramic Video for Guiding Robots
论文作者
论文摘要
指导机器人旨在有效地将人们带到他们可能未知的环境中的特定位置。在此操作期间,机器人应该能够检测和跟踪伴随的人,而永远不要忽视她/他。最小化此事件的一种解决方案是使用全向相机:其360°视场(FOV)保证,如果不遮挡或离传感器很远,则任何框架对象都不会离开FOV。但是,获得的全景视频在人们的探测和跟踪等感知任务中介绍了新的挑战,包括要处理的图像的大尺寸,圆柱形投影引起的失真效应以及全景图像的周期性。在本文中,我们提出了一组目标方法,允许有效地适应全景视频,标准人员检测和跟踪管道最初是为透视摄像机设计的。我们的方法已经在基于深度学习的人的检测和跟踪框架内通过商业360°相机实施和测试。在专门用于指导机器人应用程序和真实服务机器人上进行的数据集执行的实验显示了拟议方法比其他最新系统的有效性。我们在本文中发布了获得和注释的数据集以及我们方法的开源实现。
A guiding robot aims to effectively bring people to and from specific places within environments that are possibly unknown to them. During this operation the robot should be able to detect and track the accompanied person, trying never to lose sight of her/him. A solution to minimize this event is to use an omnidirectional camera: its 360° Field of View (FoV) guarantees that any framed object cannot leave the FoV if not occluded or very far from the sensor. However, the acquired panoramic videos introduce new challenges in perception tasks such as people detection and tracking, including the large size of the images to be processed, the distortion effects introduced by the cylindrical projection and the periodic nature of panoramic images. In this paper, we propose a set of targeted methods that allow to effectively adapt to panoramic videos a standard people detection and tracking pipeline originally designed for perspective cameras. Our methods have been implemented and tested inside a deep learning-based people detection and tracking framework with a commercial 360° camera. Experiments performed on datasets specifically acquired for guiding robot applications and on a real service robot show the effectiveness of the proposed approach over other state-of-the-art systems. We release with this paper the acquired and annotated datasets and the open-source implementation of our method.