论文标题

属性下降:在内容级别及以后模拟以对象为中心的数据集

Attribute Descent: Simulating Object-Centric Datasets on the Content Level and Beyond

论文作者

Yao, Yue, Zheng, Liang, Yang, Xiaodong, Napthade, Milind, Gedeon, Tom

论文摘要

本文旨在使用图形引擎来模拟大量具有免费注释且可能非常类似于现实世界数据的训练数据。在综合和真实之间,存在两个级别的域间隙,涉及内容水平和外观水平。尽管后者与外观样式有关,但前者问题源于不同的机制,即诸如摄像机视点,对象放置和照明条件等属性中的内容不匹配。与广泛研究的外观级别的间隙相反,内容级差异尚未得到广泛研究。为了解决内容级的未对准,我们提出了一种属性下降方法,该方法自动优化了发动机属性以启用合成数据以近似现实世界数据。我们在以对象为中心的任务上验证我们的方法,其中一个对象占用图像的大部分。在这些任务中,搜索空间相对较小,每个属性的优化产生了足够明显的监督信号。我们收集了新的合成资产车辆,并重新格式化并重新使用现有的合成资产对象X和Personx。关于图像分类和对象重新识别的广泛实验证实,可以在三种情况下有效地使用改编的合成数据:仅使用合成数据培训,培训数据增强和数值了解数据集内容。

This article aims to use graphic engines to simulate a large number of training data that have free annotations and possibly strongly resemble to real-world data. Between synthetic and real, a two-level domain gap exists, involving content level and appearance level. While the latter is concerned with appearance style, the former problem arises from a different mechanism, i.e, content mismatch in attributes such as camera viewpoint, object placement and lighting conditions. In contrast to the widely-studied appearance-level gap, the content-level discrepancy has not been broadly studied. To address the content-level misalignment, we propose an attribute descent approach that automatically optimizes engine attributes to enable synthetic data to approximate real-world data. We verify our method on object-centric tasks, wherein an object takes up a major portion of an image. In these tasks, the search space is relatively small, and the optimization of each attribute yields sufficiently obvious supervision signals. We collect a new synthetic asset VehicleX, and reformat and reuse existing the synthetic assets ObjectX and PersonX. Extensive experiments on image classification and object re-identification confirm that adapted synthetic data can be effectively used in three scenarios: training with synthetic data only, training data augmentation and numerically understanding dataset content.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源