论文标题
通过模型增强加速数据集蒸馏
Accelerating Dataset Distillation via Model Augmentation
论文作者
论文摘要
新兴领域的数据集蒸馏(DD)旨在从大型培训数据集中生成更小但有效的合成训练数据集。基于梯度匹配的现有DD方法实现领先的性能;但是,它们在计算上是极为密集的,因为它们需要在数千个随机初始化的模型中连续优化数据集。在本文中,我们假设训练与不同模型的合成数据会导致更好的泛化性能。因此,我们提出了两种模型增强技术,即使用早期阶段模型和参数扰动来学习一个信息丰富的合成集,并大大降低了培训成本。广泛的实验表明,我们的方法在与最先进的方法中达到了20倍的速度和可比的性能。
Dataset Distillation (DD), a newly emerging field, aims at generating much smaller but efficient synthetic training datasets from large ones. Existing DD methods based on gradient matching achieve leading performance; however, they are extremely computationally intensive as they require continuously optimizing a dataset among thousands of randomly initialized models. In this paper, we assume that training the synthetic data with diverse models leads to better generalization performance. Thus we propose two model augmentation techniques, i.e. using early-stage models and parameter perturbation to learn an informative synthetic set with significantly reduced training cost. Extensive experiments demonstrate that our method achieves up to 20x speedup and comparable performance on par with state-of-the-art methods.