论文标题

变形金刚作为隐式神经表示的元学习者

Transformers as Meta-Learners for Implicit Neural Representations

论文作者

Chen, Yinbo, Wang, Xiaolong

论文摘要

近年来,隐含的神经表示(INR)出现并显示了其比离散表示的好处。但是,将INR拟合到给定的观察结果通常需要从头开始优化梯度下降,这是效率低下的,并且在稀疏观测值中并不能很好地概括。为了解决这个问题,大多数先前的作品都会训练一个生成单个向量来调节INR权重的超网络,其中单个向量成为一个信息瓶颈,该信息瓶颈限制了输出INR的重建精度。最近的工作表明,可以通过基于梯度的元学习来精确地推断出INR中的整个权重。由基于梯度的元学习的广义公式的动机,我们提出了一种使用变压器作为INRS的超网络的公式,它可以直接用专门为设置为集合的变压器构建整个INR权重。我们演示了我们在不同任务和域中构建INR的方法的有效性,包括2D图像回归和查看3D对象的合成。我们的工作吸引了变压器超网与基于梯度的元学习算法之间的连接,我们提供了进一步的分析,以了解生成的INRS。

Implicit Neural Representations (INRs) have emerged and shown their benefits over discrete representations in recent years. However, fitting an INR to the given observations usually requires optimization with gradient descent from scratch, which is inefficient and does not generalize well with sparse observations. To address this problem, most of the prior works train a hypernetwork that generates a single vector to modulate the INR weights, where the single vector becomes an information bottleneck that limits the reconstruction precision of the output INR. Recent work shows that the whole set of weights in INR can be precisely inferred without the single-vector bottleneck by gradient-based meta-learning. Motivated by a generalized formulation of gradient-based meta-learning, we propose a formulation that uses Transformers as hypernetworks for INRs, where it can directly build the whole set of INR weights with Transformers specialized as set-to-set mapping. We demonstrate the effectiveness of our method for building INRs in different tasks and domains, including 2D image regression and view synthesis for 3D objects. Our work draws connections between the Transformer hypernetworks and gradient-based meta-learning algorithms and we provide further analysis for understanding the generated INRs.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源