论文标题

Spikformer:尖峰神经网络会遇到变压器

Spikformer: When Spiking Neural Network Meets Transformer

论文作者

Zhou, Zhaokun, Zhu, Yuesheng, He, Chao, Wang, Yaowei, Yan, Shuicheng, Tian, Yonghong, Yuan, Li

论文摘要

我们考虑了两个生物学上合理的结构,即尖峰神经网络(SNN)和自我发项机制。前者为深度学习提供了节能和事件驱动的范式,而后者具有捕获功能依赖性的能力,从而使变压器能够实现良好的性能。直观地希望探索他们之间的婚姻。在本文中,我们考虑利用SNN的自我发作能力和生物学特性,并提出一种新颖的尖峰自我注意力(SSA)以及一个强大的框架,称为Spiking Transformer(SpikFormer)。 SpikFormer中的SSA机制通过使用Spike-form查询,键和值而无需软max来建模稀疏视觉特征。由于其计算很少并且避免了乘法,因此SSA是有效的,计算能量消耗较低。结果表明,带有SSA的SpikFormer在神经形态和静态数据集中都可以胜过图像分类中的最先进的SNNS样框架。 SPIKFORMER(6630万参数)具有与SEW-Resnet-152(60.2m,69.26%)相当的尺寸的SpikFormer可以使用4个时间步长实现Imagenet上的74.81%TOP1精度,这是直接训练的SNNS模型中最新的。

We consider two biologically plausible structures, the Spiking Neural Network (SNN) and the self-attention mechanism. The former offers an energy-efficient and event-driven paradigm for deep learning, while the latter has the ability to capture feature dependencies, enabling Transformer to achieve good performance. It is intuitively promising to explore the marriage between them. In this paper, we consider leveraging both self-attention capability and biological properties of SNNs, and propose a novel Spiking Self Attention (SSA) as well as a powerful framework, named Spiking Transformer (Spikformer). The SSA mechanism in Spikformer models the sparse visual feature by using spike-form Query, Key, and Value without softmax. Since its computation is sparse and avoids multiplication, SSA is efficient and has low computational energy consumption. It is shown that Spikformer with SSA can outperform the state-of-the-art SNNs-like frameworks in image classification on both neuromorphic and static datasets. Spikformer (66.3M parameters) with comparable size to SEW-ResNet-152 (60.2M,69.26%) can achieve 74.81% top1 accuracy on ImageNet using 4 time steps, which is the state-of-the-art in directly trained SNNs models.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源