论文标题
通过解决尖峰神经网络中降解的症结来推进深层学习
Advancing Deep Residual Learning by Solving the Crux of Degradation in Spiking Neural Networks
论文作者
论文摘要
尽管神经形态计算取得了迅速的进展,但深度不足以及峰值神经网络(SNN)的代表力不足,在实践中严重限制了其应用程序范围。残留学习和快捷方式已被证明是训练深神经网络的重要方法,但以前的工作很少评估其对基于尖峰的通信和时空动态特征的适用性。这种疏忽导致了阻碍的信息流和随附的退化问题。在本文中,我们确定了Crux,然后提出了一个新型的SNN残差块,该块能够显着扩展直接训练的SNN的深度,例如,在Imagenet上,在CIFAR-10上最多可达482层,并且在Imainet上的104层,而无需观察到任何轻微的降解问题。我们验证方法对基于帧的基于框架和神经形态数据集的有效性,而我们的SRM-Resnet104在Imagenet上获得了76.02%的精度,这是第一次在直接训练的SNN的域中。估计了巨大的能源效率,并且所得的网络平均每个神经元仅需要一个尖峰才能对输入样本进行分类。我们认为,我们强大而可扩展的建模将为进一步探索SNN提供大力支持。
Despite the rapid progress of neuromorphic computing, the inadequate depth and the resulting insufficient representation power of spiking neural networks (SNNs) severely restrict their application scope in practice. Residual learning and shortcuts have been evidenced as an important approach for training deep neural networks, but rarely did previous work assess their applicability to the characteristics of spike-based communication and spatiotemporal dynamics. This negligence leads to impeded information flow and the accompanying degradation problem. In this paper, we identify the crux and then propose a novel residual block for SNNs, which is able to significantly extend the depth of directly trained SNNs, e.g., up to 482 layers on CIFAR-10 and 104 layers on ImageNet, without observing any slight degradation problem. We validate the effectiveness of our methods on both frame-based and neuromorphic datasets, and our SRM-ResNet104 achieves a superior result of 76.02% accuracy on ImageNet, the first time in the domain of directly trained SNNs. The great energy efficiency is estimated and the resulting networks need on average only one spike per neuron for classifying an input sample. We believe our powerful and scalable modeling will provide a strong support for further exploration of SNNs.