Spiking Neural Networks promise brain-inspired and energy-efficient computation by transmitting information through binary (0/1) spikes. Yet, their performance still lags behind that of artificial neural networks, often assumed to result from information loss caused by sparse and binary activations. In this work, we challenge this long-standing assumption and reveal a previously overlooked frequency bias: spiking neurons inherently suppress high-frequency components and preferentially propagate low-frequency information. This frequency-domain imbalance, we argue, is the root cause of degraded feature representation in SNNs. Empirically, on Spiking Transformers, adopting Avg-Pooling (low-pass) for token mixing lowers performance to 76.73% on Cifar-100, whereas replacing it with Max-Pool (high-pass) pushes the top-1 accuracy to 79.12%. Accordingly, we introduce Max-Former that restores high-frequency signals through two frequency-enhancing operators: (1) extra Max-Pool in patch embedding, and (2) Depth-Wise Convolution in place of self-attention. Notably, Max-Former attains 82.39% top-1 accuracy on ImageNet using only 63.99M parameters, surpassing Spikformer (74.81%, 66.34M) by +7.58%. Extending our insight beyond transformers, our Max-ResNet-18 achieves state-of-the-art performance on convolution-based benchmarks: 97.17% on CIFAR-10 and 83.06\% on CIFAR-100. We hope this simple yet effective solution inspires future research to explore the distinctive nature of spiking neural networks. Code is available: https://github.com/bic-L/MaxFormer.
翻译:脉冲神经网络通过二进制(0/1)脉冲传递信息,有望实现类脑且高效节能的计算。然而,其性能仍落后于人工神经网络,这一差距通常被归因于稀疏二进制激活导致的信息损失。在本研究中,我们挑战了这一长期存在的假设,并揭示了一个先前被忽视的频率偏好:脉冲神经元固有地抑制高频分量,并优先传播低频信息。我们认为,这种频域不平衡是导致SNN特征表示能力下降的根本原因。实验表明,在脉冲Transformer中,采用平均池化(低通)进行令牌混合会使Cifar-100上的性能降至76.73%,而将其替换为最大池化(高通)可将top-1准确率提升至79.12%。为此,我们提出了Max-Former,它通过两个频率增强算子来恢复高频信号:(1)在补丁嵌入层中引入额外的最大池化,(2)用深度卷积替代自注意力机制。值得注意的是,Max-Former仅使用63.99M参数便在ImageNet上达到了82.39%的top-1准确率,较Spikformer(74.81%,66.34M)提升了+7.58%。将我们的洞见拓展至Transformer架构之外,所提出的Max-ResNet-18在基于卷积的基准测试中取得了最先进的性能:CIFAR-10上为97.17%,CIFAR-100上为83.06%。我们希望这一简单而有效的解决方案能启发未来研究深入探索脉冲神经网络的独特性质。代码已开源:https://github.com/bic-L/MaxFormer。