Pretraining with Re-parametrized Self-Attention: Unlocking Generalizationin SNN-Based Neural Decoding Across Time, Brains, and Tasks

ICLR 2026 Conference SubmissionAnonymous Authors
Brain-Machine InterfaceNeural Spike DecodingSpiking Neural NetworkFoundation Model
Abstract:

The emergence of large-scale neural activity datasets provides new opportunities to enhance the generalization of neural decoding models. However, it remains a practical challenge to design neural decoders for fully implantable brain-machine interfaces (iBMIs) that achieve high accuracy, strong generalization, and low computational cost, which are essential for reliable, long-term deployment under strict power and hardware constraints. To address this, we propose the Re-parametrized self-Attention Spiking Neural Network (RAT SNN) with a cross-condition pretraining framework to integrate neural variability and adapt to stringent computational constraints. Specifically, our approach introduces multi-timescale dynamic spiking neurons to capture the complex temporal variability of neural activity. And we refine spike-driven attention within a lightweight, re-parameterized architecture that enables accumulate-only operations between spiking neurons without sacrificing decoding accuracy. Furthermore, we develop a stepwise training pipeline to systematically integrate neural variability across conditions, including neural temporal drift, subjects and tasks. Building on these advances, we construct a pretrained model capable of rapid generalization to unseen conditions with high performance. We demonstrate that RAT SNN consistently outperforms leading SNN baselines and matches the performance of state-of-the-art artificial neural network (ANN) models in terms of decoding accuracy with much lower computational cost under both seen and unseen conditions across various datasets. Collectively, Pretrained-RAT SNN represents a high-performance, highly generalizable, and energy-efficient prototype of an SNN foundation model for fully iBMI. Code is available at RAT SNN GitHub.

Disclaimer
This report is AI-GENERATED using Large Language Models and WisPaper (A scholar search engine). It analyzes academic papers' tasks and contributions against retrieved prior work. While this system identifies POTENTIAL overlaps and novel directions, ITS COVERAGE IS NOT EXHAUSTIVE AND JUDGMENTS ARE APPROXIMATE. These results are intended to assist human reviewers and SHOULD NOT be relied upon as a definitive verdict on novelty.
NOTE that some papers exist in multiple, slightly different versions (e.g., with different titles or URLs). The system may retrieve several versions of the same underlying work. The current automated pipeline does not reliably align or distinguish these cases, so human reviewers will need to disambiguate them manually.
If you have any questions, please contact: mingzhang23@m.fudan.edu.cn

Overview

Overall Novelty Assessment

The paper introduces a Re-parametrized self-Attention Spiking Neural Network (RAT SNN) with cross-condition pretraining for neural decoding in implantable brain-machine interfaces. It resides in the Spiking Neural Network Decoders leaf, which contains nine papers—a moderately populated category within the broader Decoding Algorithms and Computational Methods branch. This positions the work in an active but not overcrowded research direction, where spiking architectures are explored for their event-driven efficiency and biological plausibility in BMI applications.

The taxonomy reveals that spiking decoders sit alongside Classical and Statistical Decoding Methods (four papers using Kalman filters and Bayesian approaches) and Deep Learning and Artificial Neural Network Decoders (five papers employing transformers and recurrent networks). Neighboring branches address Hardware Implementation and System Integration, including FPGA-Based Real-Time Decoding Systems and Low-Power Decoding ASICs, which share the paper's concern for computational constraints. The scope notes clarify that spiking methods emphasize event-driven computation, distinguishing them from standard backpropagation-trained networks and classical statistical models.

Among twenty-two candidates examined across three contributions, none were flagged as clearly refuting the proposed work. The Re-parametrized self-Attention SNN examined ten candidates with zero refutable overlaps, as did the multi-timescale dynamic spiking neurons component. The cross-condition pretraining framework reviewed two candidates, also without refutation. This limited search scope—focused on top semantic matches rather than exhaustive coverage—suggests that within the examined literature, the specific combination of re-parameterized attention, multi-timescale dynamics, and cross-condition pretraining appears distinct, though the analysis does not rule out relevant prior work beyond these twenty-two papers.

Based on the top-twenty-two semantic matches, the work appears to occupy a recognizable niche within spiking neural network decoders, combining architectural innovations with a training pipeline tailored to neural variability. The absence of refutable candidates in this limited sample indicates that the specific technical choices may be novel, but the search scope leaves open the possibility of related approaches in the broader literature. The taxonomy context shows that spiking decoders remain an active area with ongoing exploration of efficiency-accuracy trade-offs.

Taxonomy

Core-task Taxonomy Papers
50
3
Claimed Contributions
22
Contribution Candidate Papers Compared
0
Refutable Paper

Research Landscape Overview

Core task: neural decoding from cortical spike trains for brain-machine interfaces. The field organizes around several major branches that reflect both methodological diversity and application scope. Decoding Algorithms and Computational Methods encompasses a spectrum from classical statistical approaches to modern deep learning architectures and biologically inspired spiking neural networks, with works like Deep Neural Decoder[14] and motorSRNN[18] exemplifying data-driven strategies. Neural Signal Sources and Recording Modalities addresses the variety of input signals—ranging from single-unit spikes to local field potentials and multimodal recordings—while Hardware Implementation and System Integration focuses on real-time, power-efficient deployment using FPGAs and custom processors, as seen in Hardware Decoding Benchmarking[13] and RISC-V SNN Decoder[39]. Meanwhile, Decoding Targets and Application Domains spans motor control, speech synthesis, and cognitive state estimation, and Learning, Adaptation, and Generalization tackles challenges like calibration-free operation and continual learning across sessions. Clinical Translation and System Evaluation emphasizes long-term reliability and user-centered performance metrics. Within the algorithmic landscape, a particularly active line of work explores spiking neural network decoders that leverage event-driven computation for efficiency and biological plausibility. Reparametrized SNN Decoding[0] sits squarely in this cluster, proposing novel training strategies to improve gradient flow and learning stability in SNNs applied to BMI tasks. This contrasts with earlier efforts like SNN Decoder BMI[30], which established foundational architectures, and complements recent hybrid approaches such as Hybrid Spiking Networks[42] that blend spiking and non-spiking components. Compared to purely deep learning methods like Neural Data Transformer[49], spiking decoders trade off representational flexibility for lower power consumption and closer alignment with neural dynamics. Open questions remain around how to best balance biological realism, computational efficiency, and decoding accuracy, especially as hardware platforms like FPGA Spiking Networks[33] enable increasingly sophisticated real-time implementations.

Claimed Contributions

Re-parametrized self-Attention Spiking Neural Network (RAT SNN)

The authors introduce RAT SNN, a lightweight spiking neural network architecture that integrates re-parameterized spike-driven self-attention with multi-timescale dynamics. The architecture maintains accumulate-only operations between spiking neurons while achieving high decoding accuracy for brain-machine interfaces.

10 retrieved papers
Cross-condition pretraining framework with subject-specific batch normalization

The authors develop a stepwise training pipeline that systematically integrates neural variability across conditions including temporal drift, subjects, and tasks. This framework uses subject-specific batch normalization to enable rapid generalization to unseen conditions.

2 retrieved papers
Multi-timescale dynamic spiking neurons with recurrent connections

The authors propose recurrently connected leaky integrate-and-fire neurons with dynamic synapses to capture multi-timescale temporal dynamics in neural activity, mimicking biological neural systems with both long-range projections and local microcircuits.

10 retrieved papers

Core Task Comparisons

Comparisons with papers in the same taxonomy category

Contribution Analysis

Detailed comparisons for each claimed contribution

Contribution

Re-parametrized self-Attention Spiking Neural Network (RAT SNN)

The authors introduce RAT SNN, a lightweight spiking neural network architecture that integrates re-parameterized spike-driven self-attention with multi-timescale dynamics. The architecture maintains accumulate-only operations between spiking neurons while achieving high decoding accuracy for brain-machine interfaces.

Contribution

Cross-condition pretraining framework with subject-specific batch normalization

The authors develop a stepwise training pipeline that systematically integrates neural variability across conditions including temporal drift, subjects, and tasks. This framework uses subject-specific batch normalization to enable rapid generalization to unseen conditions.

Contribution

Multi-timescale dynamic spiking neurons with recurrent connections

The authors propose recurrently connected leaky integrate-and-fire neurons with dynamic synapses to capture multi-timescale temporal dynamics in neural activity, mimicking biological neural systems with both long-range projections and local microcircuits.