Parallel Spiking Neurons with High Efficiency and Ability to Learn Long-term Dependencies

April 25, 2023 Β· Declared Dead Β· πŸ› Neural Information Processing Systems

πŸ’€ CAUSE OF DEATH: 404 Not Found
Code link is broken/dead
Authors Wei Fang, Zhaofei Yu, Zhaokun Zhou, Ding Chen, Yanqi Chen, Zhengyu Ma, TimothΓ©e Masquelier, Yonghong Tian arXiv ID 2304.12760 Category cs.NE: Neural & Evolutionary Cross-listed cs.CV Citations 74 Venue Neural Information Processing Systems Repository https://github.com/fangwei123456/Parallel-Spiking-Neuron} Last Checked 1 month ago
Abstract
Vanilla spiking neurons in Spiking Neural Networks (SNNs) use charge-fire-reset neuronal dynamics, which can only be simulated serially and can hardly learn long-time dependencies. We find that when removing reset, the neuronal dynamics can be reformulated in a non-iterative form and parallelized. By rewriting neuronal dynamics without reset to a general formulation, we propose the Parallel Spiking Neuron (PSN), which generates hidden states that are independent of their predecessors, resulting in parallelizable neuronal dynamics and extremely high simulation speed. The weights of inputs in the PSN are fully connected, which maximizes the utilization of temporal information. To avoid the use of future inputs for step-by-step inference, the weights of the PSN can be masked, resulting in the masked PSN. By sharing weights across time-steps based on the masked PSN, the sliding PSN is proposed to handle sequences of varying lengths. We evaluate the PSN family on simulation speed and temporal/static data classification, and the results show the overwhelming advantage of the PSN family in efficiency and accuracy. To the best of our knowledge, this is the first study about parallelizing spiking neurons and can be a cornerstone for the spiking deep learning research. Our codes are available at \url{https://github.com/fangwei123456/Parallel-Spiking-Neuron}.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

πŸ“œ Similar Papers

In the same crypt β€” Neural & Evolutionary

R.I.P. πŸ‘» Ghosted

LSTM: A Search Space Odyssey

Klaus Greff, Rupesh Kumar Srivastava, ... (+3 more)

cs.NE πŸ› IEEE TNNLS πŸ“š 6.0K cites 11 years ago

Died the same way β€” πŸ’€ 404 Not Found