S4NN: temporal backpropagation for spiking neural networks with one spike per neuron

October 21, 2019 Β· Entered Twilight Β· πŸ› International Journal of Neural Systems

πŸŒ… TWILIGHT: Old Age
Predates the code-sharing era β€” a pioneer of its time

"Last commit was 5.0 years ago (β‰₯5 year threshold)"

Evidence collected by the PWNC Scanner

Repo contents: LICENSE, MNIST.zip, README.md, S4NN.ipynb, S4NN.py, weights_pretrained.npy

Authors Saeed Reza Kheradpisheh, Timothée Masquelier arXiv ID 1910.09495 Category cs.NE: Neural & Evolutionary Cross-listed cs.CV, cs.LG, q-bio.NC Citations 205 Venue International Journal of Neural Systems Repository https://github.com/SRKH/S4NN ⭐ 65 Last Checked 1 month ago
Abstract
We propose a new supervised learning rule for multilayer spiking neural networks (SNNs) that use a form of temporal coding known as rank-order-coding. With this coding scheme, all neurons fire exactly one spike per stimulus, but the firing order carries information. In particular, in the readout layer, the first neuron to fire determines the class of the stimulus. We derive a new learning rule for this sort of network, named S4NN, akin to traditional error backpropagation, yet based on latencies. We show how approximated error gradients can be computed backward in a feedforward network with any number of layers. This approach reaches state-of-the-art performance with supervised multi fully-connected layer SNNs: test accuracy of 97.4% for the MNIST dataset, and 99.2% for the Caltech Face/Motorbike dataset. Yet, the neuron model that we use, non-leaky integrate-and-fire, is much simpler than the one used in all previous works. The source codes of the proposed S4NN are publicly available at https://github.com/SRKH/S4NN.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

πŸ“œ Similar Papers

In the same crypt β€” Neural & Evolutionary

R.I.P. πŸ‘» Ghosted

LSTM: A Search Space Odyssey

Klaus Greff, Rupesh Kumar Srivastava, ... (+3 more)

cs.NE πŸ› IEEE TNNLS πŸ“š 6.0K cites 11 years ago