Analog Spiking Neural Network Synthesis for the MNIST


  • Thomas Soupizet GeePs, CentraleSupelec
  • Zalfa Jouni GeePs,CentraleSupélec
  • Siqi Wang Geeps, Sorbonne Université
  • Aziz Benlarbi-Delai Geeps, Sorbonne Université
  • Pietro M. Ferreira GeePs, CentraleSupelec



neuromorphic computing, spiking neural network, deep learning, MNIST


Different from classical artificial neural network which processes digital data, the spiking neural network (SNN) processes spike trains. Indeed, its event-driven property helps to capture the rich dynamics the neurons have within the brain, and the sparsity of collected spikes helps reducing computational power. Novel synthesis framework is proposed and an algorithm is detailed to guide designers into deep learning and energy-efficient analog SNN using MNIST. An analog SNN composed of 86 electronic neurons (eNeuron) and 1238 synapses interacting through two hidden layers is illustrated. Three different models of eNeurons implementations are tested, being (Leaky) Integrate-and-Fire (LIF), Morris Lecar (ML) simplified (simp.) and biomimetic (bio.). The proposed SNN, coupling deep learning and ultra-low power, is trained using a common machine learning system (Tensor- Flow) for the MNIST. LIF eNeurons implementations present some limitations and weakness in terms of dynamic range. Both ML eNeurons achieve robust accuracy which is approximately of 0.82.






Selected Papers from SBCCI 2022