Feb. 12, 2024, 5:46 a.m. | Ali Safa Vikrant Jaltare Samira Sebt Kameron Gano Johannes Leugering Georges Gielen Gert Cauwenberghs

cs.CV updates on arXiv.org arxiv.org

This paper studies the use of Metropolis-Hastings sampling for training Spiking Neural Network (SNN) hardware subject to strong unknown non-idealities, and compares the proposed approach to the common use of the backpropagation of error (backprop) algorithm and surrogate gradients, widely used to train SNNs in literature. Simulations are conducted within a chip-in-the-loop training context, where an SNN subject to unknown distortion must be trained to detect cancer from measurements, within a biomedical application context. Our results show that the proposed …

algorithm backpropagation chip cs.cv cs.ne error hardware literature loop metropolis network network training neural network paper sampling simulations snn spiking neural network studies train training via

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne