Loading [MathJax]/extensions/MathZoom.js
Spiking-Diffusion: Vector Quantized Discrete Diffusion Model with Spiking Neural Networks | IEEE Conference Publication | IEEE Xplore

Spiking-Diffusion: Vector Quantized Discrete Diffusion Model with Spiking Neural Networks


Abstract:

Spiking neural networks (SNNs) have tremendous potential for energy-efficient neuromorphic chips due to their binary and event-driven architecture. SNNs have been primari...Show More

Abstract:

Spiking neural networks (SNNs) have tremendous potential for energy-efficient neuromorphic chips due to their binary and event-driven architecture. SNNs have been primarily used in classification tasks, but limited exploration on image generation tasks. To fill the gap, we propose a Spiking-Diffusion model, which is based on the vector quantized discrete diffusion model. First, we develop a vector quantized variational autoencoder with SNNs (VQ-SVAE) to learn a discrete latent space for images. In VQ-SVAE, image features are encoded using both the spike firing rate and postsynaptic potential, and an adaptive spike generator is designed to restore embedding features in the form of spike trains. Next, we perform absorbing state diffusion in the discrete latent space and construct a spiking diffusion image decoder (SDID) with SNNs to denoise the image. Our work is the first to build the diffusion model entirely from SNN layers. Experimental results on MNIST, FMNIST, KMNIST, Letters, and Cifar10 demonstrate that Spiking-Diffusion outperforms the existing SNN-based generation model. We achieve FIDs of 37.50, 91.98, 59.23, 67.41, and 120.5 on the above datasets respectively, with reductions of 58.60%, 18.75%, 64.51%, 29.75%, and 44.88% in FIDs compared with the state-of-art work. Our code will be available at https://github.com/Arktis2022/Spiking-Diffusion.
Date of Conference: 01-03 November 2024
Date Added to IEEE Xplore: 26 March 2025
ISBN Information:
Conference Location: Jingdezhen, China

Funding Agency:


I. Introduction

Spiking neural networks (SNNs) are the third generation neural networks with biological plausibility that encode and transmit information in form of spikes by mimicking the dynamics of neurons in the brain. Compared with Artificial neural networks (ANNs), the event-driven nature of SNNs allows for significant reduction in energy consumption when running on neuromorphic chips [1].

Contact IEEE to Subscribe

References

References is not available for this document.