GCD-DDPM: A Generative Change Detection Model Based on Difference-Feature-Guided DDPM | IEEE Journals & Magazine | IEEE Xplore

GCD-DDPM: A Generative Change Detection Model Based on Difference-Feature-Guided DDPM


Abstract:

Deep learning (DL)-based methods have recently shown great promise in bitemporal change detection (CD). Existing discriminative methods based on convolutional neural netw...Show More

Abstract:

Deep learning (DL)-based methods have recently shown great promise in bitemporal change detection (CD). Existing discriminative methods based on convolutional neural networks (CNNs) and Transformers rely on discriminative representation learning for change recognition while struggling with exploring local and long-range contextual dependencies. As a result, it is still challenging to obtain fine-grained and robust CD maps in diverse ground scenes. To cope with this challenge, this work proposes a generative CD model called GCD-DDPM to directly generate CD maps by exploiting the denoising diffusion probabilistic model (DDPM), instead of classifying each pixel into changed or unchanged categories. Furthermore, the difference conditional encoder (DCE), is designed to guide the generation of CD maps by exploiting multilevel difference features. Leveraging the variational inference (VI) procedure, GCD-DDPM can adaptively recalibrate the CD results through an iterative inference process, while accurately distinguishing subtle and irregular changes in diverse scenes. Finally, a noise suppression-based semantic enhancer (NSSE) is specifically designed to mitigate noise in the current step’s change-aware feature representations from the CD Encoder. This refinement, serving as an attention map, can guide subsequent iterations while enhancing CD accuracy. Extensive experiments on four high-resolution CD datasets (CDD) confirm the superior performance of the proposed GCD-DDPM. The code for this work will be available at https://github.com/udrs/GCD.
Article Sequence Number: 5404416
Date of Publication: 27 March 2024

ISSN Information:

Funding Agency:


Contact IEEE to Subscribe

References

References is not available for this document.