Processing math: 100%
Real-Time Audio-Guided Multi-Face Reenactment | IEEE Journals & Magazine | IEEE Xplore

Real-Time Audio-Guided Multi-Face Reenactment


Abstract:

Audio-guided face reenactment aims to generate authentic target faces that have matched facial expression of the input audio, and many learning-based methods have success...Show More

Abstract:

Audio-guided face reenactment aims to generate authentic target faces that have matched facial expression of the input audio, and many learning-based methods have successfully achieved this. However, most methods can only reenact a particular person once trained or suffer from the low-quality generation of the target images. Also, nearly none of the current reenactment works consider the model size and running speed that are important for practical use. To solve the above challenges, we propose an efficient Audio-guided Multi-face reenactment model named AMNet, which can reenact target faces among multiple persons with corresponding source faces and drive signals as inputs. Concretely, we design a Geometric Controller (GC) module to inject the drive signals so that the model can be optimized in an end-to-end manner and generate more authentic images. Also, we adopt a lightweight network for our face reenactor so that the model can run in real-time on both CPU and GPU devices. Abundant experiments prove our approach’s superiority over existing methods, e.g., averagely decreasing FID by 0.12\downarrow and increasing SSIM by 0.031\uparrow than APB2Face, while owning fewer parameters (\times 4 \downarrow) and faster CPU speed (\times 4 \uparrow).
Published in: IEEE Signal Processing Letters ( Volume: 29)
Page(s): 1 - 5
Date of Publication: 29 September 2021

ISSN Information:

Funding Agency:


Contact IEEE to Subscribe

References

References is not available for this document.