End To End Learning For Convolutive Multi-Channel Wiener Filtering | IEEE Conference Publication | IEEE Xplore

End To End Learning For Convolutive Multi-Channel Wiener Filtering


Abstract:

In this paper, we propose a dereverberation and speech source separation method based on deep neural network (DNN). Unlike the cascade connection of dereverberation and s...Show More

Abstract:

In this paper, we propose a dereverberation and speech source separation method based on deep neural network (DNN). Unlike the cascade connection of dereverberation and speech source separation, the proposed method performs dereverberation and speech source separation jointly by a unified convolutive multi-channel Wiener filtering (CMWF). The proposed method adopts a time-varying CMWF to achieve more dereverberation and separation performance than a time-invariant CMWF. The time-varying CMWF requires time-frequency masks and time-frequency activities. These variables are inferred via a unified DNN. The DNN is trained to optimize the output signal of the time-varying CMWF with a loss function based on a negative log-posterior probability density function. We also reveal that the time-varying CMWF can be obtained efficiently based on the Sherman-Morrison-Woodbury equation. Experimental results show that the proposed time-varying CMWF can separate speech sources under reverberant environments better than the cascade-connection based method and the time-invariant CMWF.
Date of Conference: 06-11 June 2021
Date Added to IEEE Xplore: 13 May 2021
ISBN Information:

ISSN Information:

Conference Location: Toronto, ON, Canada
No metrics found for this document.

1. INTRODUCTION

Speech source separation and dereverberation [1]–[3] are fundamental techniques in automatic speech recognition(ASR) and teleconferencing systems. Dereverberation techniques based on statistical modeling have been actively studied, e.g., Weighted Prediction Error (WPE) [4]. Simultaneous optimization of speech source separation and dereverberation has been also actively studied based on statistical modeling [5]–[8]. These techniques rely on speech source models based on super-Gaussian distributions, e.g., Laplacian distribution [9],[10] and time-varying Gaussian distribution [11]. However, the expression capability of these speech source models is not enough for expressing a complicated speech source spectrum. Recently, a deep neural network (DNN) is utilized for expressing a complicated speech source spectrum [12]–[20]. The expression capability of the DNN based speech source model is higher than that of statistical models.

No metrics found for this document.
Contact IEEE to Subscribe

References

References is not available for this document.