Publications

CONFERENCE (INTERNATIONAL) Variational Bayesian Multi-channel Speech Dereverberation under Noisy Environments with Probabilistic Convolutive Transfer Function

Masahito Togami, Tatsuya Komatsu

The 20th Annual Conference of the International Speech Communication Association (INTERSPEECH 2019)

September 15, 2019

In this paper, we propose a multi-channel speech dereverberation method which can reduce reverberation even when acoustic transfer functions (ATFs) are time varying under noisy environments. The microphone input signal is modeled as a convolutive mixture in a time-frequency domain so as to incorporate late reverberation whose tap length is longer than frame size of short term Fourier transform. To reduce reverberation effectively under the time-varying ATF conditions, the proposed method extends the deterministic convolutive transfer function (D-CTF) into a probabilistic convolutive transfer function (P-CTF). A variational Bayesian framework was applied to approximation of a joint posterior probability density functions of a speech source signal and the ATFs. Variational posterior probability density functions and the other parameters are iteratively updated so as to maximize an evidence lower bound (ELBO). Experimental results when the ATFs are time-varying and there is background noise showed that the proposed method can reduce reverberation more accurately than the Weighted Prediction error (WPE) and the Kalman-EM for dereverberation (KEMD).

Paper : Variational Bayesian Multi-channel Speech Dereverberation under Noisy Environments with Probabilistic Convolutive Transfer Functionopen into new tab or window (external link)