We consider the problem of blind multi-channel speech dereverberation without the knowledge of room acoustics. The dereverberated speech component is estimated by subtracting the undesired component, estimated using multi-channel linear prediction (MCLP), from the reference microphone signal. In this paper we present a framework for MCLP-based speech dereverberation by exploiting sparsity in the time-frequency domain. The presented framework uses a wideband or a narrowband signal model and a sparse analysis or synthesis model for the desired speech component. The proposed problems involving a reweighted $\ell_1$-norm, are solved in a flexible optimization framework. The obtained results are comparable to the state of the art, motivating further extensions exploiting sparsity and speech structure.
https://www.aes.org/e-lib/browse.cfm?elib=18089
Click to purchase paper as a non-member or login as an AES member. If your company or school subscribes to the E-Library then switch to the institutional version. If you are not an AES member and would like to subscribe to the E-Library then Join the AES!
This paper costs $33 for non-members and is free for AES members and E-Library subscribers.
Learn more about the AES E-Library
Start a discussion about this paper!