Computational Auditory Scene Analysis (CASA) is typically achieved by using statistical models that have been trained offline on available data. Their performance relies heavily on the assumption that the process generating the data along with the recording conditions are stationary over time. Nowadays, the focus of CASA is moving from structured, well-defined scenarios to unrestricted scenes with realistic characteristics where the stationarity assumption might not be true. Therefore, there is a high demand for methodologies and tools dealing with a series of problems tightly coupled with such nonstationary conditions, such as changes in the recording conditions, reverberant effects, etc. This paper formulates these obstacles under the concept drift framework and explores two fundamental adaptation approaches: active and passive. The overall aim is to learn online the statistical properties of the evolving data distribution and incorporate them into the recognition mechanism for boosting its performance. The proposed CASA system encompasses a concept drift detector and an online adaptation module. The proposed framework was evaluated in the auditory analysis of three environments (office, meeting room, and lecture hall) with diverse characteristics (dimensions, reverberation times, etc.) The results are encouraging in terms of classification rate, false positive rate, false negative rate, and detection delay.
Click to purchase paper as a non-member or login as an AES member. If your company or school subscribes to the E-Library then switch to the institutional version. If you are not an AES member and would like to subscribe to the E-Library then Join the AES!
This paper costs $33 for non-members and is free for AES members and E-Library subscribers.