Clean Audio for TV broadcast: An Object-Based Approach for Hearing-Impaired Viewers - April 2015
Audibility of a CD-Standard A/DA/A Loop Inserted into High-Resolution Audio Playback - September 2007
Sound Board: Food for Thought, Aesthetics in Orchestra Recording - April 2015
Extraction of Speech Transmission Index from Speech Signals Using Artificial Neural Networks
This paper presents a novel method to extract Speech Transmission Index (STI) from reverberated speech utterances using an artificial neural network. The convolutions of anechoic speech signals and simulated impulse responses of rooms of various kinds are used to train the artificial neural network. A time to frequency domain transformation algorithm is proposed as the pre-processor. A multi-layered feed forward neural network trained by back-propagation is adopted. Once trained, the neural network can accurately estimate Speech Transmission Index from speech signals received by a microphone in rooms. This approach utilises a naturalistic sound source, speech, and hence has potential to facilitate occupied measurement.
Click to purchase paper or login as an AES member. If your company or school subscribes to the E-Library then switch to the institutional version. If you are not an AES member and would like to subscribe to the E-Library then Join the AES!
This paper costs $20 for non-members, $5 for AES members and is free for E-Library subscribers.