Head-related transfer functions (HRTF) are used for creating the perception of a virtual sound source at horizontal angle ø and vertical angle ?. Publicly available databases use a subset of a full-grid of angular directions due to time and complexity to acquire and deconvolve responses. In this paper we build up on our prior research [5] by extending the technique to HRTF synthesis, using the IRCAM dataset, while reducing the computational complexity of the autoencoder (AE)+fully-connected-neural-network (FCNN) architecture by ˜ 60% using Bayesian optimization. We also present listening test results, demonstrating the performance of the presented approach, from a pilot study that was designed for assessing the directional cues of the proposed architecture.
https://www.aes.org/e-lib/browse.cfm?elib=20295
Click to purchase paper as a non-member or login as an AES member. If your company or school subscribes to the E-Library then switch to the institutional version. If you are not an AES member and would like to subscribe to the E-Library then Join the AES!
This paper costs $33 for non-members and is free for AES members and E-Library subscribers.
Learn more about the AES E-Library
Start a discussion about this paper!