Efficient Binaural Rendering of Spatially Extended Sound Sources
×
Cite This
Citation & Abstract
C. Anemüller, A. Adami, and J. Herre, "Efficient Binaural Rendering of Spatially Extended Sound Sources," J. Audio Eng. Soc., vol. 71, no. 5, pp. 281-292, (2023 May.). doi: https://doi.org/10.17743/jaes.2022.0069
C. Anemüller, A. Adami, and J. Herre, "Efficient Binaural Rendering of Spatially Extended Sound Sources," J. Audio Eng. Soc., vol. 71 Issue 5 pp. 281-292, (2023 May.). doi: https://doi.org/10.17743/jaes.2022.0069
Abstract: In virtual/augmented reality or 3D applications with binaural audio, it is often desired to render sound sources with a certain spatial extent in a realistic way. A common approach is to distribute multiple correlated or decorrelated point sources over the desired spatial extent range, possibly derived from the original source signal by applying suitable decorrelation filters. Based on this basic model, a novel method for efficient and realistic binaural rendering of spatially extended sound sources is proposed. Instead of rendering each point source individually, the target auditory cues are synthesized directly from just two decorrelated input signals. This procedure comes with the advantage of low computational complexity and relaxed requirements for decorrelation filters. An objective evaluation shows that the proposed method matches the basic rendering model well in terms of perceptually relevant objective metrics. A subjective listening test shows, furthermore, that the output of the proposed method is perceptually almost identical to the output of the basic rendering model. The technique is part of the Reference Model architecture of the upcoming MPEG-I Immersive Audio standard.
@article{anemüller2023efficient,
author={anemüller, carlotta and adami, alexander and herre, jürgen},
journal={journal of the audio engineering society},
title={efficient binaural rendering of spatially extended sound sources},
year={2023},
volume={71},
number={5},
pages={281-292},
doi={https://doi.org/10.17743/jaes.2022.0069},
month={may},}
@article{anemüller2023efficient,
author={anemüller, carlotta and adami, alexander and herre, jürgen},
journal={journal of the audio engineering society},
title={efficient binaural rendering of spatially extended sound sources},
year={2023},
volume={71},
number={5},
pages={281-292},
doi={https://doi.org/10.17743/jaes.2022.0069},
month={may},
abstract={in virtual/augmented reality or 3d applications with binaural audio, it is often desired to render sound sources with a certain spatial extent in a realistic way. a common approach is to distribute multiple correlated or decorrelated point sources over the desired spatial extent range, possibly derived from the original source signal by applying suitable decorrelation filters. based on this basic model, a novel method for efficient and realistic binaural rendering of spatially extended sound sources is proposed. instead of rendering each point source individually, the target auditory cues are synthesized directly from just two decorrelated input signals. this procedure comes with the advantage of low computational complexity and relaxed requirements for decorrelation filters. an objective evaluation shows that the proposed method matches the basic rendering model well in terms of perceptually relevant objective metrics. a subjective listening test shows, furthermore, that the output of the proposed method is perceptually almost identical to the output of the basic rendering model. the technique is part of the reference model architecture of the upcoming mpeg-i immersive audio standard.},}
TY - paper
TI - Efficient Binaural Rendering of Spatially Extended Sound Sources
SP - 281
EP - 292
AU - Anemüller, Carlotta
AU - Adami, Alexander
AU - Herre, Jürgen
PY - 2023
JO - Journal of the Audio Engineering Society
IS - 5
VO - 71
VL - 71
Y1 - May 2023
TY - paper
TI - Efficient Binaural Rendering of Spatially Extended Sound Sources
SP - 281
EP - 292
AU - Anemüller, Carlotta
AU - Adami, Alexander
AU - Herre, Jürgen
PY - 2023
JO - Journal of the Audio Engineering Society
IS - 5
VO - 71
VL - 71
Y1 - May 2023
AB - In virtual/augmented reality or 3D applications with binaural audio, it is often desired to render sound sources with a certain spatial extent in a realistic way. A common approach is to distribute multiple correlated or decorrelated point sources over the desired spatial extent range, possibly derived from the original source signal by applying suitable decorrelation filters. Based on this basic model, a novel method for efficient and realistic binaural rendering of spatially extended sound sources is proposed. Instead of rendering each point source individually, the target auditory cues are synthesized directly from just two decorrelated input signals. This procedure comes with the advantage of low computational complexity and relaxed requirements for decorrelation filters. An objective evaluation shows that the proposed method matches the basic rendering model well in terms of perceptually relevant objective metrics. A subjective listening test shows, furthermore, that the output of the proposed method is perceptually almost identical to the output of the basic rendering model. The technique is part of the Reference Model architecture of the upcoming MPEG-I Immersive Audio standard.
In virtual/augmented reality or 3D applications with binaural audio, it is often desired to render sound sources with a certain spatial extent in a realistic way. A common approach is to distribute multiple correlated or decorrelated point sources over the desired spatial extent range, possibly derived from the original source signal by applying suitable decorrelation filters. Based on this basic model, a novel method for efficient and realistic binaural rendering of spatially extended sound sources is proposed. Instead of rendering each point source individually, the target auditory cues are synthesized directly from just two decorrelated input signals. This procedure comes with the advantage of low computational complexity and relaxed requirements for decorrelation filters. An objective evaluation shows that the proposed method matches the basic rendering model well in terms of perceptually relevant objective metrics. A subjective listening test shows, furthermore, that the output of the proposed method is perceptually almost identical to the output of the basic rendering model. The technique is part of the Reference Model architecture of the upcoming MPEG-I Immersive Audio standard.
Authors:
Anemüller, Carlotta; Adami, Alexander; Herre, Jürgen
Affiliation:
International Audio Laboratories Erlangen, Germany. A joint institution of the Friedrich-Alexander-Universität Erlangen-Nürnberg (FAU) and Fraunhofer IIS JAES Volume 71 Issue 5 pp. 281-292; May 2023
Publication Date:
May 9, 2023Import into BibTeX
Permalink:
http://www.aes.org/e-lib/browse.cfm?elib=22131