Object-Based Six-Degrees-of-Freedom Rendering of Sound Scenes Captured with Multiple Ambisonic Receivers
×
Cite This
Citation & Abstract
L. McCormack, A. Politis, T. McKenzie, C. Hold, and V. Pulkki, "Object-Based Six-Degrees-of-Freedom Rendering of Sound Scenes Captured with Multiple Ambisonic Receivers," J. Audio Eng. Soc., vol. 70, no. 5, pp. 355-372, (2022 May.). doi: https://doi.org/10.17743/jaes.2022.0010
L. McCormack, A. Politis, T. McKenzie, C. Hold, and V. Pulkki, "Object-Based Six-Degrees-of-Freedom Rendering of Sound Scenes Captured with Multiple Ambisonic Receivers," J. Audio Eng. Soc., vol. 70 Issue 5 pp. 355-372, (2022 May.). doi: https://doi.org/10.17743/jaes.2022.0010
Abstract: This article proposes a system for object-based six-degrees-of-freedom (6DoF) rendering of spatial sound scenes that are captured using a distributed arrangement of multiple Ambisonic receivers. The approach is based on first identifying and tracking the positions of sound sources within the scene, followed by the isolation of their signals through the use of beamformers. These sound objects are subsequently spatialized over the target playback setup, with respect to both the head orientation and position of the listener. The diffuse ambience of the scene is rendered separately by first spatially subtracting the source signals from the receivers located nearest to the listener position. The resultant residual Ambisonic signals are then spatialized, decorrelated, and summed together with suitable interpolation weights. The proposed system is evaluated through an in situ listening test conducted in 6DoF virtual reality,whereby real-world sound sources are compared with the auralization achieved through the proposed rendering method. The results of 15 participants suggest that in comparison to a linear interpolation-based alternative, the proposed object-based approach is perceived as being more realistic.
@article{mccormack2022object-based,
author={mccormack, leo and politis, archontis and mckenzie, thomas and hold, christoph and pulkki, ville},
journal={journal of the audio engineering society},
title={object-based six-degrees-of-freedom rendering of sound scenes captured with multiple ambisonic receivers},
year={2022},
volume={70},
number={5},
pages={355-372},
doi={https://doi.org/10.17743/jaes.2022.0010},
month={may},}
@article{mccormack2022object-based,
author={mccormack, leo and politis, archontis and mckenzie, thomas and hold, christoph and pulkki, ville},
journal={journal of the audio engineering society},
title={object-based six-degrees-of-freedom rendering of sound scenes captured with multiple ambisonic receivers},
year={2022},
volume={70},
number={5},
pages={355-372},
doi={https://doi.org/10.17743/jaes.2022.0010},
month={may},
abstract={this article proposes a system for object-based six-degrees-of-freedom (6dof) rendering of spatial sound scenes that are captured using a distributed arrangement of multiple ambisonic receivers. the approach is based on first identifying and tracking the positions of sound sources within the scene, followed by the isolation of their signals through the use of beamformers. these sound objects are subsequently spatialized over the target playback setup, with respect to both the head orientation and position of the listener. the diffuse ambience of the scene is rendered separately by first spatially subtracting the source signals from the receivers located nearest to the listener position. the resultant residual ambisonic signals are then spatialized, decorrelated, and summed together with suitable interpolation weights. the proposed system is evaluated through an in situ listening test conducted in 6dof virtual reality,whereby real-world sound sources are compared with the auralization achieved through the proposed rendering method. the results of 15 participants suggest that in comparison to a linear interpolation-based alternative, the proposed object-based approach is perceived as being more realistic.},}
TY - paper
TI - Object-Based Six-Degrees-of-Freedom Rendering of Sound Scenes Captured with Multiple Ambisonic Receivers
SP - 355
EP - 372
AU - McCormack, Leo
AU - Politis, Archontis
AU - McKenzie, Thomas
AU - Hold, Christoph
AU - Pulkki, Ville
PY - 2022
JO - Journal of the Audio Engineering Society
IS - 5
VO - 70
VL - 70
Y1 - May 2022
TY - paper
TI - Object-Based Six-Degrees-of-Freedom Rendering of Sound Scenes Captured with Multiple Ambisonic Receivers
SP - 355
EP - 372
AU - McCormack, Leo
AU - Politis, Archontis
AU - McKenzie, Thomas
AU - Hold, Christoph
AU - Pulkki, Ville
PY - 2022
JO - Journal of the Audio Engineering Society
IS - 5
VO - 70
VL - 70
Y1 - May 2022
AB - This article proposes a system for object-based six-degrees-of-freedom (6DoF) rendering of spatial sound scenes that are captured using a distributed arrangement of multiple Ambisonic receivers. The approach is based on first identifying and tracking the positions of sound sources within the scene, followed by the isolation of their signals through the use of beamformers. These sound objects are subsequently spatialized over the target playback setup, with respect to both the head orientation and position of the listener. The diffuse ambience of the scene is rendered separately by first spatially subtracting the source signals from the receivers located nearest to the listener position. The resultant residual Ambisonic signals are then spatialized, decorrelated, and summed together with suitable interpolation weights. The proposed system is evaluated through an in situ listening test conducted in 6DoF virtual reality,whereby real-world sound sources are compared with the auralization achieved through the proposed rendering method. The results of 15 participants suggest that in comparison to a linear interpolation-based alternative, the proposed object-based approach is perceived as being more realistic.
This article proposes a system for object-based six-degrees-of-freedom (6DoF) rendering of spatial sound scenes that are captured using a distributed arrangement of multiple Ambisonic receivers. The approach is based on first identifying and tracking the positions of sound sources within the scene, followed by the isolation of their signals through the use of beamformers. These sound objects are subsequently spatialized over the target playback setup, with respect to both the head orientation and position of the listener. The diffuse ambience of the scene is rendered separately by first spatially subtracting the source signals from the receivers located nearest to the listener position. The resultant residual Ambisonic signals are then spatialized, decorrelated, and summed together with suitable interpolation weights. The proposed system is evaluated through an in situ listening test conducted in 6DoF virtual reality,whereby real-world sound sources are compared with the auralization achieved through the proposed rendering method. The results of 15 participants suggest that in comparison to a linear interpolation-based alternative, the proposed object-based approach is perceived as being more realistic.
Open Access
Authors:
McCormack, Leo; Politis, Archontis; McKenzie, Thomas; Hold, Christoph; Pulkki, Ville
Affiliations:
Department of Signal Processing and Acoustics, Aalto University, Espoo, Finland; Department of Information Technology and Communication Sciences, Tampere University, Finland; Department of Signal Processing and Acoustics, Aalto University, Espoo, Finland; Department of Signal Processing and Acoustics, Aalto University, Espoo, Finland; Department of Signal Processing and Acoustics, Aalto University, Espoo, Finland(See document for exact affiliation information.) JAES Volume 70 Issue 5 pp. 355-372; May 2022
Publication Date:
May 11, 2022Import into BibTeX
Permalink:
http://www.aes.org/e-lib/browse.cfm?elib=21739