Enhancement of Ambisonics Signals using time-frequency masking
×
Cite This
Citation & Abstract
M. Lugasi, and B. Rafaely, "Enhancement of Ambisonics Signals using time-frequency masking," Paper 2-2, (2020 August.). doi:
M. Lugasi, and B. Rafaely, "Enhancement of Ambisonics Signals using time-frequency masking," Paper 2-2, (2020 August.). doi:
Abstract: Spatial audio is an essential part of virtual reality. Unlike synthesized signals, spatial audio captured in the real world may suffer from background noise which degrades the quality of the signals. While some previous works have addressed this problem, and suggested methods to attenuate the undesired signals while preserving the desired signals with minimum distortion, these only succeed partially. Recently, methods aiming to achieve preservation of the desired signal in its entirety have been proposed, and in this work we study such methods that are based on time-frequency masking. Two masks were investigated: one in the spherical harmonics (SH) domain, and the other in the plane wave density (PWD) function domain, referred to here as the spatial domain. These two methods were compared with a low-end reference method that uses a single maximum directivity beamformer followed by a single channel time-frequency mask. A subjective investigation was conducted to estimate the performance of these methods, and showed that the spatial mask preserves the desired sound field better, while the SH mask preserves the spatial cues of the residual noise better.
@article{lugasi2020enhancement,
author={lugasi, moti and rafaely, boaz},
journal={journal of the audio engineering society},
title={enhancement of ambisonics signals using time-frequency masking},
year={2020},
volume={},
number={},
pages={},
doi={},
month={august},}
@article{lugasi2020enhancement,
author={lugasi, moti and rafaely, boaz},
journal={journal of the audio engineering society},
title={enhancement of ambisonics signals using time-frequency masking},
year={2020},
volume={},
number={},
pages={},
doi={},
month={august},
abstract={spatial audio is an essential part of virtual reality. unlike synthesized signals, spatial audio captured in the real world may suffer from background noise which degrades the quality of the signals. while some previous works have addressed this problem, and suggested methods to attenuate the undesired signals while preserving the desired signals with minimum distortion, these only succeed partially. recently, methods aiming to achieve preservation of the desired signal in its entirety have been proposed, and in this work we study such methods that are based on time-frequency masking. two masks were investigated: one in the spherical harmonics (sh) domain, and the other in the plane wave density (pwd) function domain, referred to here as the spatial domain. these two methods were compared with a low-end reference method that uses a single maximum directivity beamformer followed by a single channel time-frequency mask. a subjective investigation was conducted to estimate the performance of these methods, and showed that the spatial mask preserves the desired sound field better, while the sh mask preserves the spatial cues of the residual noise better.},}
TY - paper
TI - Enhancement of Ambisonics Signals using time-frequency masking
SP -
EP -
AU - Lugasi, Moti
AU - Rafaely, Boaz
PY - 2020
JO - Journal of the Audio Engineering Society
IS -
VO -
VL -
Y1 - August 2020
TY - paper
TI - Enhancement of Ambisonics Signals using time-frequency masking
SP -
EP -
AU - Lugasi, Moti
AU - Rafaely, Boaz
PY - 2020
JO - Journal of the Audio Engineering Society
IS -
VO -
VL -
Y1 - August 2020
AB - Spatial audio is an essential part of virtual reality. Unlike synthesized signals, spatial audio captured in the real world may suffer from background noise which degrades the quality of the signals. While some previous works have addressed this problem, and suggested methods to attenuate the undesired signals while preserving the desired signals with minimum distortion, these only succeed partially. Recently, methods aiming to achieve preservation of the desired signal in its entirety have been proposed, and in this work we study such methods that are based on time-frequency masking. Two masks were investigated: one in the spherical harmonics (SH) domain, and the other in the plane wave density (PWD) function domain, referred to here as the spatial domain. These two methods were compared with a low-end reference method that uses a single maximum directivity beamformer followed by a single channel time-frequency mask. A subjective investigation was conducted to estimate the performance of these methods, and showed that the spatial mask preserves the desired sound field better, while the SH mask preserves the spatial cues of the residual noise better.
Spatial audio is an essential part of virtual reality. Unlike synthesized signals, spatial audio captured in the real world may suffer from background noise which degrades the quality of the signals. While some previous works have addressed this problem, and suggested methods to attenuate the undesired signals while preserving the desired signals with minimum distortion, these only succeed partially. Recently, methods aiming to achieve preservation of the desired signal in its entirety have been proposed, and in this work we study such methods that are based on time-frequency masking. Two masks were investigated: one in the spherical harmonics (SH) domain, and the other in the plane wave density (PWD) function domain, referred to here as the spatial domain. These two methods were compared with a low-end reference method that uses a single maximum directivity beamformer followed by a single channel time-frequency mask. A subjective investigation was conducted to estimate the performance of these methods, and showed that the spatial mask preserves the desired sound field better, while the SH mask preserves the spatial cues of the residual noise better.
Open Access
Authors:
Lugasi, Moti; Rafaely, Boaz
Affiliation:
Ben-Gurion University of the Negev
AES Conference:
2020 AES International Conference on Audio for Virtual and Augmented Reality (August 2020)
Paper Number:
2-2
Publication Date:
August 13, 2020Import into BibTeX
Permalink:
http://www.aes.org/e-lib/browse.cfm?elib=20872