P. Eibl, Y. Grewe, D. Rieger, and U. Scuda, "Production Tools for the MPEG-H Audio System," Engineering Brief 651, (2021 October.). doi:
P. Eibl, Y. Grewe, D. Rieger, and U. Scuda, "Production Tools for the MPEG-H Audio System," Engineering Brief 651, (2021 October.). doi:
Abstract: Next Generation Audio (NGA) systems like MPEG-H Audio rely on metadata to enable a wide variety of features. Information such as channel layouts, the position and properties of audio objects, and user interactivity options are only some of the data that can be used to improve consumer experience. Creating these metadata requires suitable tools, which are used in a process known as "authoring", where interactive features and the options for 3D immersive sound rendering are defined by the content creator. Different types of productions impose specific requirements on these authoring tools, which leads to a number of solutions appearing in the market. Using the example of MPEG-H Audio, this paper will detail some of the latest developments and authoring solutions designed to enable immersive and interactive live and post-productions.
@article{eibl2021production,
author={eibl, philipp and grewe, yannik and rieger, daniela and scuda, ulli},
journal={journal of the audio engineering society},
title={production tools for the mpeg-h audio system},
year={2021},
volume={},
number={},
pages={},
doi={},
month={october},}
@article{eibl2021production,
author={eibl, philipp and grewe, yannik and rieger, daniela and scuda, ulli},
journal={journal of the audio engineering society},
title={production tools for the mpeg-h audio system},
year={2021},
volume={},
number={},
pages={},
doi={},
month={october},
abstract={next generation audio (nga) systems like mpeg-h audio rely on metadata to enable a wide variety of features. information such as channel layouts, the position and properties of audio objects, and user interactivity options are only some of the data that can be used to improve consumer experience. creating these metadata requires suitable tools, which are used in a process known as "authoring", where interactive features and the options for 3d immersive sound rendering are defined by the content creator. different types of productions impose specific requirements on these authoring tools, which leads to a number of solutions appearing in the market. using the example of mpeg-h audio, this paper will detail some of the latest developments and authoring solutions designed to enable immersive and interactive live and post-productions.},}
TY - paper
TI - Production Tools for the MPEG-H Audio System
SP -
EP -
AU - Eibl, Philipp
AU - Grewe, Yannik
AU - Rieger, Daniela
AU - Scuda, Ulli
PY - 2021
JO - Journal of the Audio Engineering Society
IS -
VO -
VL -
Y1 - October 2021
TY - paper
TI - Production Tools for the MPEG-H Audio System
SP -
EP -
AU - Eibl, Philipp
AU - Grewe, Yannik
AU - Rieger, Daniela
AU - Scuda, Ulli
PY - 2021
JO - Journal of the Audio Engineering Society
IS -
VO -
VL -
Y1 - October 2021
AB - Next Generation Audio (NGA) systems like MPEG-H Audio rely on metadata to enable a wide variety of features. Information such as channel layouts, the position and properties of audio objects, and user interactivity options are only some of the data that can be used to improve consumer experience. Creating these metadata requires suitable tools, which are used in a process known as "authoring", where interactive features and the options for 3D immersive sound rendering are defined by the content creator. Different types of productions impose specific requirements on these authoring tools, which leads to a number of solutions appearing in the market. Using the example of MPEG-H Audio, this paper will detail some of the latest developments and authoring solutions designed to enable immersive and interactive live and post-productions.
Next Generation Audio (NGA) systems like MPEG-H Audio rely on metadata to enable a wide variety of features. Information such as channel layouts, the position and properties of audio objects, and user interactivity options are only some of the data that can be used to improve consumer experience. Creating these metadata requires suitable tools, which are used in a process known as "authoring", where interactive features and the options for 3D immersive sound rendering are defined by the content creator. Different types of productions impose specific requirements on these authoring tools, which leads to a number of solutions appearing in the market. Using the example of MPEG-H Audio, this paper will detail some of the latest developments and authoring solutions designed to enable immersive and interactive live and post-productions.
Open Access
Authors:
Eibl, Philipp; Grewe, Yannik; Rieger, Daniela; Scuda, Ulli
Affiliation:
Fraunhofer Institute for Integrated Circuits IIS, Erlangen, Germany
AES Convention:
151 (October 2021)eBrief:651
Publication Date:
October 13, 2021Import into BibTeX
Subject:
Spatial audio applications
Permalink:
http://www.aes.org/e-lib/browse.cfm?elib=21514
The Engineering Briefs at this Convention were
selected on the basis of a submitted synopsis,
ensuring that they are of interest to AES members,
and are not overly commercial. These briefs have
been reproduced from the authors' advance
manuscripts, without editing, corrections, or
consideration by the Review Board. The AES takes no
responsibility for their contents. Paper copies are
not available, but any member can freely access
these briefs. Members are encouraged to provide
comments that enhance their usefulness.