Immersive technologies with binaural audio are commonly seen as virtual reality headsets for applications such as gaming or 3D film. However, audio-led augmented reality applications where the view of the wearer is not obscured by a headset have been under-explored. This is partially because of the hardware and software requirements for rendering binaural audio in a wearable device. In this workshop we present an open source, wearable system for rendering interactive, head-tracked binaural audio scenes.
This workshop is a hands-on session in which participants will learn how to use the embedded Linux platform Bela (http://bela.io/) to render head-tracked binaural signals. Bela is designed for ultra-low latency audio signal processing, making it an ideal platform for head-tracked audio. Using a combination of Pure Data and C++, participants will be shown how to use an open source library implementing a virtual loudspeaker approach along with how to work with an inertial measurement unit (IMU).
This session includes coffee break.