The Physics of Auditory Proximity and its Effects on Intelligibility and Recall
×
Cite This
Citation & Abstract
D. Griesinger, "The Physics of Auditory Proximity and its Effects on Intelligibility and Recall," Paper 9659, (2016 September.). doi:
D. Griesinger, "The Physics of Auditory Proximity and its Effects on Intelligibility and Recall," Paper 9659, (2016 September.). doi:
Abstract: Cutthroat evolution has given us seemingly magical abilities to hear speech in complex environments. We can tell instantly, independent of timbre or loudness, if a sound is close to us, and in a crowded room we can switch attention at will between at least three different simultaneous conversations. And we involuntarily switch attention if our name is spoken. These feats are only possible if, without conscious attention, each voice has been separated into an independent neural stream. We believe the separation process relies on the phase relationships between the harmonics above 1000 Hz that encode speech information, and the neurology of the inner ear that has evolved to detect them. When phase is undisturbed, once in each fundamental period harmonic phases align to create massive peaks in the sound pressure at the fundamental frequency. Pitch-sensitive filters can detect and separate these peaks from each other and from noise with amazing acuity. But reflections and sound systems randomize phases, with serious effects on attention, source separation, and intelligibility. This talk will detail the many ways ears and speech have co-evolved, and recent work on the importance of phase in acoustics and sound design.
@article{griesinger2016the,
author={griesinger, david},
journal={journal of the audio engineering society},
title={the physics of auditory proximity and its effects on intelligibility and recall},
year={2016},
volume={},
number={},
pages={},
doi={},
month={september},}
@article{griesinger2016the,
author={griesinger, david},
journal={journal of the audio engineering society},
title={the physics of auditory proximity and its effects on intelligibility and recall},
year={2016},
volume={},
number={},
pages={},
doi={},
month={september},
abstract={cutthroat evolution has given us seemingly magical abilities to hear speech in complex environments. we can tell instantly, independent of timbre or loudness, if a sound is close to us, and in a crowded room we can switch attention at will between at least three different simultaneous conversations. and we involuntarily switch attention if our name is spoken. these feats are only possible if, without conscious attention, each voice has been separated into an independent neural stream. we believe the separation process relies on the phase relationships between the harmonics above 1000 hz that encode speech information, and the neurology of the inner ear that has evolved to detect them. when phase is undisturbed, once in each fundamental period harmonic phases align to create massive peaks in the sound pressure at the fundamental frequency. pitch-sensitive filters can detect and separate these peaks from each other and from noise with amazing acuity. but reflections and sound systems randomize phases, with serious effects on attention, source separation, and intelligibility. this talk will detail the many ways ears and speech have co-evolved, and recent work on the importance of phase in acoustics and sound design.},}
TY - paper
TI - The Physics of Auditory Proximity and its Effects on Intelligibility and Recall
SP -
EP -
AU - Griesinger, David
PY - 2016
JO - Journal of the Audio Engineering Society
IS -
VO -
VL -
Y1 - September 2016
TY - paper
TI - The Physics of Auditory Proximity and its Effects on Intelligibility and Recall
SP -
EP -
AU - Griesinger, David
PY - 2016
JO - Journal of the Audio Engineering Society
IS -
VO -
VL -
Y1 - September 2016
AB - Cutthroat evolution has given us seemingly magical abilities to hear speech in complex environments. We can tell instantly, independent of timbre or loudness, if a sound is close to us, and in a crowded room we can switch attention at will between at least three different simultaneous conversations. And we involuntarily switch attention if our name is spoken. These feats are only possible if, without conscious attention, each voice has been separated into an independent neural stream. We believe the separation process relies on the phase relationships between the harmonics above 1000 Hz that encode speech information, and the neurology of the inner ear that has evolved to detect them. When phase is undisturbed, once in each fundamental period harmonic phases align to create massive peaks in the sound pressure at the fundamental frequency. Pitch-sensitive filters can detect and separate these peaks from each other and from noise with amazing acuity. But reflections and sound systems randomize phases, with serious effects on attention, source separation, and intelligibility. This talk will detail the many ways ears and speech have co-evolved, and recent work on the importance of phase in acoustics and sound design.
Cutthroat evolution has given us seemingly magical abilities to hear speech in complex environments. We can tell instantly, independent of timbre or loudness, if a sound is close to us, and in a crowded room we can switch attention at will between at least three different simultaneous conversations. And we involuntarily switch attention if our name is spoken. These feats are only possible if, without conscious attention, each voice has been separated into an independent neural stream. We believe the separation process relies on the phase relationships between the harmonics above 1000 Hz that encode speech information, and the neurology of the inner ear that has evolved to detect them. When phase is undisturbed, once in each fundamental period harmonic phases align to create massive peaks in the sound pressure at the fundamental frequency. Pitch-sensitive filters can detect and separate these peaks from each other and from noise with amazing acuity. But reflections and sound systems randomize phases, with serious effects on attention, source separation, and intelligibility. This talk will detail the many ways ears and speech have co-evolved, and recent work on the importance of phase in acoustics and sound design.
Open Access
Author:
Griesinger, David
Affiliation:
David Griesinger Acoustics, Cambridge, MA, USA
AES Convention:
141 (September 2016)
Paper Number:
9659
Publication Date:
September 20, 2016Import into BibTeX
Subject:
Perception
Permalink:
http://www.aes.org/e-lib/browse.cfm?elib=18463