An Analysis of Low-Arousal Piano Music Ratings to Uncover What Makes Calm and Sad Music So Difficult to Distinguish in Music Emotion Recognition
×
Cite This
Citation & Abstract
Y. Hong, C. Chau, and A. Horner, "An Analysis of Low-Arousal Piano Music Ratings to Uncover What Makes Calm and Sad Music So Difficult to Distinguish in Music Emotion Recognition," J. Audio Eng. Soc., vol. 65, no. 4, pp. 304-320, (2017 April.). doi: https://doi.org/10.17743/jaes.2017.0001
Y. Hong, C. Chau, and A. Horner, "An Analysis of Low-Arousal Piano Music Ratings to Uncover What Makes Calm and Sad Music So Difficult to Distinguish in Music Emotion Recognition," J. Audio Eng. Soc., vol. 65 Issue 4 pp. 304-320, (2017 April.). doi: https://doi.org/10.17743/jaes.2017.0001
Abstract: Systems that recognize the emotional content of music and systems that provide music recommendations often use a simplified 4-quadrant model with categories such as Happy, Sad, Angry, and Calm. Previous research has shown that both listeners and automated systems often have difficulty distinguishing low-arousal categories such as Calm and Sad. This paper explores what makes these categories difficult to distinguish. 300 low-arousal excerpts from the classical piano repertoire were used to determine the coverage of the categories Calm and Sad in the low-arousal space, their overlap, and their balance to one another. Results show that Calm was 40% bigger in terms of coverage than Sad, but on average, Sad excerpts were significantly more negative in mood than Calm excerpts that were positive. Calm and Sad overlapped in nearly 20% of the excerpts, meaning 20% of the excerpts were about equally Calm and Sad. Calm and Sad covered about 92% of the low-arousal space. The largest holes were for excerpts considered Mysterious and Doubtful. Due to the holes in the coverage, the overlaps, and imbalances, the Calm-Sad model adds about 6% more errors when compared to asking users directly whether the mood of the music is positive or negative. Nevertheless, the Calm-Sad model is still useful and appropriate for many applications.
@article{hong2017an,
author={hong, yu and chau, chuck-jee and horner, andrew},
journal={journal of the audio engineering society},
title={an analysis of low-arousal piano music ratings to uncover what makes calm and sad music so difficult to distinguish in music emotion recognition},
year={2017},
volume={65},
number={4},
pages={304-320},
doi={https://doi.org/10.17743/jaes.2017.0001},
month={april},}
@article{hong2017an,
author={hong, yu and chau, chuck-jee and horner, andrew},
journal={journal of the audio engineering society},
title={an analysis of low-arousal piano music ratings to uncover what makes calm and sad music so difficult to distinguish in music emotion recognition},
year={2017},
volume={65},
number={4},
pages={304-320},
doi={https://doi.org/10.17743/jaes.2017.0001},
month={april},
abstract={systems that recognize the emotional content of music and systems that provide music recommendations often use a simplified 4-quadrant model with categories such as happy, sad, angry, and calm. previous research has shown that both listeners and automated systems often have difficulty distinguishing low-arousal categories such as calm and sad. this paper explores what makes these categories difficult to distinguish. 300 low-arousal excerpts from the classical piano repertoire were used to determine the coverage of the categories calm and sad in the low-arousal space, their overlap, and their balance to one another. results show that calm was 40% bigger in terms of coverage than sad, but on average, sad excerpts were significantly more negative in mood than calm excerpts that were positive. calm and sad overlapped in nearly 20% of the excerpts, meaning 20% of the excerpts were about equally calm and sad. calm and sad covered about 92% of the low-arousal space. the largest holes were for excerpts considered mysterious and doubtful. due to the holes in the coverage, the overlaps, and imbalances, the calm-sad model adds about 6% more errors when compared to asking users directly whether the mood of the music is positive or negative. nevertheless, the calm-sad model is still useful and appropriate for many applications.},}
TY - paper
TI - An Analysis of Low-Arousal Piano Music Ratings to Uncover What Makes Calm and Sad Music So Difficult to Distinguish in Music Emotion Recognition
SP - 304
EP - 320
AU - Hong, Yu
AU - Chau, Chuck-Jee
AU - Horner, Andrew
PY - 2017
JO - Journal of the Audio Engineering Society
IS - 4
VO - 65
VL - 65
Y1 - April 2017
TY - paper
TI - An Analysis of Low-Arousal Piano Music Ratings to Uncover What Makes Calm and Sad Music So Difficult to Distinguish in Music Emotion Recognition
SP - 304
EP - 320
AU - Hong, Yu
AU - Chau, Chuck-Jee
AU - Horner, Andrew
PY - 2017
JO - Journal of the Audio Engineering Society
IS - 4
VO - 65
VL - 65
Y1 - April 2017
AB - Systems that recognize the emotional content of music and systems that provide music recommendations often use a simplified 4-quadrant model with categories such as Happy, Sad, Angry, and Calm. Previous research has shown that both listeners and automated systems often have difficulty distinguishing low-arousal categories such as Calm and Sad. This paper explores what makes these categories difficult to distinguish. 300 low-arousal excerpts from the classical piano repertoire were used to determine the coverage of the categories Calm and Sad in the low-arousal space, their overlap, and their balance to one another. Results show that Calm was 40% bigger in terms of coverage than Sad, but on average, Sad excerpts were significantly more negative in mood than Calm excerpts that were positive. Calm and Sad overlapped in nearly 20% of the excerpts, meaning 20% of the excerpts were about equally Calm and Sad. Calm and Sad covered about 92% of the low-arousal space. The largest holes were for excerpts considered Mysterious and Doubtful. Due to the holes in the coverage, the overlaps, and imbalances, the Calm-Sad model adds about 6% more errors when compared to asking users directly whether the mood of the music is positive or negative. Nevertheless, the Calm-Sad model is still useful and appropriate for many applications.
Systems that recognize the emotional content of music and systems that provide music recommendations often use a simplified 4-quadrant model with categories such as Happy, Sad, Angry, and Calm. Previous research has shown that both listeners and automated systems often have difficulty distinguishing low-arousal categories such as Calm and Sad. This paper explores what makes these categories difficult to distinguish. 300 low-arousal excerpts from the classical piano repertoire were used to determine the coverage of the categories Calm and Sad in the low-arousal space, their overlap, and their balance to one another. Results show that Calm was 40% bigger in terms of coverage than Sad, but on average, Sad excerpts were significantly more negative in mood than Calm excerpts that were positive. Calm and Sad overlapped in nearly 20% of the excerpts, meaning 20% of the excerpts were about equally Calm and Sad. Calm and Sad covered about 92% of the low-arousal space. The largest holes were for excerpts considered Mysterious and Doubtful. Due to the holes in the coverage, the overlaps, and imbalances, the Calm-Sad model adds about 6% more errors when compared to asking users directly whether the mood of the music is positive or negative. Nevertheless, the Calm-Sad model is still useful and appropriate for many applications.
Open Access
Authors:
Hong, Yu; Chau, Chuck-Jee; Horner, Andrew
Affiliation:
Department of Computer Science and Engineering Hong Kong University of Science and Technology, Clear Water Bay, Kowloon, Hong Kong JAES Volume 65 Issue 4 pp. 304-320; April 2017
Publication Date:
April 28, 2017Import into BibTeX
Permalink:
http://www.aes.org/e-lib/browse.cfm?elib=18563