A Mid-Level Approach to Local Tonality Analysis: Extracting Key Signatures from Audio
×
Cite This
Citation & Abstract
C. Weiss, E. Cano, and H. Lukashevich, "A Mid-Level Approach to Local Tonality Analysis: Extracting Key Signatures from Audio," Paper P1-2, (2014 January.). doi:
C. Weiss, E. Cano, and H. Lukashevich, "A Mid-Level Approach to Local Tonality Analysis: Extracting Key Signatures from Audio," Paper P1-2, (2014 January.). doi:
Abstract: We propose a new method to automatically determine key signature changes. In automatic music transcription, sections in distantly related keys may lead to music scores that are hard to read due to a high number of notated accidentals. The problem of key change is commonly addressed by finding the correct local key out of the 24 major and minor keys. However, to provide the best matching key signature, choosing the right mode (major or minor) is not necessary and thus, we only estimate the local underlying diatonic scale. After extracting chroma features and a beat grid from the audio data, we calculate local probabilities for the different diatonic scales. For this purpose, we present a multiplicative procedure that shows promising results for visualizing complex tonal structures. From the obtained diatonic scale estimates, we identify candidates for key signature changes. By clustering similar segments and applying minimum segment length constraints, we get the tonal segmentation. We test our method on a dataset containing 30 hand-annotated pop songs. To evaluate our results, we calculate scores based on the number of frames correctly annotated, as well as segment border F-measures and perform a cross-validation study. Our rule-based method yields up to 90% class accuracy and up to 70% F-measure score for segment borders. These results are promising and qualify the approach to be applied for automatic music transcription.
@article{weiss2014a,
author={weiss, christof and cano, estefania and lukashevich, hanna},
journal={journal of the audio engineering society},
title={a mid-level approach to local tonality analysis: extracting key signatures from audio},
year={2014},
volume={},
number={},
pages={},
doi={},
month={january},}
@article{weiss2014a,
author={weiss, christof and cano, estefania and lukashevich, hanna},
journal={journal of the audio engineering society},
title={a mid-level approach to local tonality analysis: extracting key signatures from audio},
year={2014},
volume={},
number={},
pages={},
doi={},
month={january},
abstract={we propose a new method to automatically determine key signature changes. in automatic music transcription, sections in distantly related keys may lead to music scores that are hard to read due to a high number of notated accidentals. the problem of key change is commonly addressed by finding the correct local key out of the 24 major and minor keys. however, to provide the best matching key signature, choosing the right mode (major or minor) is not necessary and thus, we only estimate the local underlying diatonic scale. after extracting chroma features and a beat grid from the audio data, we calculate local probabilities for the different diatonic scales. for this purpose, we present a multiplicative procedure that shows promising results for visualizing complex tonal structures. from the obtained diatonic scale estimates, we identify candidates for key signature changes. by clustering similar segments and applying minimum segment length constraints, we get the tonal segmentation. we test our method on a dataset containing 30 hand-annotated pop songs. to evaluate our results, we calculate scores based on the number of frames correctly annotated, as well as segment border f-measures and perform a cross-validation study. our rule-based method yields up to 90% class accuracy and up to 70% f-measure score for segment borders. these results are promising and qualify the approach to be applied for automatic music transcription.},}
TY - paper
TI - A Mid-Level Approach to Local Tonality Analysis: Extracting Key Signatures from Audio
SP -
EP -
AU - Weiss, Christof
AU - Cano, Estefania
AU - Lukashevich, Hanna
PY - 2014
JO - Journal of the Audio Engineering Society
IS -
VO -
VL -
Y1 - January 2014
TY - paper
TI - A Mid-Level Approach to Local Tonality Analysis: Extracting Key Signatures from Audio
SP -
EP -
AU - Weiss, Christof
AU - Cano, Estefania
AU - Lukashevich, Hanna
PY - 2014
JO - Journal of the Audio Engineering Society
IS -
VO -
VL -
Y1 - January 2014
AB - We propose a new method to automatically determine key signature changes. In automatic music transcription, sections in distantly related keys may lead to music scores that are hard to read due to a high number of notated accidentals. The problem of key change is commonly addressed by finding the correct local key out of the 24 major and minor keys. However, to provide the best matching key signature, choosing the right mode (major or minor) is not necessary and thus, we only estimate the local underlying diatonic scale. After extracting chroma features and a beat grid from the audio data, we calculate local probabilities for the different diatonic scales. For this purpose, we present a multiplicative procedure that shows promising results for visualizing complex tonal structures. From the obtained diatonic scale estimates, we identify candidates for key signature changes. By clustering similar segments and applying minimum segment length constraints, we get the tonal segmentation. We test our method on a dataset containing 30 hand-annotated pop songs. To evaluate our results, we calculate scores based on the number of frames correctly annotated, as well as segment border F-measures and perform a cross-validation study. Our rule-based method yields up to 90% class accuracy and up to 70% F-measure score for segment borders. These results are promising and qualify the approach to be applied for automatic music transcription.
We propose a new method to automatically determine key signature changes. In automatic music transcription, sections in distantly related keys may lead to music scores that are hard to read due to a high number of notated accidentals. The problem of key change is commonly addressed by finding the correct local key out of the 24 major and minor keys. However, to provide the best matching key signature, choosing the right mode (major or minor) is not necessary and thus, we only estimate the local underlying diatonic scale. After extracting chroma features and a beat grid from the audio data, we calculate local probabilities for the different diatonic scales. For this purpose, we present a multiplicative procedure that shows promising results for visualizing complex tonal structures. From the obtained diatonic scale estimates, we identify candidates for key signature changes. By clustering similar segments and applying minimum segment length constraints, we get the tonal segmentation. We test our method on a dataset containing 30 hand-annotated pop songs. To evaluate our results, we calculate scores based on the number of frames correctly annotated, as well as segment border F-measures and perform a cross-validation study. Our rule-based method yields up to 90% class accuracy and up to 70% F-measure score for segment borders. These results are promising and qualify the approach to be applied for automatic music transcription.
Authors:
Weiss, Christof; Cano, Estefania; Lukashevich, Hanna
Affiliation:
Fraunhofer Institute for Digital Media Ttechnology, Ilmenau, Germany
AES Conference:
53rd International Conference: Semantic Audio (January 2014)
Paper Number:
P1-2
Publication Date:
January 27, 2014Import into BibTeX
Subject:
Music Informatics and Retrieval
Permalink:
http://www.aes.org/e-lib/browse.cfm?elib=17105