Abstract
For music genre prediction typically low-level audio signal features from time, spectral or cepstral domains are taken into account. Another way is to use community-based statistics such as Last.FM tags. Whereas the first feature group often can not be clearly interpreted by listeners, the second one lacks in erroneous or not available data for less popular songs. We propose a two-level approach combining the specific advantages of the both groups: at first we create high-level descriptors which describe instrumental and harmonic characteristics of music content, some of them derived from low-level features by supervised classification or from analysis of extended chroma and chord features. The experiments show that each categorization task requires its own feature set.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Ahrendt, P. (2006). Music Genre Classification Systems—A Computational Approach (Ph.D. thesis). Informatics and Mathematical Modelling, Technical University of Denmark, Kongens Lyngby.
Berenzweig, A., Ellis, D. P. W., & Lawrence, S. (2003). Anchor space for classification and similarity measurement of music. In Proceedings of the 2003 International Conference on Multimedia and Expo (ICME) (pp. 29–32). New York: IEEE.
Bertin-Mahieux, T., Eck, D., Maillet, F., & Lamere, P. (2008). Autotagger: A model for predicting social tags from acoustic features on large music databases. Journal of New Music Research, 37(2), 115–135.
Lartillot, O., & Toivainen, P. (2007). MIR in Matlab (ii): A toolbox for musical feature extraction from audio. In Proceedings of the 8th International Conference on Music Information Retrieval (ISMIR) (pp. 127–130). Austrian Computer Society (OCG).
Mauch, M., & Dixon, S. (2010). Approximate note transcription for the improved identification of difficult chords. In Proceedings of the 11th International Society for Music Information Retrieval Conference (ISMIR) (pp. 135–140). International Society for Music Information Retrieval.
Mckay, C., & Fujinaga, I. (2008). Combining features extracted from audio, symbolic and cultural sources. In Proceedings of the 9th International Conference on Music Information Retrieval (ISMIR) (pp. 597–602). Philadelphia: Drexel University.
Müller, M., & Ewert, S. (2010). Towards timbre-invariant audio features for harmony-based music. IEEE Transactions on Audio, Speech, and Language Processing, 18(3), 649–662.
Rötter, G., Vatolkin, I., & Weihs, C. (2011). Computational prediction of high-level descriptors of music personal categories. In Proceedings of the 2011 GfKl 35th Annual Conference of the German Classification Society (GfKl). Berlin: Springer.
Theimer, W., Vatolkin, I., & Eronen, A. (2008). Definitions of audio features for music content description. Algorithm Engineering Report TR08-2-001, Technische Universität Dortmund.
Vatolkin, I., Preuß, M., & Rudolph, G. (2011). Multi-objective feature selection in music genre and style recognition tasks. In Proceedings of the 2011 Genetic and Evolutionary Computation Conference (GECCO) (pp. 411–418). New York: ACM.
Vatolkin, I., Preuß, M., Rudolph, G., Eichhoff, M., & Weihs, C. (2012). Multi-objective evolutionary feature selection for instrument recognition in polyphonic audio mixtures. Soft Computing, 16(12), 2027–2047.
Vatolkin, I., Theimer, M., & Botteck, M. (2010). AMUSE (Advanced MUSic Explorer): A multitool framework for music data analysis. In Proceedings of the 11th International Society on Music Information Retrieval Conference (ISMIR) (pp. 33–38). International Society for Music Information Retrieval.
Weihs, C., Ligges, F., Mörchen, F., & Müllensiefen, D. (2007). Classification in music research. Advances in Data Analysis and Classification, 1(3), 255–291.
Acknowledgements
We thank the Klaus Tschira Foundation for the financial support.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2014 Springer International Publishing Switzerland
About this paper
Cite this paper
Vatolkin, I., Rötter, G., Weihs, C. (2014). Music Genre Prediction by Low-Level and High-Level Characteristics. In: Spiliopoulou, M., Schmidt-Thieme, L., Janning, R. (eds) Data Analysis, Machine Learning and Knowledge Discovery. Studies in Classification, Data Analysis, and Knowledge Organization. Springer, Cham. https://doi.org/10.1007/978-3-319-01595-8_46
Download citation
DOI: https://doi.org/10.1007/978-3-319-01595-8_46
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-01594-1
Online ISBN: 978-3-319-01595-8
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)