Abstract
Skilled musicians are able to shape a given piece of music (by continuously modulating aspects like tempo, loudness, etc.) to communicate high level information such as musical structure and emotion. This activity is commonly referred to as expressive music performance. The present paper presents another step towards the automatic high-level analysis of this elusive phenomenon with AI methods. A system is presented that is able to measure tempo and dynamics of a musical performance and to track their development over time. The system accepts raw audio input, tracks tempo and dynamics changes in real time, and displays the development of these expressive parameters in an intuitive and aesthetically appealing graphical format which provides insight into the expressive patterns applied by skilled artists. The paper describes the tempo tracking algorithm (based on a new clustering method) in detail, and then presents an application of the system to the analysis of performances by different pianists.
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
A.T. Cemgil, B. Kappen, P. Desain, and H. Honing, ‘On tempo tracking: Tempogram representation and Kalman filtering’, in Proceedings of the 2000 International Computer Music Conference, pp. 352–355, San Francisco CA, (2000). International Computer Music Association.
P. Desain and H. Honing, ‘Quantization of musical time: A connectionist approach’, Computer Music Journal, 13(3), 56–66, (1989).
S. Dixon, ‘Automatic extraction of tempo and beat from expressive performances’, Journal of New Music Research, 30(1), 39–58, (2001).
S. Dixon and E. Cambouropoulos, ‘Beat tracking with musical knowledge’, in ECAI 2000: Proceedings of the 14th European Conference on Artificial Intelligence, pp. 626–630, Amsterdam, (2000). IOS Press.
M. Goto and Y. Muraoka, ‘Areal-time beat tracking system for audio signals’, in Proceedings of the International Computer Music Conference, pp. 171–174, San Francisco CA, (1995). International Computer Music Association.
M. Goto and Y. Muraoka, ‘Real-time beat tracking for drumless audio signals’, Speech Communication, 27(3–4), 331–335, (1999).
J. Langner and W. Goebl, ‘Representing expressive performance in tempo-loudness space’, in Proceedings of the ESCOM 10th Anniversary Conference on Musical Creativity, Liège, Belgium, (2002).
E.W. Large and J.F. Kolen, ‘Resonance and the perception of musical meter’, Connection Science, 6, 177–208, (1994).
C.S. Lee, ‘The perception of metrical structure: Experimental evidence and a model’, in Representing Musical Structure, eds., P. Howell, R. West, and I. Cross, 59–127, Academic Press, San Diego CA, (1991).
H.C. Longuet-Higgins, Mental Processes, MIT Press, Cambridge MA, 1987.
D. Rosenthal, ‘Emulation of human rhythm perception’, Computer Music Journal, 16(1), 64–76, (1992).
E.D. Scheirer, ‘Tempo and beat analysis of acoustic musical signals’, Journal of the Acoustical Society of America, 103(1), 588–601, (1998).
G. Widmer, ‘What is it that makes it a Horowitz? Empirical musicology via machine learning’, in Proceedings of the 12th European Conference on Artificial Intelligence (ECAI-96), Chichester, UK, (1996).Wiley & Sons.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2002 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Dixon, S., Goebl, W., Widmer, G. (2002). Real Time Tracking and Visualisation of Musical Expression. In: Anagnostopoulou, C., Ferrand, M., Smaill, A. (eds) Music and Artificial Intelligence. ICMAI 2002. Lecture Notes in Computer Science(), vol 2445. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-45722-4_7
Download citation
DOI: https://doi.org/10.1007/3-540-45722-4_7
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-44145-8
Online ISBN: 978-3-540-45722-0
eBook Packages: Springer Book Archive