Abstract
Identification of particular voices in polyphonic and polytimbral music is a task often performed by musicians in their everyday life. However, the automation of this task is very challenging, because of high complexity of audio data. Usually additional information is supplied, and the results are far from satisfactory. In this paper, we focus on classical music recordings, without requiring the user to submit additional information. Our goal is to identify musical instruments playing in short audio frames of polyphonic recordings of classical music. Additionally, we extract pitches (or pitch ranges) which combined with instrument information can be used in score-following and audio alignment, see e.g. [9, 20], or in works towards automatic score extraction, which are a motivation behind this work. Also, since instrument timbre changes with pitch, separate classifiers are trained for various pitch ranges for each instrument. Four instruments are investigated, representing stringed and wind instruments. The influence of adding harmonic (pitch-based) features to the feature set on the results is also investigated. Random forests are applied as a classification tool, and the results are presented and discussed.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Boulanger-Lewandowski, N., Bengio, Y., Vincent, P.: Discriminative non-negative matrix factorization for multiple pitch estimation. In: 13th International Society for Music Information Retrieval Conference (ISMIR), pp. 205–210 (2012)
Breiman, L.: Random forests. Mach. Learn. 45, 5–32 (2001)
Essid, S., Richard, G., David, B.: Instrument recognition in polyphonic music based on automatic taxonomies. IEEE Trans. Audio Speech Lang. Process. 14(1), 68–80 (2006)
Goto, M., Hashiguchi, H., Nishimura, T., Oka, R.: RWC music database: popular, classical, and jazz music databases. In: 3rd International Conference on Music Information Retrieval, pp. 287–288 (2002)
Goto, M., Hashiguchi, H., Nishimura, T., Oka, R.: RWC music database: music genre database and musical instrument sound database. In: 4th International Conference on Music Information Retrieval, pp. 229–230 (2003)
Heittola, T., Klapuri, A., Virtanen, A.: Musical instrument recognition in polyphonic audio using source-filter model for sound separation. In: 10th International Society for Music Information Retrieval Conference (2009)
Herrera-Boyer, P., Klapuri, A., Davy, M.: Automatic classification of pitched musical instrument sounds. In: Klapuri, A., Davy, M. (eds.) Signal Processing Methods for Music Transcription. Springer Science+Business Media LLC, US (2006)
ISO: MPEG-7 overview. http://www.chiariglione.org/mpeg/
Izmirli, O., Sharma, G.: Bridging printed music and audio through alignment using a mid-level score representation. In: 13th International Society for Music Information Retrieval Conference (ISMIR), pp. 61–66 (2012)
Kameoka, H., Nishimoto, T., Sagayama, S.: Multi-pitch detection algorithm using constrained gaussian mixture model and information criterion for simultaneous speech. In: Speech Prosody 2004, pp. 533–536 (2004)
Kirchhoff, H., Dixon, S., Klapuri, A.: Multi-template shift-variant non-negative matrix deconvolution for semi-automatic music transcription. In: 13th International Society for Music Information Retrieval Conference (ISMIR), pp. 415–420 (2012)
Kitahara, T., Goto, M., Komatani, K., Ogata, T., Okuno, H.G.: Instrument identification in polyphonic music: feature weighting to minimize influence of sound overlaps. EURASIP J. Appl. Signal Process. 2007, 1–15 (2007)
Kubera, E., Wieczorkowska, A., Raś, Z., Skrzypiec, M.: Recognition of instrument timbres in real polytimbral audio recordings. In: Balcázar, J.L., Bonchi, F., Gionis, A., Sebag, M. (eds.) ECML PKDD 2010, Part II. LNCS, vol. 6322, pp. 97–110. Springer, Heidelberg (2010)
Kubera, E., Wieczorkowska, A.A.: Mining audio data for multiple instrument recognition in classical music. In: Appice, A., Ceci, M., Loglisci, C., Manco, G., Masciari, E., Ras, Z.W. (eds.) NFMCP 2013. LNCS, vol. 8399, pp. 246–260. Springer, Heidelberg (2014)
Kubera, E., Wieczorkowska, A.A., Skrzypiec, M.: Influence of feature sets on precision, recall, and accuracy of identification of musical instruments in audio recordings. In: Andreasen, T., Christiansen, H., Cubero, J.-C., Raś, Z.W. (eds.) ISMIS 2014. LNCS, vol. 8502, pp. 204–213. Springer, Heidelberg (2014)
Kursa, M., Rudnicki, W., Wieczorkowska, A., Kubera, E., Kubik-Komar, A.: Musical instruments in random forest. In: Rauch, J., Raś, Z.W., Berka, P., Elomaa, T. (eds.) ISMIS 2009. LNCS, vol. 5722, pp. 281–290. Springer, Heidelberg (2009)
Martins, L.G., Burred, J.J., Tzanetakis, G., Lagrange, M.: Polyphonic instrument recognition using spectral clustering. In: 8th International Society for Music Information Retrieval Conference (ISMIR) (2007)
Max-Planck-Institut Informatik: chroma toolbox: pitch, chroma, CENS, CRP. http://www.mpi-inf.mpg.de/resources/MIR/chromatoolbox/
MIDOMI: Search for music using your voice by singing or humming. http://www.midomi.com/
Miotto, R., Montecchio, N., Orio, N.: Statistical music modeling aimed at identification and alignment. In: Raś, Z.W., Wieczorkowska, A.A. (eds.) Adv. Music Inform. Retrieval. SCI, vol. 274, pp. 187–212. Springer, Heidelberg (2010)
Müller, M.: Information Retrieval for Music and Motion. Springer, Heidelberg (2007)
Niewiadomy, D., Pelikant, A.: Implementation of MFCC vector generation in classification context. J. Appl. Comput. Sci. 16(2), 55–65 (2008)
Opolko, F., Wapnick, J.: MUMS – McGill University master samples: CD’s (1987)
Oxford University press: Oxford dictionaries. http://www.oxforddictionaries.com/
Sakaue, D., Otsuka, T., Itoyama, K., Okuno, H.G.: Bayesian nonnegative harmonic-temporal factorization and its application to multipitch analysis. In: 13th International Society for Music Information Retrieval Conference (ISMIR), pp. 91–96 (2012)
Shazam entertainment ltd. http://www.shazam.com/
Subrahmanian, V.S.: Principles of Multimedia Database Systems. Morgan Kaufmann, San Francisco (1998)
The University of IOWA electronic music studios: musical instrument samples. http://theremin.music.uiowa.edu/MIS.html
TrackID. https://play.google.com/store/apps/details?id=com.sonyericsson.trackid
Vincent, E., Rodet, X.: Music transcription with ISA and HMM. In: Puntonet, C.G., Prieto, A.G. (eds.) ICA 2004. LNCS, vol. 3195, pp. 1197–1204. Springer, Heidelberg (2004)
Zhang, X., Marasek, K., Ras, Z.W.: Maximum likelihood study for sound pattern separation and recognition. In: IEEE CS International Conference on Multimedia and Ubiquitous Engineering (MUE 2007), Seoul, Korea, pp. 807–812 (2007)
Zweig, G., Nguyen, P.: A segmental CRF approach to large vocabulary continuous speech recognition. In: ASRU 2009: Automatic Speech Recognition and Understanding (2009)
Acknowledgments
This work was partially supported by the Research Center of PJAIT, supported by the Ministry of Science and Higher Education in Poland.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2015 Springer International Publishing Switzerland
About this paper
Cite this paper
Kubera, E., Wieczorkowska, A.A. (2015). Pitch-Related Identification of Instruments in Classical Music Recordings. In: Appice, A., Ceci, M., Loglisci, C., Manco, G., Masciari, E., Ras, Z. (eds) New Frontiers in Mining Complex Patterns. NFMCP 2014. Lecture Notes in Computer Science(), vol 8983. Springer, Cham. https://doi.org/10.1007/978-3-319-17876-9_13
Download citation
DOI: https://doi.org/10.1007/978-3-319-17876-9_13
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-17875-2
Online ISBN: 978-3-319-17876-9
eBook Packages: Computer ScienceComputer Science (R0)