Less acoustic features means more statistical relevance: Disclosing the clustering behavior in music stimuli

Authors

Keywords:

music, audio analysis, cluster analysis

Abstract

Identification of appropriate content-based features for the description of audio signals can provide a better repre-
sentation of naturalistic music stimuli which, in recent years, have been used to understand how the human brain processes such information. In this work, an extensive clustering analysis has been carried out on a large and benchmark audio dataset to assess whether features commonly extracted in the literature are in fact statistically relevant. Our results show that not all of these well-known acoustic features might be statistically necessary. We also demonstrate quantitatively that, regardless of the musical genre, the same acoustic feature is selected to represent each cluster. This finding discloses that there is a general redundancy among the set of audio descriptors used, that does not depend on a particular music track or genre, allowing an expressive reduction of the number of features necessary to identify apropriate time instants on the audio for further brain signal processing of music stimuli.

Downloads

Download data is not yet available.

Author Biographies

Estela Ribeiro, Centro Universitario FEI, Sao Bernardo do Campo, Brazil

Estela Ribeiro received in 2015 the B. Sc. degree in mechanical engineering from FSA University Center, São Paulo, Brazil. Obtained the M.Sc. degree in electrical engineering from FEI University Center in 2017. Obtained the ph.D. degree in electrical engineering at FEI University Center, São Paulo, Brazil, in 2020. Since 2016, she has received a research fellowship from FEI, LNCT and CAPES to develop research activities on signal processing and pattern recognition. Her research interests include pattern recognition, cognitive perception and machine learning.

Carlos Eduardo Thomaz, FEI University Center

Carlos Thomaz received in 1993 the B.Sc. degree in electronic engineering from Pontifical Catholic University of Rio de Janeiro (PUC-Rio), Brazil. After working for six years in industry, he obtained the M.Sc. degree in electrical engineering from PUC-Rio in 1999. In October 2000, he joined the Department of Computing at Imperial College London where he obtained the Ph.D. degree in statistical pattern recognition in 2004. He joined the Department of Electrical Engineering, FEI University Center, São Paulo, Brazil, in 2005, as an Associate Professor, where he has been, since 2006, head of the Image Processing Laboratory. Since 2014 he has been Professor of Statistical Pattern Recognition at FEI. His research interests include pattern recognition, cognitive perception and machine learning. From 2015 to 2018, Professor Thomaz was awarded a Newton Advanced Fellowship from the Royal Society, UK.

References

D. Mitrovic, M. Zeppelzauer, and C. Breiteneder, “Features for content based audio retrieval,” in Advances in Computers Volume 78 Improving the Web, pp. 71–150, Elsevier, 2010.

V. Alluri, P. Toiviainen, I. P. Jaaskelainen, E. Glerean, M. Sams, and E. Brattico, “Large-scale brain networks emerge from dynamics

processing of musical timbre, key and rhythm,” NeuroImage, vol. 59,no. 4, pp. 3677–3689, 2012.

V. Alluri, P. Toiviainen, T. E. Lund, M. Wallentin, P. Vuust, K. Nandi, Asoke, T. Ristaniemi, and E. Brattico, “From vivaldi to beatles and back:Predicting lateralized brain responses to music,” NeuroImage, vol. 83, pp. 627–636, Dec. 2013.

H. Poikonen, V. Alluri, E. Brattico, O. Lartillot, M. Tervaniemi, and M. Huotilainen, “Event-related brain responses while listening to entire pieces of music,” Neuroscience, vol. 312, pp. 58–73, Jan. 2016.

P. Saari, I. Burunat, E. Brattico, and P. Toiviainen, “Decoding musical training from dynamic processing of musical features in the brain,”Scientific Report, vol. 708, no. 8, pp. 1–12, 2018.

E. Ribeiro and C. E. Thomaz, “A whole brain eeg analysis of musicianship,” Music Perception: An Interdisciplinary Journal, vol. 37, no. 1,

pp. 42–56, 2019.

N. T. Haumann, M. Lumaca, M. Kliuchko, J. L. Santacruz, P. Vuust, and E. Brattico, “Extracting human cortical responses to sound onsets and acoustic feature changes in real music, and their relation to event rate,” Brain Research, p. 147248, 2021.

L. A. Ferreira, E. Ribeiro, and C. E. Thomaz, “A multivariate statistical analysis of eeg signals for differentiation of musicians and nonmusicians,”in 17th Brazilian Symposium on Computer Music (SBCM) (J. T. Araújo and F. L. Schiavoni, eds.), (Sao Paulo, Brazil), pp. 80–85,Brazilian Computer Society (SBC), 10 2019.

G. Tzanetakis and P. Cook, “Musical genre classification of audio signals,” IEEE Transactions on Speech and Audio Processing, vol. 10,

pp. 293–302, 7 2002.

B. L. Sturm, “The state of the art ten years after a state of the art: Future research in music information retrieval,” Journal of New Music Research, vol. 43, p. 147–172, 4 2014.

O. Lartillot, MIRtoolbox 1.6.1 Users Manual, 2014.

A. Lerch, An Introduction to audio content analysis. Applications in signal processing and music informatics. IEEE Press, 1st ed., 2012.

P. Kness and M. Schedl, Music Similarity and Retrieval: An introduction to audio and web based strategies. Springer, 1st ed., 2016.

H. Poikonen, P. Toiviainen, and M. Tervaniemi, “Early auditory processing in musicians and dancers during a contemporary dance piece,” Scientific Reports, vol. 6, Sept. 2016.

R. A. Johnson and D. W. Wichern, Applied Multivariate Statistical Analysis. Pearson, 6th ed., 2007.

H. F. Kaiser, “The varimax criterion for analytic rotation in factor analysis,” Psychometrika, vol. 23, pp. 187–200, 1958.

P. J. Rousseeuw, “Silhouettes: A graphical aid to the interpretation and validation of cluster analysis,” Journal of Computational and Applied Mathematics, vol. 20, pp. 53 – 65, 1987.

D. L. Davies and D. W. Bouldin, “A cluster separation measure,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. PAMI-1, no. 2, pp. 224–227, 1979.

Published

2022-01-06

How to Cite

Ribeiro, E., & Thomaz, C. E. (2022). Less acoustic features means more statistical relevance: Disclosing the clustering behavior in music stimuli. IEEE Latin America Transactions, 20(4), 686–692. Retrieved from https://latamt.ieeer9.org/index.php/transactions/article/view/5908