AUTOMATIC IDENTIFICATION OF SOME VIETNAMESE FOLK SONGS CHEO AND QUANHO USING CONVOLUTIONAL NEURAL NETWORKS
Author affiliations
DOI:
https://doi.org/10.15625/1813-9663/38/1/15961Keywords:
Identification, classification, folk songs, Vietnamese, Cheo, Quanho, CNNAbstract
We can say that music in general is an indispensable spiritual food in human life. For Vietnamese people, folk music plays a very important role, it has entered the minds of every Vietnamese person right from the moment of birth through lullabies for children. In Vietnam, there are many different types of folk songs that everyone loves, and each has many different melodies. In order to archive and search music works with a very large quantity, including folk songs, it is necessary to automatically classify and identify those works. This paper presents the method of determining the feature parameters and then using the convolution neural network (CNN) to classify and identify some Vietnamese folk tunes as Quanho and Cheo. Our experimental results show that the average highest classification and identification accuracy are 99.92% and 97.67%, respectivel.
Metrics
References
[1] Cunningham, Padraig, and S. J. Delany, “k-Nearest neighbor classifiers,” Multiple Classifier Systems, vol. 34, no. 8, pp. 1-17, 2007.
[2] Y. Sazaki, A. Aramadhan, “Rock genre classification using k-nearest neighbor,” Proceeding of The 1st International Conference on Computer Science and Engineering, pp. 81-84, 2014.
[3] Ghahramani, Zoubin. “An introduction to hidden Markov models and Bayesian networks,” International Journal of Pattern Recognition and Artificial Intelligence, vol. 15, no. 1, pp. 9-42, 2001. https://doi.org/10.1142/9789812797605 0002 DOI: https://doi.org/10.1142/S0218001401000836
[4] X. Shao, C. Xu and M. S. Kankanhalli, “Unsupervised classification of music genre using hidden Markov model,” 2004 IEEE International Conference on Multimedia and Expo (ICME) (IEEE Cat. No.04TH8763), 2004, pp. 2023-2026 vol.3. Doi: 10.1109/ICME.2004.1394661 DOI: https://doi.org/10.1109/ICME.2004.1394661
[5] J. Reed and C.H. Lee. “A study on music genre classification based on universal acoustic models,” In Proceedings of the International Conference on Music Information Retrieval, pages 89-94, 2006.
[6] Ba˘gcı, Ula¸s, and Engin Erzin. “Boosting classifiers for music genre classification,” International Symposium on Computer and Information Sciences. Springer Berlin Heidelberg, 2005. DOI: https://doi.org/10.1007/11569596_60
[7] Christopher M. Bishop, “Pattern Recognition and Machine Learning”, Springer, 2013.
[8] K. Markov and T. Matsui, “Music genre and emotion recognition using gaussian processes,” in IEEE Access, vol. 2, pp. 688-697, 2014. Doi: 10.1109/ACCESS.2014.2333095 DOI: https://doi.org/10.1109/ACCESS.2014.2333095
[9] A. Meng and J. Shawe-Taylor, “An investigation of feature models for music genre classification using the support vector classifier,” In Proceedings of the International Conference on Music Information Retrieval, pages 604-609, 2005.
[10] M. Li and R. Sleep, “Genre classification via an LZ78-based string kernel,” in Proceedings of the International Conference on Music Information Retrieval, pages 252-259, 2005. DOI: https://doi.org/10.1007/11527503_80
[11] Y. Wang, X. Lin, L. Wu, W. Zhang, “Effective multi-query expansions: Collaborative deep networks for robust landmark retrieval,” IEEE Transactions on Image Processing, vol. 26, pp. 1393–1404, 2017. DOI: https://doi.org/10.1109/TIP.2017.2655449
[12] L. Wu, Y. Wang, L. Shao, “Cycle-consistent deep generative hashing for cross-modal retrieval,” IEEE Transactions on Image Processing 28, pp. 1602–1612, 2019. DOI: https://doi.org/10.1109/TIP.2018.2878970
[13] A. Krizhevsky, I. Sutskever and G.E. Hinton, “ImageNet classification with deep convolutional neural networks,” in Advances in Neural Information Processing Systems, vol. 25, no. 2, pp. 1097–110, 2012.
[14] B. Matityaho and M. Furst, “Neural network based model for classification of music type,” in Proceedings of the Convention of Electrical and Electronics Engineers in Israel, pp. 1–5, March, 1995.
[15] [Online]. Available: http://ismir.net/conferences/
[16] G. Tzanetakis and P. Cook, “Musical genre classification of audio signals,” IEEE Transactions on Speech and Audio Processing, vol. 10, no. 5, pp. 293-302, 2002. DOI: https://doi.org/10.1109/TSA.2002.800560
[17] Z. Fu, G. Lu, K. M. Ting, and D. Zhang, “A survey of audio-based music classification and annotation,” in IEEE Transactions on Multimedia, vol.13, no.2, pp. 303-319, April 2011. DOI: https://doi.org/10.1109/TMM.2010.2098858
[18] Knees, Peter, and Markus Schedl, “A survey of music similarity and recommendation from music context data.” ACM Transactions on Multimedia Computing, Communications, and Applications (TOMM), vol.10, no.1, pp. 1-21, 2013. DOI: https://doi.org/10.1145/2542205.2542206
[19] Corrˆea, C. D´ebora, and Francisco Ap Rodrigues, “A survey on symbolic data-based music genre classification,” Expert Systems with Applications, vol. 60, pp. 190-210, 2016. DOI: https://doi.org/10.1016/j.eswa.2016.04.008
[20] B.L. Sturm, “A survey of evaluation in music genre recognition,” International Workshop on Adaptive Multimedia Retrieval, Springer, pp 29–66, 2012. DOI: https://doi.org/10.1007/978-3-319-12093-5_2
[21] Aucouturier, Jean-Julien, and F. Pachet, “Representing musical genre: A state of the art.” Journal of New Music Research, vol. 32, no.1, pp. 83-93, 2003. DOI: https://doi.org/10.1076/jnmr.32.1.83.16801
[22] A. Anglade, Q. Mary, R. Ramirez, and S. Dixon, “Genre classification using harmony rules induced from automatic chord transcriptions,” In Proceedings of the International Conference on Music Information Retrieval, pages 669-674, 2009.
[23] Y. Panagakis, C. Kotropoulos, G. R. Arce, “Music genre classification using locality preserving non-negative tensor factorization and sparse representations,” in Proc. of the 10th International Society for Music Information Retrieval Conference, ISMIR, pp 249–254, 2009. DOI: https://doi.org/10.1109/ICASSP.2010.5495984
[24] Ghosal, Deepanway, and Maheshkumar H. Kolekar, “Music genre recognition using deep neural networks and transfer learning ,” Interspeech, 2018. DOI: https://doi.org/10.21437/Interspeech.2018-2045
[25] Bahuleyan, Hareesh, “Music genre classification using machine learning techniques.” arXiv preprint arXiv:1804.01149 (2018).
[26] Oramas, Sergio, et al., “Multimodal deep learning for music genre classification,” Transactions of the International Society for Music Information Retrieval, vol. 1, no. 1, pp. 4-21, 2018. DOI: https://doi.org/10.5334/tismir.10
[27] Elbir, Ahmet, and Nizamettin Aydin, “Music genre classification and music recommendation by using deep learning,” Electronics Letters, vol. 56, no.12, pp. 627-629, 2020. DOI: https://doi.org/10.1049/el.2019.4202
[28] S. Vishnupriya, and K. Meenakshi, “Automatic music genre classification using convolution neural network,” International Conference on Computer Communication and Informatics (ICCCI), IEEE, 2018. DOI: https://doi.org/10.1109/ICCCI.2018.8441340
[29] [Online]. Available: https://challenge.zalo.ai/
[30] Quang H. Nguyen, Trang T. T. Do, Thanh B. Chu, Loan V. Trinh, Dung H. Nguyen, Cuong V. Phan, Tuan A. Phan, Dung V. Doan, Hung N. Pham, Binh P. Nguyen, and Matthew C. H. Chua, “Music genre classification using residual attention network,” 2019 International Conference on System Science and Engineering (ICSSE 2019), Dong Hoi, Vietnam, July 20–21, 2019, pp 115-119. DOI: https://doi.org/10.1109/ICSSE.2019.8823100
[31] Chu Ba Thanh, Trinh Van Loan, Nguyen Hong Quang, “Automatic identification of some Vietnamese folk songs,” in Proceedings of the 19th National Symposium of Selected ICT Problems, Ha Noi, 2016. pages 92-97.
[32] Chu Ba Thanh, Trinh Van Loan, Nguyen Hong Quang, “GMM for automatic identification of some Quanho Bac Ninh folk songs,” In Proceedings of Fundamental and Applied IT Research (FAIR), Da Nang, 2017. pages 416-421.
[33] Chu Ba Thanh, Trinh Van Loan, Nguyen Hong Quang, “Classification and identification of Cheo and Quanho Bac Ninh folk songs,” in Proceedings of Fundamental and Applied IT Research (FAIR), Ha Noi, 2018. pages 395-403.
[34] Thanh Chu Ba, Trinh Van Loan, and Nguyen Hong Quang, “Some new results on automatic identification of Vietnamese folk songs Cheo and Quanho,” Journal of Computer Science and Cybernetics, vol. 36, no.4, pp. 325-345, 2020. DOI: https://doi.org/10.15625/1813-9663/36/4/14424
[35] Librosa, “Python package for music and audio analysis,” https://librosa.github.io/librosa
[36] [Online]. Available: https://en.wikipedia.org/wiki/Tonnetz
[37] Jevti´c, Filip D., and Rade T.ˇZivaljevi´c. “Generalized Tonnetz and discrete Abel-Jacobi map,” arXiv preprint arXiv:2002.09184 (2020). DOI: https://doi.org/10.12775/TMNA.2020.049
[38] Jiang, Dan-Ning, et al., “Music type classification by spectral contrast feature,” Proceedings IEEE International Conference on Multimedia and Expo, vol. 1, 2002.
[39] [Online]. Available: https://handwiki.org/wiki/Chroma feature
[40] Cho, Taemin, and Juan P. Bello, “On the relative importance of individual components of chord recognition systems,” IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 22, no. 2, pp. 477-492, 2013. DOI: https://doi.org/10.1109/TASLP.2013.2295926
[41] Mauch, Matthias, and Simon Dixon, “Simultaneous estimation of chords and musical context from audio,” IEEE Transactions on Audio, Speech, and Language Processing, vol. 18, no.6, pp. 1280-1289, 2009. DOI: https://doi.org/10.1109/TASL.2009.2032947
[42] Fujishima, Takuya, “Real-time chord recognition of musical sound: A system using common lisp music,” Proc. ICMC, Oct. , 1999, pp. 464-467.
[43] Jiang, Nanzhu, et al., “Analyzing chroma feature types for automated chord recognition,” Audio Engineering Society Conference: 42nd International Conference: Semantic Audio, Audio Engineering Society, 2011.
[44] Ellis, Daniel PW, and Graham E. Poliner, “Identifyingcover songs’ with chroma features and dynamic programming beat tracking,” 2007 IEEE International Conference on Acoustics, Speech and Signal Processing-ICASSP’07, vol. 4. IEEE, 2007. DOI: https://doi.org/10.1109/ICASSP.2007.367348
[45] Serra, Joan, et al., “Chroma binary similarity and local alignment applied to cover song identification,” IEEE Transactions on Audio, Speech, and Language Processing, vol. 16, no. 6, pp. 1138-1151, 2008. DOI: https://doi.org/10.1109/TASL.2008.924595
[46] M¨uller, Meinard, Frank Kurth, and Michael Clausen, “Audio matching via chroma-based statistical features,” ISMIR, vol. 2005, 2005.
[47] Kurth, Frank, and Meinard Muller, “Efficient index-based audio matching,” IEEE Transactions on Audio, Speech, and Language Processing, vol. 16, no. 2, pp. 382-395, 2008. DOI: https://doi.org/10.1109/TASL.2007.911552
[48] M¨uller, Meinard, “Music Synchronization,” in Fundamentals of Music Processing, Chapter 3, pp. 115-166. Springer, 2015. ISBN 978-3-319-21944-8. DOI: https://doi.org/10.1007/978-3-319-21945-5_3
[49] Yu, Yi, et al., “Combining multi-probe histogram and order-statistics based lsh for scalable audio content retrieval,” Proceedings of the 18th ACM International Conference on Multimedia, 2010. https://doi.org/10.1145/1873951.1874004 DOI: https://doi.org/10.1145/1873951.1874004
[50] Yu, Yi, et al., “Local summarization and multi-level LSH for retrieving multi-variant audio tracks,” Proceedings of the 17th ACM International Conference on Multimedia, 2009.https://doi.org/10.1145/1631272.1631320 DOI: https://doi.org/10.1145/1631272.1631320
[51] R. Girshick, J. Donahue, T. Darrell, and J. Malik, “Rich feature hierarchies for accurate object detection and semantic segmentation,” in Proceedings of the IEEE Conference On Computer Vision and Pattern Recognition, 2014, pp. 580-587. DOI: https://doi.org/10.1109/CVPR.2014.81
[52] Ross Girshick, Microsoft Research, “Fast R-NN,” Proceedings of the IEEE International Conference on Computer Vision (ICCV), 2015, pp. 1440-1448. DOI: https://doi.org/10.1109/ICCV.2015.169
[53] Ren, Shaoqing, Kaiming He, Ross Girshick, and Jian Sun, “Faster R-CNN: towards real-time object detection with region proposal networks,” IEEE Transactions on Pattern Analysis and Machine Intelligence, 39, no. 6, pp. 1137-1149, 2016. DOI: https://doi.org/10.1109/TPAMI.2016.2577031
[54] D. Rumelhart, G. Hinton, and R. Williams, “Learning representations by back-propagating errors,” Nature, vol. 323, pp. 533–536, 1986. https://doi.org/10.1038/323533a0 DOI: https://doi.org/10.1038/323533a0
[55] Hochreiter, Sepp, and J¨urgen Schmidhuber. “Long short-term memory,” Neural Computation, vol. 9, no. 8, pp. 1735-1780, 1997. DOI: https://doi.org/10.1162/neco.1997.9.8.1735
[56] Keren, Gil, and Bj¨orn Schuller, “Convolutional RNN: An enhanced model for extracting features from sequential data,” in 2016 International Joint Conference on Neural Networks (IJCNN), IEEE, 2016, pp. 3412-3419. DOI: https://doi.org/10.1109/IJCNN.2016.7727636
[57] Choi, Keunwoo, Gy¨orgy Fazekas, Mark Sandler, and Kyunghyun Cho, “Convolutional recurrent neural networks for music classification,” in 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), IEEE, 2017, pp. 2392-2396. DOI: https://doi.org/10.1109/ICASSP.2017.7952585
[58] Fukushima, Kunihiko, “A self-organizing neural network model for a mechanism of pattern recognition unaffected by shift in position,” Biol. Cybern, vol. 36, pp. 193-202, 1980. DOI: https://doi.org/10.1007/BF00344251
[59] LeCun, Yann, et al., “Gradient-based learning applied to document recognition,” Proceedings of the IEEE, vol. 86, no. 11, pp. 2278-2324, 1998. DOI: https://doi.org/10.1109/5.726791
[60] Ciregan, Dan, Ueli Meier, and J¨urgen Schmidhuber, “Multi-column deep neural networks for image classification,” 2012 IEEE Conference on Computer Vision and Pattern Recognition, IEEE, 2012. DOI: https://doi.org/10.1109/CVPR.2012.6248110
[61] Cire¸san, Dan, and Ueli Meier, “Multi-column deep neural networks for offline handwritten Chinese character classification,” 2015 International Joint Conference On Neural Networks (IJCNN), IEEE, 2015. DOI: https://doi.org/10.1109/IJCNN.2015.7280516
[62] Keunwoo Choi, George Fazekas, Mark Sandler, and Jeonghee Kim, “Auralisation of deep convolutional neural networks: Listening to learned features,” Proceedings of the 16th International Society for Music Information Retrieval Conference, ISMIR, 2015, pp. 26–30.
[63] Paulo Chiliguano and Gyorgy Fazekas, “Hybrid music recommender using content-based and social information,” 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), IEEE, 2016, pp. 2618–2622. DOI: https://doi.org/10.1109/ICASSP.2016.7472151
[64] Matocha, Mateusz, and S. K. Zieli´nski, “Music genre recognition using convolutional neural networks,” Advances in Computer Science Research, vol. 14, pp. 125-142, 2018.
[65] Pelchat, Nikki, and Craig M. Gelowitz, “Neural network music genre classification,” Canadian Journal of Electrical and Computer Engineering, vol. 43, no. 3, pp. 170-173, 2020. DOI: https://doi.org/10.1109/CJECE.2020.2970144
[66] Dao Thi Le Thuy, “Emotional recognition for Vietnamese spoken,” Doctoral Thesis - Hanoi University of Science and Technology, 2019.
[67] [Online]. Available: https://nttuan8.com/bai-5-gioi-thieu-ve-xu-ly-anh/
[68] Clevert, Djork-Arn´e, Thomas Unterthiner, and Sepp Hochreiter, “Fast and accurate deep network learning by exponential linear units (elus),” arXiv preprint arXiv:1511.07289 (2015).
[69] Hinton, E. Geoffrey, “Rectified linear units improve restricted boltzmann machines vinod nair,” ICML 2010, 2010.
[70] Sergey Ioffe, Christian Szegedy, “Batch normalization: Accelerating deep network training by reducing internal covariate shift,” in ICML’15 Proceedings of the 32nd International Conference on Machine Learning, France, July 06 – 11, 2015, pp. 448-456.
[71] [Online]. Available: https://www.programmersought.com/article/41731094913/
[72] [Online]. Available: https://medium.com/techspace-usict/normalization-techniques-in-deepneural- networks-9121bf100d8
[73] [Online]. Available: https://github.com/lutzroeder/netron
[74] Dao Thi Le Thuy, Loan Trinh Van, and Quang Nguyen Hong, “Deep convolutional neural networks for emotion recognition of Vietnamese,” International Journal of Machine Learning and Computing, vol. 10, no. 5, 2020. DOI: https://doi.org/10.18178/ijmlc.2020.10.5.992
[75] Sherstinsky, Alex, “Fundamentals of recurrent neural network (RNN) and long shortterm memory (LSTM) network,” Physica D: Nonlinear Phenomena, vol. 404, 2020. https://doi.org/10.1016/j.physd.2019.132306 DOI: https://doi.org/10.1016/j.physd.2019.132306
[76] Hochreiter, Sepp, “Untersuchungen zu dynamischen neuronalen Netzen,” Diploma, Technische Universit¨at M¨unchen, vol. 91, no. 1, 1991.
[77] Bengio, Yoshua, Patrice Simard, and Paolo Frasconi, “Learning long-term dependencies with gradient descent is difficult,” IEEE Transactions on Neural Networks, vol. 5, no. 2, pp. 157-166, 1994. DOI: https://doi.org/10.1109/72.279181
[78] Mikolov, Tom´aˇs, et al., “Recurrent neural network based language model,” Eleventh Annual Conference of the International Speech Communication Association, 2010. DOI: https://doi.org/10.21437/Interspeech.2010-343
[79] Mikolov, Tom´aˇs, et al., “Extensions of recurrent neural network language model,” 2011 IEEE International Conference On Acoustics, Speech And Signal Processing (ICASSP), IEEE, 2011. DOI: https://doi.org/10.1109/ICASSP.2011.5947611
[80] Sutskever, Ilya, James Martens, and Geoffrey E. Hinton, “Generating text with recurrent neural networks,” ICML, 2011.
[81] Sutskever, Ilya, Oriol Vinyals, and Quoc V. Le, “Sequence to sequence learning with neural networks,” Advances in Neural Information Processing Systems, 2014.
[82] Graves, Alex, and Navdeep Jaitly, “Towards end-to-end speech recognition with recurrent neural networks,” Proceedings of the 31st International Conference on Machine Learning, PMLR, vol. 32, no. 2, pp. 1764-1772, 2014.
[83] Duyu Tang, Bing Qin, and Ting Liu, “Document modeling with gated recurrent neural network for sentiment classification,” in Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, 2015, pp. 1422–1432. DOI: https://doi.org/10.18653/v1/D15-1167
[84] Zhen Zuo, Bing Shuai, Gang Wang, Xiao Liu, Xingxing Wang, Bing Wang, and Yushi Chen, “Convolutional recurrent neural networks: Learning spatial dependencies for image representation,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, 2015, pp. 18–26. DOI: https://doi.org/10.1109/CVPRW.2015.7301268
[85] Sigtia, Siddharth, Emmanouil Benetos, and Simon Dixon. “An end-to-end neural network for polyphonic piano music transcription,” IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 24, no. 5, pp. 927-939, 2016. DOI: https://doi.org/10.1109/TASLP.2016.2533858
[86] [Online]. Available:https://www.analyticsvidhya.com/blog/2020/09/overfitting-in-cnn-show-totreat-overfitting-in-convolutional-neural-networks
[87] S. Nainan, V. Kulkarni, “Enhancement in speaker recognition for optimized speech features using GMM, SVM and 1-D CNN,” Int J Speech Technol, vol. 24, pp. 809–822, 2021. https://doi.org/10.1007/s10772-020-09771-2 DOI: https://doi.org/10.1007/s10772-020-09771-2
[88] Dao Thi Le Thuy, Trinh Van Loan, Nguyen Hong Quang, “GMM for emotion recognition of Vietnamese,” Journal of Computer Science and Cybernetics, vol.33, no.3, pp.229-246, 2017. DOI: https://doi.org/10.15625/1813-9663/33/3/11017
[89] Hung Pham Ngoc, Trinh Van Loan, and Nguyen Hong Quang, “Automatic identification of vietnamese dialects,” Journal of Computer Science and Cybernetics, vol. 32, no. 1, pp. 19-30, 2016. DOI: https://doi.org/10.15625/1813-9663/32/1/7905
Downloads
Published
How to Cite
Issue
Section
License
1. We hereby assign copyright of our article (the Work) in all forms of media, whether now known or hereafter developed, to the Journal of Computer Science and Cybernetics. We understand that the Journal of Computer Science and Cybernetics will act on my/our behalf to publish, reproduce, distribute and transmit the Work.2. This assignment of copyright to the Journal of Computer Science and Cybernetics is done so on the understanding that permission from the Journal of Computer Science and Cybernetics is not required for me/us to reproduce, republish or distribute copies of the Work in whole or in part. We will ensure that all such copies carry a notice of copyright ownership and reference to the original journal publication.
3. We warrant that the Work is our results and has not been published before in its current or a substantially similar form and is not under consideration for another publication, does not contain any unlawful statements and does not infringe any existing copyright.
4. We also warrant that We have obtained the necessary permission from the copyright holder/s to reproduce in the article any materials including tables, diagrams or photographs not owned by me/us.