Applying Deep Learning Techniques to Estimate Patterns of Musical Gesture
- PMID: 33469435
- PMCID: PMC7813937
- DOI: 10.3389/fpsyg.2020.575971
Applying Deep Learning Techniques to Estimate Patterns of Musical Gesture
Abstract
Repetitive practice is one of the most important factors in improving the performance of motor skills. This paper focuses on the analysis and classification of forearm gestures in the context of violin playing. We recorded five experts and three students performing eight traditional classical violin bow-strokes: martelé, staccato, detaché, ricochet, legato, trémolo, collé, and col legno. To record inertial motion information, we utilized the Myo sensor, which reports a multidimensional time-series signal. We synchronized inertial motion recordings with audio data to extract the spatiotemporal dynamics of each gesture. Applying state-of-the-art deep neural networks, we implemented and compared different architectures where convolutional neural networks (CNN) models demonstrated recognition rates of 97.147%, 3DMultiHeaded_CNN models showed rates of 98.553%, and rates of 99.234% were demonstrated by CNN_LSTM models. The collected data (quaternion of the bowing arm of a violinist) contained sufficient information to distinguish the bowing techniques studied, and deep learning methods were capable of learning the movement patterns that distinguish these techniques. Each of the learning algorithms investigated (CNN, 3DMultiHeaded_CNN, and CNN_LSTM) produced high classification accuracies which supported the feasibility of training classifiers. The resulting classifiers may provide the foundation of a digital assistant to enhance musicians' time spent practicing alone, providing real-time feedback on the accuracy and consistency of their musical gestures in performance.
Keywords: CNN; CNN_LSTM; ConvLSTM; LSTM; bow-strokes; gesture recognition; music education; music interaction.
Copyright © 2021 Dalmazzo, Waddell and Ramírez.
Conflict of interest statement
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Figures









Similar articles
-
Bowing Gestures Classification in Violin Performance: A Machine Learning Approach.Front Psychol. 2019 Mar 4;10:344. doi: 10.3389/fpsyg.2019.00344. eCollection 2019. Front Psychol. 2019. PMID: 30886595 Free PMC article.
-
An Investigation of Deep Learning Models for EEG-Based Emotion Recognition.Front Neurosci. 2020 Dec 23;14:622759. doi: 10.3389/fnins.2020.622759. eCollection 2020. Front Neurosci. 2020. PMID: 33424547 Free PMC article.
-
Deep Learning-Based Violin Bowing Action Recognition.Sensors (Basel). 2020 Oct 9;20(20):5732. doi: 10.3390/s20205732. Sensors (Basel). 2020. PMID: 33050164 Free PMC article.
-
Hand Gesture Recognition based on Surface Electromyography using Convolutional Neural Network with Transfer Learning Method.IEEE J Biomed Health Inform. 2021 Apr;25(4):1292-1304. doi: 10.1109/JBHI.2020.3009383. Epub 2021 Apr 6. IEEE J Biomed Health Inform. 2021. PMID: 32750962
-
Gesture, Music and Computer: The Centro di Sonologia Computazionale at Padova University, a 50-Year History.Sensors (Basel). 2022 May 2;22(9):3465. doi: 10.3390/s22093465. Sensors (Basel). 2022. PMID: 35591156 Free PMC article. Review.
References
-
- Ahmed S. H., Kim D. (2016). Named data networking-based smart home. ICT Express 2, 130–134. 10.1016/j.icte.2016.08.007 - DOI
-
- Anguita D., Ghio A., Oneto L., Parra X., Reyes-Ortiz J. L. (2013). A public domain dataset for human activity recognition using smartphones, in ESANN 2013 Proceedings, 21st European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning (Bruges: ), 437–442.
-
- Caramiaux B., Bevilacqua F., Tanaka A. (2013). Beyond recognition, in CHI '13 Extended Abstracts on Human Factors in Computing Systems–CHI EA '13 (Seoul: ), 2109 10.1145/2468356.2468730 - DOI
-
- Caramiaux B., Montecchio N., Tanaka A., Bevilacqua F. (2015). Adaptive gesture recognition with variation estimation for interactive systems. ACM Trans. Interact. Intell. Syst. 4:18 10.1145/2643204 - DOI
LinkOut - more resources
Full Text Sources
Other Literature Sources