Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2022 Oct 11:253:109539.
doi: 10.1016/j.knosys.2022.109539. Epub 2022 Jul 28.

Deep learning and machine learning-based voice analysis for the detection of COVID-19: A proposal and comparison of architectures

Affiliations

Deep learning and machine learning-based voice analysis for the detection of COVID-19: A proposal and comparison of architectures

Giovanni Costantini et al. Knowl Based Syst. .

Abstract

Alongside the currently used nasal swab testing, the COVID-19 pandemic situation would gain noticeable advantages from low-cost tests that are available at any-time, anywhere, at a large-scale, and with real time answers. A novel approach for COVID-19 assessment is adopted here, discriminating negative subjects versus positive or recovered subjects. The scope is to identify potential discriminating features, highlight mid and short-term effects of COVID on the voice and compare two custom algorithms. A pool of 310 subjects took part in the study; recordings were collected in a low-noise, controlled setting employing three different vocal tasks. Binary classifications followed, using two different custom algorithms. The first was based on the coupling of boosting and bagging, with an AdaBoost classifier using Random Forest learners. A feature selection process was employed for the training, identifying a subset of features acting as clinically relevant biomarkers. The other approach was centered on two custom CNN architectures applied to mel-Spectrograms, with a custom knowledge-based data augmentation. Performances, evaluated on an independent test set, were comparable: Adaboost and CNN differentiated COVID-19 positive from negative with accuracies of 100% and 95% respectively, and recovered from negative individuals with accuracies of 86.1% and 75% respectively. This study highlights the possibility to identify COVID-19 positive subjects, foreseeing a tool for on-site screening, while also considering recovered subjects and the effects of COVID-19 on the voice. The two proposed novel architectures allow for the identification of biomarkers and demonstrate the ongoing relevance of traditional ML versus deep learning in speech analysis.

Keywords: 1E, Vowel /e/ vocal task; 2S, Sentence vocal task; 3C, Cough vocal task; Adaboost; CFS, Correlation-based Feature Selection; CNN, Convolutional Neural Network; COVID-19; Classification; DL, Deep Learning; Deep learning; H, Healthy control subjects; MFCC, Mel-frequency Cepstral Coefficients; ML, Machine Learning; NS, Nasal Swab; P, Positive subjects; PCR, Polymerase Chain Reaction-based molecular swabs; PvsH, Positive versus Healthy subjects comparison; R, Recovered subjects; RF, Random Forest; ROC, Receiver-Operating Curve; ReLu, Rectified Linear Unit; RvsH, Recovered versus Healthy subjects comparison; SVM, Support Vector Machine; Speech processing.

PubMed Disclaimer

Conflict of interest statement

The authors declare the following financial interests/personal relationships which may be considered as potential competing interests: Authors Giovanni Costantini, Giovanni Saggio, and Antonio Pisani are advisory members of VoiceWise S.r.l., spin-off company of University of Rome Tor Vergata (Rome, Italy) developing voice analysis solutions for diagnostic purposes; Valerio Cesarini cooperates with VoiceWise and is employed by CloudWise S.r.l., a company developing cloud data storage and software solutions.

Figures

Fig. 1
Fig. 1
Flowchart describing the complete pipeline of the Machine Learning approach based on the Adaboost classifier (exemplified for the PvsH comparison).
Fig. 2
Fig. 2
Visualization of the data augmentation techniques applied to mel-frequency spectrograms. Top left: original sample spectrogram, top right: pink noise addition, bottom left: time masking, bottom right: frequency masking. .
Fig. 3
Fig. 3
CNN1 architecture. A “Conv Block” is described in the higher box and is comprised of a convolutional layer followed by a batch normalization layer and a ReLu (Rectified Linear Unit) activation function. The number after “Conv Block” indicates the number of parallel convolutional filters/neurons in the layer. Max Pool: max pooling layer; FC: Fully connected layer: the number in the round brackets indicates the number of neurons.
Fig. 4
Fig. 4
CNN2 architecture (for the sole 3C — Cough vocal task). A “Conv Block” is described in the higher box and is comprised of a convolutional layer followed by a batch normalization layer and a ReLu (Rectified Linear Unit) activation function. The number after “Conv Block” indicates the number of parallel convolutional filters/neurons in the layer. Max Pool: max pooling layer; FC: Fully connected layer: the number in the round brackets indicates the number of neurons. .
Fig. 5
Fig. 5
ROC curves. Above: ROC curve for the PvsH (Positive versus Healthy) comparison. Below: ROC curve for the RvsH (Recovered versus Healthy) comparison. Red line refers to the 1E — vowel/e/ vocal task sub-classifier; blue line refers to the 2S — sentence vocal task sub-classifier; green line refers to the 3C — cough vocal task sub-classifier. Axes span from 0 to 1. AUC (Area Under the Curve) values are reported in the manuscript.
Fig. 6
Fig. 6
Radar plot for the PvsH-3C sub-classifier. PvsH: Positive versus Healthy; 3C: Cough vocal task. Radar plot was built on the top 20 features (as ranked by the linear wrapped SVM ranker), averaged over all the subjects, and normalized by the H class. Blue unit circle (colored area) represents the H class, red curve represents the P class. .

Similar articles

Cited by

References

    1. Suppa A., Asci F., Saggio G., Di Leo P., Zarezadeh Z., Ferrazzano G., Ruoppolo G., Berardelli A., Costantini G. Voice analysis with machine learning: one step closer to an objective diagnosis of essential tremor. Mov. Disorders: Off. J. Mov. Disorder Soc. 2021;36(6):1401–1410. doi: 10.1002/mds.28508. - DOI - PubMed
    1. J.P. Teixeira, J. Fernandes, F. Teixeira, P.O. Fernandes, Acoustic analysis of chronic laryngitis-statistical analysis of sustained speech parameters, in: 11th International Joint Conference on Biomedical Engineering Systems and Technologies, 2018, pp. 168–175.
    1. S.C. Costa, B.G.A. Neto, J.M. Fechine, M. Muppa, Short-Term Cepstral Analysis Applied to Vocal Fold Edema Detection, in: BIOSIGNALS (2), 2008, pp. 110–115.
    1. Petrovic-Lazic M., Jovanovic N., Kulic M., Babac S., Jurisic V. Acoustic and perceptual characteristics of the voice in patients with vocal polyps after surgery and voice therapy. J. Voice. 2015;29(2):241–246. - PubMed
    1. Alves M., Krüger E., Pillay B., Van Lierde K., Van der Linde J. The effect of hydration on voice quality in adults: A systematic review. J. Voice. 2019;33(1):125–e13. - PubMed

LinkOut - more resources