Evaluating Listening Performance for COVID-19 Detection by Clinicians and Machine Learning: Comparative Study

BACKGROUND To date, performance comparisons between men and machines have been performed in many health domains. Yet, machine learning models and human performance comparisons in audio-based respiratory diagnosis remain largely unexplored. OBJECTIVE The primary objective of this study is to compare human clinicians and a machine learning model in predicting COVID-19 from respiratory sound recordings. METHODS In this study, we compare human clinicians and a machine learning model in predicting COVID-19 from respiratory sound recordings. Prediction performance on 24 audio samples (12 tested positive) made by 36 clinicians with experience in treating COVID-19 or other respiratory illnesses is compared with predictions made by a machine learning model trained on 1,162 samples. Each sample consists of voice, cough, and breathing sound recordings from one subject, and the length of each sample is around 20 seconds. We also investigated whether combining the predictions of the model and human experts could further enhance the performance, in terms of both accuracy and confidence. RESULTS The machine learning model outperformed the clinicians, yielding a sensitivity of 0.75 and a specificity of 0.83, while the best performance achieved by the clinician was 0.67 in terms of sensitivity and 0.75 in terms of specificity. Integrating clinicians' and model's predictions, however, could enhance performance further, achieving a sensitivity of 0.83 and a specificity of 0.92. CONCLUSIONS Our findings suggest that the clinicians and the machine learning model could make better clinical decisions via a cooperative approach and achieve higher confidence in audio-based respiratory diagnosis.

[1]  A. C. Villa-Parra,et al.  Towards Multimodal Equipment to Help in the Diagnosis of COVID-19 Using Machine Learning Algorithms , 2022, Sensors.

[2]  Lorenzo A. Rossi,et al.  Prospective Comparison of Medical Oncologists and a Machine Learning Model to Predict 3-Month Mortality in Patients With Metastatic Solid Tumors , 2022, JAMA network open.

[3]  Cecilia Mascolo,et al.  Sounds of COVID-19: exploring realistic performance of audio-based digital testing , 2021, npj Digital Medicine.

[4]  Björn W. Schuller,et al.  CovNet: A Transfer Learning Framework for Automatic COVID-19 Detection From Crowd-Sourced Cough Sounds , 2022, Frontiers in Digital Health.

[5]  Florian B. Pokorny,et al.  Fitbeat: COVID-19 estimation based on wristband heart rate using a contrastive convolutional auto-encoder , 2021, Pattern Recognition.

[6]  Thomas Niesler,et al.  COVID-19 detection in cough, breath and speech using deep transfer learning and bottleneck features , 2021, Computers in Biology and Medicine.

[7]  Alice Baird,et al.  End-to-end convolutional neural network enables COVID-19 detection from breath and cough audio: a pilot study , 2021, BMJ Innovations.

[8]  Tao Wang,et al.  Pre-symptomatic detection of COVID-19 from smartwatch data , 2020, Nature Biomedical Engineering.

[9]  Jianjiang Feng,et al.  Development and evaluation of an artificial intelligence system for COVID-19 diagnosis , 2020, Nature Communications.

[10]  Daniel A. Hashimoto,et al.  Current applications of artificial intelligence for intraoperative decision support in surgery , 2020, Frontiers of Medicine.

[11]  Cecilia Mascolo,et al.  Exploring Automatic Diagnosis of COVID-19 from Crowdsourced Respiratory Sound Data , 2020, KDD.

[12]  David A. Drew,et al.  Real-time tracking of self-reported symptoms to predict potential COVID-19 , 2020, Nature Medicine.

[13]  P. M. van de Ven,et al.  The diagnostic accuracy of lung auscultation in adult patients with acute pulmonary pathologies: a meta-analysis , 2020, Scientific Reports.

[14]  A. Barabasi,et al.  Network medicine framework for identifying drug-repurposing opportunities for COVID-19 , 2020, Proceedings of the National Academy of Sciences.

[15]  A. Bohadana,et al.  Influence of observer preferences and auscultatory skill on the choice of terms to describe lung sounds: a survey of staff physicians, residents and medical students , 2020, BMJ Open Respiratory Research.

[16]  João J. Ferreira,et al.  Digital transformation in the area of health: systematic review of 45 years of evolution , 2019, Health and Technology.

[17]  S. Han,et al.  Keratinocytic Skin Cancer Detection on the Face Using Region-Based Convolutional Neural Network. , 2019, JAMA dermatology.

[18]  A. Bręborowicz,et al.  The accuracy of lung auscultation in the practice of physicians and medical students , 2019, PloS one.

[19]  R. Roseby,et al.  Digital stethoscopes compared to standard auscultation for detecting abnormal paediatric breath sounds , 2017, European Journal of Pediatrics.

[20]  Jérémie F. Cohen,et al.  STARD 2015: An Updated List of Essential Items for Reporting Diagnostic Accuracy Studies. , 2015, Clinical chemistry.

[21]  Yi-Ping Phoebe Chen,et al.  Image based computer aided diagnosis system for cancer detection , 2015, Expert Syst. Appl..

[22]  M. Sarkar,et al.  Auscultation of the respiratory system , 2015, Annals of thoracic medicine.

[23]  J. Fleiss Measuring nominal scale agreement among many raters. , 1971 .