Deep Learning for Chest Radiograph Diagnosis in the Emergency Department

被引:125
作者
Hwang, Eui Jin [1 ]
Nam, Ju Gang [1 ]
Lim, Woo Hyeon [1 ]
Park, Sae Jin [1 ]
Jeong, Yun Soo [1 ]
Kang, Ji Hee [1 ]
Hong, Eun Kyoung [1 ]
Kim, Taek Min [1 ]
Goo, Jin Mo [1 ]
Park, Sunggyun [2 ]
Kim, Ki Hwan [2 ]
Park, Chang Min [1 ]
机构
[1] Seoul Natl Univ, Coll Med, Dept Radiol, 101 Daehak Ro, Seoul 03080, South Korea
[2] Lunit, Seoul, South Korea
关键词
DIABETIC-RETINOPATHY; VALIDATION;
D O I
10.1148/radiol.2019191225
中图分类号
R8 [特种医学]; R445 [影像诊断学];
学科分类号
1002 ; 100207 ; 1009 ;
摘要
Background: The performance of a deep learning (DL) algorithm should be validated in actual clinical situations, before its clinical implementation. Purpose: To evaluate the performance of a DL algorithm for identifying chest radiographs with clinically relevant abnormalities in the emergency department (ED) setting. Materials and Methods: This single-center retrospective study included consecutive patients who visited the ED and underwent initial chest radiography between January 1 and March 31, 2017. Chest radiographs were analyzed with a commercially available DL algorithm.The performance of the algorithm was evaluated by determining the area under the receiver operating characteristic curve (AUC), sensitivity, and specificity at predefined operating cutoffs (high-sensitivity and high-specificity cutoffs). The sensitivities and specificities of the algorithm were compared with those of the on-call radiology residents who interpreted the chest radiographs in the actual practice by using McNemar tests. If there were discordant findings between the algorithm and resident, the residents reinterpreted the chest radiographs by using the algorithm's output. Results: A total of 1135 patients (mean age, 53 years +/- 18; 582 men) were evaluated. In the identification of abnormal chest radiographs,the algorithm showed an AUC of 0.95 (95% confidence interval [CI]: 0.93, 0.96), a sensitivity of 88.7% (227 of 256 radiographs; 95% CI: 84.1%, 92.3%), and a specificity of 69.6% (612 of 879 radiographs; 95% CI: 66.5%, 72.7%) at the high sensitivity cutoff and a sensitivity of 81.6% (209 of 256 radiographs; 95% CI: 76.3%, 86.2%) and specificity of 90.3% (794 of 879 radiographs; 95% CI: 88.2%, 92.2%) at the high-specificity cutoff. Radiology residents showed lower sensitivity (65.6% [168 of 256 radiographs; 95% CI: 59.5%, 71.4%], P < .001) and higher specificity (98.1% [862 of 879 radiographs; 95% CI: 96.9%,98.9%], P < .001) compared with the algorithm. After reinterpretation of chest radiographs with use of the algorithm's outputs,the sensitivity of the residents improved (73.4% [188 of 256 radiographs; 95% CI: 68.0%, 78.8%], P = .003), whereas specificity was reduced (94.3% [829 of 879 radiographs; 95% CI: 92.8%, 95.8%], P < .001). Conclusion: A deep learning algorithm used with emergency department chest radiographs showed diagnostic performance for identifying clinically relevant abnormalities and helped improve the sensitivity of radiology residents' evaluation. Published under a CC BY 4.0 license.
引用
收藏
页码:573 / 580
页数:8
相关论文
共 23 条
[1]   Accuracy of chest radiograph interpretation by emergency physicians [J].
Al Aseri Z. .
Emergency Radiology, 2009, 16 (2) :111-114
[2]   Diagnostic Assessment of Deep Learning Algorithms for Detection of Lymph Node Metastases in Women With Breast Cancer [J].
Bejnordi, Babak Ehteshami ;
Veta, Mitko ;
van Diest, Paul Johannes ;
van Ginneken, Bram ;
Karssemeijer, Nico ;
Litjens, Geert ;
van der Laak, Jeroen A. W. M. .
JAMA-JOURNAL OF THE AMERICAN MEDICAL ASSOCIATION, 2017, 318 (22) :2199-2210
[3]   Increasing Utilization of Chest Imaging in US Emergency Departments From 1994 to 2015 [J].
Chung, Jonathan H. ;
Duszak, Richard, Jr. ;
Hemingway, Jennifer ;
Hughes, Danny R. ;
Rosenkrantz, Andrew B. .
JOURNAL OF THE AMERICAN COLLEGE OF RADIOLOGY, 2019, 16 (05) :674-682
[4]   ACR Approprteness Criteria Blunt Chest Trauma [J].
Chung, Jonathan H. ;
Cox, Christian W. ;
Mohammed, Tan-Lucien H. ;
Kirsch, Jacobo ;
Brown, Kathleen ;
Dyer, Debra Sue ;
Ginsburg, Mark E. ;
Heitkamp, Darel E. ;
Kanne, Jeffrey P. ;
Kazerooni, Ella A. ;
Ketai, Loren H. ;
Ravenel, James G. ;
Saleh, Anthony G. ;
Shah, Rakesh D. ;
Steiner, Robert M. ;
Suh, Robert D. .
JOURNAL OF THE AMERICAN COLLEGE OF RADIOLOGY, 2014, 11 (04) :345-351
[5]   COMPARING THE AREAS UNDER 2 OR MORE CORRELATED RECEIVER OPERATING CHARACTERISTIC CURVES - A NONPARAMETRIC APPROACH [J].
DELONG, ER ;
DELONG, DM ;
CLARKEPEARSON, DI .
BIOMETRICS, 1988, 44 (03) :837-845
[6]   Interpretation of emergency department radiographs: A comparison of emergency medicine physicians with radiologists, residents with faculty, and film with digital display [J].
Eng, J ;
Mysko, WK ;
Weller, GER ;
Renard, R ;
Gitlin, JN ;
Bluemke, DA ;
Magid, D ;
Kelen, GD ;
Scott, WW .
AMERICAN JOURNAL OF ROENTGENOLOGY, 2000, 175 (05) :1233-1238
[7]  
Expert Panel on Thoracic Imaging:, 2018, J Am Coll Radiol, V15, pS240, DOI 10.1016/j.jacr.2018.09.012
[8]   Chest radiographs in the emergency department: is the radiologist really necessary? [J].
Gatt, ME ;
Spectre, G ;
Paltiel, O ;
Hiller, N ;
Stalnikowicz, R .
POSTGRADUATE MEDICAL JOURNAL, 2003, 79 (930) :214-217
[9]   Development and Validation of a Deep Learning Algorithm for Detection of Diabetic Retinopathy in Retinal Fundus Photographs [J].
Gulshan, Varun ;
Peng, Lily ;
Coram, Marc ;
Stumpe, Martin C. ;
Wu, Derek ;
Narayanaswamy, Arunachalam ;
Venugopalan, Subhashini ;
Widner, Kasumi ;
Madams, Tom ;
Cuadros, Jorge ;
Kim, Ramasamy ;
Raman, Rajiv ;
Nelson, Philip C. ;
Mega, Jessica L. ;
Webster, R. .
JAMA-JOURNAL OF THE AMERICAN MEDICAL ASSOCIATION, 2016, 316 (22) :2402-2410
[10]  
Heitkamp DE, 2015, J THORAC IMAG, V30, pW2, DOI 10.1097/RTI.0000000000000153