A Methodological Approach to Validate Pneumonia Encounters from Radiology Reports Using Natural Language Processing

被引:4
作者
Panny, AlokSagar [1 ]
Hegde, Harshad [1 ]
Glurich, Ingrid [1 ]
Scannapieco, Frank A. [2 ]
Vedre, Jayanth G. [3 ]
VanWormer, Jeffrey J. [4 ]
Miecznikowski, Jeffrey [5 ]
Acharya, Amit [1 ,6 ]
机构
[1] Marshfield Clin Res Inst, Ctr Oral Syst Hlth, Marshfield, WI USA
[2] SUNY Buffalo, Sch Dent Med, Dept Oral Biol, Buffalo, NY USA
[3] Marshfield Clin Hlth Syst, Dept Crit Care Med, Marshfield, WI USA
[4] Marshfield Clin Res Inst, Ctr Clin Epidemiol & Populat Hlth, Marshfield, WI USA
[5] SUNY Buffalo, Sch Publ Hlth & Hlth Profess, Dept Biostat, Buffalo, NY USA
[6] Advocate Aurora Hlth, Advocate Aurora Res Inst, Downers Grove, IL 60515 USA
基金
美国国家卫生研究院;
关键词
pneumonia; natural language processing; knowledge bases;
D O I
10.1055/a-1817-7008
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Introduction Pneumonia is caused by microbes that establish an infectious process in the lungs. The gold standard for pneumonia diagnosis is radiologist-documented pneumonia-related features in radiology notes that are captured in electronic health records in an unstructured format. Objective The study objective was to develop a methodological approach for assessing validity of a pneumonia diagnosis based on identifying presence or absence of key radiographic features in radiology reports with subsequent rendering of diagnostic decisions into a structured format. Methods A pneumonia-specific natural language processing (NLP) pipeline was strategically developed applying Clinical Text Analysis and Knowledge Extraction System (cTAKES) to validate pneumonia diagnoses following development of a pneumonia feature-specific lexicon. Radiographic reports of study-eligible subjects identified by International Classification of Diseases (ICD) codes were parsed through the NLP pipeline. Classification rules were developed to assign each pneumonia episode into one of three categories: "positive," "negative," or "not classified: requires manual review" based on tagged concepts that support or refute diagnostic codes. Results A total of 91,998 pneumonia episodes diagnosed in 65,904 patients were retrieved retrospectively. Approximately 89% (81,707/91,998) of the total pneumonia episodes were documented by 225,893 chest X-ray reports. NLP classified and validated 33% (26,800/81,707) of pneumonia episodes classified as "Pneumonia-positive," 19% as (15401/81,707) as "Pneumonia-negative," and 48% (39,209/81,707) as "episode classification pending further manual review." NLP pipeline performance metrics included accuracy (76.3%), sensitivity (88%), and specificity (75%). Conclusion The pneumonia-specific NLP pipeline exhibited good performance comparable to other pneumonia-specific NLP systems developed to date.
引用
收藏
页码:38 / 45
页数:8
相关论文
共 50 条
  • [31] Review of Natural Language Processing in Radiology
    Luo, Jack W.
    Chong, Jaron J. R.
    NEUROIMAGING CLINICS OF NORTH AMERICA, 2020, 30 (04) : 447 - +
  • [32] Automated anonymization of radiology reports: comparison of publicly available natural language processing and large language models
    Langenbach, Marcel C.
    Foldyna, Borek
    Hadzic, Ibrahim
    Langenbach, Isabel L.
    Raghu, Vineet K.
    Lu, Michael T.
    Neilan, Tomas G.
    Heemelaar, Julius C.
    EUROPEAN RADIOLOGY, 2024, : 2634 - 2641
  • [33] Natural language processing of radiology reports for identification of skeletal site-specific fractures
    Wang, Yanshan
    Mehrabi, Saeed
    Sohn, Sunghwan
    Atkinson, Elizabeth J.
    Amin, Shreyasee
    Liu, Hongfang
    BMC MEDICAL INFORMATICS AND DECISION MAKING, 2019, 19 (Suppl 3)
  • [34] The reporting quality of natural language processing studies: systematic review of studies of radiology reports
    Emma M. Davidson
    Michael T. C. Poon
    Arlene Casey
    Andreas Grivas
    Daniel Duma
    Hang Dong
    Víctor Suárez-Paniagua
    Claire Grover
    Richard Tobin
    Heather Whalley
    Honghan Wu
    Beatrice Alex
    William Whiteley
    BMC Medical Imaging, 21
  • [35] Identification of Long Bone Fractures in Radiology Reports Using Natural Language Processing to support Healthcare Quality Improvement
    Grundmeier, Robert W.
    Masino, Aaron J.
    Casper, T. Charles
    Dean, Jonathan M.
    Bell, Jamie
    Enriquez, Rene
    Deakyne, Sara
    Chamberlain, James M.
    Alpern, Elizabeth R.
    APPLIED CLINICAL INFORMATICS, 2016, 7 (04): : 1051 - 1068
  • [36] Natural language processing of radiology reports for identification of skeletal site-specific fractures
    Yanshan Wang
    Saeed Mehrabi
    Sunghwan Sohn
    Elizabeth J. Atkinson
    Shreyasee Amin
    Hongfang Liu
    BMC Medical Informatics and Decision Making, 19
  • [37] Natural Language Processing Techniques for Extracting and Categorizing Finding Measurements in Narrative Radiology Reports
    Sevenster, M.
    Buurman, J.
    Liu, P.
    Peters, J. F.
    Chang, P. J.
    APPLIED CLINICAL INFORMATICS, 2015, 6 (03): : 600 - 610
  • [38] The reporting quality of natural language processing studies: systematic review of studies of radiology reports
    Davidson, Emma M.
    Poon, Michael T. C.
    Casey, Arlene
    Grivas, Andreas
    Duma, Daniel
    Dong, Hang
    Suarez-Paniagua, Victor
    Grover, Claire
    Tobin, Richard
    Whalley, Heather
    Wu, Honghan
    Alex, Beatrice
    Whiteley, William
    BMC MEDICAL IMAGING, 2021, 21 (01)
  • [39] Characterization of Change and Significance for Clinical Findings in Radiology Reports Through Natural Language Processing
    Saeed Hassanpour
    Graham Bay
    Curtis P. Langlotz
    Journal of Digital Imaging, 2017, 30 : 314 - 322
  • [40] Automated Extraction of BI-RADS Final Assessment Categories from Radiology Reports with Natural Language Processing
    Sippo, Dorothy A.
    Warden, Graham I.
    Andriole, Katherine P.
    Lacson, Ronilda
    Ikuta, Ichiro
    Birdwell, Robyn L.
    Khorasani, Ramin
    JOURNAL OF DIGITAL IMAGING, 2013, 26 (05) : 989 - 994