arXiv Analytics

Sign in

arXiv:2212.06040 [cs.CL]AbstractReferencesReviewsResources

Semantic Decomposition Improves Learning of Large Language Models on EHR Data

David A. Bloore, Romane Gauriau, Anna L. Decker, Jacob Oppenheim

Published 2022-11-14Version 1

Electronic health records (EHR) are widely believed to hold a profusion of actionable insights, encrypted in an irregular, semi-structured format, amidst a loud noise background. To simplify learning patterns of health and disease, medical codes in EHR can be decomposed into semantic units connected by hierarchical graphs. Building on earlier synergy between Bidirectional Encoder Representations from Transformers (BERT) and Graph Attention Networks (GAT), we present H-BERT, which ingests complete graph tree expansions of hierarchical medical codes as opposed to only ingesting the leaves and pushes patient-level labels down to each visit. This methodology significantly improves prediction of patient membership in over 500 medical diagnosis classes as measured by aggregated AUC and APS, and creates distinct representations of patients in closely related but clinically distinct phenotypes.

Comments: Extended Abstract presented at Machine Learning for Health (ML4H) symposium 2022, November 28th, 2022, New Orleans, United States & Virtual, http://www.ml4h.cc, 9 pages
Categories: cs.CL, cs.LG
Related articles: Most relevant | Search more
arXiv:2401.06088 [cs.CL] (Published 2024-01-11)
Autocompletion of Chief Complaints in the Electronic Health Records using Large Language Models
arXiv:2308.06354 [cs.CL] (Published 2023-08-11)
Large Language Models to Identify Social Determinants of Health in Electronic Health Records
arXiv:2407.00242 [cs.CL] (Published 2024-06-28)
EHRmonize: A Framework for Medical Concept Abstraction from Electronic Health Records using Large Language Models