TY - JOUR
T1 - EHR-KnowGen: Knowledge-enhanced multimodal learning for disease diagnosis generation
AU - Niu, Shuai
AU - Ma, Jing
AU - Bai, Liang
AU - Wang, Zhihua
AU - Guo, Li
AU - Yang, Xian
PY - 2024/2/1
Y1 - 2024/2/1
N2 - Electronic health records (EHRs) contain diverse patient information, including medical notes, clinical events, and laboratory test results. Integrating this multimodal data can improve disease diagnoses using deep learning models. However, effectively combining different modalities for diagnosis remains challenging. Previous approaches, such as attention mechanisms and contrastive learning, have attempted to address this but do not fully integrate the modalities into a unified feature space. This paper presents EHR-KnowGen, a multimodal learning model enhanced with external domain knowledge, for improved disease diagnosis generation from diverse patient information in EHRs. Unlike previous approaches, our model integrates different modalities into a unified feature space with soft prompts learning and leverages large language models (LLMs) to generate disease diagnoses. By incorporating external domain knowledge from different levels of granularity, we enhance the extraction and fusion of multimodal information, resulting in more accurate diagnosis generation. Experimental results on real-world EHR datasets demonstrate the superiority of our generative model over comparative methods, providing explainable evidence to enhance the understanding of diagnosis results.
AB - Electronic health records (EHRs) contain diverse patient information, including medical notes, clinical events, and laboratory test results. Integrating this multimodal data can improve disease diagnoses using deep learning models. However, effectively combining different modalities for diagnosis remains challenging. Previous approaches, such as attention mechanisms and contrastive learning, have attempted to address this but do not fully integrate the modalities into a unified feature space. This paper presents EHR-KnowGen, a multimodal learning model enhanced with external domain knowledge, for improved disease diagnosis generation from diverse patient information in EHRs. Unlike previous approaches, our model integrates different modalities into a unified feature space with soft prompts learning and leverages large language models (LLMs) to generate disease diagnoses. By incorporating external domain knowledge from different levels of granularity, we enhance the extraction and fusion of multimodal information, resulting in more accurate diagnosis generation. Experimental results on real-world EHR datasets demonstrate the superiority of our generative model over comparative methods, providing explainable evidence to enhance the understanding of diagnosis results.
KW - Disease diagnosis
KW - Generative large language model
KW - Knowledge enhancement
KW - Multimodal electronic health records
KW - Multimodal learning
UR - http://www.scopus.com/inward/record.url?scp=85174416655&partnerID=8YFLogxK
UR - https://www.mendeley.com/catalogue/e6760a68-f924-3958-87d0-ef291dd4f74e/
U2 - 10.1016/j.inffus.2023.102069
DO - 10.1016/j.inffus.2023.102069
M3 - Article
SN - 1566-2535
VL - 102
JO - Information Fusion
JF - Information Fusion
M1 - 102069
ER -