Leveraging Medical Knowledge Graphs and Large Language Models for Enhanced Mental Disorder Information Extraction
Abstract
:1. Introduction
- Automated and enhanced information extraction accuracy: By leveraging LLMs for zero-shot learning, we automate the traditional manual information extraction process, thereby enabling faster and more accurate data handling.
- Improved data accuracy through entity linking: By employing entity linking techniques, we ensure accurate connections between textual information and entities within the knowledge graph, leading to increased data consistency and reliability.
- Expanded practical applications in medicine: The developed knowledge graph has broader applicability beyond depression, potentially encompassing other mental disorders.
2. Related Works
2.1. Medical Knowledge Graphs
2.2. Zero-Shot Information Extraction
2.3. Entity Linking
3. Proposed Method
3.1. Schema Definition and Annotation Guidelines
3.2. Information Extraction and Entity Linking
- Context-aware recognition: The model analyzes the surrounding text to determine the precise scope and types of entities. For instance, it can distinguish whether “depression” refers to a clinical condition or a general emotional state [31].
- Entity classification: The extracted information is categorized based on predefined classes (e.g., symptoms, diagnoses, and treatments) aligned with the established schema.
- Symptom Integration and Standardization: Similar or redundant symptoms are integrated and standardized to ensure consistency. This step addresses the issue of varied expressions for the same concept, converting data into a unified format.
- Triple Construction: New triples are constructed using the standardized symptoms and durations. This process generates new triples based on integrated symptoms, maintaining data consistency.
- Entity Linking: The constructed triples are linked to standardized terms, ensuring data consistency and reliability. Entity linking connects the extracted information with standardized terms, enhancing data consistency and improving interoperability among data collected from various sources.
Case Study: Entity Linking for Major Depressive Disorder
“A diagnosis of Major Depressive Disorder requires that the patient experiences profound sadness or a loss of interest or pleasure most of the time for at least two weeks”.
- (“Major Depressive Disorder”, “requires”, “patient experiences profound sadness”)
- (“Major Depressive Disorder”, “requires”, “patient experiences a loss of interest”)
- (“Major Depressive Disorder”, “requires”, “patient experiences a loss of pleasure”)
- (“Major Depressive Disorder”, “requires”, “symptoms last at least two weeks”)
- “profound sadness” is standardized to “Profound Sadness”
- “loss of interest” and “loss of pleasure” are integrated into “Loss of Interest or Pleasure”
- “symptoms last at least two weeks” is standardized to “At Least Two Weeks”
- (“Major Depressive Disorder”, “manifests as”, “Profound Sadness”)
- (“Major Depressive Disorder”, “manifests as”, “Loss of Interest or Pleasure”)
- (“Major Depressive Disorder”, “lasts”, “At Least Two Weeks”)
3.3. Guideline-Based Model
3.4. Triple Extraction and Output
3.5. Structure of the Knowledge Graph
- Subject nodes: These nodes represent specific medical conditions or pathological states. They categorize various forms of depression according to the DSM-5 criteria, including major depressive disorder, persistent depressive disorder, premenstrual dysphoric disorder, and substance/medication-induced depressive disorder. Each subject node is labeled with the name of the disorder and a unique identifier.
- Object nodes: These nodes depict the symptoms or characteristics that a subject node may exhibit. Examples include emotional or behavioral responses such as depressed mood, fatigue, and loss of interest or pleasure. Each object node includes the name of the symptom and a unique identifier.
- Interaction between nodes: Within the knowledge graph, each subject node is linked to one or more object nodes that represent the manifestation of disease traits or symptoms. For instance, the “manifests as” relationship indicates how major depressive disorder might manifest as irritability, while the “is characterized by” relationship suggests that it may be characterized by a depressed mood.
4. Experiments
- Zero-shot information extraction on healthcare datasets: This experiment evaluated the effectiveness of zero-shot information extraction on various biomedical datasets.
- Zero-shot relationship extraction at the document level: This experiment focused on extracting relationships from unstructured data at the document level using a zero-shot approach.
4.1. Datasets and Settings
4.1.1. BioCreative Datasets
- BC5-Chemical and BC5-Disease [38]: Derived from the BioCreative V chemical-disease relation corpus, these datasets focus on exploring interactions between drugs and diseases. Each dataset includes 1500 PubMed abstracts (evenly split) for training, development, and testing. We used a preprocessed version by Crichton et al., focusing on the named entity recognition (NER) of chemicals and diseases without relationship labels.
- NCBI Disease [39]: Provided by the National Center for Biotechnology Information (NCBI), this dataset includes 793 PubMed abstracts with 6,892 disease mentions linked to 790 unique disease entities. We used a preprocessed version by Crichton et al. for training, development, and testing splits.
- BC2GM [40]: Originating from the BioCreative II gene mention corpus, this dataset consists of sentences from PubMed abstracts with manually tagged genes and alternative gene entities. Our study focused on gene entity annotations using a version of the dataset separated for development by Crichton et al.
- JNLPBA [41]: Designed for applications in molecular biology, this dataset focuses on NER for entity types such as proteins, DNA, RNA, cell lines, and cell types. We focused on entity mention detection without differentiating between entity types, using the same splits as those of Crichton et al.
4.1.2. Document-Level Relationship Extraction Datasets
- DocRED [36]: This dataset was constructed using Wikipedia and Wikidata for document-level relationship extraction. It contains 9228 documents and 57,263 relationship triples, covering 96 predefined relationship types. DocRED was used to evaluate the ability to extract relationships from complex texts spanning multiple sentences.
- Re-DocRED [37]: Re-DocRED, which is an expanded version of DocRED, includes additional positive cases (11,854 documents and 70,608 relationship triples) and incorporates relationship types and scenarios that are not addressed by DocRED. This dataset is useful for research aimed at identifying diverse and in-depth relationship patterns within documents.
4.2. Experimental Results
4.2.1. Zero-Shot Information Extraction on Healthcare Datasets
4.2.2. Zero-Shot Relationship Extraction at the Document Level
5. Discussion
- Performance evaluation and interpretationThe results confirm that incorporating entity linking with LLMs significantly enhances zero-shot information extraction for medical data. The performance of our model surpassed that of traditional zero-shot approaches, demonstrating that these techniques can compete with conventional supervised learning methods. This is particularly valuable in healthcare, where data labeling is expensive and data privacy is paramount.
- Importance of entity linkingEntity linking plays a vital role in ensuring data consistency and boosting the model performance. In this study, this went beyond simple identification tasks. By significantly improving the overall data accuracy, entity linking underscores its importance in maintaining the integrity and usefulness of medical information systems.
- Methodological limitations and future directionsThis study used a limited set of datasets, which potentially affected the generalizability of the findings. Future studies should address this issue by exploring a broader range of medical datasets and incorporating a wider variety of entity types. This will help to validate and extend the applicability of the proposed method.
- Potential applications in healthcareThe constructed medical knowledge graph serves as a critical tool for the systematic analysis of complex medical data and disease states. It has the potential to be integrated into real-time patient management systems to improve both diagnosis and ongoing patient care. Furthermore, the model of synergized LLMs and knowledge graphs suggests the potential benefits of integrating LLMs with knowledge graphs [24]. This approach can leverage the NLP capabilities of LLMs to interpret complex medical data and provide more accurate disease diagnosis and treatment predictions. This holds promise for more precise analysis of the diverse manifestations of depression and the development of effective personalized treatment plans.
6. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Gutierrez, B.J.; McNeal, N.; Washington, C.; Chen, Y.; Li, L.; Sun, H.; Su, Y. Thinking about gpt-3 in-context learning for biomedical ie? think again. arXiv 2022, arXiv:2203.08410. [Google Scholar]
- Wang, Y.; Zhao, Y.; Petzold, L. Are large language models ready for healthcare? A comparative study on clinical language understanding. In Machine Learning for Healthcare Conference; PMLR: New York, NY, USA, 2023; pp. 804–823. [Google Scholar]
- Li, Q.; Wang, Y.; You, T.; Lu, Y. BioKnowPrompt: Incorporating imprecise knowledge into prompt-tuning verbalizer with biomedical text for relation extraction. Inf. Sci. 2022, 617, 346–358. [Google Scholar] [CrossRef]
- Kartchner, D.; Ramalingam, S.; Al-Hussaini, I.; Kronick, O.; Mitchell, C. Zero-Shot Information Extraction for Clinical Meta-Analysis using Large Language Models. In Proceedings of the 22nd Workshop on Biomedical Natural Language Processing and BioNLP Shared Tasks, Toronto, ON, Canada, 13 July 2023; pp. 396–405. [Google Scholar]
- Chia, Y.K.; Bing, L.; Poria, S.; Si, L. RelationPrompt: Leveraging prompts to generate synthetic data for zero-shot relation triplet extraction. arXiv 2022, arXiv:2203.09101. [Google Scholar]
- Wang, C.; Liu, X.; Chen, Z.; Hong, H.; Tang, J.; Song, D. Zero-shot information extraction as a unified text-to-triple translation. arXiv 2021, arXiv:2109.11171. [Google Scholar]
- Li, J.; Jia, Z.; Zheng, Z. Semi-automatic data enhancement for document-level relation extraction with distant supervision from large language models. arXiv 2023, arXiv:2311.07314. [Google Scholar]
- Gyrard, A.; Boudaoud, K. Interdisciplinary iot and emotion knowledge graph-based recommendation system to boost mental health. Appl. Sci. 2022, 12, 9712. [Google Scholar] [CrossRef]
- Svenaeus, F. Diagnosing mental disorders and saving the normal: American Psychiatric Association, 2013. Diagnostic and statistical manual of mental disorders, American Psychiatric Publishing: Washington, DC. 991 pp., ISBN: 978-0890425558. Price: $122.70. Med. Health Care Philos. 2014, 17, 241–244. [Google Scholar] [CrossRef]
- Finlayson, S.G.; LePendu, P.; Shah, N.H. Building the graph of medicine from millions of clinical narratives. Sci. Data 2014, 1, 140032. [Google Scholar] [CrossRef] [PubMed]
- Rotmensch, M.; Halpern, Y.; Tlimat, A.; Horng, S.; Sontag, D. Learning a health knowledge graph from electronic medical records. Sci. Rep. 2017, 7, 5994. [Google Scholar] [CrossRef] [PubMed]
- Zhao, C.; Jiang, J.; Xu, Z.; Guan, Y. A study of EMR-based medical knowledge network and its applications. Comput. Methods Programs Biomed. 2017, 143, 13–23. [Google Scholar] [CrossRef] [PubMed]
- Zhang, K.; Li, K.; Ma, H.; Yue, D.; Zhuang, L. Construction of MeSH-like obstetric knowledge graph. In Proceedings of the 2018 International Conference on Cyber-Enabled Distributed Computing and Knowledge Discovery (CyberC), Zhengzhou, China, 18–20 October 2018; IEEE: New York, NY, USA, 2018; pp. 160–1608. [Google Scholar]
- He, K.; Yao, L.; Zhang, J.W.; Li, Y.; Li, C. Construction of genealogical knowledge graphs from obituaries: Multitask neural network extraction system. J. Med. Internet Res. 2021, 23, e25670. [Google Scholar] [CrossRef]
- Sun, H.; Xiao, J.; Zhu, W.; He, Y.; Zhang, S.; Xu, X.; Hou, L.; Li, J.; Ni, Y.; Xie, G.; et al. Medical knowledge graph to enhance fraud, waste, and abuse detection on claim data: Model development and performance evaluation. JMIR Med. Inform. 2020, 8, e17653. [Google Scholar] [CrossRef] [PubMed]
- Sainz, O.; de Lacalle, O.L.; Labaka, G.; Barrena, A.; Agirre, E. Label verbalization and entailment for effective zero-and few-shot relation extraction. arXiv 2021, arXiv:2109.03659. [Google Scholar]
- Sainz, O.; Gonzalez-Dios, I.; de Lacalle, O.L.; Min, B.; Agirre, E. Textual entailment for event argument extraction: Zero-and few-shot with multi-source learning. arXiv 2022, arXiv:2205.01376. [Google Scholar]
- Wei, X.; Cui, X.; Cheng, N.; Wang, X.; Zhang, X.; Huang, S.; Xie, P.; Xu, J.; Chen, Y.; Zhang, M.; et al. Zero-shot information extraction via chatting with chatgpt. arXiv 2023, arXiv:2302.10205. [Google Scholar]
- Jeblick, K.; Schachtner, B.; Dexl, J.; Mittermeier, A.; Stüber, A.T.; Topalis, J.; Weber, T.; Wesp, P.; Sabel, B.O.; Ricke, J.; et al. ChatGPT makes medicine easy to swallow: An exploratory case study on simplified radiology reports. Eur. Radiol. 2023, 34, 2817–2825. [Google Scholar] [CrossRef] [PubMed]
- Hu, D.; Liu, B.; Zhu, X.; Lu, X.; Wu, N. Zero-shot information extraction from radiological reports using ChatGPT. Int. J. Med. Inform. 2024, 183, 105321. [Google Scholar] [CrossRef] [PubMed]
- Al-Moslmi, T.; Ocaña, M.G.; Opdahl, A.L.; Veres, C. Named entity extraction for knowledge graphs: A literature overview. IEEE Access 2020, 8, 32862–32881. [Google Scholar] [CrossRef]
- Peeters, R.; Bizer, C. Using chatgpt for entity matching. In Proceedings of the European Conference on Advances in Databases and Information Systems, Barcelona, Spain, 4–7 September 2023; Springer: Cham, Switzerland, 2023; pp. 221–230. [Google Scholar]
- Pan, S.; Luo, L.; Wang, Y.; Chen, C.; Wang, J.; Wu, X. Unifying large language models and knowledge graphs: A roadmap. IEEE Trans. Knowl. Data Eng. 2024, 36, 3580–3599. [Google Scholar] [CrossRef]
- Ye, J.; Chen, X.; Xu, N.; Zu, C.; Shao, Z.; Liu, S.; Cui, Y.; Zhou, Z.; Gong, C.; Shen, Y.; et al. A comprehensive capability analysis of gpt-3 and gpt-3.5 series models. arXiv 2023, arXiv:2303.10420. [Google Scholar]
- Sainz, O.; García-Ferrero, I.; Agerri, R.; de Lacalle, O.L.; Rigau, G.; Agirre, E. Gollie: Annotation guidelines improve zero-shot information-extraction. arXiv 2023, arXiv:2310.03668. [Google Scholar]
- Wang, X.; Zhou, W.; Zu, C.; Xia, H.; Chen, T.; Zhang, Y.; Zheng, R.; Ye, J.; Zhang, Q.; Gui, T.; et al. InstructUIE: Multi-task instruction tuning for unified information extraction. arXiv 2023, arXiv:2304.08085. [Google Scholar]
- Zhang, X.; Peng, B.; Li, K.; Zhou, J.; Meng, H. Sgp-tod: Building task bots effortlessly via schema-guided llm prompting. arXiv 2023, arXiv:2305.09067. [Google Scholar]
- Wei, J.; Bosma, M.; Zhao, V.Y.; Guu, K.; Yu, A.W.; Lester, B.; Du, N.; Dai, A.M.; Le, Q.V. Finetuned language models are zero-shot learners. arXiv 2021, arXiv:2109.01652. [Google Scholar]
- Zhou, W.; Zhang, S.; Gu, Y.; Chen, M.; Poon, H. Universalner: Targeted distillation from large language models for open named entity recognition. arXiv 2023, arXiv:2308.03279. [Google Scholar]
- Chen, Y.; Jiang, H.; Liu, L.; Shi, S.; Fan, C.; Yang, M.; Xu, R. An empirical study on multiple information sources for zero-shot fine-grained entity typing. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, Virtual, 7–11 November 2021; pp. 2668–2678. [Google Scholar]
- Carta, S.; Giuliani, A.; Piano, L.; Podda, A.S.; Pompianu, L.; Tiddia, S.G. Iterative zero-shot llm prompting for knowledge graph construction. arXiv 2023, arXiv:2307.01128. [Google Scholar]
- McCusker, J. LOKE: Linked Open Knowledge Extraction for Automated Knowledge Graph Construction. arXiv 2023, arXiv:2311.09366. [Google Scholar]
- Li, P.; Sun, T.; Tang, Q.; Yan, H.; Wu, Y.; Huang, X.; Qiu, X. Codeie: Large code generation models are better few-shot information extractors. arXiv 2023, arXiv:2305.05711. [Google Scholar]
- Papaluca, A.; Krefl, D.; Rodriguez, S.M.; Lensky, A.; Suominen, H. Zero-and Few-Shots Knowledge Graph Triplet Extraction with Large Language Models. arXiv 2023, arXiv:2312.01954. [Google Scholar]
- Wu, X.; Duan, J.; Pan, Y.; Li, M. Medical knowledge graph: Data sources, construction, reasoning, and applications. Big Data Min. Anal. 2023, 6, 201–217. [Google Scholar] [CrossRef]
- Yao, Y.; Ye, D.; Li, P.; Han, X.; Lin, Y.; Liu, Z.; Liu, Z.; Huang, L.; Zhou, J.; Sun, M. DocRED: A large-scale document-level relation extraction dataset. arXiv 2019, arXiv:1906.06127. [Google Scholar]
- Tan, Q.; Xu, L.; Bing, L.; Ng, H.T.; Aljunied, S.M. Revisiting DocRED–Addressing the False Negative Problem in Relation Extraction. arXiv 2022, arXiv:2205.12696. [Google Scholar]
- Li, J.; Sun, Y.; Johnson, R.J.; Sciaky, D.; Wei, C.-H.; Leaman, R.; Davis, A.P.; Mattingly, C.J.; Wiegers, T.C.; Lu, Z. BioCreative V CDR task corpus: A resource for chemical disease relation extraction. Database 2016, 2016, baw068. [Google Scholar] [CrossRef] [PubMed]
- Doğan, R.I.; Leaman, R.; Lu, Z. NCBI disease corpus: A resource for disease name recognition and concept normalization. J. Biomed. Inform. 2014, 47, 1–10. [Google Scholar] [CrossRef] [PubMed]
- Smith, L.; Tanabe, L.K.; Ando, R.J.n.; Kuo, C.-J.; Chung, I.-F.; Hsu, C.-N.; Lin, Y.-S.; Klinger, R.; Friedrich, C.M.; Ganchev, K.; et al. Overview of BioCreative II gene mention recognition. Genome Biol. 2008, 9, S2. [Google Scholar] [CrossRef] [PubMed]
- Collier, N.; Ohta, T.; Tsuruoka, Y.; Tateisi, Y.; Kim, J.-D. Introduction to the bio-entity recognition task at JNLPBA. In Proceedings of the International Joint Workshop on Natural Language Processing in Biomedicine and Its Applications (NLPBA/BioNLP), Geneva, Switzerland, 28–29 August 2004; pp. 73–78. [Google Scholar]
- Chung, H.W.; Hou, L.; Longpre, S.; Zoph, B.; Tay, Y.; Fedus, W.; Li, Y.; Wang, X.; Dehghani, M.; Brahma, S.; et al. Scaling instruction-finetuned language models. J. Mach. Learn. Res. 2024, 25, 1–53. [Google Scholar]
- Gu, Y.; Tinn, R.; Cheng, H.; Lucas, M.; Usuyama, N.; Liu, X.; Naumann, T.; Gao, J.; Poon, H. Domain-specific language model pretraining for biomedical natural language processing. ACM Trans. Comput. Healthc. (HEALTH) 2021, 3, 1–23. [Google Scholar] [CrossRef]
- Touvron, H.; Martin, L.; Stone, K.; Albert, P.; Almahairi, A.; Babaei, Y.; Bashlykov, N.; Batra, S.; Bhargava, P.; Bhosale, S.; et al. Llama 2: Open foundation and fine-tuned chat models. arXiv 2023, arXiv:2307.09288. [Google Scholar]
Number | Subject | Object | Relation |
---|---|---|---|
1 | Major depressive disorder | Irritability | manifests as |
2 | Major depressive disorder | Depressed mood | manifests as |
3 | Major depressive disorder | Loss of interest or pleasure | manifests as |
4 | Major depressive disorder | Changes in sleep patterns | manifests as |
5 | Major depressive disorder | Decreased energy levels | manifests as |
… | … | … | … |
483 | Unspecified depressive disorder | Appetite change | lasts |
484 | Unspecified depressive disorder | Weight change | lasts |
485 | Unspecified depressive disorder | Sexual interest or desire | lasts |
486 | Unspecified depressive disorder | Sleep disturbance | includes |
487 | Unspecified depressive disorder | Psychomotor changes | includes |
Category | Dataset Name | Source | Document Count | Primary Entity Type | Entity Count |
---|---|---|---|---|---|
Biomedical Dataset | BC5-Chemical [38] | PubMed Abstracts | 1500 | Chemicals (Drugs) | N/A |
BC5-Disease [38] | PubMed Abstracts | 1500 | Diseases | N/A | |
NCBI-Disease [39] | PubMed Abstracts | 793 | Diseases | 6,892 | |
BC2GM [40] | PubMed Abstracts | N/A | Genes and Alternative Gene Products | N/A | |
JNLPBA [41] | PubMed Abstracts | N/A | Proteins, DNA, RNA, Cell Lines, Cell Types | N/A | |
Document-level Information Extraction | DocRED [36] | Wikipedia | 9228 | Relationships | 57,263 Triples |
Re-DocRED [37] | Wikipedia | 11,854 | Relationships | 70,608 Triples |
Task-Specific SOTA | Zero-Shot Evaluation | ||||
---|---|---|---|---|---|
PubMedBERT | GPT3 | GPT-3.5-Turbo | Flan-T5-XXL | Proposed Method | |
NCBI | 87.8 | 51.4 | 47.5 | 51.8 | 85.4 |
BC5-disease | 85.6 | 73.0 | 67.2 | 54.7 | 87.3 |
BC5-chem | 93.3 | 43.6 | 66.5 | 67.3 | 88.5 |
BC2GM | 84.5 | 41.1 | 47.7 | 42.4 | 67.2 |
JNLPBA | 79.1 | 48.3 | 42.0 | 38.9 | 49.7 |
Existing Large Language Model | Proposed Method | ||||
---|---|---|---|---|---|
LLaMA2-7B | Flan-T5-XXL | LLaMA2-13B | Without Entity Linking | With Entity Linking | |
DocRED | 1.2 | 4.4 | 4.0 | 7.8 | 9.8 |
RE-DocRED | 1.9 | 4.3 | 3.5 | 7.5 | 9.2 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Park, C.; Lee, H.; Jeong, O.-r. Leveraging Medical Knowledge Graphs and Large Language Models for Enhanced Mental Disorder Information Extraction. Future Internet 2024, 16, 260. https://doi.org/10.3390/fi16080260
Park C, Lee H, Jeong O-r. Leveraging Medical Knowledge Graphs and Large Language Models for Enhanced Mental Disorder Information Extraction. Future Internet. 2024; 16(8):260. https://doi.org/10.3390/fi16080260
Chicago/Turabian StylePark, Chaelim, Hayoung Lee, and Ok-ran Jeong. 2024. "Leveraging Medical Knowledge Graphs and Large Language Models for Enhanced Mental Disorder Information Extraction" Future Internet 16, no. 8: 260. https://doi.org/10.3390/fi16080260
APA StylePark, C., Lee, H., & Jeong, O. -r. (2024). Leveraging Medical Knowledge Graphs and Large Language Models for Enhanced Mental Disorder Information Extraction. Future Internet, 16(8), 260. https://doi.org/10.3390/fi16080260