Specific Relation Attention-Guided Graph Neural Networks for Joint Entity and Relation Extraction in Chinese EMR
Abstract
:1. Introduction
1.1. Pipelined Extraction Model of Entity and Relation
1.2. Joint Extraction Model of Entity and Relation
1.3. Graph Neural Networks Based extraction Model
- To propose a Specific Relation Attention-guided Graph Neural Networks (SRAGNNs) model for the joint extraction of Chinese medical entities and relations. This model captures the fine-grain semantic features between relation and Chinese word characters and can extract overlapping triplets. The model uses relations to guide entity recognition.
- The model learns the correlation between various relations through attention-guided GNNs. At the same time, by combining the sentence context features encoded by the transformer, it fully captures the relevant semantics and transforms the relation extraction into a multi-label classification task.
- For the Chinese medical datasets, we use the encoding strategy of Chinese character embedding combined with word embedding to obtain the latest results on a medical evaluation dataset and a manually labeled Chinese EMR dataset, which proved to be effective.
2. Methodology
2.1. Task Description and Overview of Method
2.2. Word-Character Embedding Layer
2.3. Transformer Layer
2.4. Relation-Based Attention-Guided Graph Neural Networks Layer
2.4.1. Graph Representation of Relation
2.4.2. Adjacency Matrix Updating Mechanism in GNNs
2.5. Specific Relation Feature Vector Generation Layer
2.6. Entity Decoding Layer
3. Experiments
3.1. Experimental Setup
3.1.1. Datasets
3.1.2. Experimental Settings
3.1.3. Comparison Methods
- CopyRe is a seq2seq model with a copy mechanism, and uses the copy mechanism to generate triplets in a sentence in order. The model can jointly extract entities and relations from sentences of any type.
- GraphRel is an end-to-end relation extraction model. The model constructs a complete word graph for each sentence, using graph convolutional networks (GCNs) to jointly extract entities and relations.
- HRL applies a hierarchical paradigm. The entire extraction process is decomposed into a two-level reinforcement learning strategy hierarchy, which is used for relation detection and entity extraction, respectively. It first performs relation detection as a high-level reinforcement learning process, and then identifies the entity as a low-level learning process.
- ETL-Span decomposes the joint extraction task into two interrelated subtasks. Firstly, distinguishing all head entities, and then identifying the corresponding tail entities and relations.
- WDec designed a new representation scheme and uses the seq2seq model to generate the entire relation triplet.
- RSAN uses the relation-aware attention mechanism to construct a specific sentence representation for each relation, and then performs sequence labeling to extract its corresponding head and tail entities.
- BERT-JEORE uses a BERT-based parameter sharing layer to capture joint features of entities and overlapping relations, assigning entity labels to each token in a sentence, and then extracting entity-relation triplets.
3.2. Results
3.3. Ablation Study
3.4. Analysis on Overlapping Cases
4. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Chen, J.; Dai, X.; Yuan, Q.; Lu, C.; Huang, H. Towards interpretable clinical diagnosis with Bayesian network ensembles stacked on entity-aware CNNs. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 6–8 July 2020; pp. 3143–3153. [Google Scholar]
- Ren, P.; Hou, W.; Sheng, M.; Li, X. MKGB: A Medical Knowledge Graph Construction Framework Based on Data Lake and Active Learning. In International Conference on Health Information Science; Springer: Cham, Switzerland, 2021; pp. 245–253. [Google Scholar]
- Song, Y.; Cai, L.; Zhang, K.; Zan, H.; Liu, T.; Ren, X. Construction of Chinese Pediatric Medical Knowledge Graph. In Joint International Semantic Technology Conference; Springer: Singapore, 2019; pp. 213–220. [Google Scholar]
- Huang, X.; Zhang, J.; Xu, Z.; Ou, L.; Tong, J. A knowledge graph-based question answering method for the medical domain. PeerJ Comput. Sci. 2021, 7, e667. [Google Scholar] [CrossRef]
- Zelenko, D.; Aone, C.; Richardella, A. Kernel methods for relation extraction. J. Mach. Learn. Res. 2003, 3, 1083–1106. [Google Scholar]
- Chan, Y.; Roth, D. Exploiting syntactico-semantic structures for relation extraction. In Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, Portland, OR, USA, 19–24 June 2011; pp. 551–560. [Google Scholar]
- Gormley, M.; Yu, M.; Dredze, M. Improved relation extraction with feature-rich compositional embedding models. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, Lisbon, Portugal, 17–21 September 2015; pp. 1774–1784. [Google Scholar]
- Nadeau, D.; Sekine, S. A survey of named entity recognition and classification. Lingvisticae Investig. 2007, 30, 3–26. [Google Scholar] [CrossRef]
- Hasegawa, T.; Sekine, S.; Grishman, R. Discovering relations among named entities from large corpora. In Proceedings of the 42nd Annual Meeting of the Association for Computational Linguistics, Barcelona, Spain, 21–26 July 2004; pp. 415–422. [Google Scholar]
- Li, Q.; Ji, H. Incremental joint extraction of entity mentions and relations. In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics, Baltimore, MD, USA, 22–27 June 2014; pp. 402–412. [Google Scholar]
- Yu, B.; Zhang, Z.; Shu, X.; Liu, T.; Wang, Y.; Wang, B.; Li, W. Joint Extraction of Entities and Relations Based on a Novel Decomposition Strategy. In Proceedings of the ECAI 2020, Santiago de Compostela, Spain, 29 August–8 September 2020; IOS Press: Amsterdam, The Netherlands, 2020; pp. 2282–2289. [Google Scholar]
- Nayak, T.; Ng, H. Effective modeling of encoder-decoder architecture for joint entity and relation extraction. Proc. AAAI Conf. Artif. Intell. 2020, 34, 8528–8535. [Google Scholar] [CrossRef]
- Zheng, S.; Wang, F.; Bao, H.; Hao, Y.; Zhou, P.; Xu, B. Joint extraction of entities and relations based on a novel tagging scheme. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, Vancouver, Canada, 30 July–4 August 2017; pp. 1227–1236. [Google Scholar]
- Sun, C.; Gong, Y.; Wu, Y.; Gong, M.; Jiang, D.; Lan, M.; Sun, S.; Duan, N. Joint type inference on entities and relations via graph convolutional networks. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019; pp. 1361–1370. [Google Scholar]
- Hang, T.; Feng, J.; Wu, Y.; Yan, L.; Wang, Y. Joint extraction of entities and overlapping relations using source-target entity labeling. Expert Syst. Appl. 2021, 177, 114853. [Google Scholar] [CrossRef]
- Zeng, X.; Zeng, D.; He, S.; Liu, K.; Zhao, J. Extracting relational facts by an end-to-end neural model with copy mechanism. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, Melbourne, Australia, 15–20 July 2018; pp. 506–514. [Google Scholar]
- Takanobu, R.; Zhang, T.; Liu, J.; Huang, M. A hierarchical framework for relation extraction with reinforcement learning. Proc. AAAI Conf. Artif. Intell. 2020, 33, 7072–7079. [Google Scholar] [CrossRef]
- Yuan, Y.; Zhou, X.; Pan, S.; Zhu, Q.; Song, Z.; Guo, L. A Relation-Specific Attention Network for Joint Entity and Relation Extraction. In Proceedings of the International Joint Conference on Artificial Intelligence 2020, Yokohama, Japan, 11–17 July 2020; Association for the Advancement of Artificial Intelligence: Menlo Park, CA, USA, 2020; pp. 4054–4060. [Google Scholar]
- Scarselli, F.; Gori, M.; Tsoi, A.; Hagenbuchner, M.; Monfardini, G. The graph neural network model. IEEE Trans. Neural Netw. 2008, 20, 61–80. [Google Scholar] [CrossRef] [PubMed]
- Li, Y.; Zemel, R.; Brockschmidt, M.; Zemel, R. Gated Graph Sequence Neural Networks. arXiv 2016, arXiv:1511.05493. [Google Scholar]
- Henaff, M.; Bruna, J.; LeCun, Y. Deep convolutional networks on graph-structured data. arXiv 2015, arXiv:1506.05163. [Google Scholar]
- Defferrard, M.; Bresson, X.; Vandergheynst, P. Convolutional neural networks on graphs with fast localized spectral filtering. In Proceedings of the Advances in Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; Volume 29. [Google Scholar]
- Zhu, H.; Lin, Y.; Liu, Z.; Fu, J.; Chua, T.; Sun, M. Graph neural networks with generated parameters for relation extraction. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019; pp. 1331–1339. [Google Scholar]
- Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Liò, P.; Bengio, Y. Graph Attention Networks. Int. Conf. Learn. Represent. 2018, 1050, 20. [Google Scholar]
- Zhang, Y.; Guo, Z.; Lu, W. Attention guided graph convolutional networks for relation extraction. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019; pp. 241–251. [Google Scholar]
- Fu, T.; Li, P.; Ma, W. GraphRel: Modeling Text as Relational Graphs for Joint Entity and Relation Extraction. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019; pp. 1409–1418. [Google Scholar]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS2017), Long Beach, CA, USA, 4–9 December 2017; pp. 6000–6010. [Google Scholar]
- Stubbs, A.; Uzuner, Ö. Annotating risk factors for heart disease in clinical narratives for diabetic patients. J. Biomed. Inform. 2015, 58, S78–S91. [Google Scholar] [CrossRef] [PubMed]
- Guan, T.; Zan, H.; Zhou, X.; Xu, H.; Zhang, K. CMeIE: Construction and Evaluation of Chinese Medical Information Extraction Dataset. In Proceedings of the Natural Language Processing and Chinese Computing, 9th CCF International Conference, NLPCC 2020, Zhengzhou, China, 14–18 October 2020. [Google Scholar]
- Hotho, A.; Nürnberger, A.; Paaß, G. A brief survey of text mining. Ldv Forum. 2005, 20, 19–62. [Google Scholar]
Sentence Type | Sentence | Entity Pairs and Relation Triplets |
---|---|---|
EPO | 铅中毒最主要毒性表现为高血压 (The primary toxicity of lead poisoning is hypertension.) | (铅中毒, 临床表现, 高血压) (lead poisoning, clinical manifestations, hypertension) (铅中毒, 并发症, 高血压) (lead poisoning, complications, hypertension) |
SEO | 空肠弯曲杆菌是引起急性胃肠炎的主要病因, 也是最常见CBS前驱感染源 (Campylobacter jejuni is the leading cause of acute gastroenteritis and the most common source of CBS precursor infection.) | (空肠弯曲杆, 病因, 急性肠胃炎) (Campylobacter jejuni, Cause, acute gastroenteritis) (空肠弯曲杆菌, 高危因素, GBS) (Campylobacter jejuni, high-risk factors, GBS) |
Normal | 食物中毒首选氟喹诺酮类, 对大肠杆菌感染所致腹泻有一定疗效 (Fluoroquinolones are the first choice for food poisoning, which has a specific effect on diarrhea caused by Escherichia coli infection.) | (氟喹诺酮类, 药物治疗, 食物中毒) (Fluoroquinolones, drug therapy, food Poisoning) (腹泻, 临床表现, 大肠杆菌感染) (diarrhea, clinical manifestations, Escherichia coli infection) |
Datasets | Relation Types | Train Sentences | Dev Sentences | Test Sentences |
---|---|---|---|---|
CEMR | 15 | 35,146 | 4222 | 4389 |
CMeIE | 44 | 14,320 | 1790 | 1814 |
Parameter Description | Value |
---|---|
Number of transformer encoder blocks | 6 |
Layers of GAT | 3 |
Dimension of character embedding | 100 |
Dimension of word embedding | 100 |
Dimension of relational embedding | 200 |
Learning rate | 0.001 |
Batch size | 32 |
Number of epochs | 50 |
Dropout rate | 0.1 |
Model | CMeIE | CEMR | ||||
---|---|---|---|---|---|---|
Precision (%) | Recall (%) | F1 (%) | Precision (%) | Recall (%) | F1 (%) | |
CopyRe | 45.81 | 36.50 | 40.63 | 51.05 | 47.50 | 49.21 |
GraphRel | 53.26 | 50.02 | 51.59 | 56.86 | 53.52 | 55.14 |
HRL | 58.93 | 52.32 | 55.43 | 62.33 | 58.57 | 60.39 |
ETL-Span | 66.59 | 52.50 | 58.71 | 72.93 | 55.37 | 62.95 |
WDec | 68.05 | 56.37 | 61.66 | 77.04 | 61.04 | 68.11 |
RSAN | 65.02 | 61.44 | 63.18 | 73.05 | 69.80 | 71.39 |
BERT-JEORE | 66.21 | 63.31 | 64.73 | 72.53 | 74.33 | 73.42 |
SRAGNNs(#layers = 2) | 67.58 | 64.77 | 66.14 | 76.44 | 74.35 | 75.30 |
Element | Precision (%) | Recall (%) | F1 (%) |
---|---|---|---|
H | 84.49 | 85.98 | 85.23 |
T | 83.86 | 85.29 | 84.57 |
R | 88.02 | 88.80 | 88.41 |
(H, T) | 75.26 | 84.76 | 79.73 |
(H, R) | 77.28 | 80.38 | 78.80 |
(R, T) | 74.53 | 84.58 | 79.24 |
(H, R, T) | 76.44 | 74.35 | 75.30 |
Model | N = 1(%) | N = 2(%) | N = 3(%) | N = 4(%) | N ≥ 5 (%) |
---|---|---|---|---|---|
CopyRe | 59.37 | 49.33 | 43.62 | 46.19 | 25.49 |
GraphRel | 64.61 | 54.83 | 49.37 | 48.31 | 32.86 |
HRL | 70.87 | 62.21 | 56.58 | 58.93 | 39.87 |
ETL-Span | 71.83 | 68.12 | 57.64 | 69.44 | 57.73 |
WDec | 67.66 | 70.20 | 68.35 | 75.67 | 43.50 |
RSAN | 67.83 | 71.54 | 74.22 | 80.92 | 62.41 |
BERT-JEORE | 74.21 | 72.36 | 75.47 | 80.34 | 64.72 |
SRAGNNs(#layers = 2) | 76.18 | 74.53 | 77.46 | 81.40 | 66.93 |
Datasets | Train Sentences | Dev Sentences | Test Sentences |
---|---|---|---|
SRAGNNs(#layers = 2) | 76.44 | 74.35 | 75.30 |
- Word Embedding | 74.59 | 72.99 | 73.78 |
- Attention | 69.42 | 73.20 | 71.26 |
- RAGNN Layers 1 | 63.30 | 67.66 | 65.41 |
SRAGNNs (#layers = 1) | 75.81 | 73.82 | 74.80 |
SRAGNNs (#layers = 3) | 75.70 | 74.15 | 74.92 |
Sentence S1 | 脑梗死遗留尿便失禁 |
(Urinary and fecal incontinence after cerebral infarction.) | |
SRAGNNs | (脑梗死, 临床表现, 尿便失禁) |
(Cerebral infarction, clinical manifestations, Urinary and fecal incontinence) | |
Ground truth | (脑梗死, 临床表现, 尿便失禁) |
(Cerebral infarction, clinical manifestations, Urinary and fecal incontinence) | |
Sentence S2 | 铅中毒最主要毒性表现为高血压 |
(The main toxicity of lead poisoning is hypertension.) | |
SRAGNNs | (铅中毒, 临床表现, 高血压) |
(lead poisoning, clinical manifestations, hypertension) | |
Ground truth | (铅中毒, 并发症, 高血压) |
(lead poisoning, complications, hypertension) | |
(铅中毒, 临床表现, 高血压) | |
(lead poisoning, clinical manifestations, hypertension) | |
Sentence S3 | ADV肺炎可发展为闭塞性细支气管炎(B0), 导致反复喘息 |
(ADV pneumonia may progress to bronchiolitis obliterans (B0), leading to repeated wheezing.) | |
SRAGNNs | (ADV肺炎, 转化, 闭塞性细支气管炎) |
(ADV pneumonia, transformation, bronchiolitis obliterans) | |
(闭塞性细支气管炎, 同义词, BO) | |
(Bronchiolitis obliterans, synonym, BO) | |
(ADV肺炎, 临床表现, 反复喘息) | |
(ADV pneumonia, clinical manifestations, recurrent wheezing) | |
Ground truth | (ADV肺炎, 转化, 闭塞性细支气管炎) |
(ADV pneumonia, transformation, bronchiolitis obliterans) | |
(闭塞性细支气管炎, 同义词, BO) | |
(Bronchiolitis obliterans, synonym, BO) | |
(ADV肺炎, 临床表现, 反复喘息) | |
(ADV pneumonia, clinical manifestations, recurrent wheezing) |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Pang, Y.; Qin, X.; Zhang, Z. Specific Relation Attention-Guided Graph Neural Networks for Joint Entity and Relation Extraction in Chinese EMR. Appl. Sci. 2022, 12, 8493. https://doi.org/10.3390/app12178493
Pang Y, Qin X, Zhang Z. Specific Relation Attention-Guided Graph Neural Networks for Joint Entity and Relation Extraction in Chinese EMR. Applied Sciences. 2022; 12(17):8493. https://doi.org/10.3390/app12178493
Chicago/Turabian StylePang, Yali, Xiaohui Qin, and Zhichang Zhang. 2022. "Specific Relation Attention-Guided Graph Neural Networks for Joint Entity and Relation Extraction in Chinese EMR" Applied Sciences 12, no. 17: 8493. https://doi.org/10.3390/app12178493
APA StylePang, Y., Qin, X., & Zhang, Z. (2022). Specific Relation Attention-Guided Graph Neural Networks for Joint Entity and Relation Extraction in Chinese EMR. Applied Sciences, 12(17), 8493. https://doi.org/10.3390/app12178493