Named Entity Recognition Model Based on Feature Fusion
Abstract
:1. Introduction
2. Related Works
3. Our Methodology
3.1. Embedded Layer
3.1.1. Enhanced Word Vectors
- (1)
- Extract word vectors
- (2)
- Extract word boundary vectors
3.1.2. Albert Extracts Word Vectors
3.1.3. Vector Fusion
3.2. Coding Layer
3.3. Multiple Layers of Attention
3.4. Output Layer
4. Results and Discussion
4.1. The Data Set
4.2. Data Set Annotation and Evaluation Criteria
4.3. Comparative Experiment and Result Analysis
4.3.1. Comparative Experiment
4.3.2. Ablation Experiment
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Dinarelli, M.; Tellier, I. Improving Recurrent Neural Networks For Sequence Labelling. arXiv 2016, arXiv:1606.02555. [Google Scholar]
- Gui, T.; Ma, R.; Zhang, Q.; Zhao, L.; Jiang, Y.-G.; Huang, X. CNN-Based Chinese NER with Lexicon Rethinking. In Proceedings of the International Joint Conference on Artificial Intelligence IJCAI, Macao, China, 10–16 August 2019; pp. 4982–4988. [Google Scholar]
- Hu, D.; Wei, L. SLK-NER: Exploiting second-order lexicon knowledge for Chinese NER. arXiv 2020, arXiv:08416.2007. [Google Scholar]
- Jia, C.; Shi, Y.; Yang, Q. Entity enhanced BERT pre-training for Chinese NER. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), Online, 16–20 November 2020; pp. 6384–6396. [Google Scholar]
- Yan, S.; Chai, J.; Wu, L. Bidirectional GRU with multi-head attention for Chinese NER. In Proceedings of the 2020 IEEE 5th Information Technology and Mechatronics Engineering Conference (ITOEC), Chongqing, China, 12–14 June 2020; pp. 1160–1164. [Google Scholar]
- Kenton, J.D.M.W.C.; Toutanova, L.K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of NAACL-HLT, Minneapolis, MN, USA, 2–7 June 2019; pp. 4171–4186. [Google Scholar]
- Zhang, K.; Ren, W.; Zhang, Y. Attention-based Bi-LSTM for Chinese named entity recognition. In Proceedings of the Workshop on Chinese Lexical Semantics, Chiayi, China, 26–28 May 2018; pp. 643–652. [Google Scholar]
- Wang, W.; Chang, L.; Bin, C. ESN-NER: Entity storage network using attention mechanism for chinese NER. In Proceedings of the International Conference on Artificial Intelligence, Information Processing and Cloud Computing, New York, NY, USA, 19–21 December 2019; pp. 1–8. [Google Scholar]
- Yin, J.; Luo, S.L.; Wu, Z.T. Chinese named entity recognition with character-level BiLSTM and soft attention model. J. Beijing Inst. Technol. 2020, 29, 63–74. [Google Scholar]
- Li, P.H.; Fu, T.J.; Ma, W.Y. Why attention? Analyze BiLSTM deficiency and its remedies in the case of NER. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 29 January 2020; pp. 8236–8244. [Google Scholar]
- Zhu, Y.; Wang, G. CAN-NER: Convolutional Attention Network for Chinese Named Entity Recognition. In Proceedings of NAACL-HLT, Minneapolis, MN, USA, 2–7 June 2019; pp. 3384–3393. [Google Scholar]
- Vaswani, A.; Shazeer, N.; Parmar, N. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 1–8. [Google Scholar]
- Tran, Q.H.; MacKinlay, A.; Yepes, A.J. Named Entity Recognition with Stack Residual LSTM and Trainable Bias Decoding. In Proceedings of the Eighth International Joint Conference on Natural Language Processing, Taipei, China, 27 November 2017; pp. 566–575. [Google Scholar]
- Yu, A.W.; Dohan, D.; Luong, M.T. QANet: Combining Local Convolution with Global Self-Attention for Reading Comprehension. arXiv 2018, arXiv:1804.09541. [Google Scholar]
- Liu, Z.; Yang, M.; Wang, X. Entity recognition from clinical texts via recurrent neural network. BMC Med. Inform. Decis. Mak. 2017, 17, 67. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Fan, M.; Zhou, Q.; Zheng, T.F. Distributed representation learning for knowledge graphs with entity descriptions. Pattern Recognit. Lett. 2017, 93, 31–37. [Google Scholar] [CrossRef]
- Collins, M.; Singer, Y. Unsupervised models for named entity classification. In Proceedings of the 1999 Joint SIGDAT Conference on Empirical Methods in Natural Language Processing and Very Large Corpora, College Park, MA, USA, 21–22 June 1999. [Google Scholar]
- Safranchik, E.; Luo, S.; Bach, S. Weakly supervised sequence tagging from noisy rules. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 5570–5578. [Google Scholar]
- Rabiner, L.R. A tutorial on hidden Markov models and selected applications in speech recognition. In Proceedings of the IEEE, San Diego, CA, USA, 15–18 May 1989; pp. 257–286. [Google Scholar]
- Settles, B. Biomedical named entity recognition using conditional random fields and rich feature sets. In Proceedings of the International Joint Workshop on Natural Language Processing in Biomedicine and Its Applications (NLPBA/BioNLP), Geneva, Switzerland, 28–29 August 2004; pp. 107–110. [Google Scholar]
- Isozaki, H.; Kazawa, H. Efficient support vector classifiers for named entity recognition. In Proceedings of the COLING 2002: The 19th International Conference on Computational Linguistics, Taipei, China, 26–30 August 2002; pp. 1–7. [Google Scholar]
- Zhu, W.; Cheung, D. Lex-BERT: Enhancing BERT based NER with lexicons. arXiv 2021, arXiv:2101.00396. [Google Scholar]
- Unanue, I.J.; Borzeshi, E.Z.; Piccardi, M. Recurrent neural networks with specialized word embeddings for health-domain named-entity recognition. J. Biomed. Inform. 2017, 76, 102–109. [Google Scholar] [CrossRef] [PubMed]
- Collobert, R.; Weston, J.; Bottou, L. Natural language processing (almost) from scratch. J. Mach. Learn. Res. 2011, 12, 2493–2537. [Google Scholar]
- Mikolov, T.; Chen, K.; Corrado, G.; Dean, J. Efficient Estimation of Word Representations in Vector Space. arXiv 2013, arXiv:1301.3781. [Google Scholar]
- Zhang, Y.; Yang, J. Chinese NER using lattice LSTM. arXiv 2018, arXiv:1805.02023. [Google Scholar]
- Yang, Y.; Chen, W.; Li, Z. Distantly supervised NER with partial annotation learning and reinforcement learning. In Proceedings of the 27th International Conference on Computational Linguistics, Santa Fe, NM, USA, 1 August 2018; pp. 2159–2169. [Google Scholar]
- Jia, Y.; Ma, X. Attention in character-Based BiLSTM-CRF for Chinese named entity recognition. In Proceedings of the 2019 4th International Conference on Mathematics and Artificial Intelligence (ICMAI), New York, NY, USA, 12–15 April 2019; pp. 1–4. [Google Scholar]
- Lan, Z.; Chen, M.; Goodman, S. Albert: A lite bert for self-supervised learning of language representations. arXiv 2019, arXiv:1909.11942. [Google Scholar]
- Raffel, C.; Ellis, D.P.W. Feed-forward networks with attention can solve some long-term memory problems. arXiv 2015, arXiv:1512.08756. [Google Scholar]
- Sui, D.; Chen, Y.; Liu, K. Leverage lexical knowledge for Chinese named entity recognition via collaborative graph network. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, 1 November 2019; pp. 3830–3840. [Google Scholar]
- Li, X.; Yan, H.; Qiu, X. FLAT: Chinese NER Using Flat-Lattice Transformer. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Vancouver, BC, Canada, 1 July 2020; pp. 6836–6842. [Google Scholar]
- Meng, Y.; Wu, W.; Wang, F. Glyce: Glyph-vectors for Chinese character representations. In Proceedings of the 33rd International Conference on Neural Information Processing Systems (NIPS), online, 8 December 2019; pp. 2746–2757. [Google Scholar]
Data Set | Type | The Training Set | The Test Set | Validation Set |
---|---|---|---|---|
WeiboNER | sentence | 1.4 k | 0.3 k | 0.3 k |
character | 73.5 k | 14.4 k | 14.8 k | |
Microsoft MSRA | sentence | 46.1 k | 4.4 k | - |
character | 2169.9 k | 4.4 k | ||
Cluener2020 | sentence | 10.7 k | 13.4 k | 13.4 k |
character | 41.3 k | 13.4 k | 51.6 k |
Model | MSRA | WeiboNER | ||||
---|---|---|---|---|---|---|
P | R | F1 | P | R | F1 | |
Lattice-LSTM | 93.57% | 92.79% | 93.18% | - | - | 58.79% |
LR-CNN | 94.50% | 92.93% | 93.71% | - | - | 59.92% |
CGN | 94.01% | 92.93% | 93.47% | - | - | 63.09% |
FLAT | - | - | 94.35% | - | - | 67.71% |
Glyce | 95.57% | 95.51% | 95.54% | 67.78% | 67.71% | 67.60% |
Our model | 95.81% | 91.97% | 94.78% | 72.57% | 70.31% | 71.42% |
Model | CLUENER2020 | WeiboNER | MSRA |
---|---|---|---|
BiLSTM-CRF | 70.41% | 58.76% | 83.41% |
ALBERT-BiLSTM-CRF | 79.63% | 68.59% | 92.61% |
Stitching vector | 80.32% | 68.75% | 93.52% |
Fusion vector | 81.59% | 70.35% | 93.57% |
Our model | 82.47% | 71.42% | 94.78% |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Sun, Z.; Li, X. Named Entity Recognition Model Based on Feature Fusion. Information 2023, 14, 133. https://doi.org/10.3390/info14020133
Sun Z, Li X. Named Entity Recognition Model Based on Feature Fusion. Information. 2023; 14(2):133. https://doi.org/10.3390/info14020133
Chicago/Turabian StyleSun, Zhen, and Xinfu Li. 2023. "Named Entity Recognition Model Based on Feature Fusion" Information 14, no. 2: 133. https://doi.org/10.3390/info14020133
APA StyleSun, Z., & Li, X. (2023). Named Entity Recognition Model Based on Feature Fusion. Information, 14(2), 133. https://doi.org/10.3390/info14020133