Similarity Calculation via Passage-Level Event Connection Graph
Abstract
:1. Introduction
2. Related Work
3. Model Details
3.1. Task Overview
3.2. Graph Construction
3.3. Node Evaluation
4. Two Improvements Made on Our Event Connection Graph
4.1. Tuning Trigger Words
4.2. Node Representation via Graph Embedding
- Taking for example, the event connection graph formed from , we treat one node in as the starting point and choose the successive node via randomly jumping to one of the adjacent nodes. Repeat this jump for times. A path of length can be obtained.
- Repeat step 1 for times on each node in . We then get a path set (noted as ) whose size is , where n denotes the size of .
- Each path in PH is treated as one training sample.
5. Experiments and Analyses
5.1. Experimental Setting
- (1)
- Average: the mean of all the word vectors in the input text.
- (2)
- TextRank+Average: take TextRank to choose keywords from input text and then treat the mean of the chosen keyword vectors as representation.
- (3)
- TextRank+Concatenation: take TextRank to choose keywords and concatenate the word vectors of the chosen keywords to form a long vector.
5.2. Experimental Results
5.2.1. Testing on Threshold
5.2.2. Comparison of Different Event Extraction Methods
5.2.3. Case Study
5.2.4. Comparison of Different Algorithms
5.2.5. Ablation Results
5.2.6. Task Transferring
6. Conclusions and Future Work
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Ji, H.; Grishman, R. Refining event extraction through cross-document inference. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Columbus, OH, USA, 16–18 June 2008; pp. 254–262. [Google Scholar]
- Baker, C.; Fillmore, C.; Lowe, J. The berkeley framenet project. In Proceedings of the 36th Annual Meeting of the Association for Computational Linguistics and 17th International Conference on Computational Linguistics, Montreal, QC, Canada, 10–14 August 1998; pp. 86–90. [Google Scholar]
- Jacksi, K.; Ibrahim, R.; Zeebaree, S.; Zebari, R.; Sadeeq, M. Clustering documents based on semantic similarity using HAC and K-mean algorithms. In Proceedings of the 2020 International Conference on Advanced Science and Engineering, Duhok, Iraq, 23–24 December 2020; pp. 205–210. [Google Scholar]
- Huang, X.; Qi, J.; Sun, Y.; Zhang, R. Mala: Cross-domain dialogue generation with action learning. In Proceedings of the 34th AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 7977–7984. [Google Scholar]
- Kieu, B.; Unanue, I.; Pham, S.; Phan, H.; Piccardi, M. Learning neural textual representations for citation recommendation. In Proceedings of the 25th International Conference on Pattern Recognition, Milan, Italy, 10–15 January 2021; pp. 4145–4152. [Google Scholar]
- Chen, Y.; Zhou, M.; Wang, S. Reranking answers for definitional QA using language modeling. In Proceedings of the 21st International Conference on Computational Linguistics and 44th Annual Meeting of the Association for Computational Linguistics, Sydney, Australia, 17–21 July 2006; pp. 1081–1088. [Google Scholar]
- Turian, J.; Ratinov, L.; Bengio, Y. Word representations: A simple and general method for semi-supervised learning. In Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics, Uppsala, Sweden, 11–16 July 2010; pp. 384–394. [Google Scholar]
- Mikolov, T.; Sutskever, I.; Chen, K.; Corrado, G.; Dean, J. Distributed representations of words and phrases and their compositionality. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 5–10 December 2013; pp. 3111–3119. [Google Scholar]
- Mikolov, T.; Chen, K.; Corrado, G.; Dean, J. Efficient estimation of word representations in vector space. In Proceedings of the 1st International Conference on Learning Representations, Scottsdale, AZ, USA, 2–4 May 2013; pp. 1–12. [Google Scholar]
- Pennington, J.; Socher, R.; Manning, C. GloVe: Global vectors for word representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing, Doha, Qatar, 25–29 October 2014; pp. 1532–1543. [Google Scholar]
- Peters, M.; Neumann, M.; Iyyer, M.; Gardner, M.; Clark, C.; Lee, K.; Zettlemoyer, L. Deep contextualized word representations. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, New Orleans, LA, USA, 1–6 June 2018; pp. 2227–2237. [Google Scholar]
- Chen, J.; Dai, X.; Yuan, Q.; Lu, C.; Huang, H. Towards interpretable clinical diagnosis with Bayesian network ensembles stacked on entity-aware CNNs. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 3143–3153. [Google Scholar]
- Cho, K.; van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning Phrase Representations Using RNN Encoder-Decoder for Statistical Machine Translation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing, Doha, Qatar, 25–29 October 2014; pp. 1724–1734. [Google Scholar]
- Behera, R.K.; Jena, M.; Rath, S.K.; Misra, S. Co-LSTM: Convolutional LSTM model for sentiment analysis in social big data. Inf. Process. Manag. 2020, 58, 102435. [Google Scholar] [CrossRef]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.; Kaiser, L.; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 5998–6008.
- Radford, A.; Narasimhan, K.; Salimans, T.; Sutskever, I. Improving Language Understanding by Generative Pre-Training. 2018. Available online: https://S3-Us-West-2.Amazonaws.Com (accessed on 31 December 2021).
- Devlin, J.; Chang, W.; Lee, K.; Toutanova, K. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Minneapolis, MN, USA, 2–7 June 2019; pp. 4171–4186. [Google Scholar]
- Yang, Z.; Dai, Z.; Yang, Y.; Carbonell, J.; Salakhutdinov, R.; Le, V. XLNET: Generalized autoregressive pretraining for language understanding. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; pp. 5753–5763. [Google Scholar]
- Liu, Y.; Ott, M.; Goyal, N.; Du, J.; Joshi, M.; Chen, D.; Levy, O.; Lewis, M.; Zettlemoyer, L.; Stoyanov, V. RoBERTa: A robustly optimized BERT pretraining approach. arXiv 2019, arXiv:1907.11692. [Google Scholar]
- Pan, Y.; Yao, T.; Li, Y.; Mei, T. X-Linear attention networks for image captioning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 10971–10980. [Google Scholar]
- Liu, Y.; Zhang, X.; Zhang, Q.; Li, C.; Huang, F.; Tang, X.; Li, Z. Dual self-attention with co-attention networks for visual question answering. Pattern Recognit. 2021, 117, 107956. [Google Scholar] [CrossRef]
- Lee, L.H.; Wan, C.H.; Rajkumar, R.; Isa, D. An enhanced Support Vector Machine classification framework by using Euclidean distance function for text document categorization. Appl. Intell. 2011, 37, 80–99. [Google Scholar] [CrossRef]
- Zhu, H.; Zhang, P.; Gao, Z. K-means text dynamic clustering algorithm based on KL divergence. In Proceedings of the 17th IEEE/ACIS International Conference on Computer and Information Science, Singapore, 6–8 June 2018; pp. 659–663. [Google Scholar]
- Huang, A. Similarity measures for text document clustering. In Proceedings of the 6th New Zealand Computer Science Research Student Conference, Hamilton, New Zealand, 14–18 April 2008; pp. 9–56. [Google Scholar]
- Atoum, I.; Otoom, A. Efficient Hybrid Semantic Text Similarity using Wordnet and a Corpus. Int. J. Adv. Comput. Sci. Appl. 2016, 7, 124–130. [Google Scholar] [CrossRef] [Green Version]
- Gomaa, W.; Fahmy, A. A survey of text similarity approaches. Int. J. Comput. Appl. 2013, 68, 13–18. [Google Scholar]
- Robertson, S. Understanding inverse document frequency: On theoretical arguments for IDF. J. Doc. 2004, 60, 503–520. [Google Scholar] [CrossRef] [Green Version]
- Mihalcea, R.; Tarau, P. Textrank: Bringing order into text. In Proceedings of the 2004 Conference on Empirical Methods in Natural Language Processing, Barcelona, Spain, 25–26 July 2004; pp. 404–411. [Google Scholar]
- Blei, D.; Ng, A.; Jordan, M. Latent dirichlet allocation. J. Mach. Learn. Res. 2003, 3, 993–1022. [Google Scholar]
- Pavlinek, M.; Podgorelec, V. Text classification method based on self-training and LDA topic models. Expert Syst. Appl. 2017, 80, 83–93. [Google Scholar] [CrossRef]
- Sharif, U.; Ghada, E.; Atlam, E.; Fuketa, M.; Morita, K.; Aoe, J.-I. Improvement of building field association term dictionary using passage retrieval. Inf. Process. Manag. 2007, 43, 1793–1807. [Google Scholar] [CrossRef]
- Dorji, T.C.; Atlam, E.-S.; Yata, S.; Fuketa, M.; Morita, K.; Aoe, J.-I. Extraction, selection and ranking of Field Association (FA) Terms from domain-specific corpora for building a comprehensive FA terms dictionary. Knowl. Inf. Syst. 2010, 27, 141–161. [Google Scholar] [CrossRef]
- Malkiel, I.; Ginzburg, D.; Barkan, O.; Caciularu, A.; Weill, J.; Koenigstein, N. Interpreting BERT-based Text Similarity via Activation and Saliency Maps. In Proceedings of the 2022 ACM Web Conference, Lyon, France, 25–29 April 2022; pp. 3259–3268. [Google Scholar]
- Grishman, R.; Sundheim, B. Message understanding conference-6: A brief history. In Proceedings of the 16th International Conference on Computational Linguistics, Copenhagen, Denmark, 5–9 August 1996; pp. 466–471. [Google Scholar]
- Doddington, G.; Mitchell, A.; Przybocki, M.; Ramshaw, L.; Strassel, S.; Weischedel, R. The Automatic Content Extraction (ACE) program-tasks, data, and evaluation. In Proceedings of the 4th International Conference on Language Resources and Evaluation, Centro Cultural de Belem, Lisbon, 24–30 May 2004; pp. 837–840. [Google Scholar]
- Yang, S.; Feng, D.; Qiao, L.; Kan, Z.; Li, D. Exploring pre-trained language models for event extraction and generation. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019; pp. 169–175. [Google Scholar]
- Wang, Z.; Wang, X.; Han, X.; Lin, Y.; Hou, L.; Liu, Z.; Li, P.; Li, J.; Zhou, J. CLEVE: Contrastive pre-training for event extraction. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, Online, 1–6 August 2021; pp. 6283–6297. [Google Scholar]
- Du, X.; Cardie, C. Document-Level event role filler extraction using multi-granularity contextualized encoding. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Washington, DC, USA, 5–10 July 2020; pp. 8010–8020. [Google Scholar]
- Hu, Z.; Liu, M.; Wu, Y.; Xu, J.; Qin, B.; Li, J. Document-level event subject pair recognition. In Proceedings of the 9th CCF International Conference on Natural Language Processing and Chinese Computing, Zhengzhou, China, 14–18 October 2020; pp. 283–293. [Google Scholar]
- Lin, Y.; Ji, H.; Huang, F.; Wu, L. A joint neural model for information extraction with global features. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Washington, DC, USA, 5–10 July 2020; pp. 7999–8009. [Google Scholar]
- Liao, S.; Grishman, R. Using document level cross-event inference to improve event extraction. In Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics, Uppsala, Sweden, 11–16 July 2010; pp. 789–797. [Google Scholar]
- Atlam, E.-S.; Elmarhomy, G.; Morita, K.; Fuketa, M.; Aoe, J.-I. Automatic building of new Field Association word candidates using search engine. Inf. Process. Manag. 2006, 42, 951–962. [Google Scholar] [CrossRef]
- Lai, D.; Lu, H.; Nardini, C. Finding communities in directed networks by PageRank random walk induced network embedding. Phys. A Stat. Mech. Its Appl. 2010, 389, 2443–2454. [Google Scholar] [CrossRef]
- Li, P.; Zhou, G.; Zhu, Q.; Hou, L. Employing compositional semantics and discourse consistency in Chinese event extraction. In Proceedings of the 2012 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning, Jeju Island, Korea, 12–14 July 2012; pp. 1006–1016. [Google Scholar]
- Amir, H.; Béatrice, D. Word embedding approach for synonym extraction of multi-word terms. In Proceedings of the Eleventh International Conference on Language Resources and Evaluation, Miyazaki, Japan, 7–12 May 2018; pp. 297–303. [Google Scholar]
- Davenport, E.; Cronin, B. Knowledge management: Semantic drift or conceptual shift? J. Educ. Libr. Inf. Sci. 2000, 1, 294–306. [Google Scholar] [CrossRef]
- Grover, A.; Leskovec, J. Node2vec: Scalable feature learning for networks. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 855–864. [Google Scholar]
- Sasano, R.; Korhonen, A. Investigating word-class distributions in word vector spaces. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 3657–3666. [Google Scholar]
- Kennington, C. Enriching language models with visually-grounded word vectors and the lancaster sensorimotor norms. In Proceedings of the 25th Conference on Computational Natural Language Learning, Online, 10–11 November 2021; pp. 148–157. [Google Scholar]
- Mysore, S.; Cohan, A.; Hope, T. Multi-vector models with textual guidance for fine-grained scientific document similarity. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Seattle, Washington, DC, USA, 10–15 July 2022; pp. 4453–4470. [Google Scholar]
- Sun, X.; Meng, Y.; Ao, X.; Wu, F.; Zhang, T.; Li, J.; Fan, C. Sentence similarity based on contexts. Trans. Assoc. Comput. Linguist. 2022, 10, 573–588. [Google Scholar] [CrossRef]
- Lee, S.; Lee, D.; Jang, S.; Yu, H. Toward interpretable semantic textual similarity via optimal transport-based contrastive sentence learning. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Dublin, Ireland, 22–27 May 2022; pp. 5969–5979. [Google Scholar]
- Li, Q.; Ji, H.; Huang, L. Joint event extraction via structured prediction with global features. In Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics, Sofia, Bulgaria, 4–9 August 2013; pp. 73–82. [Google Scholar]
- Zhang, J.; Qin, Y.; Zhang, Y.; Liu, M.; Ji, D. Extracting entities and events as a single task using a transition-based neural model. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, Macao, China, 10–16 August 2019; pp. 5422–5428. [Google Scholar]
- Liu, X.; Huang, H.; Zhang, Y. Open domain event extraction using neural latent variable models. In Proceedings of the 57th Conference of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019; pp. 2860–2871. [Google Scholar]
- Levy, O.; Søgaard, A.; Goldberg, Y. A strong baseline for learning cross-lingual word embeddings from sentence alignments. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics, Valencia, Spain, 3–7 April 2017; pp. 765–774. [Google Scholar]
Feature/Method | Supervised | Unsupervised |
---|---|---|
Training data | Results derived from training data | Free from training data |
Performance | Higher performance | Lower performance |
W/O semantics | Semantic embedding via representation | Lack of understanding semantics |
Domain transfer | Hard | Easy |
Methods | English | Chinese | Spanish | |||
---|---|---|---|---|---|---|
Para | Q&Q | MA | Q&Q | MA | MA | |
BeemSearch | 0.81 | 0.84 | 0.66 | 0.85 | 0.62 | 0.67 |
JointTransition | 0.79 | 0.83 | 0.64 | 0.84 | 0.63 | 0.68 |
ODEE-FER | 0.82 | 0.85 | 0.68 | 0.86 | 0.65 | 0.71 |
OneIE | 0.82 | 0.86 | 0.68 | 0.89 | 0.67 | 0.71 |
Extreme | 0.77 | 0.82 | 0.63 | 0.82 | 0.61 | 0.66 |
Methods | TextRank | LDA | Ours |
---|---|---|---|
Long Texts | |||
A1 | ZecOps, Apple, mail, leak, hacker | Apple, mobile, ZecOps, bug, hacker | Apple, mail, software, leak, * |
A2 | Rooney, soccer, Bayern Munich, injury, champion | soccer, Bayern Munich, beat, Man Utd, Rooney | Bayern Munich, beats, Man Utd, * |
A3 | Barclay, bank, economic, coronavirus, profit | Barclay, coronavirus, bank, pandemic, work | coronavirus, pandemic, costs, £2.1bn, * |
Short Texts | |||
S1 | Carmaker, Tesla, build, factory, Shanghai | Carmaker, Tesla, build, factory, Shanghai | Tesla, build, factory, Shanghai, * |
S2 | Kobe Bryant, death, BBC, TV news, mistake | BBC, apologize, footage, mistake, * | BBC, apologize, footage, mistake, * |
S3 | Coronavirus, economy, sink, pandemic, shutdown | Coronavirus, economy, sink, shutdown, * | Economy, sink, pandemic, shutdown, * |
Methods | English | Chinese | Spanish | ||||
---|---|---|---|---|---|---|---|
Para | Q&Q | MA | Q&Q | MA | MA | ||
supervised | TextCNN | 0.84 *** | 0.82 ** | --- | 0.86 *** | --- | --- |
LSTM | 0.83 ** | 0.84 ** | --- | 0.83 ** | --- | --- | |
LSTM+BIA | 0.83 *** | 0.87 ** | --- | 0.89 *** | --- | --- | |
BERT-base | 0.85 *** | 0.89 *** | --- | 0.91 *** | --- | --- | |
Con-SIM | 0.87 *** | 0.90 *** | --- | 0.92 *** | --- | --- | |
RCMD | 0.90 *** | 0.92 *** | --- | 0.93 *** | --- | --- | |
unsupervised | AVE | 0.58 * | 0.61 ** | 0.42 * | 0.63 ** | 0.49 * | 0.48 ** |
TR+AVE | 0.65 ** | 0.66 ** | 0.49 ** | 0.68 ** | 0.55 ** | 0.59 ** | |
TR+CON | 0.69 * | 0.71 *** | 0.53 ** | 0.66 ** | 0.53 ** | 0.57 ** | |
Ours | 0.82 *** | 0.86 *** | 0.68 *** | 0.89 *** | 0.67 *** | 0.71 *** |
Methods | English | Chinese | Spanish | |||
---|---|---|---|---|---|---|
Para | Q&Q | MA | Q&Q | MA | MA | |
Only via graph | 0.76 | 0.80 | 0.55 | 0.81 | 0.56 | 0.57 |
+linking triggers | 0.78 | 0.82 | 0.59 | 0.85 | 0.60 | 0.62 |
+node representation | 0.81 | 0.84 | 0.63 | 0.87 | 0.65 | 0.69 |
+linking triggers and node representation | 0.82 | 0.86 | 0.68 | 0.89 | 0.67 | 0.71 |
Methods | English | Chinese | C-E | E-C | ||||
---|---|---|---|---|---|---|---|---|
Para | Q&Q | MA | Q&Q | MA | Q&Q | Q&Q | ||
supervised | TextCNN | 0.61 *** | 0.60 ** | 0.49 ** | 0.47 * | 0.51 * | 0.31 ** | 0.27 ** |
LSTM | 0.56 ** | 0.59 ** | 0.51 * | 0.54 ** | 0.52 ** | 0.23 * | 0.22 * | |
LSTM+BIA | 0.54 ** | 0.53 ** | 0.47 * | 0.55 ** | 0.56 ** | 0.26 * | 0.24 * | |
BERT-base | 0.57 *** | 0.55 *** | 0.53 ** | 0.49 *** | 0.51 ** | 0.48 ** | 0.41 ** | |
Con-SIM | 0.52 *** | 0.53 *** | 0.51 * | 0.47 *** | 0.50 * | 0.44 ** | 0.38 ** | |
RCMD | 0.53 *** | 0.56 *** | 0.54 * | 0.46 *** | 0.53 ** | 0.42 ** | 0.43 ** | |
unsupervised | AVE | 0.58 * | 0.61 ** | 0.42 * | 0.63 ** | 0.49 * | 0.61 ** | 0.63 ** |
TR+AVE | 0.65 ** | 0.66 ** | 0.49 ** | 0.68 ** | 0.55 ** | 0.66 ** | 0.68 ** | |
TR+CON | 0.69 * | 0.71 *** | 0.53 ** | 0.66 ** | 0.53 ** | 0.71 *** | 0.66 ** | |
Ours | 0.82 *** | 0.86 *** | 0.68 *** | 0.89 *** | 0.67 *** | 0.86 *** | 0.89 *** |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Liu, M.; Chen, L.; Zheng, Z. Similarity Calculation via Passage-Level Event Connection Graph. Appl. Sci. 2022, 12, 9887. https://doi.org/10.3390/app12199887
Liu M, Chen L, Zheng Z. Similarity Calculation via Passage-Level Event Connection Graph. Applied Sciences. 2022; 12(19):9887. https://doi.org/10.3390/app12199887
Chicago/Turabian StyleLiu, Ming, Lei Chen, and Zihao Zheng. 2022. "Similarity Calculation via Passage-Level Event Connection Graph" Applied Sciences 12, no. 19: 9887. https://doi.org/10.3390/app12199887
APA StyleLiu, M., Chen, L., & Zheng, Z. (2022). Similarity Calculation via Passage-Level Event Connection Graph. Applied Sciences, 12(19), 9887. https://doi.org/10.3390/app12199887