Role Knowledge Prompting for Document-Level Event Argument Extraction
Abstract
:1. Introduction
- (1)
- We propose a role knowledge-prompting model for DEAE capable of extracting all arguments by slot span filling, activating PLMs’ potential with a prompt-tuning paradigm to guide the template precisely.
- (2)
- We designed a role knowledge-prompting model to enhance the representation of roles through the interaction of the text, template, and role knowledge, which helps to better capture the deep semantic relationships. This gives our model a good ability to cope with arguments and roles overlapping.
- (3)
- The experimental results show that our method achieves the state of the art on two document-level benchmark datasets—RAMS and WIKIEVENTS—with a 3.2% and 1.4% F1 improvement in argument classification, respectively.
2. Related Work
3. Methodology
3.1. Problem Definition
3.2. Template Design for DEAE
3.3. Role Knowledge Guidance
3.4. Argument Generation
3.4.1. Text Span Querier
3.4.2. Argument Generation
4. Experiments
4.1. Datasets and Evaluation Metrics
4.2. Experimental Settings
4.3. Baselines
- SpanSel [7], a span ranking-based model, enumerates all possible spans in a document to identify event arguments.
- DocMRC [12] is an approach that models EAE as a machine reading comprehension (MRC) task, using BERT as a PLM.
- DocMRC(IDA) [12], which uses an implicit data augmentation approach for EAE, is an enhanced version of DocMRC.
- BERT-QA [13] is an approach that translates EAE into a QA task, using BERT as a PLM.
- BART-Gen [16] is an approach for generating the event arguments by sequence-to-sequence modeling and prompt learning, using the BART-large generative model as a PLM.
- SpanSel [7]: We report the results from the original paper.
- DocMRC [12]: We report the results from the original paper.
- DocMRC(IDA) [12]: We report the results from the original paper.
- BERT-QA [13]: We used their code (https://github.com/xinyadu/eeqa, accessed on 26 October 2020) to test its performance. We set the question template as “What is the ROLE in TRIGGER WORD?”, following the second template setting in the original paper.
- BART-Gen [16]: For the BART-base model, we used their code (https://github.com/raspberryice/gen-arg, accessed on 13 April 2021) to test its performance. For the BART-large model, we report the results from the original paper.
4.4. Main Results and Analysis
- (1)
- When comparing the approaches for the MRC and QA tasks (where MRC means machine reading comprehesion while QA means question and answer), MRC is usually in the form of fill-in-the-blank completion for the model to determine which hidden word is most likely, depending on the context, while the questions of QA are answered one by one in accordance with the set question templates. For DEAE, these approaches both use the BERT model, which is applicable to natural language understanding. DocMRC and BERT-QA had comparable performance on the RAMS dataset, but BERT-QA had a substantial improvement over DocMRC on the WIKIEVENTS dataset. Our approach is similar to that of reading comprehension, which is carried out in the form of span extraction and has a nearly 10 point performance improvement over the former. We conjecture that the performance improvement mainly lies in adding templates to the pretrained model.
- (2)
- In comparing the approaches of generative tasks, the generative model BART-Gen and our model both used prompt templates and had similar performances, but our performance was a bit better. We found that role knowledge guidance plays an active role in the templates. Unlike the prompt of BART-Gen, we designed two sub-prompts (i.e., role definition and event description) to help the model understand the semantics better and introduce role concept knowledge and role interaction knowledge into the template to guide the template for more precise prompting. The results demonstrate that role knowledge prompting for DEAE not only helps to locate the arguments accurately, but it also can cope with arguments and roles overlapping.
4.5. Ablation Experiments
- (1)
- Without eventdes: Removing the event description from the template resulted in a 2.1% and 5.8% decrease in the F1 score of Arg-C on the two respective datasets. Intuitively, the description of the event arguments according to the logic of event occurrence enhanced the semantic expressiveness of the template.
- (2)
- Without rolekg: To verify the effectiveness of the role knowledge guidance module, we removed the role knowledge module and used only templates for prompting. Without the role concept knowledge to add descriptions to the roles in the template, and without the role interaction knowledge to guide the role relationships, the performance of the model decreased by 0.8% and 1.4% on Arg-C F1. This shows the importance of role knowledge, and it acts as a guide for the prompt.
- (3)
- Without prompt: To demonstrate the necessity of the prompt template module, the entire prompt was removed, and DEAE was performed with only the original text as input without giving any prompt to the model. This resulted in a significant decline in performance. This illustrates that the prompt templates can stimulate the learning ability of the PLMs and the precise prompting promotes more accurate prediction of the model.
4.6. Impact of the Input Encoder Content
4.7. Effectiveness of Role Knowledge
4.8. Case Study
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
Abbreviations
RKDE | Role knowledge prompting for document-level event argument extraction |
PLMs | Pretrained language models |
EE | Event extraction |
EAE | Event argument extraction |
SEAE | Sentence-level event argument extraction |
DEAE | Document-level event argument extraction |
References
- Liu, K.; Chen, Y.; Liu, J.; Zuo, X.; Zhao, J. Extracting events and their relations from texts: A survey on recent research progress and challenges. AI Open 2020, 1, 22–39. [Google Scholar] [CrossRef]
- Xi, X.; Ye, W.; Zhang, T.; Wang, Q.; Zhang, S.; Jiang, H.; Wu, W. Improving event detection by exploiting label hierarchy. In Proceedings of the ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 7688–7692. [Google Scholar]
- Chen, Y.; Chen, T.; Van Durme, B. Joint modeling of arguments for event understanding. In Proceedings of the First Workshop on Computational Approaches to Discourse, online, 17 June 2020; pp. 96–101. [Google Scholar]
- Bosselut, A.; Choi, Y. Dynamic knowledge graph construction for zero-shot commonsense question answering. arXiv 2019, arXiv:1911.03876. [Google Scholar] [CrossRef]
- Gao, L.; Wu, J.; Qiao, Z.; Zhou, C.; Yang, H.; Hu, Y. Collaborative social group influence for event recommendation. In Proceedings of the 25th ACM International on Conference on Information and Knowledge Management, Indianapolis, IN, USA, 24–28 October 2016; pp. 1941–1944. [Google Scholar]
- Boyd-Graber, J.; Börschinger, B. What question answering can learn from trivia nerds. arXiv 2019, arXiv:1910.14464. [Google Scholar]
- Ebner, S.; Xia, P.; Culkin, R.; Rawlins, K.; Van Durme, B. Multi-sentence argument linking. arXiv 2019, arXiv:1911.03766. [Google Scholar]
- Chen, Y.; Xu, L.; Liu, K.; Zeng, D.; Zhao, J. Event extraction via dynamic multi-pooling convolutional neural networks. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), Beijing, China, 26–31 July 2015; pp. 167–176. [Google Scholar]
- Nguyen, T.H.; Cho, K.; Grishman, R. Joint event extraction via recurrent neural networks. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, San Diego, CA, USA, 13–15 June 2016; pp. 300–309. [Google Scholar]
- Liu, X.; Luo, Z.; Huang, H. Jointly multiple events extraction via attention-based graph information aggregation. arXiv 2018, arXiv:1809.09078. [Google Scholar]
- Lin, Y.; Ji, H.; Huang, F.; Wu, L. A joint neural model for information extraction with global features. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 7999–8009. [Google Scholar]
- Liu, J.; Chen, Y.; Xu, J. Machine reading comprehension as data augmentation: A case study on implicit event argument extraction. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, Online, 7–11 November 2021; pp. 2716–2725. [Google Scholar]
- Du, X.; Cardie, C. Event extraction by answering (almost) natural questions. arXiv 2020, arXiv:2004.13625. [Google Scholar]
- Li, F.; Peng, W.; Chen, Y.; Wang, Q.; Pan, L.; Lyu, Y.; Zhu, Y. Event extraction as multi-turn question answering. In Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2020, Online, 16–20 November 2020; pp. 829–838. [Google Scholar]
- Lu, Y.; Lin, H.; Xu, J.; Han, X.; Tang, J.; Li, A.; Sun, L.; Liao, M.; Chen, S. Text2Event: Controllable sequence-to-structure generation for end-to-end event extraction. arXiv 2021, arXiv:2106.09232. [Google Scholar]
- Li, S.; Ji, H.; Han, J. Document-level event argument extraction by conditional generation. arXiv 2021, arXiv:2104.05919. [Google Scholar]
- Grishman, R.; Sundheim, B.M. Message understanding conference-6: A brief history. In Proceedings of the COLING 1996 Volume 1: The 16th International Conference on Computational Linguistics, Copenhagen, Denmark, 5–9 August 1996. [Google Scholar]
- Liu, S.; Chen, Y.; He, S.; Liu, K.; Zhao, J. Leveraging framenet to improve automatic event detection. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Berlin, Germany, 7–12 August 2016; pp. 2134–2143. [Google Scholar]
- Patwardhan, S.; Riloff, E. 56-A unified model of phrasal and sentential evidence for information extraction. In Proceedings of the 2009 Conference on Empirical Methods in Natural Language Processing, Singapore, 6–7 August 2009; pp. 151–160. [Google Scholar]
- Liao, S.; Grishman, R. Using document level cross-event inference to improve event extraction. In Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics, Uppsala, Sweden, 11–16 July 2010; pp. 789–797. [Google Scholar]
- Huang, R.; Riloff, E. Modeling textual cohesion for event extraction. In Proceedings of the AAAI Conference on Artificial Intelligence, Toronto, ON, Canada, 22–26 July 2012; Volume 26, pp. 1664–1670. [Google Scholar]
- Li, Q.; Ji, H.; Huang, L. Joint event extraction via structured prediction with global features. In Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Sofia, Bulgaria, 4–9 August 2013; pp. 73–82. [Google Scholar]
- Nguyen, T.H.; Grishman, R. Event detection and domain adaptation with convolutional neural networks. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 2: Short Papers), Beijing, China, 26–31 July 2015; pp. 365–371. [Google Scholar]
- Sha, L.; Qian, F.; Chang, B.; Sui, Z. Jointly extracting event triggers and arguments by dependency-bridge RNN and tensor-based argument interaction. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32. [Google Scholar]
- Feng, X.; Huang, L.; Tang, D.; Ji, H.; Qin, B.; Liu, T. A language-independent neural network for event detection. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), Berlin, Germany, 7–12 August 2016; pp. 66–71. [Google Scholar]
- Huang, L.; Ji, H.; Cho, K.; Voss, C.R. Zero-shot transfer learning for event extraction. arXiv 2017, arXiv:1707.01066. [Google Scholar]
- Zhang, T.; Whitehead, S.; Zhang, H.; Li, H.; Ellis, J.; Huang, L.; Liu, W.; Ji, H.; Chang, S.F. Improving event extraction via multimodal integration. In Proceedings of the 25th ACM International Conference on Multimedia, Mountain View, CA, USA, 23–27 October 2017; pp. 270–278. [Google Scholar]
- Chen, Y.; Liu, S.; Zhang, X.; Liu, K.; Zhao, J. Automatically labeled data generation for large scale event extraction. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Vancouver, BC, Canada, 30 July–4 August 2017; pp. 409–419. [Google Scholar]
- Zhu, Z.; Li, S.; Zhou, G.; Xia, R. Bilingual event extraction: A case study on trigger type determination. In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), Baltimore, MD, USA, 22–27 June 2014; pp. 842–847. [Google Scholar]
- Liu, S.; Chen, Y.; Liu, K.; Zhao, J. Exploiting argument information to improve event detection via supervised attention mechanisms. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Vancouver, BC, Canada, 30 July–4 August 2017; pp. 1789–1798. [Google Scholar]
- Lou, D.; Liao, Z.; Deng, S.; Zhang, N.; Chen, H. MLBiNet: A cross-sentence collective event detection network. arXiv 2021, arXiv:2105.09458. [Google Scholar]
- Yang, S.; Feng, D.; Qiao, L.; Kan, Z.; Li, D. Exploring pre-trained language models for event extraction and generation. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019; pp. 5284–5294. [Google Scholar]
- Wadden, D.; Wennberg, U.; Luan, Y.; Hajishirzi, H. Entity, relation, and event extraction with contextualized span representations. arXiv 2019, arXiv:1909.03546. [Google Scholar]
- Du, X.; Cardie, C. Document-level event role filler extraction using multi-granularity contextualized encoding. arXiv 2020, arXiv:2005.06579. [Google Scholar]
- Xu, R.; Liu, T.; Li, L.; Chang, B. Document-level event extraction via heterogeneous graph-based interaction model with a tracker. arXiv 2021, arXiv:2105.14924. [Google Scholar]
- Tao, J.; Pan, Y.; Li, X.; Hu, B.; Peng, W.; Han, C.; Wang, X. Multi-Role Event Argument Extraction as Machine Reading Comprehension with Argument Match Optimization. In Proceedings of the ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 23–27 May 2022; pp. 6347–6351. [Google Scholar]
- Wang, X.D.; Leser, U.; Weber, L. BEEDS: Large-Scale Biomedical Event Extraction using Distant Supervision and Question Answering. In Proceedings of the 21st Workshop on Biomedical Language Processing, Dublin, Ireland, 26 May 2022; pp. 298–309. [Google Scholar]
- Raffel, C.; Shazeer, N.; Roberts, A.; Lee, K.; Narang, S.; Matena, M.; Zhou, Y.; Li, W.; Liu, P.J. Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 2020, 21, 1–67. [Google Scholar]
- Lewis, M.; Liu, Y.; Goyal, N.; Ghazvininejad, M.; Mohamed, A.; Levy, O.; Stoyanov, V.; Zettlemoyer, L. Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. arXiv 2019, arXiv:1910.13461. [Google Scholar]
- Paolini, G.; Athiwaratkun, B.; Krone, J.; Ma, J.; Achille, A.; Anubhai, R.; Santos, C.N.d.; Xiang, B.; Soatto, S. Structured prediction as translation between augmented natural languages. arXiv 2021, arXiv:2101.05779. [Google Scholar]
- Schick, T.; Schütze, H. Exploiting cloze questions for few shot text classification and natural language inference. arXiv 2020, arXiv:2001.07676. [Google Scholar]
- Han, X.; Zhao, W.; Ding, N.; Liu, Z.; Sun, M. Ptr: Prompt tuning with rules for text classification. AI Open 2022, 3, 182–192. [Google Scholar] [CrossRef]
- Hsu, I.; Huang, K.H.; Boschee, E.; Miller, S.; Natarajan, P.; Chang, K.W.; Peng, N. Event extraction as natural language generation. arXiv 2021, arXiv:2108.12724. [Google Scholar]
- Ye, H.; Zhang, N.; Bi, Z.; Deng, S.; Tan, C.; Chen, H.; Huang, F.; Chen, H. Learning to Ask for Data-Efficient Event Argument Extraction (Student Abstract). Argument 2022, 80, 100. [Google Scholar] [CrossRef]
- Ding, N.; Chen, Y.; Han, X.; Xu, G.; Xie, P.; Zheng, H.T.; Liu, Z.; Li, J.; Kim, H.G. Prompt-learning for fine-grained entity typing. arXiv 2021, arXiv:2108.10604. [Google Scholar]
- Zuo, X.; Cao, P.; Chen, Y.; Liu, K.; Zhao, J.; Peng, W.; Chen, Y. LearnDA: Learnable knowledge-guided data augmentation for event causality identification. arXiv 2021, arXiv:2106.01649. [Google Scholar]
- Shin, T.; Razeghi, Y.; Logan IV, R.L.; Wallace, E.; Singh, S. Autoprompt: Eliciting knowledge from language models with automatically generated prompts. arXiv 2020, arXiv:2010.15980. [Google Scholar]
- Liu, X.; Zheng, Y.; Du, Z.; Ding, M.; Qian, Y.; Yang, Z.; Tang, J. GPT understands, too. arXiv 2021, arXiv:2103.10385. [Google Scholar]
- Li, X.L.; Liang, P. Prefix-tuning: Optimizing continuous prompts for generation. arXiv 2021, arXiv:2101.00190. [Google Scholar]
- Hu, S.; Ding, N.; Wang, H.; Liu, Z.; Li, J.; Sun, M. Knowledgeable prompt-tuning: Incorporating knowledge into prompt verbalizer for text classification. arXiv 2021, arXiv:2108.02035. [Google Scholar]
- Zhang, N.; Deng, S.; Cheng, X.; Chen, X.; Zhang, Y.; Zhang, W.; Chen, H.; Center, H.I. Drop Redundant, Shrink Irrelevant: Selective Knowledge Injection for Language Pretraining. In Proceedings of the IJCAI, Virtual, 19–27 August 2021; pp. 4007–4014. [Google Scholar]
- Gao, L.; Choubey, P.K.; Huang, R. Modeling document-level causal structures for event causal relation identification. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), Minneapolis, MN, USA, 2–7 June 2019. [Google Scholar]
- Speer, R.; Chin, J.; Havasi, C.C. 5.5: An open multilingual graph of general knowledge. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence (December 2016), Francisco, CA, USA, 4–9 February 2017; pp. 4444–4451. [Google Scholar]
- Ye, H.; Zhang, N.; Deng, S.; Chen, X.; Chen, H.; Xiong, F.; Chen, X.; Chen, H. Ontology-enhanced Prompt-tuning for Few-shot Learning. In Proceedings of the ACM Web Conference 2022, Lyon, France, 25–29 April 2022; pp. 778–787. [Google Scholar]
- Ma, Y.; Wang, Z.; Cao, Y.; Li, M.; Chen, M.; Wang, K.; Shao, J. Prompt for Extraction? PAIE: Prompting Argument Interaction for Event Argument Extraction. arXiv 2022, arXiv:2202.12109. [Google Scholar]
Datasets | RAMS | WIKIEVENTS |
---|---|---|
Sentences: Train | 7329 | 5262 |
Sentences: Dev | 924 | 378 |
Sentences: Test | 871 | 492 |
Arguments: Train | 17,026 | 4552 |
Arguments: Dev | 2188 | 428 |
Arguments: Test | 2023 | 566 |
Method | RAMS | WIKIEVENTS | |||
---|---|---|---|---|---|
Arg-I | Arg-C | Arg-I | Arg-C | Head-C | |
SpanSel [7] | - | 40.7 | - | 38.3 | - |
DocMRC [12] | - | 44.3 | - | 42.1 | - |
DocMRC(IDA) [12] | - | 45.7 | - | 43.3 | - |
BERT-QA [13] | 46.4 | 44.0 | 54.3 | 53.2 | 56.9 |
BART-Gen (BART-Base) [16] | 50.9 | 44.9 | 47.5 | 41.7 | 44.2 |
BART-Gen (BART-Large) [16] | 51.2 | 47.1 | 66.8 | 62.4 | 65.4 |
RKDE (ours) | 55.1 | 50.3 | 69.1 | 63.8 | 66.8 |
Method | RAMS | WIKIEVENTS | |||||
---|---|---|---|---|---|---|---|
Arg-I | Arg-C | ∇(Arg-C) | Arg-I | Arg-C | Head-C | ∇(Arg-C) | |
RKDE (ours) | 55.1 | 50.3 | - | 69.1 | 63.8 | 66.8 | - |
Without eventdes | 54.6 | 48.2 | −2.1 | 63.7 | 57.6 | 60.9 | −5.8 |
Without rolekg | 54.7 | 49.5 | −0.8 | 68.9 | 62.4 | 65.5 | −1.4 |
Without prompt | 54.0 | 47.6 | −2.7 | 63.2 | 57.3 | 60.1 | −6.5 |
No. | Samples | Event Type | Role: Argument | Nop | TEAE | +Role |
---|---|---|---|---|---|---|
1 | According to investigators, police seized a sum of some $120 million and €2 million ($2.2 million). “The final amount is unknown. Police<t> confiscated </t> a cache only in Zakharchenko ’s house; it is not yet known how many more assets he has, but we are working in this direction, ” …. | Transaction. Transaction. Transfercontrol. | Recipient: a cache | ✗ | ✔ | ✔ |
Beneficiary: Police | ✗ | ✗ | ✔ | |||
Giver: Police | ✔ | ✔ | ✔ | |||
Place: Zakharchenko’s house | ✗ | ✗ | ✗ | |||
2 | Merkel has faced growing criticism in her country for allowing an unprece- dented number of asylum seekers (almost 1.1 million last year) to enter Germany. In her weekly video pod- cast Saturday, she <t> urged </t> refugees from Iraq and Syria to integr- ate in the country and learn German … | Contact. Requestadvise. Correspondence. | Place: Germany | ✗ | ✔ | ✔ |
Recipient: refugees from Iraq and Syria | ✗ | ✗ | ✔ | |||
Communicator: she | ✔ | ✔ | ✔ | |||
3 | In February 1975, the Provisional Irish Republican Army <t> agreed </t> to a truce and ceasefire with the British government and the Northern Ireland Office. Seven “incident centres” were established in Irish nationalist areas in Northern Ireland to monitor the cease- fire and the activity of the security forces. | Contact. Contact. Unspecified. | Participant: Provisional Irish Republican Army | ✔ | ✔ | ✔ |
Participant:government | ✗ | ✔ | ✗ | |||
Participant: Northern Ireland Office | ✗ | ✔ | ✔ |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Hu, R.; Liu, H.; Zhou, H. Role Knowledge Prompting for Document-Level Event Argument Extraction. Appl. Sci. 2023, 13, 3041. https://doi.org/10.3390/app13053041
Hu R, Liu H, Zhou H. Role Knowledge Prompting for Document-Level Event Argument Extraction. Applied Sciences. 2023; 13(5):3041. https://doi.org/10.3390/app13053041
Chicago/Turabian StyleHu, Ruijuan, Haiyan Liu, and Huijuan Zhou. 2023. "Role Knowledge Prompting for Document-Level Event Argument Extraction" Applied Sciences 13, no. 5: 3041. https://doi.org/10.3390/app13053041
APA StyleHu, R., Liu, H., & Zhou, H. (2023). Role Knowledge Prompting for Document-Level Event Argument Extraction. Applied Sciences, 13(5), 3041. https://doi.org/10.3390/app13053041