An Improved Nested Named-Entity Recognition Model for Subject Recognition Task under Knowledge Base Question Answering
Abstract
:1. Introduction
- We propose an Improved Nested NER model rather than a flat NER model for the SR task under KBQA. In the case that a general flat NER model fails to recognize the golden entity, our model could still have an opportunity to recognize the golden entity, which shows more effectiveness than baseline flat NER models.
- We employ an approximate matching strategy rather than a strict matching strategy in our model. This strategy shows better effectiveness and robustness especially to noisy questions where the golden entity is different from the golden subject. Experimental results show that our model is more robust than baseline flat NER models.
- Our model is effective to both single−relation questions and complex questions. Experimental results show that our model outperforms the baseline flat NER model by a margin of 3.3% accuracy on the SimpleQuestions dataset (single−relation questions) and a margin of 11.0% accuracy on the WebQuestionsSP dataset (complex questions).
2. Related Work
3. Approach
3.1. Overview
- is the start token of the golden subject;
- is the end token of the golden subject;
- All tokens between and could be strictly matched to corresponding golden subject tokens.
3.2. Token Labeling
3.3. Entity Candidate Generation
3.4. Approximate Matching
4. Experiments
4.1. Dataset
4.2. Experiment Setting
- BERT-CRF/GP/EGP (Figure 1a): BERT-CRF/GP/EGP, which is achieved by Bert4keras, is employed to generate X and the best-matched entity, which is strictly matched to the best-matched subject. In this case, GP and EGP work as flat NER models.
- BERT-CRF + AM (Figure 1b): BERT-CRF is employed to generate X, A, and B in TL. Then, the best-matched subject is generated by our ECG and AM.
- INNM-I (Figure 1c): EGP with is employed to generate X, A, and B in TL. Then, the best-matched subject is generated by our ECG and AM.
- INNM-II/INNM-III (Figure 1d): BERT-CRF/EGP is employed to generate X and the best-matched entity. If the entity could be strictly matched to a subject in the KB, it would be considered as the best-matched subject. Otherwise, INNM-I/BERT-CRF + AM is employed to recognize the best-matched subject.
4.3. Experimental Results
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
Abbreviations
KBQA | Knowledge Base Question Answering |
SR | Subject Recognition |
NER | Named−Entity Recognition |
KB | Knowledge Base |
INNM | Improved Nested NER Model |
NLP | Natural Language Processing |
RE | Relation Extraction |
GP | GlobalPointer |
EGP | Efficient GlobalPointer |
SQ | SimpleQuestions |
WQSP | WebQuestionsSP |
TL | Token Labeling |
ECG | Entity Candidate Generation |
AM | Approximate Matching |
References
- Dai, Z.; Li, L.; Xu, W. CFO: Conditional Focused Neural Question Answering with Large-scale Knowledge Bases. In Proceedings of the Annual Meeting of the Association for Computational Linguistics, Berlin, Germany, 7–12 August 2016; pp. 800–810. [Google Scholar]
- Liang, P.; Jordan, M.I.; Klein, D. Learning dependency-based compositional semantics. Comput. Linguist. 2013, 39, 389–446. [Google Scholar] [CrossRef]
- Bollacker, K.; Evans, C.; Paritosh, P.; Sturge, T.; Taylor, J. Freebase: A collaboratively created graph database for structuring human knowledge. In Proceedings of the 2008 ACM SIGMOD International Conference on Management of Data, Vancouver, BC, Canada, 10–12 June 2008; pp. 1247–1250. [Google Scholar]
- Yao, X.; Durme, B.V. Information extraction over structured data: Question answering with freebase. In Proceedings of the Annual Meeting of the Association for Computational Linguistics, Baltimore, MD, USA, 22–27 June 2014; pp. 956–966. [Google Scholar]
- Wang, R.; Ling, Z.; Hu, Y. Knowledge base question answering with attentive pooling for question representation. IEEE Access 2019, 7, 46773–46784. [Google Scholar] [CrossRef]
- Qu, Y.; Liu, J.; Kang, L.; Shi, Q.; Ye, D. Question answering over freebase via attentive rnn with similarity matrix based cnn. arXiv 2018, arXiv:1804.03317. [Google Scholar]
- Zhao, W.; Chung, T.; Goyal, A.; Metallinou, A. Simple question answering with subgraph ranking and joint-scoring. arXiv 2019, arXiv:1904.04049. [Google Scholar]
- Jin, H.; Luo, Y.; Gao, C.; Tang, X.; Yuan, P. Comqa: Question answering over knowledge base via semantic matching. IEEE Access 2019, 7, 75235–75246. [Google Scholar] [CrossRef]
- Wei, M.; Zhang, Y. Natural answer generation with attention over instances. IEEE Access 2019, 7, 61008–61017. [Google Scholar] [CrossRef]
- Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Minneapolis, MN, USA, 2–7 June 2019; Association for Computational Linguistics: Minneapolis, MN, USA, 2019; pp. 4171–4186. [Google Scholar]
- Lan, Z.; Chen, M.; Goodman, S.; Gimpel, K.; Sharma, P.; Soricut, R. ALBERT: A Lite BERT for Self-supervised Learning of Language Representations. arXiv 2019, arXiv:1909.11942. [Google Scholar]
- Yang, Z.; Dai, Z.; Yang, Y.; Carbonell, J.; Salakhutdinov, R.R.; Le, Q.V. XLNet: Generalized Autoregressive Pretraining for Language Understanding. In Proceedings of the 2019 Annual Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; pp. 5754–5764. [Google Scholar]
- Clark, K.; Luong, M.T.; Le, Q.V.; Manning, C.D. ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators. In Proceedings of the International Conference on Learning Representations, Addis Ababa, Ethiopia, 26–30 April 2020. [Google Scholar]
- Martinc, M.; Škrlj, B.; Pollak, S. TNT-KID: Transformer-based neural tagger for keyword identification. Nat. Lang. Eng. 2022, 28, 409–448. [Google Scholar] [CrossRef]
- Blšták, M.; Rozinajová, V. Automatic question generation based on sentence structure analysis using machine learning approach. Nat. Lang. Eng. 2022, 28, 487–517. [Google Scholar] [CrossRef]
- Wysocki, O.; Zhou, Z.; O’Regan, P.; Ferreira, D.; Wysocka, M.; Landers, D.; Freitas, A. Transformers and the Representation of Biomedical Background Knowledge. Comput. Linguist. 2023, 49, 73–115. [Google Scholar] [CrossRef]
- Laskar, M.T.R.; Hoque, E.; Huang, J.X. Domain Adaptation with pre-trained Transformers for Query-Focused Abstractive Text Summarization. Comput. Linguist. 2022, 48, 279–320. [Google Scholar] [CrossRef]
- Gangwar, A.; Jain, S.; Sourav, S.; Modi, A. Counts@IITK at SemEval-2021 Task 8: SciBERT Based Entity And Semantic Relation Extraction For Scientific Data. In Proceedings of the 15th International Workshop on Semantic Evaluation, Bangkok, Thailand, 5–6 August 2021; pp. 1232–1238. [Google Scholar]
- Luo, D.; Su, J.; Yu, S. A BERT-based Approach with Relation-aware Attention for Knowledge Base Question Answering. In Proceedings of the International Joint Conference on Neural Networks, Glasgow, UK, 19–24 July 2020; pp. 1–8. [Google Scholar]
- Zhu, W. AutoRC: Improving BERT Based Relation Classification Models via Architecture Search. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing: Student Research Workshop, Virtual Event, 1–6 August 2021; pp. 33–43. [Google Scholar]
- Wang, A.; Singh, A.; Michael, J.; Hill, F.; Levy, O.; Bowman, S.R. GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding. In Proceedings of the 7th International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
- Zhang, J.; Zhang, X.; Yu, J.; Tang, J.; Tang, J.; Li, C.; Chen, H. Subgraph Retrieval Enhanced Model for Multi-hop Knowledge Base Question Answering. arXiv 2022, arXiv:2202.13296. [Google Scholar]
- Gu, Y.; Su, Y. ArcaneQA: Dynamic Program Induction and Contextualized Encoding for Knowledge Base Question Answering. arXiv 2022, arXiv:2204.08109. [Google Scholar]
- Shi, J.; Cao, S.; Hou, L.; Li, J.; Zhang, H. TransferNet: An Effective and Transparent Framework for Multi-hop Question Answering over Relation Graph. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, Virtual Event, 7–11 November 2021; pp. 4149–4158. [Google Scholar]
- Su, W.; Zhu, X.; Cao, Y.; Li, B.; Lu, L.; Wei, F.; Dai, J. VL-BERT: Pre-training of Generic Visual-Linguistic Representations. In Proceedings of the International Conference on Learning Representations, Addis Ababa, Ethiopia, 30 April 2020. [Google Scholar]
- Tan, H.; Bansal, M. LXMERT: Learning Cross-Modality Encoder Representations from Transformers. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, Hong Kong, China, 3–7 November 2019; pp. 5100–5111. [Google Scholar]
- Sun, C.; Myers, A.; Vondrick, C.; Murphy, K.; Schmid, C. VideoBERT: A Joint Model for Video and Language Representation Learning. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 7463–7472. [Google Scholar]
- Lei, J.; Li, L.; Zhou, L.; Gan, Z.; Berg, T.L.; Bansal, M.; Liu, J. Less Is More: ClipBERT for Video-and-Language Learning via Sparse Sampling. In Proceedings of the 2021 IEEE Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 7331–7341. [Google Scholar]
- Schneider, S.; Baevski, A.; Collobert, R.; Auli, M. wav2vec: Unsupervised Pre-Training for Speech Recognition. In Proceedings of the 20th Annual Conference of the International Speech Communication Association, Graz, Austria, 15–19 September 2019; pp. 3465–3469. [Google Scholar]
- Chuang, Y.S.; Liu, C.L.; Lee, H.Y.; Lee, L.S. SpeechBERT: An Audio-and-text Jointly Learned Language Model for End-to-end Spoken Question Answering. arXiv 2019, arXiv:1910.11559. [Google Scholar]
- Yang, J.; Gupta, A.; Upadhyay, S.; He, L.; Goel, R.; Paul, S. TableFormer: Robust Transformer Modeling for Table-Text Encoding. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics, Dublin, Ireland, 22–27 May 2022; pp. 528–537. [Google Scholar]
- Su, H.; Shi, W.; Shen, X.; Xiao, Z.; Ji, T.; Fang, J.; Zhou, J. RoCBert: Robust Chinese Bert with Multimodal Contrastive Pretraining. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics, Dublin, Ireland, 22–27 May 2022; pp. 921–931. [Google Scholar]
- Liu, S.; Li, K.; Li, Z. A Robustly Optimized BMRC for Aspect Sentiment Triplet Extraction. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Seattle, WA, USA, 10–15 July 2022; pp. 272–278. [Google Scholar]
- Zhu, D.; Hedderich, M.A.; Zhai, F.; Adelani, D.I.; Klakow, D. Is BERT Robust to Label Noise? A Study on Learning with Noisy Labels in Text Classification. In Proceedings of the Third Workshop on Insights from Negative Results in NLP, Dublin, Ireland, 26–27 May 2022; pp. 62–67. [Google Scholar]
- Ye, F.; Feng, Y.; Yilmaz, E. ASSIST: Towards Label Noise-Robust Dialogue State Tracking. In Proceedings of the Findings of the Association for Computational Linguistics: ACL 2022, Dublin, Ireland, 22–27 May 2022; pp. 2719–2731. [Google Scholar]
- Nguyen, H.; Khatwani, D. Robust Product Classification with Instance-Dependent Noise. In Proceedings of the Fifth Workshop on e-Commerce and NLP, Dublin, Ireland, 26 May 2022; pp. 171–180. [Google Scholar]
- Su, J.; Murtadha, A.; Pan, S.; Hou, J.; Sun, J.; Huang, W.; Wen, B.; Liu, Y. Global Pointer: Novel Efficient Span-based Approach for Named Entity Recognition. arXiv 2022, arXiv:2208.03054. [Google Scholar]
- Su, J. Efficient GlobalPointer: Less Parameters, More Effectiveness. 2022. Available online: https://kexue.fm/archives/8877 (accessed on 10 June 2023).
- Levenshtein, V. Binary codes capable of correcting deletions, insertions and reversals. Dokl. Akad. Nauk SSSR 1966, 163, 845–848. [Google Scholar]
- Bordes, A.; Usunier, N.; Chopra, S.; Weston, J. Large-scale simple question answering with memory networks. arXiv 2015, arXiv:1506.02075. [Google Scholar]
- Yih, W.T.; Richardson, M.; Meek, C.; Chang, M.W.; Suh, J. The Value of Semantic Parse Labeling for Knowledge Base Question Answering. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, Berlin, Germany, 7–12 August 2016; pp. 201–206. [Google Scholar]
- Kingma, D.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2015, arXiv:1412.6980. [Google Scholar]
- Petrochuk, M.; Zettlemoyer, L. SimpleQuestions Nearly Solved: A New Upperbound and Baseline Approach. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018; pp. 554–558. [Google Scholar]
Question | Golden Entity | Golden Subject |
---|---|---|
what is iqsdirectory.com? | iqsdirectory.com | iqs directory |
in footbal, what position does Tserenjav Enkhjargal play? | Tserenjav Enkhjargal | Tserenjavyn Enkhjargal |
what area is blackwireuk from? | blackwireuk | black wire |
what country is Guanica, Puerto Rico in? | Guanica, Puerto Rico | Guánica |
what label is chrisadamsstringdriventhing under? | chrisadamsstring- driventhing | string driven thing |
Method | Dataset I | Dataset II | SQ | Dataset III | Dataset IV | WQSP |
---|---|---|---|---|---|---|
BERT-CRF | 97.0 | 0 | 90.8 | 79.8 | 0 | 60.0 |
GP | 95.5 | 0 | 89.4 | 80.3 | 0 | 60.4 |
EGP | 95.6 | 0 | 89.5 | 81.8 | 0 | 61.5 |
BERT-CRF + AM | 98.2 | 24.8 | 93.5 | 89.7 | 6.4 | 69.1 |
INNM-I | 98.0 | 27.4 | 93.5 | 91.2 | 6.4 | 70.2 |
INNM-II | 98.7 | 26.1 | 94.1 | 89.8 | 6.2 | 69.1 |
INNM-III | 98.6 | 22.2 | 93.8 | 92.2 | 6.4 | 71.0 |
Method | Accuracy (%) |
---|---|
MemNN-Ensemble [40] | 63.9 |
CFO [1] | 75.7 |
BiLSTM-CRF + BiLSTM [43] | 78.1 |
Structure Attention + MLTA [5] | 82.3 |
BERT-CRF | 82.6 |
GP [37] | 81.3 |
EGP [38] | 81.4 |
INNM-I | 85.1 |
INNM-II | 85.6 |
INNM-III | 85.3 |
Method | WQSP | Dataset V | Dataset VI | Train Time (per Epoch) |
---|---|---|---|---|
BERT-CRF | 60.0 | 30.2 | 16.6 | 469 s |
INNM-I | 70.2 | 37.9 | 21.8 | 376 s |
INNM-II | 69.1 | 41.1 | 25.1 | 469 s + 376 s |
INNM-III | 71.0 | 41.5 | 25.4 | 376 s + 469 s |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Wang, Z.; Xu, X.; Li, X.; Li, H.; Wei, X.; Huang, D. An Improved Nested Named-Entity Recognition Model for Subject Recognition Task under Knowledge Base Question Answering. Appl. Sci. 2023, 13, 11249. https://doi.org/10.3390/app132011249
Wang Z, Xu X, Li X, Li H, Wei X, Huang D. An Improved Nested Named-Entity Recognition Model for Subject Recognition Task under Knowledge Base Question Answering. Applied Sciences. 2023; 13(20):11249. https://doi.org/10.3390/app132011249
Chicago/Turabian StyleWang, Ziming, Xirong Xu, Xinzi Li, Haochen Li, Xiaopeng Wei, and Degen Huang. 2023. "An Improved Nested Named-Entity Recognition Model for Subject Recognition Task under Knowledge Base Question Answering" Applied Sciences 13, no. 20: 11249. https://doi.org/10.3390/app132011249
APA StyleWang, Z., Xu, X., Li, X., Li, H., Wei, X., & Huang, D. (2023). An Improved Nested Named-Entity Recognition Model for Subject Recognition Task under Knowledge Base Question Answering. Applied Sciences, 13(20), 11249. https://doi.org/10.3390/app132011249