Next Article in Journal
A Multi-Task Knowledge-Tracking Model with a Novel Representative Approach to Problem Difficulty and Student Ability
Next Article in Special Issue
From Discourse Relations to Network Edges: A Network Theory Approach to Discourse Analysis
Previous Article in Journal
Clinical Outcome of Class I and II Restorations with and without an Intermediary Layer of a Flowable Composite after 24 Months: A Prospective, Randomized, Split-Mouth-Designed, Controlled and Single-Blinded Clinical Trial
Previous Article in Special Issue
A Chinese Few-Shot Text Classification Method Utilizing Improved Prompt Learning and Unlabeled Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Fine-Tuning BERT-Based Pre-Trained Models for Arabic Dependency Parsing

by
Sharefah Al-Ghamdi
*,
Hend Al-Khalifa
and
Abdulmalik Al-Salman
College of Computer and Information Sciences, King Saud University, P.O. Box 2614, Riyadh 13312, Saudi Arabia
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(7), 4225; https://doi.org/10.3390/app13074225
Submission received: 14 February 2023 / Revised: 24 March 2023 / Accepted: 25 March 2023 / Published: 27 March 2023
(This article belongs to the Special Issue Natural Language Processing: Recent Development and Applications)

Abstract

With the advent of pre-trained language models, many natural language processing tasks in various languages have achieved great success. Although some research has been conducted on fine-tuning BERT-based models for syntactic parsing, and several Arabic pre-trained models have been developed, no attention has been paid to Arabic dependency parsing. In this study, we attempt to fill this gap and compare nine Arabic models, fine-tuning strategies, and encoding methods for dependency parsing. We evaluated three treebanks to highlight the best options and methods for fine-tuning Arabic BERT-based models to capture syntactic dependencies in the data. Our exploratory results show that the AraBERTv2 model provides the best scores for all treebanks and confirm that fine-tuning to the higher layers of pre-trained models is required. However, adding additional neural network layers to those models drops the accuracy. Additionally, we found that the treebanks have differences in the encoding techniques that give the highest scores. The analysis of the errors obtained by the test examples highlights four issues that have an important effect on the results: parse tree post-processing, contextualized embeddings, erroneous tokenization, and erroneous annotation. This study reveals a direction for future research to achieve enhanced Arabic BERT-based syntactic parsing.
Keywords: syntactic parsing; dependency parsing; fine-tuning methods; machine learning; neural networks; deep learning; language models; Arabic natural language processing syntactic parsing; dependency parsing; fine-tuning methods; machine learning; neural networks; deep learning; language models; Arabic natural language processing

Share and Cite

MDPI and ACS Style

Al-Ghamdi, S.; Al-Khalifa, H.; Al-Salman, A. Fine-Tuning BERT-Based Pre-Trained Models for Arabic Dependency Parsing. Appl. Sci. 2023, 13, 4225. https://doi.org/10.3390/app13074225

AMA Style

Al-Ghamdi S, Al-Khalifa H, Al-Salman A. Fine-Tuning BERT-Based Pre-Trained Models for Arabic Dependency Parsing. Applied Sciences. 2023; 13(7):4225. https://doi.org/10.3390/app13074225

Chicago/Turabian Style

Al-Ghamdi, Sharefah, Hend Al-Khalifa, and Abdulmalik Al-Salman. 2023. "Fine-Tuning BERT-Based Pre-Trained Models for Arabic Dependency Parsing" Applied Sciences 13, no. 7: 4225. https://doi.org/10.3390/app13074225

APA Style

Al-Ghamdi, S., Al-Khalifa, H., & Al-Salman, A. (2023). Fine-Tuning BERT-Based Pre-Trained Models for Arabic Dependency Parsing. Applied Sciences, 13(7), 4225. https://doi.org/10.3390/app13074225

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop