Next Article in Journal
On the Extended Adjacency Eigenvalues of a Graph
Next Article in Special Issue
The Localization of Software and Video Games: Current State and Future Perspectives
Previous Article in Journal
Detection of Random Body Movements Using Clustering-Based Methods in Bioradar Systems
Previous Article in Special Issue
Human versus Neural Machine Translation Creativity: A Study on Manipulated MWEs in Literature
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Exploring the Potential of Neural Machine Translation for Cross-Language Clinical Natural Language Processing (NLP) Resource Generation through Annotation Projection

by
Jan Rodríguez-Miret
1,
Eulàlia Farré-Maduell
1,
Salvador Lima-López
1,
Laura Vigil
1,
Vicent Briva-Iglesias
2 and
Martin Krallinger
1,*
1
Barcelona Supercomputing Center (BSC), 08034 Barcelona, Spain
2
School of Applied Languages and Intercultural Studies, Dublin City University, D09 V209 Dublin, Ireland
*
Author to whom correspondence should be addressed.
Information 2024, 15(10), 585; https://doi.org/10.3390/info15100585
Submission received: 2 August 2024 / Revised: 10 September 2024 / Accepted: 19 September 2024 / Published: 25 September 2024
(This article belongs to the Special Issue Machine Translation for Conquering Language Barriers)

Abstract

:
Recent advancements in neural machine translation (NMT) offer promising potential for generating cross-language clinical natural language processing (NLP) resources. There is a pressing need to be able to foster the development of clinical NLP tools that extract key clinical entities in a comparable way for a multitude of medical application scenarios that are hindered by lack of multilingual annotated data. This study explores the efficacy of using NMT and annotation projection techniques with expert-in-the-loop validation to develop named entity recognition (NER) systems for an under-resourced target language (Catalan) by leveraging Spanish clinical corpora annotated by domain experts. We employed a state-of-the-art NMT system to translate three clinical case corpora. The translated annotations were then projected onto the target language texts and subsequently validated and corrected by clinical domain experts. The efficacy of the resulting NER systems was evaluated against manually annotated test sets in the target language. Our findings indicate that this approach not only facilitates the generation of high-quality training data for the target language (Catalan) but also demonstrates the potential to extend this methodology to other languages, thereby enhancing multilingual clinical NLP resource development. The generated corpora and components are publicly accessible, potentially providing a valuable resource for further research and application in multilingual clinical settings.

1. Introduction

Recent advances in artificial intelligence (AI) technologies hold significant promise for the development of clinical natural language processing (NLP) applications [1]. These applications are crucial for the healthcare domain, where accurate and timely information extraction from clinical texts can significantly enhance patient care, research, and administrative processes [2,3,4,5]. Named entity recognition (NER) systems play a vital role by identifying and classifying key information types such as medications, symptoms, and diagnoses within clinical documents [6,7,8].
Clinical NER systems have demonstrated considerable impacts. For instance, Uzuner and colleagues [9] showcased the use of NER to extract medical problems, treatments, and tests from clinical narratives, facilitating the creation of comprehensive patient summaries. Similarly, Luo et al. [10] analyzed the use of NER systems to identify adverse drug events from clinical documents, improving pharmacovigilance and patient safety. These systems have helped optimize clinical workflows, enhancing decision support systems, and contributing to large-scale health data analysis.
However, the development of robust NER systems is contingent upon the availability of substantially annotated corpora. These corpora must be meticulously annotated by clinical experts to ensure accuracy, a process that is both time-consuming and costly, and this can entail extensive privacy risks [11]. Furthermore, annotated corpora are often language-specific, posing significant challenges in multilingual contexts where resources may be scarce for less widely spoken languages. Large clinical studies, the characterization of rare diseases, and international medical projects require that clinical information is extracted comparably, which is particularly challenging due to the lack of multilingual clinical corpora annotated using similar data labeling criteria. Most of the current freely accessible resources are limited to data in English [12], and the use of clinical corpora in English as a base to generate resources in other languages and speed up NLP resource generation might show certain limitations due to the differences in syntax, grammar, and morphology, as well as lexical differences between languages [13].
It is also worth stressing that clinical content is typically written in the local language or even language variants spoken in each region or country [14]. In some geographical regions, clinical records might even correspond to textual narratives written in multiple languages, therefore presenting code-switching phenomena [15]. Such is the case with some Spanish regions, like Catalonia, where clinical health records are found both in Spanish and in Catalan, resulting in multilingual content written in similar languages.
In this context, current advances in neural machine translation (NMT) might facilitate the creation of annotated corpora and the application of language technologies in multiple languages [16]. NMT systems can translate annotated corpora from one language to another while preserving the integrity and context of the original annotations [17,18]. This potential is enhanced within languages of the same family, like Romance languages, because of their similarity [19,20]. Consequently, this study explores the use of a state-of-the-art NMT system (Softcatalà Translator) to translate three Spanish clinical case corpora into Catalan [21]. The original Spanish corpora, meticulously annotated by clinical experts, serve as the basis for this translation. To facilitate the transfer of annotations from Spanish to Catalan, we employ an annotation projection technique (explained in detail in Section 3). This method bypasses the need for extensive manual annotation of the Catalan corpora, thus saving significant time and resources.
This study details a strategy exploiting NMT, annotation projection, and human-in-the-loop expert validation to foster the development of comparable clinical NER systems and the generation of annotated corpora, together with a detailed evaluation and comparison of different NER systems. Under this application scenario, the use of NMT can effectively save time in manually annotating corpora. As a result, we have generated the first clinical NER components for three high-impact entity types in Catalan. Through this study, we aim to provide insights into the feasibility and benefits of leveraging NMT and annotation projection for developing robust NER systems in multilingual clinical settings.

2. Neural Machine Translation for Annotation Projection

The advent of NMT in the past decade has revolutionized MT, utilizing deep learning techniques to enhance accuracy and contextual understanding [22,23]. This has translated into considerable effort over the past years to generate MT systems, not only for general domain purposes but also for health-related documents [24,25,26]. Most medical MT efforts focused on translations between English and other languages [27], covering a broad variety of practical use case scenarios, such as the generation of medical dictionaries or glossaries, automatic clinical coding [28], the processing of patient–physician interviews [29], or the generation of translation systems for low-resource languages [25]. MT and word alignment strategies have been used between English and German to generate German NER models for medical semantic annotation of medication mentions [16,30], and for Spanish and French in the case of tumor morphology concepts [31].
Parallel and comparable biomedical corpora have also been explored to generate bilingual medical glossaries and terminological resources by exploiting different word alignment approaches and clinical vocabularies, for instance, between content in English and French [32] or English and Spanish [33].
Despite recent advances in NMT technologies and the increasing use of large language models (LLMs) for translation-related tasks [34,35], only limited attempts have been made so far to exploit existing commonalities across medical texts in Romance languages [36]. In this paper, we propose a strategy to generate medical corpora more efficiently across languages by exploiting an NMT system for the development of clinical entity tagger resources.
The proposed approach relies on the use of NMT, followed by the expert validation of existing recently released and widely used Gold Standard clinical corpora available in Spanish to train and evaluate NER tools, in this case for Catalan. Figure 1 shows an overview of the proposed methodology. The sections below explain in detail the study conducted.

3. Materials and Methods

The number of corpora manually annotated by clinical experts available in Romance languages is scarce. This is particularly true for lower-resource languages such as Catalan, Romanian or Galician. Recent community evaluation efforts and shared tasks have resulted in the release of a number of medical corpora in Spanish [37,38,39,40]. These have been widely exploited for the development and benchmarking of clinical NER systems [41]. This section describes the exploitation of three clinical corpora with NMT and annotation projection approaches to generate annotated resources for the Romance language Catalan.

3.1. Clinical Corpora: DisTEMIST, DrugTEMIST and MEDDOPROF

Three separate datasets were chosen for the annotation projection experiments, namely DrugTEMIST [42], DisTEMIST [43], and MEDDOPROF [44]. These are all Gold Standard corpora comprising clinical case reports written in Spanish that have been annotated manually by clinical domain experts. The annotations followed an iterative process in which different annotation guidelines were created and refined to ensure the quality and consistency of the final output. In addition, these corpora have been publicly validated as part of shared tasks.
The selection of multiple datasets with varying entity types allowed us to test the methodology’s effectiveness across a broad spectrum of clinical information, thereby providing a comprehensive assessment of its potential utility in multilingual clinical NLP applications. On the one hand, DisTEMIST and DrugTEMIST cover annotations of diseases and medications, respectively, as reported in 1000 clinical case reports in Spanish per corpus that include medical, surgical, and dentistry specialties. In the case of the DisTEMIST corpus, the annotation consistency, measured through a pairwise agreement between independent annotators, reached 82.3% IAA (exact disease mention offsets) [45,46]. The DrugTEMIST corpus was annotated by the same team of clinical experts following detailed annotation guidelines with an IAA of 95.50%. On the other hand, MEDDOPROF covers information on the social determinants of health found in clinical texts (namely occupations and working statuses), an entity type that is not specific to the medical domain and that often appears in descriptive form. It was annotated using the same methodology as the other two corpora described, with an IAA of 90%. Table 1 provides some basic statistics for the three datasets.

3.2. Corpus Translation and Annotation Projection

In essence, the proposed annotation projection leverages labeled data or text-bound annotations from a source language (source corpus) into a different target language (target corpus). In the case of span-based annotations (i.e., for named entity recognition scenarios), this implies mapping the annotations of a given text in the source language into the translated text of the target language. For our experiments, we propose a lexical annotation projection, consisting of a more strict mapping strategy particularly focusing on human annotation correction efficiency because it decreases the need to correct matching boundary issues often encountered by more advanced sentence-alignment-based matching alternatives. Unlike previous approaches, which mainly relied on sentence alignment techniques to detect correspondences between source and target concept mentions [32], the source text and the source annotations are translated separately into the target language, and the annotations are then mapped onto the target text using term look-up methods. Some publicly released corpora that have tested this methodology include DisTEMIST [43], MEDDOPLACE [38], and SympTEMIST [39].
In the case of our approach, the translations from Spanish to Catalan were generated using the SoftCatala API (https://www.softcatala.org/traductor/) (accessed on 18 September 2024). The choice of Softcatalà as the NMT system was influenced by several factors. First, Softcatalà Translator is an open-source system, which allows other researchers to freely access and replicate our methodology, contributing to the collective advancement of NMT for clinical NLP applications [47]. Second, ensuring the privacy of clinical data is paramount, and we could handle sensitive datasets without exposing them to potential security risks associated with other commercial NMT services. This compliance with ethical standards and legal regulations regarding patient data is crucial. Finally, Softcatalà Translator is renowned for its high-quality translations between Romance languages, displaying state-of-the-art results comparable with major NMT systems [21]. Maintaining the accuracy of translated annotations directly impacts the effectiveness of the annotation projection method and the performance of NER systems.
Consequently, the output of the annotation projection process can be considered as a sort of Silver Standard corpus, because the generated annotations in the target language text are not perfect and may contain missing mentions not matched or mapped to the target text, as well as incorrect annotations (due to either ambiguity, as is often the case for abbreviations, or annotations with incorrect boundaries).
The quality of the annotation projection process is dependent on the underlying NMT system quality, the correctness of the source text, the medical expressions used, the potential ambiguity of the annotated entity mentions, or the similarity of the source and target languages. Even then, machine-translated corpora can be quite valuable for multiple reasons, especially for under-resourced languages. On the one hand, the automatically mapped annotations can be directly validated or corrected by native speakers with domain expertise to generate a new Gold Standard training or test set fast and efficiently. On the other hand, we wanted to test whether models trained on non-validated data (in a different language to that of the source corpora) can still provide sufficiently good results when evaluated on a test set that had been manually corrected by domain experts.
Table 2 shows an overview of the result of the annotation projection process of the three previously mentioned corpora from Spanish (source language) to Catalan (target language). The statistics related to the number and percentage of projected mentions are calculated as part of the annotation projection process to give an estimate of the completeness of the projection, but they are not necessarily representative of the quality of the final output (i.e., a mention might be marked as mapped, but that does not mean that it is correct, as it might be a false positive or have the wrong mention boundaries). Significantly, the majority of the Gold Standard annotations could be mapped onto the Catalan translations. This might be attributed to the closeness between both languages (Spanish and Catalan have 85% similarity) [48], the translation quality, as well as the existence of similar morphological, syntactic, and even etymological patterns. This is especially relevant in the medical domain, where many clinical terms in both languages share the same Latin and Greek roots (DisTEMIST corpus). In the case of medication and drug names, these were almost always substance names with identical word roots for both languages (DrugTEMIST corpus). Since occupation mentions (MEDDOPROF corpus) sometimes correspond to more descriptive expressions, it is expected that the percentage of non-projected annotations in this corpus is somewhat higher than for the other two corpora.
Crucially, and despite the seemingly good results, we should emphasize that, without human validation, it is not possible to truly assess the extent to which the resulting NMT and annotation projection are correct. It might be the case that the projected data include false positives or incorrectly translated entities. For this reason, we incorporate a final step: human validation and the correction of the annotation projection.

3.3. Corpus Validation and Correction Process

To be able to benchmark and evaluate the generated entity recognition results on systems trained on automatically projected annotations, having a test set manually validated and corrected by clinical domain experts is crucial.
For this experiment, two clinical experts, a professional translator, and one linguist were asked to validate the test set subsets of all three corpora. They had ample experience in text data annotation and are native Spanish and Catalan speakers. The validation step was performed using the annotation tool brat’s side-by-side document comparison mode [49]. In addition to the test set, to explore the difference between training systems using the annotation projection output as is or using a corrected version, the experts were also asked to validate the train subsets of the corpora.
As shown by Figure 2, annotators were shown the Gold Standard annotated documents in Spanish on the left side and the projected document in Catalan on the right side. The main task for annotators was to validate the Catalan version, adding, removing, or modifying annotations to make the projected version as close to the Gold Standard as possible. The annotation of a Gold Standard corpus from scratch requires annotators to carefully read the entire document and annotate mentions without any aids. In contrast, after NMT and annotation projection, a fast scan of the document usually suffices. The side-by-side modality further eases the annotators’ work by enabling direct comparisons between the source and target texts. This method significantly reduces the annotation effort because annotators can focus on verifying and correcting the pre-existing annotations rather than creating them de novo.
A second task given to annotators was to verify that the annotated concepts were correctly translated and idiomatic. When the concept was incorrect, they were asked to suggest alternative translations. It is important to note that only the annotated concepts were corrected, since correcting the entire machine-translated text goes beyond the scope and point of the experiment. Annotators used the notes section in brat to propose improved translations for each individual annotation. During the data post-processing step, the translated texts were reconstructed. Here, incorrectly translated concepts were replaced by their proposed alternative, adjusting the character spans of all annotations as required.
Finally, to evaluate the resulting systems on non-machine-translated data, we gathered a set of clinical case reports in Catalan to be annotated by the three experts with the same set of labels (that is, diseases, medications, and occupations). This data collection, named CataCCC, consists of case reports of various clinical specialities extracted from open journals. This method mimics the document selection used for the DisTEMIST and DrugTEMIST corpora. CataCCC includes a total of 200 documents with 72,003 tokens and 400,378 characters. It has 2557 disease annotations, 531 medication annotations, and 142 occupations. The corpus is openly available for download, along with the rest of the data discussed in this paper and the annotation projection correction guidelines, in Zenodo at https://zenodo.org/doi/10.5281/zenodo.13133124 (accessed on 18 September 2024).

3.4. Clinical Named Entity Recognition Models

Transformer-based pre-trained language models have been increasingly used for NER tasks, also in the biomedical domain [50,51]. These models are trained on general tasks, like masked or causal language modeling, in a self-supervised manner. After that, the pre-trained models can be leveraged for different downstream tasks (e.g., text classification, NER, etc.) if properly fine-tuned using supervised data. To analyze the practical usefulness of the generated annotation projection corpora, we have trained several NER systems on the original Spanish corpora and the machine-translated Catalan corpora with the corrected mentions from the annotation projection. These systems were evaluated against human-annotated/validated test set annotations in all cases: (a) their corresponding Spanish test set, (b) the Catalan corrected test set, and (c) the CataCCC corpus, which is only used for evaluation. In this section, the developed NER models will be presented, along with the data used for their training and evaluation.
We formulate the task as a token classification problem, where each token is encoded using the IOB2 schema [52]. As a simple multi-class problem (all tokens must have exactly one class), in the case of nested mentions, only the longest one is kept for training. The model is, thus, unable to produce nested entities, but the percentage of these in the datasets is low, so the simplicity of this method outweighs the presumably negligible negative impact it may have. This approach, i.e., fine-tuning a pre-trained model on a token classification task, is also used by the top-scoring teams in the MEDDOPROF and DisTEMIST shared tasks [53,54].
As a base model, we used a RoBERTa-based model (124M parameters) pre-trained on both clinical and biomedical corpora, accounting for 91M tokens and 968M words, respectively [55]. We will refer to this model as RoBERTa bsc-bio-ehr-es. For each corpus, one-third of the original training set was selected as the validation or development set (250 documents for DisTEMIST and DrugTEMIST and 500 documents for MEDDOPROF), which is used to evaluate the model at each epoch during training. The same splits are used for all experiment configurations and languages. For the loss computation, we used the cross-entropy loss implementation provided by Pytorch.
The label ACTIVIDAD (activity) was removed from the MEDDOPROF corpus after initial experimentation since the number of mentions is quite low (147 in the whole dataset) and they are semantically broad and complex. The rest of mentions and entities are kept as in the original corpora.
Notably, the dataset is mostly composed of O tags when encoded with the IOB2 schema, while some tags like B-SITUACION_LABORAL only account for 0.09% of tokens. To compensate for this imbalance, we tried different weight strategies for computing the loss:
  • none: all tokens have the same importance.
  • freq: each token has a weight inversely proportional to the frequency of its ground truth class (IOB tag) in the training split.
    w e i g h t c l a s s i = n t o k e n s n t o k e n s c l a s s i · n c l a s s e s ,
    where n t o k e n s is the total number of tokens, n t o k e n s c l a s s i is the number of tokens for class i, and n c l a s s e s is the number of classes.
  • freq_sqrt: each token has a weight inversely proportional to the square root of the frequency of its ground truth class.
    w e i g h t c l a s s i = n t o k e n s n t o k e n s c l a s s i · n c l a s s e s
Note that both freq and the smoother freq_sqrt encourage a higher recall for entities, as false negatives (FN) are more penalized than false positives (FP).

3.4.1. Model Selection

For the MEDDOPROF corpus (Spanish), we also tested a general domain pre-trained model named BETO [56], but it was discarded after early experimentation in favor of the RoBERTa bsc-bio-ehr-es, which was consistently outperforming BETO. The models were trained in an HPC environment, where each node/experiment had 2 GPUs (AMD Radeon Pro VII, 16 GB each), although their availability was not exclusively for the purpose of this study.
A hyperparameter optimization (HPO) exploration was conducted through the Weights and Biases Sweeps feature. We used the same sweep configuration for both Spanish and Catalan, running 15–30 random configurations for each corpus and language. The different possible values explored are listed below: batch size: { 8 , 16 , 32 } , learning rate: [ 1 7 , 1 4 ] , classifier dropout: [ 0 , 0.8 ] , weight decay: [ 0 , 0.03 ] , warmup ratio: { 0 , 0.1 } , weight strategy: {none, freq, freq_sqrt}, number of epochs: { 15 , 20 } .
The best model is selected as the one with the highest strict mention-wise F1-score at some epoch k in the validation set. There are two exceptions to this: in both Spanish DisTEMIST and MEDDOPROF, the model with the highest F1-score did not apply the dropout regularization technique (i.e., clasifier dropout = 0). The difference in performance regarding the second-highest F1-score was less than 1 percentage point in both cases, but these had the dropout set to a value of 0.3 and 0.8, respectively. We hypothesized that these latter models would result in a better generalization and thus a higher performance in the test set. After all the experiments, we evaluated the former two discarded models in the test set and confirmed that their performance was 1.2 percentage points below the chosen regularized ones.

3.4.2. Model Availability

To ensure transparency and reproducibility and encourage further research on the use of NMT and annotation projection in languages other than English, all models appearing in this article are publicly available at Hugging Face: https://huggingface.co/collections/BSC-NLP4BIA/clinical-nmt-ner-66a7badd9189298acbbe504f (accessed on 18 September 2024). More details on each model, licence, and how to use them can be found on each Model Card.
As the initial baseline, we trained models on the original annotation projection data without human correction. These are noted with the suffix “-v0” in the above Hugging Face collection. Associated details are provided as additional materials in Zenodo.

4. Results

4.1. Cross-Language Model Evaluation

To evaluate the transferability of the Spanish NER corpora to Catalan, we trained a transformer-based model on each original Spanish training set corpus and on the corrected training set version of the Catalan automatically transferred corpus.
All results can be seen in Table 3. As expected, the best-performing models in both the Catalan test set and CataCCC are the ones trained on the Catalan corrected version of the corpora, while the models trained on the original Spanish corpora achieved the best metrics in their corresponding test sets. Interestingly, the Spanish model for drugs obtained a higher precision than the Catalan corrected version when evaluated on the test set of the latter (0.886 vs. 0.885), though with a much lower recall (0.809 vs. 0.874).
Note that the models trained and evaluated on the original Spanish versions set a higher bound or reference in performance regarding the Catalan version. That is, it was not expected to surpass these performances for the settings trained and evaluated on the Catalan version, as the RoBERTa base model was trained on Spanish texts.
For reference, in their respective original shared tasks, using the Spanish data, the winner team for the NER sub-track of DisTEMIST obtained an F1-score of 0.77 [54]; the winner of MultiCardioNER, the shared task where DrugTEMIST was used, obtained an F1 of 0.92 [57]; and the winner of MEDDOPROF obtained an F1 of 0.81 [53]. Their results are slightly higher than those obtained by the Spanish models reported here, although it must be kept in mind that the shared tasks systems are more complex and include components such as Conditional Random Fields (CRF).
It is important to emphasize the fact that F1-scores increase largely when not considering the exact boundary of a mention but just some overlap between the prediction and the ground truth. This is especially true for DisTEMIST, which has the longest mentions (see Table 4).

4.2. MT Error Analysis

Even if our primary focus is on using NMT for data generation in clinical NLP through an annotation projection technique, understanding how MT errors can impact the quality and accuracy of the projected annotations is crucial. The three expert annotators assessed the MT quality by using the error categories in the Multidimensional Quality Metric (MQM) framework for human evaluation of translation quality, namely Terminology, Accuracy, Linguistic Conventions, and Style [58]. Following this strict methodology [59,60], we identified areas for MT improvement, which impacted the annotation projection. The MT quality analysis was conducted separately for each corpus, since the impact the MT results had on the annotation projection was different for each of the three entities (diseases, medications and occupations).
The DisTEMIST corpus contains a high number of annotated disease entities per document. Diseases have specialized, in-domain terminology, and, as expected, most MT errors in this corpus fall into the Terminology and Accuracy categories because the MT system failed to adequately translate some of these entities. One common Terminology error identified was the MT system producing literal translations, such as the case of “enfermedad de 3 vasos” (disease affecting three vessels), which was machine-translated as “malaltia de 3 gots” (disease of three glasses) instead of the adequate “malaltia de 3 vasos”, mixing up key concepts. In terms of Accuracy, one common error was not translating the entity at all, like leaving in Spanish “bifidez piélica” (bifid pelvis) instead of translating the entity into Catalan as “bifidesa pièlica”. Regarding Linguistic Conventions, there were some typos, as in “metástasis” instead of “metàstasi”. Surprisingly, the quality of the MT generally did not hinder the annotation transfer in this corpus. Despite the MT errors, the robustness of the annotation projection technique ensured that all the above disease annotations were accurately transferred.
The DrugTEMIST corpus, which contains annotations of medicines, presented unique challenges. Proper nouns and specific drug names were particularly prone to errors, often falling into the Terminology and Accuracy categories. For instance, the MT system frequently mistranslated drug names by either altering their endings or failing to detect them as proper nouns. A common Terminology error was translating nouns as verbs, such as in the case of “sales de fosfato” (phosphate salts) as “surts de fosfat” (you leave phosphate). Accuracy errors included cases where the MT system did not translate the term at all, leaving “ácido folínico” untranslated instead of converting it to “àcid folínic”. Additionally, there were instances of the inappropriate translation of abbreviations and symbols, such as “5-ASA” being incorrectly translated to “5-NANSA” (in Spanish, “asa” can refer to mesalazine in the medical domain or to a bag handle, which is the option that the MT system chose). These errors highlight the need for meticulous verification of medical terminology in MT systems to ensure accurate annotation projection.
Finally, the MEDDOPROF corpus, which deals with occupations, demonstrated different types of MT errors. Given the descriptive nature of the annotations of this latter corpus, the primary issues were related to Terminology and Linguistic Conventions. Abbreviations posed significant challenges; for example, “estudiante de 3º ESO” (student of 3rd year of secondary school) was often incorrectly translated to “estudiant de 3º ESO” instead of “estudiant de 3r ESO”. The error in the translation of the abbreviation caused every annotation to contain an abbreviation that was not properly transferred to the Catalan corpus. Terminology errors included the mistranslation of terms such as “médico” (doctor) to “mèdic” (medical) rather than the correct “metge” (doctor). Another similar case was the mistranslation of “empleado en Carpintería metálica” (worker in metallic carpentry) to “emprat en Carpinteria metàl·lica” (used in metallic carpentry). In these cases, the annotation was not projected adequately. These errors underline the importance of high-quality MT systems for accurate annotation projection, especially in clinical contexts where precision is crucial. Further implications on the MT are addressed in Section 5.

4.3. Annotation Projection and NER System Error Analysis

In this section, we analyze the errors encountered during the annotation projection and the subsequent performance of the NER systems trained on the projected annotations. This analysis aims to identify the common sources of errors and their potential impacts on the system’s overall performance.
In the qualitative analysis of the annotations, we could observe some false positives and false negatives. The most prevalent type of false positive error involved substances that could be both medicinal products and physiological substances, such as potassium and albumin. In the original Spanish version, these were only annotated when referring to medicinal products. However, in the Catalan version, the annotation projection’s lexical approach also considered these concepts when describing blood test results, leading to over-annotation. False negatives were most commonly observed with medicinal products that included specific concentrations, such as sodium chloride 0.9% solution. These cases often required manual annotation in the target language due to incomplete projection.
Using the CataCCC corpus for evaluation, we classified model predictions into correct, incorrect, partial, missing, and spurious categories to gain a more profound understanding of performance. This methodology was employed by following the MUC-5 evaluation [61]. We observed that incorrect boundaries were a frequent issue, often due to variations in the level of detail in the annotations.
For drug annotations (DrugTEMIST), the model faced challenges with special characters like the Catalan “l·l”, apostrophes, and other symbols such as percentages. These issues are likely related to inadequate tokenization for these characters. For example, al·lopurinol and meticil·lina were often incorrectly tokenized, affecting the model’s ability to correctly identify these terms.
In the DisTEMIST corpus, many predictions included extended descriptions or adjectives that were not present in the Gold Standard corpus, reflecting the difficulty in standardizing the level of detail for disease mentions. Additionally, many spurious (totally false positive) and missing (totally false negative) mentions can also be considered symptoms, which illustrates the blurred line between the two semantic classes.
For occupations (MEDDOPROF), incorrect boundary predictions were common, often due to varying levels of descriptive detail. Some mentions were particularly difficult for the model to identify as occupations or employment statuses, given their use in common language rather than specific occupational terms. For example, “diagnostic radiology technician” and “on disability benefits” required a precise understanding of context, which was often lacking.

5. Discussion

Overall, the results obtained indicate that the use of NMT, together with an annotation projection technique, might be used to generate medical NLP training data. We achieved competitive results for our target language compared with the source language. Moreover, we have tested this strategy for automatic clinical semantic annotation focusing on three distinct relevant entity types, namely medications, diseases, and occupations. Each entity type has specific linguistic characteristics in terms of MT results when considering the source and target languages of our experiment. Medications often correspond to short substance or trade name mentions with certain morphological characteristics, suffixes, and prefixes that follow an established pattern between the two Romance languages of our study. In the case of disease mentions, these often consisted of complex, long, multi-word expressions built up by words with Latin or Greek roots. Occupations, on the other hand, are not exclusive to medical content. Nevertheless, when encountered in clinical narratives, they usually appear both in the form of proper occupation names or as more descriptive expressions.
When using transformer-based clinical NER systems with biomedical RoBERTa, we observed that the performance only degraded slightly when the system was trained on the annotation projection dataset and applied to another similar language, i.e., Catalan. The best results were obtained for medication mentions.
Our results indicate that the reuse of existing clinical corpora for similar language adaptation scenarios could be a promising strategy to accelerate the annotation process and generate resources of importance under data scarcity, as is the case for clinical corpora. In our case, due to the existence of clinical data written in both Spanish and Catalan, accounting for a population of over 10 million people, unlocking clinically relevant information requires systems that work well for content in either language. For this reason, the machine translation of texts from Spanish to Catalan for the development of NLP resources and systems in Catalan might considerably enhance patient comprehension and communication by providing medical information in people’s native language within hospitals. It could also improve operational efficiency and ensure consistent medical terminology, although it requires careful review for accuracy and privacy compliance.
The three steps of the corpus construction experiment, namely, annotation projection, projection validation, and correction of translated mentions, have underscored important factors to be considered when leveraging resources between different languages. Firstly, improvements in the quality of the MT system would result in a better projection of annotations. The corpora written mainly in Castilian Spanish (DisTEMIST and DrugTEMIST) seemed to have somewhat better translation results when compared to the panhispanic MEDDOPROF corpus, which covered clinical cases not only from Spain but also from a diversity of Latin American countries. Efforts are being made to record the homogeneity and heterogeneity of the medical language shared by all Spanish speakers covering the entire human geography and language variants but are currently limited to medical terminologies rather than MT aspects [62], but a fine tuning of MT systems in the medical domain may facilitate these tasks [63].
Aspects related to language variants and the adaptation of annotation projection and automatic semantic annotation systems to similar languages would require a more in-depth analysis to determine the impact and issues to be addressed. An accurate translation is needed not only for the entity mentions per se but also for the grammatical elements dependent on it, since Romance languages use declensions universally. Other issues are associated with ambiguous words and abbreviations in the original language, which are sometimes translated incorrectly in the target language. For example, in Spanish, ampolla means ampule and blister (ampul·la and butllofa in Catalan, respectively).
The system used for annotation projection also showed some inconsistencies. For instance, the Spanish medication mention amoxicilina-ácido clavulánico was found sometimes correctly translated into amoxicil·lina-àcid clavulànic, and in other instances projected directly from Spanish without changes. Additionally, the resulting projection did not account for the actual context and annotation semantics of a particular mention. In the original Spanish text, magnesium was correctly labeled only when used as a medication, while in the target language, it was sometimes incorrectly transferred where it corresponded to the physiological ion (semantic discrepancies). The projection of annotations was more challenging for entities that consist of noun phrases. One example is the case of occupations and work statuses, diagnostic radiology technician and on disability benefits, respectively. Regarding diseases, the length of the mention can vary greatly, particularly with the addition of anatomical parts affected by the condition. Examples of shorter and longer mentions are brucellosis and right sphenoid wing en plaque meningioma.
Interestingly, this annotation projection methodology was tested in a shared task part of the BioASQ initiative, MultiCardioNER [42]. A new version of DrugTEMIST in English and Italian was released and used as part of the task, as well as a new corpus called CardioCCC, also annotated with medications. In the MultiDrug sub-track, participants were challenged to create systems to detect medications in Spanish, English, and Italian, being evaluated on a test set of cardiology clinical case reports. The multilingual data of this task were created using the same methodology reported in this paper with the help of clinicians who were native speakers of each language. In a similar fashion to our experiments, the results across all languages are quite comparable, once again pointing toward the validity of annotation projection as a means to create multilingual data.

6. Conclusions

We have shown that annotation projection via NMT, human expert correction and transformer-based NER can yield promising results to generate resources for clinical NLP for similar languages scenarios under the assumption that high-quality MT systems are in place. We have also made available one of the first clinical NER Gold Standard corpora and systems of diseases, drugs, and professions in Catalan: https://zenodo.org/doi/10.5281/zenodo.13133124 (accessed on 18 September 2024). These information types, especially drugs and diseases, are very high-impact information types in a clinical setting, and the models trained for them achieve very competitive results when compared to their Spanish counterpart.
The proposed approach could potentially also be explored for other currently accessible clinical corpora in Spanish (e.g., CANTEMIST [64], PharmaCoNER [65], MEDDOPLACE [38], MedProcNER [40], or SympTEMIST [39]) and languages similar to Spanish like Portuguese (both European and Brazilian) or Galician. As MT resources from Spanish into multiple Romance languages are currently available, including also French, Italian, Romanian, Corsican, or Haitian Creole, this could allow for systematically testing the proposed strategy with languages showing different degrees of similarity. So far, we have generated additional, non-validated annotation projection corpora versions of the described datasets for English, French, Italian, Portuguese, Romanian, Czech, Dutch, and Swedish. These versions are available to download in Zenodo together with the rest of the data described in this article.
Furthermore, to address data augmentation, the use of different MT systems or back-translation, as well as the use of generative models and synthetic clinical texts, could be tested to increase the number of training instances with a particular focus on entity-aware NMT results. It is critical to keep in mind that the proposed data augmentation strategies, also when combined with more traditional deep learning approaches applied to health-related content, needs to also account for reliability and fairness [66]. Alternative settings that are showing interesting outcomes relate to the combination of LLMs, prompt engineering [67], and human–AI collaborations (crowd-workers and domain experts) to generate expert-annotated datasets [68]. Although clinical records and case reports are a valuable source of clinical content with high impact associated practical use case scenarios, these are typically generated by clinical experts or researchers, not covering health-related content directly generated by patients. Adapting the proposed cross-language corpus annotation strategies described in this article to transcribed conversational health datasets for real-world applications could complement previously used annotation schemas and data labeling approaches for selling dialogues [69,70].
Finally, it would be interesting to explore how the proposed setting would work for other application domains, genres, or content types such as social media or user/patient-generated content, as well as for the creation of multilingual models. Different base models and architectures could also be tested to assess the generalization of this method further.

Author Contributions

Conceptualization, M.K.; methodology, J.R.-M.; software, J.R.-M.; validation, E.F.-M., L.V. and V.B.-I.; formal analysis, J.R.-M., E.F.-M., S.L.-L., L.V., V.B.-I. and M.K.; investigation, J.R.-M., E.F.-M., S.L.-L., L.V., V.B.-I. and M.K.; resources, S.L.-L., E.F.-M., L.V. and V.B.-I.; data curation, S.L.-L., E.F.-M., L.V. and V.B.-I.; writing—original draft preparation, S.L.-L. and M.K.; writing—review and editing, S.L.-L., E.F.-M. and M.K.; visualization, J.R.-M.; supervision, M.K.; project administration, M.K.; funding acquisition, M.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Ministerio de Ciencia e Innovación (MICINN) under project AI4ProfHealth (PID2020-119266RA-I00 MICIU/AEI/10.13039/501100011033) and BARITONE (TED2021-129974B-C22). This work is also supported by the European Union’s Horizon Europe Co-ordination & Support Action under Grant Agreement No 101080430 (AI4HF), as well as Grant Agreement No 101057849 (DataTool4Heartproject).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The original data presented in the study are openly available in Zenodo at https://zenodo.org/doi/10.5281/zenodo.13133124 (accessed on 18 September 2024).

Acknowledgments

We acknowledge Miguel Rodríguez Ortega, Luis Gascó Sanchez and Darryl Johan Estrada for collaborations and previous work related to neural machine translation and biomedical WMT shared task activities relevant for some of the initial experiments that motivated this work.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Al Kuwaiti, A.; Nazer, K.; Al-Reedy, A.; Al-Shehri, S.; Al-Muhanna, A.; Subbarayalu, A.V.; Al Muhanna, D.; Al-Muhanna, F.A. A Review of the Role of Artificial Intelligence in Healthcare. J. Pers. Med. 2023, 13, 951. [Google Scholar] [CrossRef] [PubMed]
  2. Houssein, E.H.; Mohamed, R.E.; Ali, A.A. Machine Learning Techniques for Biomedical Natural Language Processing: A Comprehensive Review. IEEE Access 2021, 9, 140628–140653. [Google Scholar] [CrossRef]
  3. Kundeti, S.R.; Vijayananda, J.; Mujjiga, S.; Kalyan, M. Clinical Named Entity Recognition: Challenges and Opportunities. In Proceedings of the 2016 IEEE International Conference on Big Data (Big Data), Washington, DC, USA, 5–8 December 2016; pp. 1937–1945. [Google Scholar] [CrossRef]
  4. Pagad, N.S.; Pradeep, N. Clinical Named Entity Recognition Methods: An Overview. In Proceedings of the International Conference on Innovative Computing and Communications, Delhi, India, 20–21 February 2021; Khanna, A., Gupta, D., Bhattacharyya, S., Hassanien, A.E., Anand, S., Jaiswal, A., Eds.; Springer: Singapore, 2022; pp. 151–165. [Google Scholar] [CrossRef]
  5. Miranda-Escalada, A.; Gonzalez-Agirre, A.; Armengol-Estapé, J.; Krallinger, M. Overview of Automatic Clinical Coding: Annotations, Guidelines, and Solutions for non-English Clinical Cases at CodiEsp Track of CLEF eHealth 2020. In Proceedings of the CLEF (Working Notes), Thessaloniki, Greece, 22–25 September 2020. [Google Scholar]
  6. Wu, Y.; Jiang, M.; Xu, J.; Zhi, D.; Xu, H. Clinical Named Entity Recognition Using Deep Learning Models. AMIA Annu. Symp. Proc. 2018, 2017, 1812–1819. [Google Scholar] [PubMed]
  7. Kreimeyer, K.; Foster, M.; Pandey, A.; Arya, N.; Halford, G.; Jones, S.F.; Forshee, R.; Walderhaug, M.; Botsis, T. Natural language processing systems for capturing and standardizing unstructured clinical information: A systematic review. J. Biomed. Inform. 2017, 73, 14–29. [Google Scholar] [CrossRef] [PubMed]
  8. Skeppstedt, M.; Kvist, M.; Nilsson, G.H.; Dalianis, H. Automatic recognition of disorders, findings, pharmaceuticals and body structures from clinical text: An annotation and machine learning study. J. Biomed. Inform. 2014, 49, 148–158. [Google Scholar] [CrossRef]
  9. Uzuner, Ö.; South, B.R.; Shen, S.; DuVall, S.L. 2010 I2b2/VA Challenge on Concepts, Assertions, and Relations in Clinical Text. J. Am. Med Inform. Assoc. JAMIA 2011, 18, 552–556. [Google Scholar] [CrossRef]
  10. Luo, Y.; Thompson, W.K.; Herr, T.M.; Zeng, Z.; Berendsen, M.A.; Jonnalagadda, S.R.; Carson, M.B.; Starren, J. Natural Language Processing for EHR-Based Pharmacovigilance: A Structured Review. Drug Saf. 2017, 40, 1075–1089. [Google Scholar] [CrossRef]
  11. Hovy, D.; Prabhumoye, S. Five Sources of Bias in Natural Language Processing. Lang. Linguist. Compass 2021, 15, e12432. [Google Scholar] [CrossRef]
  12. Névéol, A.; Dalianis, H.; Velupillai, S.; Savova, G.; Zweigenbaum, P. Clinical natural language processing in languages other than English: Opportunities and challenges. J. Biomed. Semant. 2018, 9, 12. [Google Scholar] [CrossRef]
  13. Schneider, E.T.R.; de Souza, J.V.A.; Knafou, J.; e Oliveira, L.E.S.; Copara, J.; Gumiel, Y.B.; de Oliveira, L.F.A.; Paraiso, E.C.; Teodoro, D.; Barra, C.M.C.M. BioBERTpt—A Portuguese Neural Language Model for Clinical Named Entity Recognition. In Proceedings of the 3rd Clinical Natural Language Processing Workshop, Online, 19 November 2020; pp. 65–72. [Google Scholar] [CrossRef]
  14. García-Izquierdo, I.; Montalt, V. Cultural Competence and the Role of the Patient’s Mother Tongue: An Exploratory Study of Health Professionals’ Perceptions. Societies 2022, 12, 53. [Google Scholar] [CrossRef]
  15. Montalt, V. Ethical Considerations in the Translation of Health Genres in Crisis Communication. In Translating Crises; Bloomsbury Publishing: London, UK, 2022; pp. 17–36. [Google Scholar]
  16. Schäfer, H.; Idrissi-Yaghir, A.; Horn, P.; Friedrich, C. Cross-Language Transfer of High-Quality Annotations: Combining Neural Machine Translation with Cross-Linguistic Span Alignment to Apply NER to Clinical Texts in a Low-Resource Language. In Proceedings of the 4th Clinical Natural Language Processing Workshop, Seattle, WA, USA, 14 July 2022; Naumann, T., Bethard, S., Roberts, K., Rumshisky, A., Eds.; Association for Computational Linguistics: New Orleans, LA, USA, 2022; pp. 53–62. [Google Scholar] [CrossRef]
  17. Xie, S.; Xia, Y.; Wu, L.; Huang, Y.; Fan, Y.; Qin, T. End-to-End Entity-Aware Neural Machine Translation. Mach. Learn. 2022, 111, 1181–1203. [Google Scholar] [CrossRef]
  18. Jain, A.; Paranjape, B.; Lipton, Z.C. Entity Projection via Machine Translation for Cross-Lingual NER. arXiv 2019, arXiv:1909.05356. [Google Scholar] [CrossRef]
  19. Mikolov, T.; Le, Q.V.; Sutskever, I. Exploiting Similarities among Languages for Machine Translation. arXiv 2013, arXiv:1309.4168. [Google Scholar] [CrossRef]
  20. Altintas, K.; Cicekli, I. A Machine Translation System Between a Pair of Closely Related Languages. In International Symposium on Computer and Information Sciences; CRC Press: Boca Raton, FL, USA, 2002. [Google Scholar]
  21. Briva-Iglesias, V. English-Catalan Neural Machine Translation: State-of-the-Art Technology, Quality, and Productivity. Tradumàtica Tecnol. Trad. 2022, 20, 149–176. [Google Scholar] [CrossRef]
  22. Castilho, S.; Moorkens, J.; Gaspari, F.; Calixto, I.; Tinsley, J.; Way, A. Is Neural Machine Translation the New State of the Art? Prague Bull. Math. Linguist. 2017, 109–120. [Google Scholar] [CrossRef]
  23. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. Adv. Neural Inf. Process. Syst. 2017, 30. [Google Scholar]
  24. Costa-Jussà, M.R.; Farrús, M.; Serrano Pons, J. Machine translation in medicine: A quality analysis of statistical machine translation in the medical domain. Presented at the Advanced Research in Scientific Areas, Bratislava, Slovakia, 3–7 December 2012. [Google Scholar]
  25. Soto, X.; Perez-de Viñaspre, O.; Oronoz, M.; Labaka, G. Development of a machine translation system for promoting the use of a low resource language in the clinical domain: The case of Basque. In Natural Language Processing in Healthcare; CRC Press: Boca Raton, FL, USA, 2022; pp. 139–158. [Google Scholar]
  26. Soares, F.; Krallinger, M. BSC participation in the WMT translation of biomedical abstracts. In Proceedings of the Fourth Conference on Machine Translation (Volume 3: Shared Task Papers, Day 2), Florence, Italy, 1–2 August 2019; pp. 175–178. [Google Scholar]
  27. Neves, M.; Yepes, A.J.; Siu, A.; Roller, R.; Thomas, P.; Navarro, M.V.; Yeganova, L.; Wiemann, D.; Di Nunzio, G.M.; Vezzani, F.; et al. Findings of the WMT 2022 biomedical translation shared task: Monolingual clinical case reports. In Proceedings of the WMT22-Seventh Conference on Machine Translation, Abu Dhabi, United Arab Emirates, 7–8 December 2022; pp. 694–723. [Google Scholar]
  28. Almagro, M.; Martínez, R.; Montalvo, S.; Fresno, V. A cross-lingual approach to automatic ICD-10 coding of death certificates by exploring machine translation. J. Biomed. Inform. 2019, 94, 103207. [Google Scholar] [CrossRef]
  29. Pilegaard, M. Translation of medical research articles. Benjamins Transl. Libr. 1997, 26, 159–184. [Google Scholar]
  30. Frei, J.; Frei-Stuber, L.; Kramer, F. GERNERMED++: Semantic annotation in German medical NLP through transfer-learning, translation and word alignment. J. Biomed. Inform. 2023, 147, 104513. [Google Scholar] [CrossRef]
  31. Zaghir, J.; Bjelogrlic, M.; Goldman, J.P.; Aananou, S.; Gaudet-Blavignac, C.; Lovis, C. FRASIMED: A Clinical French Annotated Resource Produced through Crosslingual BERT-Based Annotation Projection. arXiv 2023, arXiv:2309.10770. [Google Scholar]
  32. Deléger, L.; Merkel, M.; Zweigenbaum, P. Translating medical terminologies through word alignment in parallel text corpora. J. Biomed. Inform. 2009, 42, 692–701. [Google Scholar] [CrossRef] [PubMed]
  33. Villegas, M.; Intxaurrondo, A.; Gonzalez-Agirre, A.; Marimon, M.; Krallinger, M. The MeSpEN resource for English-Spanish medical machine translation and terminologies: Census of parallel corpora, glossaries and term translations. In Proceedings of the LREC 2018 Workshop “MultilingualBIO: Multilingual Biomedical Text Processing”, Miyazaki, Japan, 8 May 2018. [Google Scholar]
  34. Yang, R.; Tan, T.F.; Lu, W.; Thirunavukarasu, A.J.; Ting, D.S.W.; Liu, N. Large language models in health care: Development, applications, and challenges. Health Care Sci. 2023, 2, 255–263. [Google Scholar] [CrossRef] [PubMed]
  35. Briva-Iglesias, V.; Camargo, J.L.C.; Dogru, G. Large Language Models “Ad Referendum”: How Good Are They at Machine Translation in the Legal Domain? arXiv 2024, arXiv:2402.07681. [Google Scholar] [CrossRef]
  36. ten Hacken, P. The Language of Medicine in the Romance Languages; Oxford Research Encyclopedia of Linguistics: New York, NY, USA, 2023. [Google Scholar]
  37. Miranda-Escalada, A.; Farré-Maduell, E.; Lima-López, S.; Estrada, D.; Gascó, L.; Krallinger, M. Mention detection, normalization & classification of species, pathogens, humans and food in clinical documents: Overview of LivingNER shared task and resources. Proces. Leng. Nat. 2022, 69, 241–253. [Google Scholar]
  38. Lima-López, S.; Farré-Maduell, E.; Brivá-Escalada, V.; Gascó, L.; Krallinger, M. MEDDOPLACE Shared Task overview: Recognition, normalization and classification of locations and patient movement in clinical texts. Proces. Leng. Nat. 2023, 71, 301–311. [Google Scholar]
  39. Lima-López, S.; Farré-Maduell, E.; Gasco-Sánchez, L.; Rodríguez-Miret, J.; Krallinger, M. Overview of SympTEMIST at BioCreative VIII: Corpus, guidelines and evaluation of systems for the detection and normalization of symptoms, signs and findings from text. In Proceedings of the BioCreative VIII Challenge and Workshop: Curation and Evaluation in the Era of Generative Models, New Orleans, LA, USA, 12 November 2023. [Google Scholar]
  40. Lima-López, S.; Farré-Maduell, E.; Gascó, L.; Nentidis, A.; Krithara, A.; Katsimpras, G.; Paliouras, G.; Krallinger, M. Overview of MedProcNER task on medical procedure detection and entity linking at BioASQ 2023. In Proceedings of the Working Notes of CLEF 2023—Conference and Labs of the Evaluation Forum, Thessaloniki, Greece, 18–21 September 2023. [Google Scholar]
  41. Shaitarova, A.; Zaghir, J.; Lavelli, A.; Krauthammer, M.; Rinaldi, F. Exploring the Latest Highlights in Medical Natural Language Processing across Multiple Languages: A Survey. Yearb. Med Inform. 2023, 32, 230–243. [Google Scholar] [CrossRef]
  42. Lima-López, S.; Farré-Maduell, E.; Rodríguez-Miret, J.; Rodríguez-Ortega, M.; Lilli, L.; Lenkowicz, J.; Ceroni, G.; Kossoff, J.; Shah, A.; Nentidis, A.; et al. Overview of MultiCardioNER task at BioASQ 2024 on Medical Speciality and Language Adaptation of Clinical NER Systems for Spanish, English and Italian. In Proceedings of the CLEF 2024: Conference and Labs of the Evaluation Forum, Grenoble, France, 9–12 September 2024. [Google Scholar]
  43. Miranda-Escalada, A.; Gascó, L.; Lima-López, S.; Farré-Maduell, E.; Estrada, D.; Nentidis, A.; Krithara, A.; Katsimpras, G.; Paliouras, G.; Krallinger, M. Overview of DisTEMIST at BioASQ: Automatic detection and normalization of diseases from clinical texts: Results, methods, evaluation and multilingual resources. In Proceedings of the Working Notes of Conference and Labs of the Evaluation (CLEF) Forum. CEUR Workshop Proceedings, Bologna, Italy, 5–8 September 2022. [Google Scholar]
  44. Lima-López, S.; Farré-Maduell, E.; Miranda-Escalada, A.; Brivá-Iglesias, V.; Krallinger, M. NLP applied to occupational health: MEDDOPROF shared task at IberLEF 2021 on automatic recognition, classification and normalization of professions and occupations from medical texts. Proces. Leng. Nat. 2021, 67, 243–256. [Google Scholar]
  45. Artstein, R. Inter-annotator Agreement. In Handbook of Linguistic Annotation; Ide, N., Pustejovsky, J., Eds.; Springer: Dordrecht, The Netherlands, 2017; pp. 297–313. [Google Scholar] [CrossRef]
  46. Hripcsak, G.; Rothschild, A.S. Agreement, the F-Measure, and Reliability in Information Retrieval. J. Am. Med. Inform. Assoc. 2005, 12, 296–298. [Google Scholar] [CrossRef]
  47. Team, N.; Costa-jussà, M.R.; Cross, J.; Çelebi, O.; Elbayad, M.; Heafield, K.; Heffernan, K.; Kalbassi, E.; Lam, J.; Licht, D.; et al. No Language Left Behind: Scaling Human-Centered Machine Translation. arXiv 2022, arXiv:2207.04672. [Google Scholar] [CrossRef]
  48. Ethnologue. The Catalan Language. 2017. Available online: https://www.ethnologue.com/language/cat/ (accessed on 18 September 2024).
  49. Stenetorp, P.; Pyysalo, S.; Topić, G.; Ohta, T.; Ananiadou, S.; Tsujii, J. BRAT: A web-based tool for NLP-assisted text annotation. In Proceedings of the Demonstrations at the 13th Conference of the European Chapter of the Association for Computational Linguistics, Avignon, France, 23–27 April 2012; pp. 102–107. [Google Scholar]
  50. Tian, S.; Erdengasileng, A.; Yang, X.; Guo, Y.; Wu, Y.; Zhang, J.; Bian, J.; He, Z. Transformer-Based Named Entity Recognition for Parsing Clinical Trial Eligibility Criteria. In Proceedings of the 12th ACM Conference on Bioinformatics, Computational Biology, and Health Informatics, BCB’21, Gainesville, FL, USA, 1–4 August 2021; Association for Computing Machinery: New York, NY, USA, 2021; pp. 1–6. [Google Scholar] [CrossRef]
  51. Yang, F.; Wang, X.; Ma, H.; Li, J. Transformers-Sklearn: A Toolkit for Medical Language Understanding with Transformer-Based Models. BMC Med. Inform. Decis. Mak. 2021, 21, 90. [Google Scholar] [CrossRef]
  52. Ratnaparkhi, A.; Marcus, M.P. Maximum Entropy Models for Natural Language Ambiguity Resolution. Ph.D. Thesis, University of Pennsylvania, Philadelphia, PA, USA, 1998. [Google Scholar]
  53. Lange, L.; Adel, H.; Strötgen, J. Boosting Transformers for Job Expression Extraction and Classification in a Low-Resource Setting. arXiv 2021, arXiv:2109.08597. [Google Scholar] [CrossRef]
  54. Moscato, V.; Postiglione, M.; Sperlì, G. Biomedical Spanish Language Models for entity recognition and linking at BioASQ DisTEMIST. In Proceedings of the CLEF 2022: Conference and Labs of the Evaluation Forum, Bologna, Italy, 5–8 September 2022. [Google Scholar]
  55. Carrino, C.P.; Llop, J.; Pàmies, M.; Gutiérrez-Fandiño, A.; Armengol-Estapé, J.; Silveira-Ocampo, J.; Valencia, A.; Gonzalez-Agirre, A.; Villegas, M. Pretrained Biomedical Language Models for Clinical NLP in Spanish. In Proceedings of the 21st Workshop on Biomedical Language Processing, Dublin, Ireland, 6 May 2022; pp. 193–199. [Google Scholar] [CrossRef]
  56. Cañete, J.; Chaperon, G.; Fuentes, R.; Ho, J.H.; Kang, H.; Pérez, J. Spanish Pre-Trained BERT Model and Evaluation Data. In Proceedings of the PML4DC at ICLR 2020, Online, 26 April 2020. [Google Scholar]
  57. Lee, C.; Simpson, T.I.; Posma, J.M.; Lain, A.D. Comparative Analyses of Multilingual Drug Entity Recognition Systems for Clinical Case Reports in Cardiology. In Proceedings of the 5th Working Notes of the Conference and Labs of the Evaluation Forum, Grenoble, France, 9 September 2024. [Google Scholar]
  58. Lommel, A. Metrics for Translation Quality Assessment: A Case for Standardising Error Typologies. In Translation Quality Assessment: From Principles to Practice; Moorkens, J., Castilho, S., Gaspari, F., Doherty, S., Eds.; Springer International Publishing: Cham, Switzerland, 2018; pp. 109–127. [Google Scholar] [CrossRef]
  59. Freitag, M.; Foster, G.; Grangier, D.; Ratnakar, V.; Tan, Q.; Macherey, W. Experts, Errors, and Context: A Large-Scale Study of Human Evaluation for Machine Translation. Trans. Assoc. Comput. Linguist. 2021, 9, 1460–1474. [Google Scholar] [CrossRef]
  60. Moorkens, J.; Castilho, S.; Gaspari, F.; Doherty, S. (Eds.) Translation Quality Assessment: From Principles to Practice. In Machine Translation: Technologies and Applications; Springer International Publishing: Cham, Switzerland, 2018; Volume 1. [Google Scholar] [CrossRef]
  61. Chinchor, N.; Sundheim, B. MUC-5 Evaluation Metrics. In Proceedings of the Fifth Message Understanding Conference (MUC-5), Baltimore, MD, USA, 25–27 August 1993. [Google Scholar]
  62. Reyna, O.G.P. El Diccionario panhispánico de términos médicos. Rev. Soc. Peru. Med. Interna 2023, 36, 169–170. [Google Scholar]
  63. Soto, X.; Perez-de Viñaspre, O.; Labaka, G.; Oronoz, M. Neural machine translation of clinical texts between long distance languages. J. Am. Med. Inform. Assoc. 2019, 26, 1478–1487. [Google Scholar] [CrossRef] [PubMed]
  64. Miranda-Escalada, A.; Farré-Maduell, E.; Krallinger, M. Named entity recognition, concept normalization and clinical coding: Overview of the CANTEMIST track for cancer text mining in Spanish, corpus, guidelines, methods and results. In Proceedings of the Iberian Languages Evaluation Forum (IberLEF 2020), CEUR Workshop Proceedings, Online, 22 September 2020. [Google Scholar]
  65. Gonzalez-Agirre, A.; Marimon, M.; Intxaurrondo, A.; Rabal, O.; Villegas, M.; Krallinger, M. Pharmaconer: Pharmacological substances, compounds and proteins named entity recognition track. In Proceedings of the 5th Workshop on BioNLP Open Shared Tasks, Hong Kong, China, 4 November 2019; pp. 1–10. [Google Scholar]
  66. Awon, V.K.; Balloccu, S.; Wu, Z.; Reiter, E.; Helaouie, R.; Reforgiato Recupero, D.; Riboni, D. Data Augmentation for Reliability and Fairness in Counselling Quality Classification. In Proceedings of the 1st Workshop on Scarce Data in Artificial Intelligence for Healthcare (SDAIH 2022), Vienna, Austria, 23 July 2022. [Google Scholar]
  67. Hu, Y.; Chen, Q.; Du, J.; Peng, X.; Keloth, V.K.; Zuo, X.; Zhou, Y.; Li, Z.; Jiang, X.; Lu, Z.; et al. Improving large language models for clinical named entity recognition via prompt engineering. J. Am. Med. Inform. Assoc. 2024, 31, 1812–1820. [Google Scholar] [CrossRef]
  68. Balloccu, S.; Reiter, E.; Kumar, V.; Recupero, D.R.; Riboni, D. Ask the experts: Sourcing high-quality datasets for nutritional counselling through Human-AI collaboration. arXiv 2024, arXiv:2401.08420. [Google Scholar]
  69. Wu, Z.; Balloccu, S.; Kumar, V.; Helaoui, R.; Reforgiato Recupero, D.; Riboni, D. Creation, analysis and evaluation of annomi, a dataset of expert-annotated counselling dialogues. Future Internet 2023, 15, 110. [Google Scholar] [CrossRef]
  70. Wu, Z.; Balloccu, S.; Kumar, V.; Helaoui, R.; Reiter, E.; Recupero, D.R.; Riboni, D. Anno-mi: A dataset of expert-annotated counselling dialogues. In Proceedings of the ICASSP 2022–2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 23–27 May 2022; pp. 6177–6181. [Google Scholar]
Figure 1. Overview of the clinical case report translation and entity annotation projection approach.
Figure 1. Overview of the clinical case report translation and entity annotation projection approach.
Information 15 00585 g001
Figure 2. Example screenshot of the brat interface side-by-side view used for the clinical expert validation and correction of entities. By double-clicking on each annotation’s label name, annotators were able to provide alternative translations using the Notes section. The text translation in English would read “47-year-old woman with a history of glycogenosis type V (McArdle’s disease), Chiari malformation, chronic renal failure without filiation, retinal atrophy and arterial hypertension. A peritoneal catheter was implanted on 19 May 2000 in another centre, and dialysis with a 2-L infusion was started 15 days later”.
Figure 2. Example screenshot of the brat interface side-by-side view used for the clinical expert validation and correction of entities. By double-clicking on each annotation’s label name, annotators were able to provide alternative translations using the Notes section. The text translation in English would read “47-year-old woman with a history of glycogenosis type V (McArdle’s disease), Chiari malformation, chronic renal failure without filiation, retinal atrophy and arterial hypertension. A peritoneal catheter was implanted on 19 May 2000 in another centre, and dialysis with a 2-L infusion was started 15 days later”.
Information 15 00585 g002
Table 1. Characteristics of the three Gold Standard corpora used for the annotation projection experiment.
Table 1. Characteristics of the three Gold Standard corpora used for the annotation projection experiment.
DatasetTopicDocumentsTokensAnnotations
DisTEMISTDiseases1000406,31810,664
DrugTEMISTMedications1000406,3182782
MEDDOPROFOccupations18441,291,1864770
Table 2. Statistics of the projected corpora from Spanish to Catalan, divided in splits. GS stands for “Gold Standard” (i.e., original Spanish corpus); Ann. stands for “annotations”.
Table 2. Statistics of the projected corpora from Spanish to Catalan, divided in splits. GS stands for “Gold Standard” (i.e., original Spanish corpus); Ann. stands for “annotations”.
DatasetDocumentsGS Ann.Not Projected% Not Projected
DisTEMIST train+dev7508065981.21%
DisTEMIST test2502599552.11%
DrugTEMIST train+dev7502090341.62%
DrugTEMIST test250692182.60%
MEDDOPROF train+dev150036581393.79%
MEDDOPROF test3441085544.97%
Table 3. Precision (P), recall (R), and F1-score (F1) metrics for the test sets of DrugTEMIST, DisTEMIST, and MEDDOPROF (“Spa” and “Cat”), and CataCCC. Models were fine-tuned on the training set of the original Spanish version (“Spa”) and the corrected Catalan version (“Cat corr.”), with the RoBERTa bsc-bio-ehr-es as the base model in both cases. The best metrics for each corpus are marked in bold. Note that each square represents the results of the same model evaluated on the different corpora. Only the best/selected model from all hyperparameter optimization experiments is shown.
Table 3. Precision (P), recall (R), and F1-score (F1) metrics for the test sets of DrugTEMIST, DisTEMIST, and MEDDOPROF (“Spa” and “Cat”), and CataCCC. Models were fine-tuned on the training set of the original Spanish version (“Spa”) and the corrected Catalan version (“Cat corr.”), with the RoBERTa bsc-bio-ehr-es as the base model in both cases. The best metrics for each corpus are marked in bold. Note that each square represents the results of the same model evaluated on the different corpora. Only the best/selected model from all hyperparameter optimization experiments is shown.
 DrugsDiseasesOccupations
Trained on/Eval. onPRF1PRF1PRF1
Spa/Spa0.9170.9090.9130.7540.7590.7570.7850.7760.780
Spa/Cat corr.0.8860.8090.8460.6100.6080.6090.6420.3370.442
Spa/CataCCC0.9080.8570.8820.7420.7020.7210.7500.4440.558
Cat corr./Spa0.8840.8800.8820.6720.6440.6580.7230.7260.725
Cat corr./Cat corr.0.8850.8740.8790.7010.7180.7090.7430.7170.729
Cat corr./CataCCC0.9210.9040.9130.7750.8180.7960.8380.7940.815
Table 4. Strict and relaxed F1-score of the fine-tuned RoBERTa bsc-bio-ehr-es models using the MEDDOPLACE Scoring Script [38] for the test sets. The relaxed metric will count a prediction as a true positive if it has at least some overlap with a ground truth mention.
Table 4. Strict and relaxed F1-score of the fine-tuned RoBERTa bsc-bio-ehr-es models using the MEDDOPLACE Scoring Script [38] for the test sets. The relaxed metric will count a prediction as a true positive if it has at least some overlap with a ground truth mention.
 DrugsDiseasesOccupations
Trained on/ Eval. onStrictRelaxedStrictRelaxedStrictRelaxed
Spa/Spa0.9130.9540.7570.8920.7820.872
Cat corr./Cat corr.0.8790.9370.7090.8660.7290.840
Cat corr./CataCCC0.9130.9570.7960.8890.8150.875
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Rodríguez-Miret, J.; Farré-Maduell, E.; Lima-López, S.; Vigil, L.; Briva-Iglesias, V.; Krallinger, M. Exploring the Potential of Neural Machine Translation for Cross-Language Clinical Natural Language Processing (NLP) Resource Generation through Annotation Projection. Information 2024, 15, 585. https://doi.org/10.3390/info15100585

AMA Style

Rodríguez-Miret J, Farré-Maduell E, Lima-López S, Vigil L, Briva-Iglesias V, Krallinger M. Exploring the Potential of Neural Machine Translation for Cross-Language Clinical Natural Language Processing (NLP) Resource Generation through Annotation Projection. Information. 2024; 15(10):585. https://doi.org/10.3390/info15100585

Chicago/Turabian Style

Rodríguez-Miret, Jan, Eulàlia Farré-Maduell, Salvador Lima-López, Laura Vigil, Vicent Briva-Iglesias, and Martin Krallinger. 2024. "Exploring the Potential of Neural Machine Translation for Cross-Language Clinical Natural Language Processing (NLP) Resource Generation through Annotation Projection" Information 15, no. 10: 585. https://doi.org/10.3390/info15100585

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop