Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (849)

Search Parameters:
Keywords = word embedding

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
14 pages, 1592 KB  
Article
Fine-Tuning Large Language Models for Effective Nutrition Support in Residential Aged Care: A Domain Expertise Approach
by Mohammad Alkhalaf, Dinithi Vithanage, Jun Shen, Hui Chen (Rita) Chang, Chao Deng and Ping Yu
Healthcare 2025, 13(20), 2614; https://doi.org/10.3390/healthcare13202614 - 17 Oct 2025
Abstract
Background: Malnutrition is a serious health concern among older adults in residential aged care (RAC), and timely identification is critical for effective intervention. Recent advancements in transformer-based large language models (LLMs), such as RoBERTa, provide context-aware embeddings that improve predictive performance in clinical [...] Read more.
Background: Malnutrition is a serious health concern among older adults in residential aged care (RAC), and timely identification is critical for effective intervention. Recent advancements in transformer-based large language models (LLMs), such as RoBERTa, provide context-aware embeddings that improve predictive performance in clinical tasks. Fine-tuning these models on domain-specific corpora, like nursing progress notes, can further enhance their applicability in healthcare. Methodology: We developed a RAC domain-specific LLM by training RoBERTa on 500,000 nursing progress notes from RAC electronic health records (EHRs). The model’s embeddings were used for two downstream tasks: malnutrition note identification and malnutrition prediction. Long sequences were truncated and processed in segments of up to 1536 tokens to fit RoBERTa’s 512-token input limit. Performance was compared against Bag of Words, GloVe, baseline RoBERTa, BlueBERT, ClinicalBERT, BioClinicalBERT, and PubMed models. Results: Using 5-fold cross-validation, the RAC domain-specific LLM outperformed other models. For malnutrition note identification, it achieved an F1-score of 0.966, and for malnutrition prediction, it achieved an F1-score of 0.687. Conclusions: This approach demonstrates the feasibility of developing specialised LLMs for identifying and predicting malnutrition among older adults in RAC. Future work includes further optimisation of prediction performance and integration with clinical workflows to support early intervention. Full article
Show Figures

Figure 1

27 pages, 3065 KB  
Article
Chinese Financial News Analysis for Sentiment and Stock Prediction: A Comparative Framework with Language Models
by Hsiu-Min Chuang, Hsiang-Chih He and Ming-Che Hu
Big Data Cogn. Comput. 2025, 9(10), 263; https://doi.org/10.3390/bdcc9100263 - 16 Oct 2025
Viewed by 258
Abstract
Financial news has a significant impact on investor sentiment and short-term stock price trends. While many studies have applied natural language processing (NLP) techniques to financial forecasting, most have focused on single tasks or English corpora, with limited research in non-English language contexts [...] Read more.
Financial news has a significant impact on investor sentiment and short-term stock price trends. While many studies have applied natural language processing (NLP) techniques to financial forecasting, most have focused on single tasks or English corpora, with limited research in non-English language contexts such as Taiwan. This study develops a joint framework to perform sentiment classification and short-term stock price prediction using Chinese financial news from Taiwan’s top 50 listed companies. Five types of word embeddings—one-hot, TF-IDF, CBOW, skip-gram, and BERT—are systematically compared across 17 traditional, deep, and Transformer models, as well as a large language model (LLaMA3) fully fine-tuned on the Chinese financial texts. To ensure annotation quality, sentiment labels were manually assigned by annotators with finance backgrounds and validated through a double-checking process. Experimental results show that a CNN using skip-gram embeddings achieves the strongest performance among deep learning models, while LLaMA3 yields the highest overall F1-score for sentiment classification. For regression, LSTM consistently provides the most reliable predictive power across different volatility groups, with Bayesian Linear Regression remaining competitive for low-volatility firms. LLaMA3 is the only Transformer-based model to achieve a positive R2 under high-volatility conditions. Furthermore, forecasting accuracy is higher for the five-day horizon than for the fifteen-day horizon, underscoring the increasing difficulty of medium-term forecasting. These findings confirm that financial news provides valuable predictive signals for emerging markets and that short-term sentiment-informed forecasts enhance real-time investment decisions. Full article
(This article belongs to the Special Issue Natural Language Processing Applications in Big Data)
Show Figures

Figure 1

29 pages, 1977 KB  
Article
From Market Volatility to Predictive Insight: An Adaptive Transformer–RL Framework for Sentiment-Driven Financial Time-Series Forecasting
by Zhicong Song, Harris Sik-Ho Tsang, Richard Tai-Chiu Hsung, Yulin Zhu and Wai-Lun Lo
Forecasting 2025, 7(4), 55; https://doi.org/10.3390/forecast7040055 - 2 Oct 2025
Viewed by 396
Abstract
Financial time-series prediction remains a significant challenge, driven by market volatility, nonlinear dynamic characteristics, and the complex interplay between quantitative indicators and investor sentiment. Traditional time-series models (e.g., ARIMA and GARCH) struggle to capture the nuanced sentiment in textual data, while static deep [...] Read more.
Financial time-series prediction remains a significant challenge, driven by market volatility, nonlinear dynamic characteristics, and the complex interplay between quantitative indicators and investor sentiment. Traditional time-series models (e.g., ARIMA and GARCH) struggle to capture the nuanced sentiment in textual data, while static deep learning integration methods fail to adapt to market regime transitions (bull markets, bear markets, and consolidation). This study proposes a hybrid framework that integrates investor forum sentiment analysis with adaptive deep reinforcement learning (DRL) for dynamic model integration. By constructing a domain-specific financial sentiment dictionary (containing 16,673 entries) based on the sentiment analysis approach and word-embedding technique, we achieved up to 97.35% accuracy in forum title classification tasks. Historical price data and investor forum sentiment information were then fed into a Support Vector Regressor (SVR) and three Transformer variants (single-layer, multi-layer, and bidirectional variants) for predictions, with a Deep Q-Network (DQN) agent dynamically fusing the prediction results. Comprehensive experiments were conducted on diverse financial datasets, including China Unicom, the CSI 100 index, corn, and Amazon (AMZN). The experimental results demonstrate that our proposed approach, combining textual sentiment with adaptive DRL integration, significantly enhances prediction robustness in volatile markets, achieving the lowest RMSEs across diverse assets. It overcomes the limitations of static methods and multi-market generalization, outperforming both benchmark and state-of-the-art models. Full article
Show Figures

Figure 1

24 pages, 2380 KB  
Article
Resisting Chauvinist Stereotypes: The Impertinence of Russian Painting at London’s International Exhibition of 1862
by Rosalind Polly Blakesley
Arts 2025, 14(5), 118; https://doi.org/10.3390/arts14050118 - 30 Sep 2025
Viewed by 354
Abstract
The Russian empire’s displays of applied and decorative art at the Great Exhibition of 1851 and its immediate successors have long galvanised scholars for their semantic complexity. By contrast, Russia’s first selection of paintings for this fiercely competitive arena, shown at London’s International [...] Read more.
The Russian empire’s displays of applied and decorative art at the Great Exhibition of 1851 and its immediate successors have long galvanised scholars for their semantic complexity. By contrast, Russia’s first selection of paintings for this fiercely competitive arena, shown at London’s International Exhibition of 1862, failed to ignite the public imagination and has largely evaded the historian’s gaze. While the three-dimensional artworks provided a recurrent source of wonderment for their superlative craftsmanship, stupendous materials, and often hyperbolic proportions, the paintings were apparently flat in every sense of the word: derivative, lacklustre, and incapable of capitalising on the opportunity that international exhibitions offered to present a national school. The dismissive comments they attracted set the tone for many later accounts, embedding the idea that Russian painting prior to the twentieth century was of limited consequence—a perception that would prove convenient to those asserting the originality of the avant-garde. Yet renewed consideration of Russia’s display of paintings in 1862 suggests that their critical reception speaks to concerns that went well beyond the pictures’ supposed obligation to represent a national school. Notably, a small but significant number of history and portrait paintings by academically trained and often well-travelled artists challenged notions of Russians as primitive and parochial. The technically adventurous of these parried the belief that Russian art was insufficiently mature to experiment in painterly effect. Most audacious of all, they broached unspoken national boundaries by daring to suggest that Imperial Russian artists could innovate in areas on which the success of British painting rested. The attitudes towards Russian painting in 1862 thus invite fresh scrutiny, revealing as they do a disruptive arena in which aesthetic rivalries and chauvinist sensibilities came to the fore. Full article
Show Figures

Figure 1

20 pages, 1724 KB  
Article
Spectral Features of Wolaytta Ejectives
by Firew Elias, Derib Ado and Feda Negesse
Languages 2025, 10(10), 250; https://doi.org/10.3390/languages10100250 - 29 Sep 2025
Viewed by 483
Abstract
This study analyzes the spectral properties of word-initial and intervocalic ejectives in Wolaytta, an Omotic language of southern Ethiopia. Using tokens embedded in three vowel contexts, we examined mean burst intensity, spectral moments, and vowel perturbation following ejection. Results show that ejectives adjacent [...] Read more.
This study analyzes the spectral properties of word-initial and intervocalic ejectives in Wolaytta, an Omotic language of southern Ethiopia. Using tokens embedded in three vowel contexts, we examined mean burst intensity, spectral moments, and vowel perturbation following ejection. Results show that ejectives adjacent to high front vowels were produced with greater intensity, supporting the hypothesis that increased oral cavity tenseness correlates with acoustic energy. Centroid and standard deviation differentiate place of articulation, while skewness and kurtosis distinguish singleton from geminate ejectives. Post-ejective vowel pitch and spectral tilt varied systematically with the ejectives’ place of articulation, indicating creaky phonation induced by ejection. Overall, the findings enhance our understanding of factors impacting acoustic features of ejectives. Full article
Show Figures

Figure 1

21 pages, 3434 KB  
Article
Deep Learning-Based Compliance Assessment for Chinese Rail Transit Dispatch Speech
by Qiuzhan Zhao, Jinbai Zou and Lingxiao Chen
Appl. Sci. 2025, 15(19), 10498; https://doi.org/10.3390/app151910498 - 28 Sep 2025
Viewed by 213
Abstract
Rail transit dispatch speech plays a critical role in ensuring the safety of urban rail operations. To enable automated and accurate compliance assessment of dispatch speech, this study proposes an improved deep learning model to address the limitations of conventional approaches in terms [...] Read more.
Rail transit dispatch speech plays a critical role in ensuring the safety of urban rail operations. To enable automated and accurate compliance assessment of dispatch speech, this study proposes an improved deep learning model to address the limitations of conventional approaches in terms of accuracy and robustness. Building upon the baseline Whisper model, two key enhancements are introduced: (1) low-rank adaptation (LoRA) fine-tuning to better adapt the model to the specific acoustic and linguistic characteristics of rail transit dispatch speech, and (2) a novel entity-aware attention mechanism that incorporates named entity recognition (NER) embeddings into the decoder. This mechanism enables attention computation between words belonging to the same entity category across different commands and recitations, which helps highlight keywords critical for compliance assessment and achieve precise inter-sentence element alignment. Experimental results on real-world test sets demonstrate that the proposed model improves recognition accuracy by 30.5% compared to the baseline model. In terms of robustness, we evaluate the relative performance retention under severe noise conditions. While Zero-shot, Full Fine-tuning, and LoRA-only models achieve robustness scores of 72.2%, 72.4%, and 72.1%, respectively, and the NER-only variant reaches 88.1%, our proposed approach further improves to 89.6%. These results validate the model’s significant robustness and its potential to provide efficient and reliable technical support for ensuring the normative use of dispatch speech in urban rail transit operations. Full article
Show Figures

Figure 1

22 pages, 1250 KB  
Article
Entity Span Suffix Classification for Nested Chinese Named Entity Recognition
by Jianfeng Deng, Ruitong Zhao, Wei Ye and Suhong Zheng
Information 2025, 16(10), 822; https://doi.org/10.3390/info16100822 - 23 Sep 2025
Viewed by 327
Abstract
Named entity recognition (NER) is one of the fundamental tasks in building knowledge graphs. For some domain-specific corpora, the text descriptions exhibit limited standardization, and some entity structures have entity nesting. The existing entity recognition methods have problems such as word matching noise [...] Read more.
Named entity recognition (NER) is one of the fundamental tasks in building knowledge graphs. For some domain-specific corpora, the text descriptions exhibit limited standardization, and some entity structures have entity nesting. The existing entity recognition methods have problems such as word matching noise interference and difficulty in distinguishing different entity labels for the same character in sequence label prediction. This paper proposes a span-based feature reuse stacked bidirectional long short term memory network (BiLSTM) nested named entity recognition (SFRSN) model, which transforms the entity recognition of sequence prediction into the problem of entity span suffix category classification. Firstly, character feature embedding is generated through bidirectional encoder representation of transformers (BERT). Secondly, a feature reuse stacked BiLSTM is proposed to obtain deep context features while alleviating the problem of deep network degradation. Thirdly, the span feature is obtained through the dilated convolution neural network (DCNN), and at the same time, a single-tail selection function is introduced to obtain the classification feature of the entity span suffix, with the aim of reducing the training parameters. Fourthly, a global feature gated attention mechanism is proposed, integrating span features and span suffix classification features to achieve span suffix classification. The experimental results on four Chinese-specific domain datasets demonstrate the effectiveness of our approach: SFRSN achieves micro-F1 scores of 83.34% on ontonotes, 73.27% on weibo, 96.90% on resume, and 86.77% on the supply chain management dataset. This represents a maximum improvement of 1.55%, 4.94%, 2.48%, and 3.47% over state-of-the-art baselines, respectively. The experimental results demonstrate the effectiveness of the model in addressing nested entities and entity label ambiguity issues. Full article
(This article belongs to the Section Artificial Intelligence)
Show Figures

Graphical abstract

15 pages, 471 KB  
Article
Bilingual Contextual Variability: Learning Words in Two Languages
by Justin Lauro and Pamela Freitas Pereira Toassi
Educ. Sci. 2025, 15(9), 1264; https://doi.org/10.3390/educsci15091264 - 22 Sep 2025
Viewed by 331
Abstract
Background. Bilingual novel word learning is shaped by both semantic context and the language in which learning occurs. According to the context variability hypothesis and instance-based learning frameworks, varied semantic contexts promote the formation of flexible lexical-semantic representations. However, the extent to [...] Read more.
Background. Bilingual novel word learning is shaped by both semantic context and the language in which learning occurs. According to the context variability hypothesis and instance-based learning frameworks, varied semantic contexts promote the formation of flexible lexical-semantic representations. However, the extent to which these benefits generalize across languages and transfer to novel contexts remains unclear. Method. Two experiments examined the effects of study language (L1, L2, or both) and semantic variability (repeated vs. varied contexts) on novel word learning in English–Spanish bilinguals. Participants studied rare words embedded in sentences and were tested via a word-stem completion task. In Experiment 1, test sentences were identical to those seen during the study. In Experiment 2, half of the test sentences were novel, requiring generalization beyond previously encountered contexts. Orthographic overlap across languages was also assessed. Results. In Experiment 1, varied semantic contexts improved recall accuracy, supporting the context variability hypothesis. Unexpectedly, words studied in L2 were recalled more accurately than those studied in L1, consistent with desirable difficulty effects. Additionally, orthographic overlap moderated learning, with greater benefits observed in mixed-language conditions. In Experiment 2, overall accuracy declined, and no main effects of language or context were observed. However, a three-way interaction showed that orthographic overlap improved recall only when words were studied in L1 and tested in novel contexts. Conclusions. Semantic and linguistic variability can enhance bilingual word learning when test conditions are consistent with the learning context. However, generalization to novel contexts may require deeper processing, extended exposure, or additional retrieval cues. Full article
(This article belongs to the Special Issue Language Learning in Multilingual, Inclusive and Immersive Contexts)
Show Figures

Figure 1

29 pages, 2935 KB  
Article
Optimising Contextual Embeddings for Meaning Conflation Deficiency Resolution in Low-Resourced Languages
by Mosima A. Masethe, Sunday O. Ojo and Hlaudi D. Masethe
Computers 2025, 14(9), 402; https://doi.org/10.3390/computers14090402 - 22 Sep 2025
Viewed by 421
Abstract
Meaning conflation deficiency (MCD) presents a continual obstacle in natural language processing (NLP), especially for low-resourced and morphologically complex languages, where polysemy and contextual ambiguity diminish model precision in word sense disambiguation (WSD) tasks. This paper examines the optimisation of contextual embedding models, [...] Read more.
Meaning conflation deficiency (MCD) presents a continual obstacle in natural language processing (NLP), especially for low-resourced and morphologically complex languages, where polysemy and contextual ambiguity diminish model precision in word sense disambiguation (WSD) tasks. This paper examines the optimisation of contextual embedding models, namely XLNet, ELMo, BART, and their improved variations, to tackle MCD in linguistic settings. Utilising Sesotho sa Leboa as a case study, researchers devised an enhanced XLNet architecture with specific hyperparameter optimisation, dynamic padding, early termination, and class-balanced training. Comparative assessments reveal that the optimised XLNet attains an accuracy of 91% and exhibits balanced precision–recall metrics of 92% and 91%, respectively, surpassing both its baseline counterpart and competing models. Optimised ELMo attained the greatest overall metrics (accuracy: 92%, F1-score: 96%), whilst optimised BART demonstrated significant accuracy improvements (96%) despite a reduced recall. The results demonstrate that fine-tuning contextual embeddings using MCD-specific methodologies significantly improves semantic disambiguation for under-represented languages. This study offers a scalable and flexible optimisation approach suitable for additional low-resource language contexts. Full article
Show Figures

Figure 1

19 pages, 1380 KB  
Article
Educational QA System-Oriented Answer Selection Model Based on Focus Fusion of Multi-Perspective Word Matching
by Xiaoli Hu, Junfei He, Zhaoyu Shou, Ziming Liu and Huibing Zhang
Computers 2025, 14(9), 399; https://doi.org/10.3390/computers14090399 - 19 Sep 2025
Viewed by 265
Abstract
Question-answering systems have become an important tool for learning and knowledge acquisition. However, current answer selection models often rely on representing features using whole sentences, which leads to neglecting individual words and losing important information. To address this challenge, the paper proposes a [...] Read more.
Question-answering systems have become an important tool for learning and knowledge acquisition. However, current answer selection models often rely on representing features using whole sentences, which leads to neglecting individual words and losing important information. To address this challenge, the paper proposes a novel answer selection model based on focus fusion of multi-perspective word matching. First, according to the different combination relationships between sentences, focus distribution in terms of words is obtained from the matching perspectives of serial, parallel, and transfer. Then, the sentence’s key position information is inferred from its focus distribution. Finally, a method of aligning key information points is designed to fuse the focus distribution for each perspective, which obtains match scores for each candidate answer to the question. Experimental results show that the proposed model significantly outperforms the Transformer encoder fine-tuned model based on contextual embedding, achieving a 4.07% and 5.51% increase in MAP and a 1.63% and 4.86% increase in MRR, respectively. Full article
Show Figures

Figure 1

20 pages, 2051 KB  
Article
A Study on the Evolution of Online Public Opinion During Major Public Health Emergencies Based on Deep Learning
by Yimin Yang, Julin Wang and Ming Liu
Mathematics 2025, 13(18), 3021; https://doi.org/10.3390/math13183021 - 18 Sep 2025
Viewed by 351
Abstract
This study investigates the evolution of online public opinion during the COVID-19 pandemic by integrating topic mining with sentiment analysis. To overcome the limitations of traditional short-text models and improve the accuracy of sentiment detection, we propose a novel hybrid framework that combines [...] Read more.
This study investigates the evolution of online public opinion during the COVID-19 pandemic by integrating topic mining with sentiment analysis. To overcome the limitations of traditional short-text models and improve the accuracy of sentiment detection, we propose a novel hybrid framework that combines a GloVe-enhanced Biterm Topic Model (BTM) for semantic-aware topic clustering with a RoBERTa-TextCNN architecture for deep, context-rich sentiment classification. The framework is specifically designed to capture both the global semantic relationships of words and the dynamic contextual nuances of social media discourse. Using a large-scale corpus of more than 550,000 Weibo posts, we conducted comprehensive experiments to evaluate the model’s effectiveness. The proposed approach achieved an accuracy of 92.45%, significantly outperforming baseline transformer-based baseline representative of advanced contextual embedding models across multiple evaluation metrics, including precision, recall, F1-score, and AUC. These results confirm the robustness and stability of the hybrid design and demonstrate its advantages in balancing precision and recall. Beyond methodological validation, the empirical analysis provides important insights into the dynamics of online public discourse. User engagement is found to be highest for the topics directly tied to daily life, with discussions about quarantine conditions alone accounting for 42.6% of total discourse. Moreover, public sentiment proves to be highly volatile and event-driven; for example, the announcement of Wuhan’s reopening produced an 11% surge in positive sentiment, reflecting a collective emotional uplift at a major turning point of the pandemic. Taken together, these findings demonstrate that online discourse evolves in close connection with both societal conditions and government interventions. The proposed topic–sentiment analysis framework not only advances methodological research in text mining and sentiment analysis, but also has the potential to serve as a practical tool for real-time monitoring online opinion. By capturing the fluctuations of public sentiment and identifying emerging themes, this study aims to provide insights that could inform policymaking by suggesting strategies to guide emotional contagion, strengthen crisis communication, and promote constructive public debate during health emergencies. Full article
(This article belongs to the Special Issue AI, Machine Learning and Optimization)
Show Figures

Figure 1

24 pages, 2616 KB  
Article
Symmetric Affix–Context Co-Attention: A Dual-Gating Framework for Robust POS Tagging in Low-Resource MRLs
by Yuan Qi, Samat Ali and Alim Murat
Symmetry 2025, 17(9), 1561; https://doi.org/10.3390/sym17091561 - 18 Sep 2025
Viewed by 468
Abstract
Part-of-speech (POS) tagging in low-resource, morphologically rich languages (LRLs/MRLs) remains challenging due to extensive affixation, high out-of-vocabulary (OOV) rates, and pervasive polysemy. We propose MRL-POS, a unified Transformer-CRF framework that dynamically selects informative affix features and integrates them with deep contextual embeddings via [...] Read more.
Part-of-speech (POS) tagging in low-resource, morphologically rich languages (LRLs/MRLs) remains challenging due to extensive affixation, high out-of-vocabulary (OOV) rates, and pervasive polysemy. We propose MRL-POS, a unified Transformer-CRF framework that dynamically selects informative affix features and integrates them with deep contextual embeddings via a novel dual-gating co-attention mechanism. First, a Dynamic Affix Selector adaptively adjusts n-gram ranges and frequency thresholds based on word length to ensure high-precision affix segmentation. Second, the Affix–Context Co-Attention Module employs two gating functions that conditionally amplify contextual dimensions with affix cues and vice versa, enabling robust disambiguation of complex and ambiguous forms. Third, Layer-Wise Attention Pooling aggregates multi-layer XLM-RoBERTa representations, emphasizing those most relevant for morphological and syntactic tagging. Evaluations on Uyghur, Kyrgyz, and Uzbek show that MRL-POS achieves an average F1 of 84.10%, OOV accuracy of 84.24%, and Poly-F1 of 72.14%, outperforming strong baselines by up to 8 F1 points. By explicitly modeling the symmetry between morphological affix cues and sentence-level context through a dual-gating co-attention mechanism, MRL-POS achieves a balanced fusion that both preserves local structure and captures global dependencies. Interpretability analyses confirm that 89.1% of the selected affixes align with linguistic expectations. This symmetric design not only enhances robustness in low-resource and agglutinative settings but also offers a general paradigm for symmetry-aware sequence labeling tasks. Full article
Show Figures

Figure 1

31 pages, 4340 KB  
Article
iAttention Transformer: An Inter-Sentence Attention Mechanism for Automated Grading
by Ibidapo Dare Dada, Adio T. Akinwale, Idowu A. Osinuga, Henry Nwagu Ogbu and Ti-Jesu Tunde-Adeleke
Mathematics 2025, 13(18), 2991; https://doi.org/10.3390/math13182991 - 16 Sep 2025
Viewed by 435
Abstract
This study developed and evaluated transformer-based models enhanced with inter-sentence attention (iAttention) mechanisms to improve the automatic grading of student responses to open-ended questions. Traditional transformer models emphasize intra-sentence relationships and often fail to capture complex semantic alignments needed for accurate assessment. To [...] Read more.
This study developed and evaluated transformer-based models enhanced with inter-sentence attention (iAttention) mechanisms to improve the automatic grading of student responses to open-ended questions. Traditional transformer models emphasize intra-sentence relationships and often fail to capture complex semantic alignments needed for accurate assessment. To overcome this limitation, three iAttention mechanisms, including iAttentionTFIDF, iAttentionword and iAttentionHW were proposed to enhance the model’s capacity to align key ideas between students and reference answers. This helps improve the model’s ability to capture important semantic relationships between words in two sentences. Unlike previous approaches that rely solely on aggregated sentence embeddings, the proposed method introduces inter-sentence attention layers that explicitly model semantic correspondence between individual sentences. This enables finer-grained matching of key concepts, reasoning, and logical structure which are crucial for fair and reliable assessment. The models were evaluated on multiple benchmark datasets, including Semantic Textual Similarity (STS), SemEval-2013 Beetle, SciEntsBank, Mohler, and a composite of university-level educational datasets (U-datasets). Experimental results demonstrated that integrating iAttention consistently outperforms baseline models, achieving higher Pearson and Spearman Correlation scores on STS, Mohler, and U-datasets, as well as superior Macro-F1, Weighted-F1, and Accuracy on the Beetle and SciEntsBank datasets. This approach contributes to the development of scalable, consistent, and fair automated grading systems by narrowing the gap between machine evaluation and human judgment, ultimately leading to more accurate and efficient assessment practices. Full article
(This article belongs to the Section E1: Mathematics and Computer Science)
Show Figures

Figure 1

21 pages, 588 KB  
Article
Research on an MOOC Recommendation Method Based on the Fusion of Behavioral Sequences and Textual Semantics
by Wenxin Zhao, Lei Zhao and Zhenbin Liu
Appl. Sci. 2025, 15(18), 10024; https://doi.org/10.3390/app151810024 - 13 Sep 2025
Viewed by 430
Abstract
To address the challenges of user behavior sparsity and insufficient utilization of course semantics on MOOC platforms, this paper proposes a personalized recommendation method that integrates user behavioral sequences with course textual semantic features. First, shallow word-level features from course titles are extracted [...] Read more.
To address the challenges of user behavior sparsity and insufficient utilization of course semantics on MOOC platforms, this paper proposes a personalized recommendation method that integrates user behavioral sequences with course textual semantic features. First, shallow word-level features from course titles are extracted using FastText, and deep contextual semantic representations from course descriptions are obtained via a fine-tuned BERT model. The two sets of semantic features are concatenated to form a multi-level semantic representation of course content. Next, the fused semantic features are mapped into the same vector space as course ID embeddings through a linear projection layer and combined with the original course ID embeddings via an additive fusion strategy, enhancing the model’s semantic perception of course content. Finally, the fused features are fed into an improved SASRec model, where a multi-head self-attention mechanism is employed to model the evolution of user interests, enabling collaborative recommendations across behavioral and semantic modalities. Experiments conducted on the MOOCCubeX dataset (1.26 million users, 632 courses) demonstrated that the proposed method achieved NDCG@10 and HR@10 scores of 0.524 and 0.818, respectively, outperforming SASRec and semantic single-modality baselines. This study offers an efficient yet semantically rich recommendation solution for MOOC scenarios. Full article
Show Figures

Figure 1

22 pages, 1579 KB  
Article
Stance Detection in Arabic Tweets: A Machine Learning Framework for Identifying Extremist Discourse
by Arwa K. Alkhraiji and Aqil M. Azmi
Mathematics 2025, 13(18), 2965; https://doi.org/10.3390/math13182965 - 13 Sep 2025
Viewed by 698
Abstract
Terrorism remains a critical global challenge, and the proliferation of social media has created new avenues for monitoring extremist discourse. This study investigates stance detection as a method to identify Arabic tweets expressing support for or opposition to specific organizations associated with extremist [...] Read more.
Terrorism remains a critical global challenge, and the proliferation of social media has created new avenues for monitoring extremist discourse. This study investigates stance detection as a method to identify Arabic tweets expressing support for or opposition to specific organizations associated with extremist activities, using Hezbollah as a case study. Thousands of relevant Arabic tweets were collected and manually annotated by expert annotators. After extensive preprocessing and feature extraction using term frequency–inverse document frequency (tf-idf), we implemented traditional machine learning (ML) classifiers—Support Vector Machines (SVMs) with multiple kernels, Multinomial Naïve Bayes, and Weighted K-Nearest Neighbors. ML models were selected over deep learning (DL) approaches due to (1) limited annotated Arabic data availability for effective DL training; (2) computational efficiency for resource-constrained environments; and (3) the critical need for interpretability in counterterrorism applications. While interpretability is not a core focus of this work, the use of traditional ML models (rather than DL) makes the system inherently more transparent and readily adaptable for future integration of interpretability techniques. Comparative experiments using FastText word embeddings and tf-idf with supervised classifiers revealed superior performance with the latter approach. Our best result achieved a macro F-score of 78.62% using SVMs with the RBF kernel, demonstrating that interpretable ML frameworks offer a viable and resource-efficient approach for monitoring extremist discourse in Arabic social media. These findings highlight the potential of such frameworks to support scalable and explainable counterterrorism tools in low-resource linguistic settings. Full article
(This article belongs to the Special Issue Machine Learning Theory and Applications)
Show Figures

Figure 1

Back to TopTop