Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (318)

Search Parameters:
Keywords = word class

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
18 pages, 2586 KB  
Article
A Comparative Study of X Data About the NHS Using Sentiment Analysis
by Saeed Ur Rehman, Obi Oluchi Blessing and Anwar Ali
Big Data Cogn. Comput. 2025, 9(10), 244; https://doi.org/10.3390/bdcc9100244 - 24 Sep 2025
Viewed by 87
Abstract
This study investigates sentiment analysis of X data about the National Health Service (NHS) during a politically charged period, using lexicon-based, machine learning, and deep learning approaches, as well as topic modelling and aspect-based sentiment analysis (ABSA). This study is distinct in its [...] Read more.
This study investigates sentiment analysis of X data about the National Health Service (NHS) during a politically charged period, using lexicon-based, machine learning, and deep learning approaches, as well as topic modelling and aspect-based sentiment analysis (ABSA). This study is distinct in its comparative evaluation of sentiment analysis techniques on NHS-related tweets during a politically sensitive period, offering insights into public opinion shaped by political discourse. A dataset of 35,000 tweets collected and analysed using various techniques, including VADER, TextBlob, Naive Bayes, Support Vector Machines, Logistic Regression, Ensemble Learning, and BERT. Unlike previous studies that focus on structured feedback or general sentiment, this research uniquely explores unstructured public discourse during an election period, capturing real-time political sentiment towards NHS policies. The sentiment distribution from lexicon-based methods depicted that the presence of stop words could affect model performance. While all models achieved high accuracy on the validation dataset, challenges such as class imbalance and limited labelled data impacted performance, with signs of overfitting observed. Topic modelling identified nine topic clusters, with “waiting list,” “service,” and “immigration” carrying negative sentiments. At the same time, words like “thank,” “support,” “care,” and “team” had the most positive sentiments, reflecting public delight in these areas. ABSA identified positive sentiments towards aspects like “useful service”. This study contributes a comparative framework for evaluating sentiment analysis techniques in politically contextualised healthcare discourse, offering insights for policymakers and researchers. The study underscores the importance of data quality in sentiment analysis. Future research should consider incorporating multilingual datasets, extending data collection periods, optimising deep learning models, and employing hybrid approaches to enhance performance. Full article
Show Figures

Figure 1

32 pages, 852 KB  
Article
Benchmarking the Responsiveness of Open-Source Text-to-Speech Systems
by Ha Pham Thien Dinh, Rutherford Agbeshi Patamia, Ming Liu and Akansel Cosgun
Computers 2025, 14(10), 406; https://doi.org/10.3390/computers14100406 - 23 Sep 2025
Viewed by 134
Abstract
Responsiveness—the speed at which a text-to-speech (TTS) system produces audible output—is critical for real-time voice assistants yet has received far less attention than perceptual quality metrics. Existing evaluations often touch on latency but do not establish reproducible, open-source standards that capture responsiveness as [...] Read more.
Responsiveness—the speed at which a text-to-speech (TTS) system produces audible output—is critical for real-time voice assistants yet has received far less attention than perceptual quality metrics. Existing evaluations often touch on latency but do not establish reproducible, open-source standards that capture responsiveness as a first-class dimension. This work introduces a baseline benchmark designed to fill that gap. Our framework unifies latency distribution, tail latency, and intelligibility within a transparent and dataset-diverse pipeline, enabling a fair and replicable comparison across 13 widely used open-source TTS models. By grounding evaluation in structured input sets ranging from single words to sentence-length utterances and adopting a methodology inspired by standardized inference benchmarks, we capture both typical and worst-case user experiences. Unlike prior studies that emphasize closed or proprietary systems, our focus is on establishing open, reproducible baselines rather than ranking against commercial references. The results reveal substantial variability across architectures, with some models delivering near-instant responses while others fail to meet interactive thresholds. By centering evaluation on responsiveness and reproducibility, this study provides an infrastructural foundation for benchmarking TTS systems and lays the groundwork for more comprehensive assessments that integrate both fidelity and speed. Full article
Show Figures

Figure 1

29 pages, 2935 KB  
Article
Optimising Contextual Embeddings for Meaning Conflation Deficiency Resolution in Low-Resourced Languages
by Mosima A. Masethe, Sunday O. Ojo and Hlaudi D. Masethe
Computers 2025, 14(9), 402; https://doi.org/10.3390/computers14090402 - 22 Sep 2025
Viewed by 203
Abstract
Meaning conflation deficiency (MCD) presents a continual obstacle in natural language processing (NLP), especially for low-resourced and morphologically complex languages, where polysemy and contextual ambiguity diminish model precision in word sense disambiguation (WSD) tasks. This paper examines the optimisation of contextual embedding models, [...] Read more.
Meaning conflation deficiency (MCD) presents a continual obstacle in natural language processing (NLP), especially for low-resourced and morphologically complex languages, where polysemy and contextual ambiguity diminish model precision in word sense disambiguation (WSD) tasks. This paper examines the optimisation of contextual embedding models, namely XLNet, ELMo, BART, and their improved variations, to tackle MCD in linguistic settings. Utilising Sesotho sa Leboa as a case study, researchers devised an enhanced XLNet architecture with specific hyperparameter optimisation, dynamic padding, early termination, and class-balanced training. Comparative assessments reveal that the optimised XLNet attains an accuracy of 91% and exhibits balanced precision–recall metrics of 92% and 91%, respectively, surpassing both its baseline counterpart and competing models. Optimised ELMo attained the greatest overall metrics (accuracy: 92%, F1-score: 96%), whilst optimised BART demonstrated significant accuracy improvements (96%) despite a reduced recall. The results demonstrate that fine-tuning contextual embeddings using MCD-specific methodologies significantly improves semantic disambiguation for under-represented languages. This study offers a scalable and flexible optimisation approach suitable for additional low-resource language contexts. Full article
Show Figures

Figure 1

18 pages, 3987 KB  
Article
Interactive Application with Virtual Reality and Artificial Intelligence for Improving Pronunciation in English Learning
by Gustavo Caiza, Carlos Villafuerte and Adriana Guanuche
Appl. Sci. 2025, 15(17), 9270; https://doi.org/10.3390/app15179270 - 23 Aug 2025
Viewed by 824
Abstract
Technological advances have enabled the development of innovative educational tools, particularly those aimed at supporting English as a Second Language (ESL) learning, with a specific focus on oral skills. However, pronunciation remains a significant challenge due to the limited availability of personalized learning [...] Read more.
Technological advances have enabled the development of innovative educational tools, particularly those aimed at supporting English as a Second Language (ESL) learning, with a specific focus on oral skills. However, pronunciation remains a significant challenge due to the limited availability of personalized learning opportunities that offer immediate feedback and contextualized practice. In this context, the present research proposes the design, implementation, and validation of an immersive application that leverages virtual reality (VR) and artificial intelligence (AI) to enhance English pronunciation. The proposed system integrates a 3D interactive environment developed in Unity, voice classification models trained using Teachable Machine, and real-time communication with Firebase, allowing users to practice and assess their pronunciation in a simulated library-like virtual setting. Through its integrated AI module, the application can analyze the pronunciation of each word in real time, detecting correct and incorrect utterances, and then providing immediate feedback to help users identify and correct their mistakes. The virtual environment was designed to be a welcoming and user-friendly, promoting active engagement with the learning process. The application’s distributed architecture enables automated feedback generation via data flow between the cloud-based AI, the database, and the visualization interface. Results demonstrate that using 400 samples per class and a confidence threshold of 99.99% for training the AI model effectively eliminated false positives, significantly increasing system accuracy and providing users with more reliable feedback. This directly contributes to enhanced learner autonomy and improved ESL acquisition outcomes. Furthermore, user surveys conducted to understand their perceptions of the application’s usefulness as a support tool for English learning yielded an average acceptance rate of 93%. This reflects the acceptance of these immersive technologies in educational contexts, as the combination of these technologies offers a realistic and user-friendly simulation environment, in addition to detailed word analysis, facilitating self-assessment and independent learning among students. Full article
Show Figures

Figure 1

26 pages, 823 KB  
Article
Reconciling Teaching and Research Tensions: A Sustainability Framework for Expert Teacher Development in Research Intensive Universities
by Yue Huang, Lin Jiang and Ruirui Zhai
Sustainability 2025, 17(15), 7113; https://doi.org/10.3390/su17157113 - 6 Aug 2025
Viewed by 634
Abstract
The sustainable development of teaching expertise in research-intensive universities remains a critical global challenge. This study investigates the distinctive characteristics of expert teachers—exemplary faculty in research universities—addressing their developmental trajectories and motivational mechanisms within prevailing incentive systems that prioritize research productivity over pedagogical [...] Read more.
The sustainable development of teaching expertise in research-intensive universities remains a critical global challenge. This study investigates the distinctive characteristics of expert teachers—exemplary faculty in research universities—addressing their developmental trajectories and motivational mechanisms within prevailing incentive systems that prioritize research productivity over pedagogical excellence. Employing grounded theory methodology, we conducted iterative coding of 20,000-word interview transcripts from 13 teaching-awarded professors at Chinese “Double First-Class” universities. Key findings reveal the following: (1) Compared to the original K-12 expert teacher model, university-level teaching experts exhibit distinctive disciplinary mastery—characterized by systematic knowledge structuring and cross-disciplinary integration capabilities. (2) Their developmental trajectory transcends linear expertise acquisition, instead manifesting as a problem-solving continuum across four nonlinear phases: career initiation, dilemma adaptation, theoretical consciousness, and leadership expansion. (3) Sustainable teaching excellence relies fundamentally on teachers’ professional passion, sustained through a virtuous cycle of high-quality instructional engagement and external validation (including positive student feedback, institutional recognition, and peer collaboration). Universities must establish comprehensive support systems—including (a) fostering a supportive and flexible learning atmosphere, (b) reforming evaluation mechanisms, and (c) facilitating interdisciplinary collaboration through teaching development communities—to institutionalize this developmental ecosystem. Full article
Show Figures

Figure 1

20 pages, 390 KB  
Article
Injective Hulls of Infinite Totally Split-Decomposable Metric Spaces
by Maël Pavón
Axioms 2025, 14(8), 606; https://doi.org/10.3390/axioms14080606 - 4 Aug 2025
Viewed by 445
Abstract
We extend the theory of splits in finite metric spaces to infinite ones. Within this more general framework, we investigate the class of spaces having metrics that are integer-valued and totally split-decomposable, as well as the polyhedral complex structure of their injective hulls. [...] Read more.
We extend the theory of splits in finite metric spaces to infinite ones. Within this more general framework, we investigate the class of spaces having metrics that are integer-valued and totally split-decomposable, as well as the polyhedral complex structure of their injective hulls. For this class, we provide a characterization for the injective hull to be combinatorially equivalent to a CAT(0) cube complex. Intermediate results include the generalization of the decomposition theory introduced by Bandelt and Dress in 1992 as well as results on the tight span of totally split-decomposable metric spaces proved by Huber, Koolen, and Moulton in 2006. Next, using results of Lang from 2013, we obtain proper actions on CAT(0) cube complexes for finitely generated groups endowed with a totally split-decomposable word metric and for which the associated splits satisfy a simple combinatorial property. In the case of Gromov hyperbolic groups, the obtained action is both proper aand co-compact. Finally, we obtain as an application that injective hulls of odd cycles are cell complexes isomorphic to CAT(0) cube complexes. Full article
(This article belongs to the Section Geometry and Topology)
Show Figures

Graphical abstract

18 pages, 839 KB  
Article
From Narratives to Diagnosis: A Machine Learning Framework for Classifying Sleep Disorders in Aging Populations: The sleepCare Platform
by Christos A. Frantzidis
Brain Sci. 2025, 15(7), 667; https://doi.org/10.3390/brainsci15070667 - 20 Jun 2025
Viewed by 1199
Abstract
Background/Objectives: Sleep disorders are prevalent among aging populations and are often linked to cognitive decline, chronic conditions, and reduced quality of life. Traditional diagnostic methods, such as polysomnography, are resource-intensive and limited in accessibility. Meanwhile, individuals frequently describe their sleep experiences through [...] Read more.
Background/Objectives: Sleep disorders are prevalent among aging populations and are often linked to cognitive decline, chronic conditions, and reduced quality of life. Traditional diagnostic methods, such as polysomnography, are resource-intensive and limited in accessibility. Meanwhile, individuals frequently describe their sleep experiences through unstructured narratives in clinical notes, online forums, and telehealth platforms. This study proposes a machine learning pipeline (sleepCare) that classifies sleep-related narratives into clinically meaningful categories, including stress-related, neurodegenerative, and breathing-related disorders. The proposed framework employs natural language processing (NLP) and machine learning techniques to support remote applications and real-time patient monitoring, offering a scalable solution for the early identification of sleep disturbances. Methods: The sleepCare consists of a three-tiered classification pipeline to analyze narrative sleep reports. First, a baseline model used a Multinomial Naïve Bayes classifier with n-gram features from a Bag-of-Words representation. Next, a Support Vector Machine (SVM) was trained on GloVe-based word embeddings to capture semantic context. Finally, a transformer-based model (BERT) was fine-tuned to extract contextual embeddings, using the [CLS] token as input for SVM classification. Each model was evaluated using stratified train-test splits and 10-fold cross-validation. Hyperparameter tuning via GridSearchCV optimized performance. The dataset contained 475 labeled sleep narratives, classified into five etiological categories relevant for clinical interpretation. Results: The transformer-based model utilizing BERT embeddings and an optimized Support Vector Machine classifier achieved an overall accuracy of 81% on the test set. Class-wise F1-scores ranged from 0.72 to 0.91, with the highest performance observed in classifying normal or improved sleep (F1 = 0.91). The macro average F1-score was 0.78, indicating balanced performance across all categories. GridSearchCV identified the optimal SVM parameters (C = 4, kernel = ‘rbf’, gamma = 0.01, degree = 2, class_weight = ‘balanced’). The confusion matrix revealed robust classification with limited misclassifications, particularly between overlapping symptom categories such as stress-related and neurodegenerative sleep disturbances. Conclusions: Unlike generic large language model applications, our approach emphasizes the personalized identification of sleep symptomatology through targeted classification of the narrative input. By integrating structured learning with contextual embeddings, the framework offers a clinically meaningful, scalable solution for early detection and differentiation of sleep disorders in diverse, real-world, and remote settings. Full article
(This article belongs to the Special Issue Perspectives of Artificial Intelligence (AI) in Aging Neuroscience)
Show Figures

Graphical abstract

24 pages, 2410 KB  
Article
UA-HSD-2025: Multi-Lingual Hate Speech Detection from Tweets Using Pre-Trained Transformers
by Muhammad Ahmad, Muhammad Waqas, Ameer Hamza, Sardar Usman, Ildar Batyrshin and Grigori Sidorov
Computers 2025, 14(6), 239; https://doi.org/10.3390/computers14060239 - 18 Jun 2025
Cited by 1 | Viewed by 2173
Abstract
The rise in social media has improved communication but also amplified the spread of hate speech, creating serious societal risks. Automated detection remains difficult due to subjectivity, linguistic diversity, and implicit language. While prior research focuses on high-resource languages, this study addresses the [...] Read more.
The rise in social media has improved communication but also amplified the spread of hate speech, creating serious societal risks. Automated detection remains difficult due to subjectivity, linguistic diversity, and implicit language. While prior research focuses on high-resource languages, this study addresses the underexplored multilingual challenges of Arabic and Urdu hate speech through a comprehensive approach. To achieve this objective, this study makes four different key contributions. First, we have created a unique multi-lingual, manually annotated binary and multi-class dataset (UA-HSD-2025) sourced from X, which contains the five most important multi-class categories of hate speech. Secondly, we created detailed annotation guidelines to make a robust and perfect hate speech dataset. Third, we explore two strategies to address the challenges of multilingual data: a joint multilingual and translation-based approach. The translation-based approach involves converting all input text into a single target language before applying a classifier. In contrast, the joint multilingual approach employs a unified model trained to handle multiple languages simultaneously, enabling it to classify text across different languages without translation. Finally, we have employed state-of-the-art 54 different experiments using different machine learning using TF-IDF, deep learning using advanced pre-trained word embeddings such as FastText and Glove, and pre-trained language-based models using advanced contextual embeddings. Based on the analysis of the results, our language-based model (XLM-R) outperformed traditional supervised learning approaches, achieving 0.99 accuracy in binary classification for Arabic, Urdu, and joint-multilingual datasets, and 0.95, 0.94, and 0.94 accuracy in multi-class classification for joint-multilingual, Arabic, and Urdu datasets, respectively. Full article
(This article belongs to the Special Issue Recent Advances in Social Networks and Social Media)
Show Figures

Figure 1

18 pages, 456 KB  
Article
Named Entity Recognition Based on Multi-Class Label Prompt Selection and Core Entity Replacement
by Di Wu, Yao Chen and Mingyue Yan
Appl. Sci. 2025, 15(11), 6171; https://doi.org/10.3390/app15116171 - 30 May 2025
Viewed by 709
Abstract
At present, researchers are showing a marked interest in the topic of few-shot named entity recognition (NER). Previous studies have demonstrated that prompt-based learning methods can effectively improve the performance of few-shot NER models and can reduce the need for annotated data. However, [...] Read more.
At present, researchers are showing a marked interest in the topic of few-shot named entity recognition (NER). Previous studies have demonstrated that prompt-based learning methods can effectively improve the performance of few-shot NER models and can reduce the need for annotated data. However, the contextual information of the relationship between core entities and a given prompt may not have been considered in these studies; moreover, research in this field continues to suffer from the negative impact of a limited amount of annotated data. A multi-class label prompt selection and core entity replacement-based named entity recognition (MPSCER-NER) model is proposed in this study. A multi-class label prompt selection strategy is presented, which can assist in the task of sentence–word representation. A long-distance dependency is formed between the sentence and the multi-class label prompt. A core entity replacement strategy is presented, which can enrich the word vectors of training data. In addition, a weighted random algorithm is used to retrieve the core entities that are to be replaced from the multi-class label prompt. The experimental results show that, when implemented on the CoNLL-2003, Ontonotes 5.0, Ontonotes 4.0, and BC5CDR datasets under 5-Way k-Shot (k = 5, 10), the MPSCER-NER model achieves minimum F1-score improvements of 1.32%, 2.14%, 1.05%, 1.32%, 0.84%, 1.46%, 1.43%, and 1.11% in comparison with NNshot, StructShot, MatchingCNN, ProtoBERT, DNER, and SRNER, respectively. Full article
Show Figures

Figure 1

27 pages, 3040 KB  
Article
Optimisation of the Production Process of Ironing Refractory Products Using the OEE Indicator as Part of Innovative Solutions for Sustainable Production
by Mariusz Niekurzak and Wojciech Lewicki
Sustainability 2025, 17(11), 4779; https://doi.org/10.3390/su17114779 - 22 May 2025
Cited by 1 | Viewed by 738
Abstract
The article addresses the problem of optimising a selected production process in a company from the refractory products industry. As part of the research, individual activities were divided, identifying key wastes occurring in the production process. In addition, the 5S (the 5S [...] Read more.
The article addresses the problem of optimising a selected production process in a company from the refractory products industry. As part of the research, individual activities were divided, identifying key wastes occurring in the production process. In addition, the 5S (the 5S methodology—Sort, Set in Order, Shine, Standardise, and Sustain) quality system was modified, its efficiency was increased, and a better work organisation was established based on it. Data from the actual production process were analysed based on total work efficiency using the OEE (Overall Equipment Effectiveness) coefficient. The use of machine working time was indicated, and key parameters were determined, i.e., availability, efficiency, and quality of the implemented production processes. The results obtained in the course of the research were compared to the Word Class OEE standards. The goal of the work is to indicate possibilities and recommendations for increasing production efficiency without increasing costs, thanks to actions reducing the number of production defects and optimal distribution of employees on the production line. The presented analyses can help assess the management processes of other manufacturing companies operating in this highly specialised manufacturing sector. At the same time, the research conclusions enable other entities to evaluate the implementation of the proposed solutions in practice without incurring unnecessary financial outlays on improving production processes. Full article
(This article belongs to the Special Issue Recent Advances in Modern Technologies for Sustainable Manufacturing)
Show Figures

Figure 1

15 pages, 819 KB  
Article
Reinforcement Learning-Based Intent Classification of Chinese Questions About Respiratory Diseases
by Hao Wu, Degen Huang and Xiaohui Lin
Appl. Sci. 2025, 15(7), 3983; https://doi.org/10.3390/app15073983 - 4 Apr 2025
Viewed by 765
Abstract
The intent classification of Chinese questions about respiratory diseases (CQRD) can not only promote the development of smart medical care, but also strengthen epidemic surveillance. The major core of the intent classification of CQRD is text representation. This paper studies how to utilize [...] Read more.
The intent classification of Chinese questions about respiratory diseases (CQRD) can not only promote the development of smart medical care, but also strengthen epidemic surveillance. The major core of the intent classification of CQRD is text representation. This paper studies how to utilize keywords to construct CQRD representation. Based on the characteristics of CQRD, we propose a keywords-based reinforcement learning model. In the reinforcement learning model based on keywords, we crafted a word frequency reward function to aid in generating the reward function and determining keyword categories. Simultaneously, to generate CQRD representations using keywords, we developed two models: keyword-driven LSTM (KD-LSTM) and keyword-driven GCN (KD-GCN). The KD-LSTM incorporates two methods: one based on word weights and the other based on category vectors. The KD-GCN employs keywords to construct a weight matrix for training. The method based on word weight achieves the best results on the CQRD_28000 dataset, which is 0.72% higher than the Bi-LSTM model. The method based on category vector outperforms the Bi-LSTM model in the CQRD_8000 dataset by 2.41%. The KD-GCN, although not attaining the optimal outcome, exhibited a superior performance of 3.12% compared to the GCN model. Both methods have significantly improved the classification results of minority classes. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

24 pages, 9841 KB  
Article
Mexican Sign Language Recognition: Dataset Creation and Performance Evaluation Using MediaPipe and Machine Learning Techniques
by Mario Rodriguez, Outmane Oubram, A. Bassam, Noureddine Lakouari and Rasikh Tariq
Electronics 2025, 14(7), 1423; https://doi.org/10.3390/electronics14071423 - 1 Apr 2025
Cited by 4 | Viewed by 1705
Abstract
In Mexico, around 2.4 million people (1.9% of the national population) are deaf, and Mexican Sign Language (MSL) support is essential for people with communication disabilities. Research and technological prototypes of sign language recognition have been developed to support public communication systems without [...] Read more.
In Mexico, around 2.4 million people (1.9% of the national population) are deaf, and Mexican Sign Language (MSL) support is essential for people with communication disabilities. Research and technological prototypes of sign language recognition have been developed to support public communication systems without human interpreters. However, most of these systems and research are closely related to American Sign Language (ASL) or other sign languages of other languages whose scope has had the highest level of accuracy and recognition of letters and words. The objective of the current study is to develop and evaluate a sign language recognition system tailored to MSL. The research aims to achieve accurate recognition of dactylology and the first ten numerical digits (1–10) in MSL. A database of sign language and numeration of MSL was created with the 29 different characters of MSL’s dactylology and the first ten digits with a camera. Then, MediaPipe was first applied for feature extraction for both hands (21 points per hand). Once the features were extracted, Machine Learning and Deep Learning Techniques were applied to recognize MSL signs. The recognition of MSL patterns in the context of static (29 classes) and continuous signs (10 classes) yielded an accuracy of 92% with Support Vector Machine (SVM) and 86% with Gated Recurrent Unit (GRU) accordingly. The trained algorithms are based on full scenarios with both hands; therefore, it will sign under these conditions. To improve the accuracy, it is suggested to amplify the number of samples. Full article
Show Figures

Figure 1

29 pages, 4979 KB  
Article
Land Cover Classification Model Using Multispectral Satellite Images Based on a Deep Learning Synergistic Semantic Segmentation Network
by Abdorreza Alavi Gharahbagh, Vahid Hajihashemi, José J. M. Machado and João Manuel R. S. Tavares
Sensors 2025, 25(7), 1988; https://doi.org/10.3390/s25071988 - 22 Mar 2025
Cited by 2 | Viewed by 2512
Abstract
Land cover classification (LCC) using satellite images is one of the rapidly expanding fields in mapping, highlighting the need for updating existing computational classification methods. Advances in technology and the increasing variety of applications have introduced challenges, such as more complex classes and [...] Read more.
Land cover classification (LCC) using satellite images is one of the rapidly expanding fields in mapping, highlighting the need for updating existing computational classification methods. Advances in technology and the increasing variety of applications have introduced challenges, such as more complex classes and a demand for greater detail. In recent years, deep learning and Convolutional Neural Networks (CNNs) have significantly enhanced the segmentation of satellite images. Since the training of CNNs requires sophisticated and expensive hardware and significant time, using pre-trained networks has become widespread in the segmentation of satellite image. This study proposes a hybrid synergistic semantic segmentation method based on the Deeplab v3+ network and a clustering-based post-processing scheme. The proposed method accurately classifies various land cover (LC) types in multispectral satellite images, including Pastures, Other Built-Up Areas, Water Bodies, Urban Areas, Grasslands, Forest, Farmland, and Others. The post-processing scheme includes a spectral bag-of-words model and K-medoids clustering to refine the Deeplab v3+ outputs and correct possible errors. The simulation results indicate that combining the post-processing scheme with deep learning improves the Matthews correlation coefficient (MCC) by approximately 5.7% compared to the baseline method. Additionally, the proposed approach is robust to data imbalance cases and can dynamically update its codewords over different seasons. Finally, the proposed synergistic semantic segmentation method was compared with several state-of-the-art segmentation methods in satellite images of Italy’s Lake Garda (Lago di Garda) region. The results showed that the proposed method outperformed the best existing techniques by at least 6% in terms of MCC. Full article
Show Figures

Figure 1

28 pages, 4947 KB  
Article
The Detection of Spurious Correlations in Public Bidding and Contract Descriptions Using Explainable Artificial Intelligence and Unsupervised Learning
by Hélcio de Abreu Soares, Raimundo Santos Moura, Vinícius Ponte Machado, Anselmo Paiva, Weslley Lima and Rodrigo Veras
Electronics 2025, 14(7), 1251; https://doi.org/10.3390/electronics14071251 - 22 Mar 2025
Cited by 1 | Viewed by 1328
Abstract
Artificial Intelligence (AI) models, including deep learning and rule-based approaches, often function as black boxes, limiting transparency and increasing uncertainty in decisions. This study addresses spurious correlations, defined as associations between patterns and classes that do not reflect causal relationships, affecting AI models’ [...] Read more.
Artificial Intelligence (AI) models, including deep learning and rule-based approaches, often function as black boxes, limiting transparency and increasing uncertainty in decisions. This study addresses spurious correlations, defined as associations between patterns and classes that do not reflect causal relationships, affecting AI models’ reliability and applicability. In Natural Language Processing (NLP), these correlations lead to inaccurate predictions, biases, and challenges in model generalization. We propose a method that employs Explainable Artificial Intelligence (XAI) techniques to detect spurious patterns in textual datasets for binary classification tasks. The method applies the K-means algorithm to cluster patterns and interprets them based on their distance from centroids. It hypothesizes that patterns farther from the centroids are more likely to be spurious than those closer to them. We apply the method to public procurement datasets from the Court of Auditors of Piauí (TCE-PI) using models based on Support Vector Machine (SVM) and Logistic Regression with text representations from TFIDF and Word Embeddings, as well as a BERT model. The analysis is extended to the IMDB movie review dataset to evaluate generalizability. The results support the hypothesis that patterns farther from centroids exhibit higher spuriousness potential and demonstrate the clustering’s consistency across models and datasets. The method operates independently of the techniques used in its stages, enabling the automatic detection and quantification of spurious patterns without prior human intervention. Full article
(This article belongs to the Special Issue Advanced Natural Language Processing Technology and Applications)
Show Figures

Figure 1

22 pages, 6014 KB  
Article
Evaluation of Industrial Water Use Efficiency on an Enterprise Scale Based on Analytic Hierarchy Process, Entropy Weight Method and Self-Organizing Map: A Case Study in Zhejiang, China
by Yimin Qian, Yingjie Zhao, Hao Qian, Junhong Xiang, Caiming Chen, Longqiang Su and Chenkai Cai
Water 2025, 17(6), 901; https://doi.org/10.3390/w17060901 - 20 Mar 2025
Cited by 1 | Viewed by 822
Abstract
The increasingly serious imbalance between the supply and demand of water resources necessitates the establishment of a scientific and reasonable comprehensive evaluation method for industrial water use efficiency (WUE). In this study, a general method for industrial WUE evaluation on an enterprise scale [...] Read more.
The increasingly serious imbalance between the supply and demand of water resources necessitates the establishment of a scientific and reasonable comprehensive evaluation method for industrial water use efficiency (WUE). In this study, a general method for industrial WUE evaluation on an enterprise scale was proposed by combining the analytic hierarchy process (AHP), entropy weight method (EWM), and self-organizing map (SOM), and it was tested in several areas of Zhejiang Province, China. The results show that the composite indexes generated using the AHP and EWM were different and were employed as the input of the SOM to divide enterprises into four categories. Most enterprises were classified as Class A, with a relatively high WUE, accounting for 82.5% of the total, while those in Class D, with a relatively low WUE, only accounted for 0.5% of the total. Furthermore, the differences in WUE for industry classification and spatial distribution were also analyzed. The classification results of several industries were more diverse, especially for those industries in which water plays an important role in production. Moreover, the spatial distribution of WUE classifications also implied that the clustering of enterprises has a positive effect on the improvement in WUE. In other words, it is feasible to improve WUE through industry clustering and sub-industry management. In summary, a comprehensive, detailed evaluation of industrial WUE was conducted on an enterprise scale, which can also be applied to other areas and used as a reference for local water resource managers for formulating targeted policies. Full article
(This article belongs to the Section Water Use and Scarcity)
Show Figures

Figure 1

Back to TopTop