Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (3,549)

Search Parameters:
Keywords = design languages

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
25 pages, 998 KB  
Article
Modeling Kinematic and Dynamic Structures with Hypergraph-Based Formalism
by Csaba Hajdu and Norbert Hegyi
Appl. Mech. 2025, 6(4), 74; https://doi.org/10.3390/applmech6040074 (registering DOI) - 9 Oct 2025
Abstract
This paper introduces a hypergraph-based formalism for modeling kinematic and dynamic structures in robotics, addressing limitations of the existing formats such as Unified Robot Description Format (URDF), MuJoCo-XML, and Simulation Description Format (SDF). Our method represents mechanical constraints and connections as hyperedges, enabling [...] Read more.
This paper introduces a hypergraph-based formalism for modeling kinematic and dynamic structures in robotics, addressing limitations of the existing formats such as Unified Robot Description Format (URDF), MuJoCo-XML, and Simulation Description Format (SDF). Our method represents mechanical constraints and connections as hyperedges, enabling the native description of multi-joint closures, tendon-driven actuation, and multi-physics coupling. We present a tensor-based representation derived via star-expansion, implemented in the Hypergraph Model Cognition Framework (HyMeKo) language. Comparative experiments show a substantial reduction in model verbosity compared to URDF while retaining expressiveness for large-language model integration. The approach is demonstrated on simple robotic arms and a quarter vehicle model, with derived state-space equations. This work suggests that hypergraph-based models can provide a modular, compact, and semantically rich alternative for the next-generation simulation and design workflows. The introduced formalism reaches 50% reduction compared to URDF descriptions and 20% reduction compared to MuJoCo-XML descriptions. Full article
16 pages, 7184 KB  
Article
Towards Robust Scene Text Recognition: A Dual Correction Mechanism with Deformable Alignment
by Yajiao Feng and Changlu Li
Electronics 2025, 14(19), 3968; https://doi.org/10.3390/electronics14193968 - 9 Oct 2025
Abstract
Scene Text Recognition (STR) faces significant challenges under complex degradation conditions, such as distortion, occlusion, and semantic ambiguity. Most existing methods rely heavily on language priors for correction, but effectively constructing language rules remains a complex problem. This paper addresses two key challenges: [...] Read more.
Scene Text Recognition (STR) faces significant challenges under complex degradation conditions, such as distortion, occlusion, and semantic ambiguity. Most existing methods rely heavily on language priors for correction, but effectively constructing language rules remains a complex problem. This paper addresses two key challenges: (1) The over-correction behavior of language models, particularly on semantically deficient input, can result in both recognition errors and loss of critical information. (2) Character misalignment in visual features, which affects recognition accuracy. To address these problems, we propose a Deformable-Alignment-based Dual Correction Mechanism (DADCM) for STR. Our method includes the following key components: (1) We propose a visually guided and language-assisted correction strategy. A dynamic confidence threshold is used to control the degree of language model intervention. (2) We designed a visual backbone network called SCRTNet. The net enhances key text regions through a channel attention module (SENet) and applies deformable convolution (DCNv4) in deep layers to better model distorted or curved text. (3) We propose a deformable alignment module (DAM). The module combines Gumbel-Softmax-based anchor sampling and geometry-aware self-attention to improve character alignment. Experiments on multiple benchmark datasets demonstrate the superiority of our approach. Especially on the Union14M-Benchmark, where the recognition accuracy surpasses previous methods by 1.1%, 1.6%, 3.0%, and 1.3% on the Curved, Multi-Oriented, Contextless, and General subsets, respectively. Full article
Show Figures

Figure 1

15 pages, 897 KB  
Article
Comparative Assessment of Large Language Models in Optics and Refractive Surgery: Performance on Multiple-Choice Questions
by Leah Attal, Elad Shvartz, Alon Gorenshtein, Shirley Pincovich and Daniel Bahir
Vision 2025, 9(4), 85; https://doi.org/10.3390/vision9040085 (registering DOI) - 9 Oct 2025
Abstract
This study aimed to evaluate the performance of seven advanced AI Large Language Models (LLMs)—ChatGPT 4o, ChatGPT O3 Mini, ChatGPT O1, DeepSeek V3, DeepSeek R1, Gemini 2.0 Flash, and Grok-3—in answering multiple-choice questions (MCQs) in optics and refractive surgery, to assess their role [...] Read more.
This study aimed to evaluate the performance of seven advanced AI Large Language Models (LLMs)—ChatGPT 4o, ChatGPT O3 Mini, ChatGPT O1, DeepSeek V3, DeepSeek R1, Gemini 2.0 Flash, and Grok-3—in answering multiple-choice questions (MCQs) in optics and refractive surgery, to assess their role in medical education for residents. The AI models were tested using 134 publicly available MCQs from national ophthalmology certification exams, categorized by the need to perform calculations, the relevant subspecialty, and the use of images. Accuracy was analyzed and compared statistically. ChatGPT O1 achieved the highest overall accuracy (83.5%), excelling in complex optical calculations (84.1%) and optics questions (82.4%). DeepSeek V3 displayed superior accuracy in refractive surgery-related questions (89.7%), followed by ChatGPT O3 Mini (88.4%). ChatGPT O3 Mini significantly outperformed others in image analysis, with 88.2% accuracy. Moreover, ChatGPT O1 demonstrated comparable accuracy rates for both calculated and non-calculated questions (84.1% vs. 83.3%). This is in stark contrast to other models, which exhibited significant discrepancies in accuracy for calculated and non-calculated questions. The findings highlight the ability of LLMs to achieve high accuracy in ophthalmology MCQs, particularly in complex optical calculations and visual items. These results suggest potential applications in exam preparation and medical training contexts, while underscoring the need for future studies designed to directly evaluate their role and impact in medical education. The findings highlight the significant potential of AI models in ophthalmology education, particularly in performing complex optical calculations and visual stem questions. Future studies should utilize larger, multilingual datasets to confirm and extend these preliminary findings. Full article
Show Figures

Figure 1

25 pages, 4379 KB  
Review
Bridging Global Perspectives: A Comparative Review of Agent-Based Modeling for Block-Level Walkability in Chinese and International Research
by Yidan Wang, Renzhang Wang, Xiaowen Xu, Bo Zhang, Marcus White and Xiaoran Huang
Buildings 2025, 15(19), 3613; https://doi.org/10.3390/buildings15193613 - 9 Oct 2025
Abstract
As cities strive for human-centered and fine-tuned development, Agent-Based Modeling (ABM) has emerged as a powerful tool for simulating pedestrian behavior and optimizing walkable neighborhood design. This study presents a comparative bibliometric analysis of ABM applications in block-scale walkability research from 2015 to [...] Read more.
As cities strive for human-centered and fine-tuned development, Agent-Based Modeling (ABM) has emerged as a powerful tool for simulating pedestrian behavior and optimizing walkable neighborhood design. This study presents a comparative bibliometric analysis of ABM applications in block-scale walkability research from 2015 to 2024, drawing on both Chinese- and English-language literature. Using visualization tools such as VOSviewer, the analysis reveals divergences in national trajectories, methodological approaches, and institutional logics. Chinese research demonstrates a policy-driven growth pattern, particularly following the introduction of the “15-Minute Community Life Circle” initiative, with an emphasis on neighborhood renewal, age-friendly design, and transit-oriented planning. In contrast, international studies show a steady output driven by technological innovation, integrating methods such as deep learning, semantic segmentation, and behavioral simulation to address climate resilience, equity, and mobility complexity. The study also classifies ABM applications into five key application domains, highlighting how Chinese and international studies differ in focus, data inputs, and implementation strategies. Despite these differences, both research streams recognize the value of ABM in transport planning, public health, and low-carbon urbanism. Key challenges identified include data scarcity, algorithmic limitations, and ethical concerns. The study concludes with future research directions, including multimodal data fusion, integration with extended reality, and the development of privacy-aware, cross-cultural modeling standards. These findings reinforce ABM’s potential as a smart urban simulation tool for advancing adaptive, human-centered, and sustainable neighborhood planning. Full article
(This article belongs to the Special Issue Sustainable Urban and Buildings: Lastest Advances and Prospects)
Show Figures

Figure 1

14 pages, 1917 KB  
Article
Moroccan Sign Language Recognition with a Sensory Glove Using Artificial Neural Networks
by Hasnae El Khoukhi, Assia Belatik, Imane El Manaa, My Abdelouahed Sabri, Yassine Abouch and Abdellah Aarab
Digital 2025, 5(4), 53; https://doi.org/10.3390/digital5040053 - 8 Oct 2025
Abstract
Every day, countless individuals with hearing or speech disabilities struggle to communicate effectively, as their conditions limit conventional verbal interaction. For them, sign language becomes an essential and often sole tool for expressing thoughts and engaging with others. However, the general public’s limited [...] Read more.
Every day, countless individuals with hearing or speech disabilities struggle to communicate effectively, as their conditions limit conventional verbal interaction. For them, sign language becomes an essential and often sole tool for expressing thoughts and engaging with others. However, the general public’s limited understanding of sign language poses a major barrier, often resulting in social, educational, and professional exclusion. To bridge this communication gap, the present study proposes a smart wearable glove system designed to translate Arabic sign language (ArSL), especially Moroccan sign language (MSL), into a written alphabet in real time. The glove integrates five MPU6050 motion sensors, one on each finger, capable of capturing detailed motion data, including angular velocity and linear acceleration. These motion signals are processed using an Artificial Neural Network (ANN), implemented directly on a Raspberry Pi Pico through embedded machine learning techniques. A custom dataset comprising labeled gestures corresponding to the MSL alphabet was developed for training the model. Following the training phase, the neural network attained a gesture recognition accuracy of 98%, reflecting strong performance in terms of reliability and classification precision. We developed an affordable and portable glove system aimed at improving daily communication for individuals with hearing impairments in Morocco, contributing to greater inclusivity and improved accessibility. Full article
Show Figures

Figure 1

18 pages, 437 KB  
Article
“It Was Horrible!” Understanding the Transition Experiences of Direct Year 2 Entry Students in Computer Science
by Mireilla Bikanga Ada
Trends High. Educ. 2025, 4(4), 57; https://doi.org/10.3390/higheredu4040057 - 8 Oct 2025
Abstract
While first-year transitions are well studied, less is known about students who enter directly into Year 2 of a four-year Scottish Computing Science degree via international foundation programmes, UK colleges, or high schools. This study investigated their academic preparedness, use of AI tools, [...] Read more.
While first-year transitions are well studied, less is known about students who enter directly into Year 2 of a four-year Scottish Computing Science degree via international foundation programmes, UK colleges, or high schools. This study investigated their academic preparedness, use of AI tools, English language confidence, and transition challenges. Using a mixed-methods design, 77 students completed a survey with Likert-scale and open-ended items. Findings indicate gaps in programming skills, independent learning, and understanding academic expectations. Many students reported feeling a sense of low social belonging after joining pre-established cohorts. AI tools were commonly used for programming support and concept clarification, but they offered limited emotional reassurance. Students recommended clearer academic alignment, a tailored induction process, compulsory social events, and peer mentoring. This study advocates for equity-driven transition models that cater to the diverse needs of direct entrants, thereby fostering inclusion, belonging, and success. Full article
Show Figures

Figure 1

22 pages, 4797 KB  
Article
Early Oral Cancer Detection with AI: Design and Implementation of a Deep Learning Image-Based Chatbot
by Pablo Ormeño-Arriagada, Gastón Márquez, Carla Taramasco, Gustavo Gatica, Juan Pablo Vasconez and Eduardo Navarro
Appl. Sci. 2025, 15(19), 10792; https://doi.org/10.3390/app151910792 - 7 Oct 2025
Abstract
Oral cancer remains a critical global health challenge, with delayed diagnosis driving high morbidity and mortality. Despite progress in artificial intelligence, computer vision, and medical imaging, early detection tools that are accessible, explainable, and designed for patient engagement remain limited. This study presents [...] Read more.
Oral cancer remains a critical global health challenge, with delayed diagnosis driving high morbidity and mortality. Despite progress in artificial intelligence, computer vision, and medical imaging, early detection tools that are accessible, explainable, and designed for patient engagement remain limited. This study presents a novel system that combines a patient-centred chatbot with a deep learning framework to support early diagnosis, symptom triage, and health education. The system integrates convolutional neural networks, class activation mapping, and natural language processing within a conversational interface. Five deep learning models were evaluated (CNN, DenseNet121, DenseNet169, DenseNet201, and InceptionV3) using two balanced public datasets. Model performance was assessed using accuracy, sensitivity, specificity, diagnostic odds ratio (DOR), and Cohen’s Kappa. InceptionV3 consistently outperformed the other models across these metrics, achieving the highest diagnostic accuracy (77.6%) and DOR (20.67), and was selected as the core engine of the chatbot’s diagnostic module. The deployed chatbot provides real-time image assessments and personalised conversational support via multilingual web and mobile platforms. By combining automated image interpretation with interactive guidance, the system promotes timely consultation and informed decision-making. It offers a prototype for a chatbot, which is scalable and serves as a low-cost solution for underserved populations and demonstrates strong potential for integration into digital health pathways. Importantly, the system is not intended to function as a formal screening tool or replace clinical diagnosis; rather, it provides preliminary guidance to encourage early medical consultation and informed health decisions. Full article
Show Figures

Figure 1

12 pages, 229 KB  
Article
Cross-Cultural Adaptation and Validation of the Mini-Eating and Drinking Ability Classification System for Korean Children with Cerebral Palsy Aged 18–36 Months
by You Gyoung Yi, Seoyon Yang, Jeong-Yi Kwon, Dong-wook Rha, Juntaek Hong, Ja Young Choi, Eun Jae Ko, Bo Young Hong and Dae-Hyun Jang
Children 2025, 12(10), 1348; https://doi.org/10.3390/children12101348 - 7 Oct 2025
Abstract
Background/Objectives: Feeding and swallowing difficulties are common in young children with cerebral palsy (CP), yet no validated tool has been available in Korea for those under 3 years. The Mini-Eating and Drinking Ability Classification System (Mini-EDACS) was designed for children aged 18–36 months. [...] Read more.
Background/Objectives: Feeding and swallowing difficulties are common in young children with cerebral palsy (CP), yet no validated tool has been available in Korea for those under 3 years. The Mini-Eating and Drinking Ability Classification System (Mini-EDACS) was designed for children aged 18–36 months. This study aimed to translate the Mini-EDACS into Korean and evaluate its reliability and validity. Methods: Translation followed international guidelines, including forward–backward translation and Delphi consensus with experts in pediatric dysphagia. Forty-eight children with CP (mean age 27.1 ± 5.0 months) were assessed. Caregivers and speech–language pathologists (SLPs) independently rated Mini-EDACS and assistance levels. Inter-rater reliability was examined using Cohen’s κ. Construct validity was tested by Spearman’s correlations with the Gross Motor Function Classification System (GMFCS), Mini-MACS, the Communication Function Classification System (CFCS), the Visual Function Classification System (VFCS), and the Functional Oral Intake Scale for Children (FOIS-C). Results: Agreement between caregivers and SLPs was excellent (κ = 0.90; weighted κ = 0.98). Assistance-level ratings also showed almost perfect concordance (κ = 0.97). Mini-EDACS correlated strongly with FOIS-C (ρ = −0.86, p < 0.001) and with assistance levels (ρ = 0.81, p < 0.001). Moderate-to-strong positive correlations were observed with GMFCS (ρ = 0.56), Mini-MACS (ρ = 0.64), CFCS (ρ = 0.61), and VFCS (ρ = 0.61), supporting construct validity. Conclusions: The Korean Mini-EDACS is a reliable and valid tool for classifying eating and drinking abilities in children with CP under 3 years. It enables standardized communication between caregivers and clinicians, complements existing functional classification systems, and may facilitate earlier identification and intervention for feeding difficulties. Full article
(This article belongs to the Special Issue Children with Cerebral Palsy and Other Developmental Disabilities)
32 pages, 2305 KB  
Article
SCEditor-Web: Bridging Model-Driven Engineering and Generative AI for Smart Contract Development
by Yassine Ait Hsain, Naziha Laaz and Samir Mbarki
Information 2025, 16(10), 870; https://doi.org/10.3390/info16100870 - 7 Oct 2025
Abstract
Smart contracts are central to blockchain ecosystems, yet their development remains technically demanding, error-prone, and tied to platform-specific programming languages. This paper introduces SCEditor-Web, a web-based modeling environment that combines model-driven engineering (MDE) with generative artificial intelligence (Gen-AI) to simplify contract design and [...] Read more.
Smart contracts are central to blockchain ecosystems, yet their development remains technically demanding, error-prone, and tied to platform-specific programming languages. This paper introduces SCEditor-Web, a web-based modeling environment that combines model-driven engineering (MDE) with generative artificial intelligence (Gen-AI) to simplify contract design and code generation. Developers specify the structural and behavioral aspects of smart contracts through a domain-specific visual language grounded in a formal metamodel. The resulting contract model is exported as structured JSON and transformed into executable, platform-specific code using large language models (LLMs) guided by a tailored prompt engineering process. A prototype implementation was evaluated on Solidity contracts as a proof of concept, using representative use cases. Experiments with state-of-the-art LLMs assessed the generated contracts for compilability, semantic alignment with the contract model, and overall code quality. Results indicate that the visual-to-code workflow reduces manual effort, mitigates common programming errors, and supports developers with varying levels of expertise. The contributions include an abstract smart contract metamodel, a structured prompt generation pipeline, and a web-based platform that bridges high-level modeling with practical multi-language code synthesis. Together, these elements advance the integration of MDE and LLMs, demonstrating a step toward more accessible and reliable smart contract engineering. Full article
(This article belongs to the Special Issue Using Generative Artificial Intelligence Within Software Engineering)
Show Figures

Figure 1

20 pages, 1174 KB  
Review
LLMs for Commit Messages: A Survey and an Agent-Based Evaluation Protocol on CommitBench
by Mohamed Mehdi Trigui and Wasfi G. Al-Khatib
Computers 2025, 14(10), 427; https://doi.org/10.3390/computers14100427 - 7 Oct 2025
Abstract
Commit messages are vital for traceability, maintenance, and onboarding in modern software projects, yet their quality is frequently inconsistent. Recent large language models (LLMs) can transform code diffs into natural language summaries, offering a path to more consistent and informative commit messages. This [...] Read more.
Commit messages are vital for traceability, maintenance, and onboarding in modern software projects, yet their quality is frequently inconsistent. Recent large language models (LLMs) can transform code diffs into natural language summaries, offering a path to more consistent and informative commit messages. This paper makes two contributions: (i) it provides a systematic survey of automated commit message generation with LLMs, critically comparing prompt-only, fine-tuned, and retrieval-augmented approaches; and (ii) it specifies a transparent, agent-based evaluation blueprint centered on CommitBench. Unlike prior reviews, we include a detailed dataset audit, preprocessing impacts, evaluation metrics, and error taxonomy. The protocol defines dataset usage and splits, prompting and context settings, scoring and selection rules, and reporting guidelines (results by project, language, and commit type), along with an error taxonomy to guide qualitative analysis. Importantly, this work emphasizes methodology and design rather than presenting new empirical benchmarking results. The blueprint is intended to support reproducibility and comparability in future studies. Full article
Show Figures

Figure 1

23 pages, 2173 KB  
Article
Prototype-Enhanced Few-Shot Relation Extraction Method Based on Cluster Loss Optimization
by Shenyi Qian, Bowen Fu, Chao Liu, Songhe Jin, Tong Sun, Zhen Chen, Daiyi Li, Yifan Sun, Yibing Chen and Yuheng Li
Symmetry 2025, 17(10), 1673; https://doi.org/10.3390/sym17101673 - 7 Oct 2025
Viewed by 63
Abstract
The purpose of few-shot relation extraction (RE) is to recognize the relationship between specific entity pairs in text when there are a limited number of labeled samples. A few-shot RE method based on a prototype network, which constructs relation prototypes by relying on [...] Read more.
The purpose of few-shot relation extraction (RE) is to recognize the relationship between specific entity pairs in text when there are a limited number of labeled samples. A few-shot RE method based on a prototype network, which constructs relation prototypes by relying on the support set to assign labels to query samples, inherently leverages the symmetry between support and query processing. Although these methods have achieved remarkable results, they still face challenges such as the misjudging of noisy samples or outliers, as well as distinguishing semantic similarity relations. To address the aforementioned challenges, we propose a novel semantic enhanced prototype network, which can integrate the semantic information of relations more effectively to promote more expressive representations of instances and relation prototypes, so as to improve the performance of the few-shot RE. Firstly, we design a prompt encoder to uniformly process different prompt templates for instance and relation information, and then utilize the powerful semantic understanding and generation capabilities of large language models (LLMs) to obtain precise semantic representations of instances, their prototypes, and conceptual prototypes. Secondly, graph attention learning techniques are introduced to effectively extract specific-relation features between conceptual prototypes and isomorphic instances while maintaining structural symmetry. Meanwhile, a prototype-level contrastive learning strategy with bidirectional feature symmetry is proposed to predict query instances by integrating the interpretable features of conceptual prototypes and the intra-class shared features captured by instance prototypes. In addition, a clustering loss function was designed to guide the model to learn a discriminative metric space with improved relational symmetry, effectively improving the accuracy of the model’s relationship recognition. Finally, the experimental results on the FewRel1.0 and FewRel2.0 datasets show that the proposed approach delivers improved performance compared to existing advanced models in the task of few-shot RE. Full article
(This article belongs to the Section Computer)
Show Figures

Figure 1

42 pages, 460 KB  
Review
Ethical Problems in the Use of Artificial Intelligence by University Educators
by Roman Chinoracky and Natalia Stalmasekova
Educ. Sci. 2025, 15(10), 1322; https://doi.org/10.3390/educsci15101322 - 6 Oct 2025
Viewed by 317
Abstract
This study examines the ethical problems of using artificial intelligence (AI) applications in higher education, focusing on activities performed by university educators. Drawing on Slovak legislation that defines educators’ responsibilities, the study classifies their activities into three categories: teaching, scientific research, and other [...] Read more.
This study examines the ethical problems of using artificial intelligence (AI) applications in higher education, focusing on activities performed by university educators. Drawing on Slovak legislation that defines educators’ responsibilities, the study classifies their activities into three categories: teaching, scientific research, and other (academic management and self-directed professional development). From standpoint of methodology, a thematic review of 42 open-access, peer-reviewed articles published between 2022 and 2025 was conducted across the Web of Science and Scopus databases. Relevant AI applications and their associated ethical issues were identified and thematically categorized. Results of this study show that AI applications are extensively used across all analysed areas of university educators’ activities. Most notably used are applications that are generative language models, editing and paraphrasing tools, learning and assessment software, management and search tools, visualizing and design tools, and analysis and management systems. Their adoption raises ethical concerns which can be thematically grouped into six categories: privacy and data protection, bias and fairness, transparency and accountability, autonomy and oversight, governance gaps, and integrity and plagiarism. The results provide universities with a structured analytical framework to assess and address ethical risks related to AI use in specific academic activities. Although the study is limited to open-access literature, it offers a conceptual foundation for future empirical research and the development of ethical, institutionally grounded AI policies in higher education. Full article
35 pages, 3451 KB  
Article
Developing Speaking Skills in Third-Grade Students Through the Analysis of Visual Material in Two Languages (Lithuanian and English)
by Daiva Jakavonytė-Staškuvienė and Guostė Streikutė
Behav. Sci. 2025, 15(10), 1362; https://doi.org/10.3390/bs15101362 - 5 Oct 2025
Viewed by 508
Abstract
In language classes, speaking skills are often taken for granted, and not enough attention is paid to developing these skills in a targeted way. In our study, the speaking skills of third-grade students (N = 46) are developed in integrated Lithuanian and English [...] Read more.
In language classes, speaking skills are often taken for granted, and not enough attention is paid to developing these skills in a targeted way. In our study, the speaking skills of third-grade students (N = 46) are developed in integrated Lithuanian and English lessons through the analysis of visual material. Visual material is an aid and a means for expanding students’ vocabulary and developing their ability to express their thoughts verbally. The students are aged 9–10 years old. The aim of the study was to investigate the development of third-grade students’ speaking skills using visual material analysis in two languages. The Action Research was conducted in a school in one of Lithuania’s major cities. During the Action Research, students completed mind maps and analyzed visual material by answering questions in two languages. The questions were designed to cover different groups of thinking skills (knowledge and understanding, drawing conclusions, interpretation, and evaluation). The students spoke their prepared answers to the questions. The accuracy and correctness of the answers, English pronunciation, and the ability to speak in complete sentences were evaluated. Full article
18 pages, 46866 KB  
Article
SATrack: Semantic-Aware Alignment Framework for Visual–Language Tracking
by Yangyang Tian, Liusen Xu, Zhe Li, Liang Jiang, Cen Chen and Huanlong Zhang
Electronics 2025, 14(19), 3935; https://doi.org/10.3390/electronics14193935 - 4 Oct 2025
Viewed by 211
Abstract
Visual–language tracking often faces challenges like target deformation and confusion caused by similar objects. These issues can disrupt the alignment between visual inputs and their textual descriptions, leading to cross-modal semantic drift and feature-matching errors. To address these issues, we propose SATrack, a [...] Read more.
Visual–language tracking often faces challenges like target deformation and confusion caused by similar objects. These issues can disrupt the alignment between visual inputs and their textual descriptions, leading to cross-modal semantic drift and feature-matching errors. To address these issues, we propose SATrack, a Semantic-Aware Alignment framework for visual–language tracking. Specifically, we first propose the Semantically Aware Contrastive Alignment module, which leverages attention-guided semantic distance modeling to identify hard negative samples that are semantically similar but carry different labels. This helps the model better distinguish confusing instances and capture fine-grained cross-modal differences. Secondly, we design the Cross-Modal Token Filtering strategy, which leverages attention responses guided by both the visual template and the textual description to filter out irrelevant or weakly related tokens in the search region. This helps the model focus more precisely on the target. Finally, we propose a Confidence-Guided Template Memory mechanism, which evaluates the prediction quality of each frame using convolutional operations and confidence thresholding. High-confidence frames are stored to selectively update the template memory, enabling the model to adapt to appearance changes over time. Extensive experiments show that SATrack achieves a 65.8% success rate on the TNL2K benchmark, surpassing the previous state-of-the-art UVLTrack by 3.1% and demonstrating superior robustness and accuracy. Full article
(This article belongs to the Special Issue Deep Perception in Autonomous Driving, 2nd Edition)
Show Figures

Figure 1

22 pages, 5925 KB  
Article
The Construction of a Design Method Knowledge Graph Driven by Multi-Source Heterogeneous Data
by Jixing Shi, Kaiyi Wang, Zhongqing Wang, Zhonghang Bai and Fei Hu
Appl. Sci. 2025, 15(19), 10702; https://doi.org/10.3390/app151910702 - 3 Oct 2025
Viewed by 206
Abstract
To address the fragmentation and weak correlation of knowledge in the design method domain, this paper proposes a framework for constructing a knowledge graph driven by multi-source heterogeneous data. The process involves collecting multi-source heterogeneous data and subsequently utilizing text mining and natural [...] Read more.
To address the fragmentation and weak correlation of knowledge in the design method domain, this paper proposes a framework for constructing a knowledge graph driven by multi-source heterogeneous data. The process involves collecting multi-source heterogeneous data and subsequently utilizing text mining and natural language processing techniques to extract design themes and method elements. A “theme–stage–attribute” three-dimensional mapping model is established to achieve semantic coupling of knowledge. The BERT-BiLSTM-CRF (Bidirectional Encoder Representations from Transformers-Bidirectional Long Short-Term Memory-Conditional Random Field) model is employed for entity recognition and relation extraction, while the Sentence-BERT (Sentence Bidirectional Encoder Representations from Transformers) model is used to perform multi-source knowledge fusion. The Neo4j graph database facilitates knowledge storage, visualization, and querying, forming the basis for developing a prototype of a design method recommendation system. The framework’s effectiveness was validated through experiments on extraction performance and knowledge graph quality. The results demonstrate that the framework achieves an F1 score of 91.2% for knowledge extraction, and an 8.44% improvement over the baseline. The resulting graph’s node and relation coverage reached 94.1% and 91.2%, respectively. In complex semantic query tasks, the framework shows a significant advantage over traditional classification systems, achieving a maximum F1 score of 0.97. It can effectively integrate dispersed knowledge in the field of design methods and support method matching throughout the entire design process. This research is of significant value for advancing knowledge management and application in innovative product design. Full article
Show Figures

Figure 1

Back to TopTop