Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (469)

Search Parameters:
Keywords = voice interaction

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
20 pages, 2732 KB  
Article
Redesigning Multimodal Interaction: Adaptive Signal Processing and Cross-Modal Interaction for Hands-Free Computer Interaction
by Bui Hong Quan, Nguyen Dinh Tuan Anh, Hoang Van Phi and Bui Trung Thanh
Sensors 2025, 25(17), 5411; https://doi.org/10.3390/s25175411 - 2 Sep 2025
Viewed by 255
Abstract
Hands-free computer interaction is a key topic in assistive technology, with camera-based and voice-based systems being the most common methods. Recent camera-based solutions leverage facial expressions or head movements to simulate mouse clicks or key presses, while voice-based systems enable control via speech [...] Read more.
Hands-free computer interaction is a key topic in assistive technology, with camera-based and voice-based systems being the most common methods. Recent camera-based solutions leverage facial expressions or head movements to simulate mouse clicks or key presses, while voice-based systems enable control via speech commands, wake-word detection, and vocal gestures. However, existing systems often suffer from limitations in responsiveness and accuracy, especially under real-world conditions. In this paper, we present 3-Modal Human-Computer Interaction (3M-HCI), a novel interaction system that dynamically integrates facial, vocal, and eye-based inputs through a new signal processing pipeline and a cross-modal coordination mechanism. This approach not only enhances recognition accuracy but also reduces interaction latency. Experimental results demonstrate that 3M-HCI outperforms several recent hands-free interaction solutions in both speed and precision, highlighting its potential as a robust assistive interface. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

26 pages, 5349 KB  
Article
Smart Forest Modeling Behavioral for a Greener Future: An AI Text-by-Voice Blockchain Approach with Citizen Involvement in Sustainable Forestry Functionality
by Dimitrios Varveris, Vasiliki Basdekidou, Chrysanthi Basdekidou and Panteleimon Xofis
FinTech 2025, 4(3), 47; https://doi.org/10.3390/fintech4030047 - 1 Sep 2025
Viewed by 188
Abstract
This paper introduces a novel approach to tree modeling architecture integrated with blockchain technology, aimed at enhancing landscape spatial planning and forest monitoring systems. The primary objective is to develop a low-cost, automated tree CAD modeling methodology combined with blockchain functionalities to support [...] Read more.
This paper introduces a novel approach to tree modeling architecture integrated with blockchain technology, aimed at enhancing landscape spatial planning and forest monitoring systems. The primary objective is to develop a low-cost, automated tree CAD modeling methodology combined with blockchain functionalities to support smart forest projects and collaborative design processes. The proposed method utilizes a parametric tree CAD model consisting of four 2D tree-frames with a 45° division angle, enriched with recorded tree-leaves’ texture and color. An “AI Text-by-Voice CAD Programming” technique is employed to create tangible tree-model NFT tokens, forming the basis of a thematic “Internet-of-Trees” blockchain. The main results demonstrate the effectiveness of the blockchain/Merkle hash tree in tracking tree geometry growth and texture changes through parametric transactions, enabling decentralized design, data validation, and planning intelligence. Comparative analysis highlights the advantages in cost, time efficiency, and flexibility over traditional 3D modeling techniques, while providing acceptable accuracy for metaverse projects in smart forests and landscape architecture. Core contributions include the integration of AI-based user voice interaction with blockchain and behavioral data for distributed and collaborative tree modeling, the introduction of a scalable and secure “Merkle hash tree” for smart forest monitoring, and the facilitation of fintech adoption in environmental projects. This framework offers significant potential for advancing metaverse-based landscape architecture, smart forest surveillance, sustainable urban planning, and the improvement of citizen involvement in sustainable forestry paving the way for a greener future. Full article
Show Figures

Figure 1

26 pages, 555 KB  
Concept Paper
Do We Need a Voice Methodology? Proposing a Voice-Centered Methodology: A Conceptual Framework in the Age of Surveillance Capitalism
by Laura Caroleo
Societies 2025, 15(9), 241; https://doi.org/10.3390/soc15090241 - 30 Aug 2025
Viewed by 229
Abstract
This paper explores the rise in voice-based social media as a pivotal transformation in digital communication, situated within the broader era of chatbots and voice AI. Platforms such as Clubhouse, X Spaces, Discord and similar ones foreground vocal interaction, reshaping norms of participation, [...] Read more.
This paper explores the rise in voice-based social media as a pivotal transformation in digital communication, situated within the broader era of chatbots and voice AI. Platforms such as Clubhouse, X Spaces, Discord and similar ones foreground vocal interaction, reshaping norms of participation, identity construction, and platform governance. This shift from text-centered communication to hybrid digital orality presents new sociological and methodological challenges, calling for the development of voice-centered analytical approaches. In response, the paper introduces a multidimensional methodological framework for analyzing voice-based social media platforms in the context of surveillance capitalism and AI-driven conversational technologies. We propose a high-level reference architecture machine learning for social science pipeline that integrates digital methods techniques, automatic speech recognition (ASR) models, and natural language processing (NLP) models within a reflexive and ethically grounded framework. To illustrate its potential, we outline possible stages of a PoC (proof of concept) audio analysis machine learning pipeline, demonstrated through a conceptual use case involving the collection, ingestion, and analysis of X Spaces. While not a comprehensive empirical study, this pipeline proposal highlights technical and ethical challenges in voice analysis. By situating the voice as a central axis of online sociality and examining it in relation to AI-driven conversational technologies, within an era of post-orality, the study contributes to ongoing debates on surveillance capitalism, platform affordances, and the evolving dynamics of digital interaction. In this rapidly evolving landscape, we urgently need a robust vocal methodology to ensure that voice is not just processed but understood. Full article
Show Figures

Figure 1

24 pages, 4245 KB  
Article
Healthy Movement Leads to Emotional Connection: Development of the Movement Poomasi “Wello!” Application Based on Digital Psychosocial Touch—A Mixed-Methods Study
by Suyoung Hwang, Hyunmoon Kim and Eun-Surk Yi
Healthcare 2025, 13(17), 2157; https://doi.org/10.3390/healthcare13172157 - 29 Aug 2025
Viewed by 286
Abstract
Background/Objective: The global acceleration of population aging presents profound challenges to the physical, psychological, and social well-being of older adults. As traditional exercise programs face limitations in accessibility, personalization, and sustained social support, there is a critical need for innovative, inclusive, and community-integrated [...] Read more.
Background/Objective: The global acceleration of population aging presents profound challenges to the physical, psychological, and social well-being of older adults. As traditional exercise programs face limitations in accessibility, personalization, and sustained social support, there is a critical need for innovative, inclusive, and community-integrated digital movement solutions. This study aimed to develop and evaluate Movement Poomasi, a hybrid digital healthcare application designed to promote physical activity, improve digital accessibility, and strengthen social connectedness among older adults. Methods: From March 2023 to November 2023, Movement Poomasi was developed through an iterative user-centered design process involving domain experts in physical therapy and sports psychology. In this study, the term UI/UX—short for user interface and user experience—refers to the overall design and interaction framework of the application, encompassing visual layout, navigation flow, accessibility features, and user engagement optimization tailored to older adults’ sensory, cognitive, and motor characteristics. The application integrates adaptive exercise modules, senior-optimized UI/UX, voice-assisted navigation, and peer-interaction features to enable both home-based and in-person movement engagement. A two-phase usability validation was conducted. A 4-week pilot test with 15 older adults assessed the prototype, followed by a formal 6-week study with 50 participants (≥65 years), stratified by digital literacy and activity background. Quantitative metrics—movement completion rates, session duration, and engagement with social features—were analyzed alongside semi-structured interviews. Statistical analysis included ANOVA and regression to examine usability and engagement outcomes. The application has continued iterative testing and refinement until May 2025, and it is scheduled for re-launch under the name Wello! in August 2025. Results: Post-implementation UI refinements significantly increased navigation success rates (from 68% to 87%, p = 0.042). ANOVA revealed that movement selection and peer-interaction tasks posed greater cognitive load (p < 0.01). A strong positive correlation was found between digital literacy and task performance (r = 0.68, p < 0.05). Weekly participation increased by 38%, with 81% of participants reporting enhanced social connectedness through group challenges and hybrid peer-led meetups. Despite high satisfaction scores (mean 4.6 ± 0.4), usability challenges remained among low-literacy users, indicating the need for further interface simplification. Conclusions: The findings underscore the potential of hybrid digital platforms tailored to older adults’ physical, cognitive, and social needs. Movement Poomasi demonstrates scalable feasibility and contributes to reducing the digital divide while fostering active aging. Future directions include AI-assisted onboarding, adaptive tutorials, and expanded integration with community care ecosystems to enhance long-term engagement and inclusivity. Full article
(This article belongs to the Special Issue Emerging Technologies for Person-Centred Healthcare)
Show Figures

Figure 1

14 pages, 685 KB  
Proceeding Paper
Predictive Analysis of Voice Pathology Using Logistic Regression: Insights and Challenges
by Divya Mathews Olakkengil and Sagaya Aurelia P
Eng. Proc. 2025, 107(1), 28; https://doi.org/10.3390/engproc2025107028 - 27 Aug 2025
Viewed by 414
Abstract
Voice pathology diagnosis is essential for the timely detection and management of voice disorders, which can significantly impact an individual’s quality of life. This study employed logistic regression to evaluate the predictive power of variables that include age, severity, loudness, breathiness, pitch, roughness, [...] Read more.
Voice pathology diagnosis is essential for the timely detection and management of voice disorders, which can significantly impact an individual’s quality of life. This study employed logistic regression to evaluate the predictive power of variables that include age, severity, loudness, breathiness, pitch, roughness, strain, and gender on a binary diagnosis outcome (Yes/No). The analysis was performed on the Perceptual Voice Qualities Database (PVQD), a comprehensive dataset containing voice samples with perceptual ratings. Two widely used voice quality assessment tools, CAPE-V (Consensus Auditory-Perceptual Evaluation of Voice) and GRBAS (Grade, Roughness, Breathiness, Asthenia, Strain), were employed to annotate voice qualities, ensuring systematic and clinically relevant perceptual evaluations. The model revealed that age (odds ratio: 1.033, p < 0.001), loudness (odds ratio: 1.071, p = 0.005), and gender (male) (odds ratio: 1.904, p = 0.043) were statistically significant predictors of voice pathology. In contrast, severity and voice quality-related features like breathiness, pitch, roughness, and strain did not show statistical significance, suggesting their limited predictive contributions within this model. While the results provide valuable insights, the study underscores notable limitations of logistic regression. The model assumes a linear relationship between the independent variables and the log odds of the outcome, which restricts its ability to capture complex, non-linear patterns within the data. Additionally, logistic regression does not inherently account for interactions between predictors or feature dependencies, potentially limiting its performance in more intricate datasets. Furthermore, a fixed classification threshold (0.5) may lead to misclassification, particularly in datasets with imbalanced classes or skewed predictor distributions. These findings highlight that although logistic regression serves as a useful tool for identifying significant predictors, its results are dataset-dependent and cannot be generalized across diverse populations. Future research should validate these findings using heterogeneous datasets and employ advanced machine learning techniques to address the limitations of logistic regression. Integrating non-linear models or feature interaction analyses may enhance diagnostic accuracy, ensuring more reliable and robust voice pathology predictions. Full article
Show Figures

Figure 1

25 pages, 19135 KB  
Article
Development of a Multi-Platform AI-Based Software Interface for the Accompaniment of Children
by Isaac León, Camila Reyes, Iesus Davila, Bryan Puruncajas, Dennys Paillacho, Nayeth Solorzano, Marcelo Fajardo-Pruna, Hyungpil Moon and Francisco Yumbla
Multimodal Technol. Interact. 2025, 9(9), 88; https://doi.org/10.3390/mti9090088 - 26 Aug 2025
Viewed by 544
Abstract
The absence of parental presence has a direct impact on the emotional stability and social routines of children, especially during extended periods of separation from their family environment, as in the case of daycare centers, hospitals, or when they remain alone at home. [...] Read more.
The absence of parental presence has a direct impact on the emotional stability and social routines of children, especially during extended periods of separation from their family environment, as in the case of daycare centers, hospitals, or when they remain alone at home. At the same time, the technology currently available to provide emotional support in these contexts remains limited. In response to the growing need for emotional support and companionship in child care, this project proposes the development of a multi-platform software architecture based on artificial intelligence (AI), designed to be integrated into humanoid robots that assist children between the ages of 6 and 14. The system enables daily verbal and non-verbal interactions intended to foster a sense of presence and personalized connection through conversations, games, and empathetic gestures. Built on the Robot Operating System (ROS), the software incorporates modular components for voice command processing, real-time facial expression generation, and joint movement control. These modules allow the robot to hold natural conversations, display dynamic facial expressions on its LCD (Liquid Crystal Display) screen, and synchronize gestures with spoken responses. Additionally, a graphical interface enhances the coherence between dialogue and movement, thereby improving the quality of human–robot interaction. Initial evaluations conducted in controlled environments assessed the system’s fluency, responsiveness, and expressive behavior. Subsequently, it was implemented in a pediatric hospital in Guayaquil, Ecuador, where it accompanied children during their recovery. It was observed that this type of artificial intelligence-based software, can significantly enhance the experience of children, opening promising opportunities for its application in clinical, educational, recreational, and other child-centered settings. Full article
Show Figures

Graphical abstract

12 pages, 249 KB  
Article
Hoarseness, Quality of Life, and Social Anxiety: A Case–Control Study
by Süleyman Dönmezdil and Serdar Ferit Toprak
Behav. Sci. 2025, 15(9), 1160; https://doi.org/10.3390/bs15091160 - 26 Aug 2025
Viewed by 335
Abstract
Hoarseness is a common voice symptom that can impair communication and lead to psychosocial difficulties. It has been hypothesized that chronic hoarseness may contribute to elevated social anxiety. This study aimed to assess the impact of hoarseness on quality of life and social [...] Read more.
Hoarseness is a common voice symptom that can impair communication and lead to psychosocial difficulties. It has been hypothesized that chronic hoarseness may contribute to elevated social anxiety. This study aimed to assess the impact of hoarseness on quality of life and social anxiety in affected individuals. Thirty-eight patients with chronic hoarseness (voice disorders) and 40 matched healthy controls were evaluated in a prospective case-control study. Quality of life was measured using the WHOQOL-BREF questionnaire (Physical, Psychological, Social, and Environmental domains). Social anxiety was assessed with the Liebowitz Social Anxiety Scale (LSAS), and general anxiety and depression with the Hospital Anxiety and Depression Scale (HADS). Group scores were compared using appropriate statistical tests, and effect sizes with 95% confidence intervals were calculated. Patients with hoarseness had significantly lower Psychological Health and Social Relationships scores on the WHOQOL-BREF than controls (p < 0.01 for both; large effect sizes), indicating worse quality of life in these domains. Physical Health and Environmental domain scores did not differ between groups. The hoarseness group also showed higher social anxiety: LSAS total scores and Social Interaction subscale scores were significantly greater than those of controls (p < 0.01 and p < 0.05, respectively; moderate-to-large effects), whereas the Performance Anxiety subscale was similar between groups. By contrast, HADS anxiety and depression scores did not differ significantly between patients and controls. Notably, mean HADS scores in both groups fell in the mild (borderline) range rather than the normal range. Chronic hoarseness is associated with reduced quality of life in emotional and social domains and with increased social anxiety symptoms, but not with elevated general anxiety or depression. These findings underscore the need to address psychosocial factors, particularly social anxiety, in the clinical management of patients with voice disorders. Full article
(This article belongs to the Special Issue Providing Emotional Support for People with Chronic Diseases)
20 pages, 2568 KB  
Article
Towards Spatial Awareness: Real-Time Sensory Augmentation with Smart Glasses for Visually Impaired Individuals
by Nadia Aloui
Electronics 2025, 14(17), 3365; https://doi.org/10.3390/electronics14173365 - 25 Aug 2025
Viewed by 475
Abstract
This research presents an innovative Internet of Things (IoT) and artificial intelligence (AI) platform designed to provide holistic assistance and foster autonomy for visually impaired individuals within the university environment. Its main novelty is real-time sensory augmentation and spatial awareness, integrating ultrasonic, LiDAR, [...] Read more.
This research presents an innovative Internet of Things (IoT) and artificial intelligence (AI) platform designed to provide holistic assistance and foster autonomy for visually impaired individuals within the university environment. Its main novelty is real-time sensory augmentation and spatial awareness, integrating ultrasonic, LiDAR, and RFID sensors for robust 360° obstacle detection, environmental perception, and precise indoor localization. A novel, optimized Dijkstra algorithm calculates optimal routes; speech and intent recognition enable intuitive voice control. The wearable smart glasses are complemented by a platform providing essential educational functionalities, including lesson reminders, timetables, and emergency assistance. Based on gamified principles of exploration and challenge, the platform includes immersive technology settings, intelligent image recognition, auditory conversion, haptic feedback, and rapid contextual awareness, delivering a sophisticated, effective navigational experience. Exhaustive technical evaluation reveals that a more autonomous and fulfilling university experience is made possible by notable improvements in navigation performance, object detection accuracy, and technical capabilities for social interaction features, according to a thorough technical audit. Full article
(This article belongs to the Section Computer Science & Engineering)
Show Figures

Figure 1

18 pages, 3987 KB  
Article
Interactive Application with Virtual Reality and Artificial Intelligence for Improving Pronunciation in English Learning
by Gustavo Caiza, Carlos Villafuerte and Adriana Guanuche
Appl. Sci. 2025, 15(17), 9270; https://doi.org/10.3390/app15179270 - 23 Aug 2025
Viewed by 601
Abstract
Technological advances have enabled the development of innovative educational tools, particularly those aimed at supporting English as a Second Language (ESL) learning, with a specific focus on oral skills. However, pronunciation remains a significant challenge due to the limited availability of personalized learning [...] Read more.
Technological advances have enabled the development of innovative educational tools, particularly those aimed at supporting English as a Second Language (ESL) learning, with a specific focus on oral skills. However, pronunciation remains a significant challenge due to the limited availability of personalized learning opportunities that offer immediate feedback and contextualized practice. In this context, the present research proposes the design, implementation, and validation of an immersive application that leverages virtual reality (VR) and artificial intelligence (AI) to enhance English pronunciation. The proposed system integrates a 3D interactive environment developed in Unity, voice classification models trained using Teachable Machine, and real-time communication with Firebase, allowing users to practice and assess their pronunciation in a simulated library-like virtual setting. Through its integrated AI module, the application can analyze the pronunciation of each word in real time, detecting correct and incorrect utterances, and then providing immediate feedback to help users identify and correct their mistakes. The virtual environment was designed to be a welcoming and user-friendly, promoting active engagement with the learning process. The application’s distributed architecture enables automated feedback generation via data flow between the cloud-based AI, the database, and the visualization interface. Results demonstrate that using 400 samples per class and a confidence threshold of 99.99% for training the AI model effectively eliminated false positives, significantly increasing system accuracy and providing users with more reliable feedback. This directly contributes to enhanced learner autonomy and improved ESL acquisition outcomes. Furthermore, user surveys conducted to understand their perceptions of the application’s usefulness as a support tool for English learning yielded an average acceptance rate of 93%. This reflects the acceptance of these immersive technologies in educational contexts, as the combination of these technologies offers a realistic and user-friendly simulation environment, in addition to detailed word analysis, facilitating self-assessment and independent learning among students. Full article
Show Figures

Figure 1

16 pages, 235 KB  
Article
Many Pros and a Little Cons: Experiences of First-Time Guide Dog Recipients
by Chalotte Glintborg, Johan Trettvik, Rasmus Holm and Tia G. B. Hansen
Animals 2025, 15(16), 2461; https://doi.org/10.3390/ani15162461 - 21 Aug 2025
Viewed by 392
Abstract
Guide dogs are primarily seen as mobility aids for people with visual impairments, but they may also influence users’ psychological well-being and social lives. Benefits of being partnered with a guide dog have been found in several countries but Scandinavian studies are sparse. [...] Read more.
Guide dogs are primarily seen as mobility aids for people with visual impairments, but they may also influence users’ psychological well-being and social lives. Benefits of being partnered with a guide dog have been found in several countries but Scandinavian studies are sparse. This qualitative study explores the lived experiences of six Danish guide dog partners through semi-structured interviews. A reflexive thematic analysis found six primary benefits: improved mobility, increased physical activity, enhanced freedom and independence, new social interactions, a strong emotional bond with the dog, and better mental well-being. Users also reported challenges, including frequent distractions caused by the public interacting with the dog, difficulties during the initial adjustment period, increased cleaning, and occasional trouble finding care for the dog. Participants unanimously emphasized the irreplaceable value of their guide dogs and positive socioemotional dimensions of guide dog partnerships. The findings add Scandinavian voices that support existing international research and note three areas of potential improvement: better public education about guide dog etiquette, enhanced support during the adjustment period, and improved care-sharing arrangements. The study underscores the importance of considering guide dogs as both functional aids and sources of emotional and social support. Full article
(This article belongs to the Special Issue Assistance Dogs—What Should We Focus On)
18 pages, 3632 KB  
Article
Multilingual Mobility: Audio-Based Language ID for Automotive Systems
by Joowon Oh and Jeaho Lee
Appl. Sci. 2025, 15(16), 9209; https://doi.org/10.3390/app15169209 - 21 Aug 2025
Viewed by 420
Abstract
With the growing demand for natural and intelligent human–machine interaction in multilingual environments, automatic language identification (LID) has emerged as a crucial component in voice-enabled systems, particularly in the automotive domain. This study proposes an audio-based LID model that identifies the spoken language [...] Read more.
With the growing demand for natural and intelligent human–machine interaction in multilingual environments, automatic language identification (LID) has emerged as a crucial component in voice-enabled systems, particularly in the automotive domain. This study proposes an audio-based LID model that identifies the spoken language directly from voice input without requiring manual language selection. The model architecture leverages two types of feature extraction pipelines: a Variational Autoencoder (VAE) and a pre-trained Wav2Vec model, both used to obtain latent speech representations. These embeddings are then fed into a multi-layer perceptron (MLP)-based classifier to determine the speaker’s language among five target languages: Korean, Japanese, Chinese, Spanish, and French. The model is trained and evaluated using a dataset preprocessed into Mel-Frequency Cepstral Coefficients (MFCCs) and raw waveform inputs. Experimental results demonstrate the effectiveness of the proposed approach in achieving accurate and real-time language detection, with potential applications in in-vehicle systems, speech translation platforms, and multilingual voice assistants. By eliminating the need for predefined language settings, this work contributes to more seamless and user-friendly multilingual voice interaction systems. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

19 pages, 1149 KB  
Article
A Gamified Digital Mental Health Intervention Across Six Sub-Saharan African Countries: A Cross-Sectional Evaluation of a Large-Scale Implementation
by Christopher K. Barkley, Charmaine N. Nyakonda, Kondwani Kuthyola, Polite Ndlovu, Devyn Lee, Andrew Dallos, Danny Kofi-Armah, Priscilla Obeng and Katherine G. Merrill
Int. J. Environ. Res. Public Health 2025, 22(8), 1281; https://doi.org/10.3390/ijerph22081281 - 15 Aug 2025
Viewed by 819
Abstract
Mental health conditions affect many young people in sub-Saharan Africa (SSA), where stigma is high and access to care is limited. Digital tools accessible on basic mobile phones offer a scalable way to promote mental health, but evidence on their effectiveness in SSA [...] Read more.
Mental health conditions affect many young people in sub-Saharan Africa (SSA), where stigma is high and access to care is limited. Digital tools accessible on basic mobile phones offer a scalable way to promote mental health, but evidence on their effectiveness in SSA is limited. This study evaluated the reach, feasibility, acceptability, and knowledge outcomes of Digital MindSKILLZ, an interactive voice response (IVR) mental health intervention implemented in the Democratic Republic of Congo, Ghana, Nigeria, Rwanda, Uganda, and Zambia. Over seven months, 700,138 people called the platform, and 425,395 (61%) listened to at least one message. Of these users, 63.6% were under 25 and 68.3% were from rural areas. The three content branches—mental health information, mental health skills, and soccer quizzes—were accessed by 36.5%, 46.4%, and 50.9% of users, respectively. Among users who accessed the mental health branch of the intervention, the mean number of messages completed was 7.6 out of 18 messages. In a follow-up survey, 91% of users understood the content, 85% would recommend the intervention, and 38% found the mental health content most helpful. Average knowledge scores were 62%, with lower scores on common disorders and stigma. The intervention showed strong reach and acceptability, but content and implementation improvements are needed to boost engagement and retention. Full article
(This article belongs to the Special Issue Advancing Youth Mental Health: Innovations, Integration, and Equity)
Show Figures

Figure 1

20 pages, 3244 KB  
Article
SOUTY: A Voice Identity-Preserving Mobile Application for Arabic-Speaking Amyotrophic Lateral Sclerosis Patients Using Eye-Tracking and Speech Synthesis
by Hessah A. Alsalamah, Leena Alhabrdi, May Alsebayel, Aljawhara Almisned, Deema Alhadlaq, Loody S. Albadrani, Seetah M. Alsalamah and Shada AlSalamah
Electronics 2025, 14(16), 3235; https://doi.org/10.3390/electronics14163235 - 14 Aug 2025
Viewed by 420
Abstract
Amyotrophic Lateral Sclerosis (ALS) is a neurodegenerative disorder that progressively impairs motor and communication abilities. Globally, the prevalence of ALS was estimated at approximately 222,800 cases in 2015 and is projected to increase by nearly 70% to 376,700 cases by 2040, primarily driven [...] Read more.
Amyotrophic Lateral Sclerosis (ALS) is a neurodegenerative disorder that progressively impairs motor and communication abilities. Globally, the prevalence of ALS was estimated at approximately 222,800 cases in 2015 and is projected to increase by nearly 70% to 376,700 cases by 2040, primarily driven by demographic shifts in aging populations, and the lifetime risk of developing ALS is 1 in 350–420. Despite international advancements in assistive technologies, a recent national survey in Saudi Arabia revealed that 100% of ALS care providers lack access to eye-tracking communication tools, and 92% reported communication aids as inconsistently available. While assistive technologies such as speech-generating devices and gaze-based control systems have made strides in recent decades, they primarily support English speakers, leaving Arabic-speaking ALS patients underserved. This paper presents SOUTY, a cost-effective, mobile-based application that empowers ALS patients to communicate using gaze-controlled interfaces combined with a text-to-speech (TTS) feature in Arabic language, which is one of the five most widely spoken languages in the world. SOUTY (i.e., “my voice”) utilizes a personalized, pre-recorded voice bank of the ALS patient and integrated eye-tracking technology to support the formation and vocalization of custom phrases in Arabic. This study describes the full development life cycle of SOUTY from conceptualization and requirements gathering to system architecture, implementation, evaluation, and refinement. Validation included expert interviews with Human–Computer Interaction (HCI) expertise and speech pathology specialty, as well as a public survey assessing awareness and technological readiness. The results support SOUTY as a culturally and linguistically relevant innovation that enhances autonomy and quality of life for Arabic-speaking ALS patients. This approach may serve as a replicable model for developing inclusive Augmentative and Alternative Communication (AAC) tools in other underrepresented languages. The system achieved 100% task completion during internal walkthroughs, with mean phrase selection times under 5 s and audio playback latency below 0.3 s. Full article
Show Figures

Figure 1

24 pages, 2716 KB  
Article
Interactive Indoor Audio-Map as a Digital Equivalent of the Tactile Map
by Dariusz Gotlib, Krzysztof Lipka and Hubert Świech
Appl. Sci. 2025, 15(16), 8975; https://doi.org/10.3390/app15168975 - 14 Aug 2025
Viewed by 325
Abstract
There are still relatively few applications that serve the function of a traditional tactile map, allowing visually impaired individuals to explore a digital map by sliding their fingers across it. Moreover, existing technological solutions either lack a spatial learning mode or provide only [...] Read more.
There are still relatively few applications that serve the function of a traditional tactile map, allowing visually impaired individuals to explore a digital map by sliding their fingers across it. Moreover, existing technological solutions either lack a spatial learning mode or provide only limited functionality, focusing primarily on navigating to a selected destination. To address these gaps, the authors have proposed an original concept for an indoor mobile application that enables map exploration by sliding a finger across the smartphone screen, using audio spatial descriptions as the primary medium for conveying information. The spatial descriptions are hierarchical and contextual, focusing on anchoring them in space and indicating their extent of influence. The basis for data management and analysis is GIS technology. The application is designed to support spatial orientation during user interaction with the digital map. The research emphasis was on creating an effective cartographic communication message, utilizing voice-based delivery of spatial information stored in a virtual building model (within a database) and tags placed in real-world buildings. Techniques such as Text-to-Speech, TalkBack, QRCode technologies were employed to achieve this. Preliminary tests conducted with both blind and sighted people demonstrated the usefulness of the proposed concept. The proposed solution supporting people with disabilities can also be useful and attractive to all users of navigation applications and may affect the development of such applications. Full article
(This article belongs to the Section Earth Sciences)
Show Figures

Figure 1

11 pages, 2092 KB  
Article
Multiplayer Virtual Labs for Electronic Circuit Design: A Digital Twin-Based Learning Approach
by Konstantinos Sakkas, Niki Eleni Ntagka, Michail Spyridakis, Andreas Miltiadous, Euripidis Glavas, Alexandros T. Tzallas and Nikolaos Giannakeas
Electronics 2025, 14(16), 3163; https://doi.org/10.3390/electronics14163163 - 8 Aug 2025
Viewed by 361
Abstract
The rapid development of digital technologies is opening up new avenues for transforming education, particularly in fields that require practical training, such as electronic circuit design. In this context, this paper presents the development of a multiplayer virtual learning platform that makes use [...] Read more.
The rapid development of digital technologies is opening up new avenues for transforming education, particularly in fields that require practical training, such as electronic circuit design. In this context, this paper presents the development of a multiplayer virtual learning platform that makes use of digital twins technology to offer a realistic, collaborative experience in a simulated environment. Users can interact in real time through synchronized avatars, voice communication, and multiple viewing angles, simulating a physical classroom. Evaluation of the platform with undergraduate students showed positive results in terms of usability, collaboration, and learning effectiveness. Despite the limitations of the sample, the findings reinforce the prospect of virtual laboratories as a modern tool in technical education. Full article
Show Figures

Figure 1

Back to TopTop