Next Article in Journal
Projection-Angle-Sensor-Assisted X-ray Computed Tomography for Cylindrical Lithium-Ion Batteries
Next Article in Special Issue
A Review on 3D Scanners Studies for Producing Customized Orthoses
Previous Article in Journal
Optimizing Epoch Length and Activity Count Threshold Parameters in Accelerometry: Enhancing Upper Extremity Use Quantification in Cerebral Palsy
Previous Article in Special Issue
Visualization of Subcutaneous Blood Vessels Based on Hyperspectral Imaging and Three-Wavelength Index Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multilingual Framework for Risk Assessment and Symptom Tracking (MRAST)

1
Faculty of Electrical Engineering and Computer Science, University of Maribor, 2000 Maribor, Slovenia
2
Science Department, Symptoma GmbH, 1030 Vienna, Austria
3
Department of Internal Medicine, Paracelsus Medical University, 5020 Salzburg, Austria
4
Department of Psychology, Faculty of Arts, University of Maribor, 2000 Maribor, Slovenia
5
Institute of Clinical and Preventive Medicine, University of Latvia, LV-1586 Riga, Latvia
6
Department of Information System Management, Centre Hospitalier Universitaire de Liège, 4000 Liège, Belgium
7
Department of Oncology, University Medical Centre Maribor, 2000 Maribor, Slovenia
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(4), 1101; https://doi.org/10.3390/s24041101
Submission received: 19 December 2023 / Revised: 2 February 2024 / Accepted: 6 February 2024 / Published: 8 February 2024
(This article belongs to the Special Issue Biomedical Data and Imaging: Sensing, Understanding and Applications)

Abstract

:
The importance and value of real-world data in healthcare cannot be overstated because it offers a valuable source of insights into patient experiences. Traditional patient-reported experience and outcomes measures (PREMs/PROMs) often fall short in addressing the complexities of these experiences due to subjectivity and their inability to precisely target the questions asked. In contrast, diary recordings offer a promising solution. They can provide a comprehensive picture of psychological well-being, encompassing both psychological and physiological symptoms. This study explores how using advanced digital technologies, i.e., automatic speech recognition and natural language processing, can efficiently capture patient insights in oncology settings. We introduce the MRAST framework, a simplified way to collect, structure, and understand patient data using questionnaires and diary recordings. The framework was validated in a prospective study with 81 colorectal and 85 breast cancer survivors, of whom 37 were male and 129 were female. Overall, the patients evaluated the solution as well made; they found it easy to use and integrate into their daily routine. The majority (75.3%) of the cancer survivors participating in the study were willing to engage in health monitoring activities using digital wearable devices daily for an extended period. Throughout the study, there was a noticeable increase in the number of participants who perceived the system as having excellent usability. Despite some negative feedback, 44.44% of patients still rated the app’s usability as above satisfactory (i.e., 7.9 on 1–10 scale) and the experience with diary recording as above satisfactory (i.e., 7.0 on 1–10 scale). Overall, these findings also underscore the significance of user testing and continuous improvement in enhancing the usability and user acceptance of solutions like the MRAST framework. Overall, the automated extraction of information from diaries represents a pivotal step toward a more patient-centered approach, where healthcare decisions are based on real-world experiences and tailored to individual needs. The potential usefulness of such data is enormous, as it enables better measurement of everyday experiences and opens new avenues for patient-centered care.

1. Introduction

With the increasingly aging population, episodic (symptom-triggered) healthcare is not meeting the needs of patients, especially those with chronic conditions [1]. In fact, in many cases it leads to poor health outcomes [2]. Namely, episodic care, even in those cases where early action has a direct impact on morbidity/survival, largely relies on the identification of relevant symptoms, and thus, on an individual to initiate the process. In fact, a radical shift toward person-centered care relies on assessing and responding to the self-reported needs of patients [3]. Patient-reported outcomes (PROMs) are increasingly used even in clinical practice [4,5]. Namely, PROMs can be used to identify problems and trace changes over time, especially in patients with (multiple) chronic conditions [6], by “monitoring” their health status or experiences using a set of standardized quantitative data collection instruments. PROMs consist of questions describing a wide variety of symptoms, side effects, functional changes, and quality of life [7,8]. Generic PROMs are designed to address a wide variety of patients and can be used to generalize/compare across multiple conditions [9]. Condition-specific PROMs have a greater validity, specificity, and responsiveness to changes in the patient’s specific condition [6].
It is obvious that the main challenge in the use of PROMs, from a systematic perspective, is to select the “right” PROMs and at the right time [9,10]. The selection must carefully reflect the intended purposes, including why the patient’s responses are collected (i.e., purpose), what is the goal of the investigation and what information it needs to collect, and how this information will be used to benefit the individual [11,12,13]. Due to the complexity and individuality of a “perfect” list of questions, patients do not all benefit equally from their use; some engage minimally, and some do not engage with PROMs at all [14]. In many cases, the PROMs do not address, or only partially address, the most apparent issues experienced by the patients or concepts that matter to underserved groups [15]. This creates a discrepancy between the perceived personal benefits and the effort required for the intervention [14]. The lack of proper translation or help with navigation leads to inaccurate captures of data and makes completion of the measures burdensome and challenging [16,17]. Furthermore, repeatedly answering the same questions about symptoms and general health can contribute to an increased focus on the disease, causing anxiety and emotional distress [18]. Finally, even when PROMs are fully completed and the completion rates are high, data validity depends on comprehension and patients’ ability to select responses that accurately reflect their experiences [19]. Furthermore, their validity is in general limited within a relatively homogenous diagnostic group. The tools become less reliable in more general populations and underrepresented subgroups (e.g., adoption groups, immigrant groups, disability groups) [20,21]. Overall, PROMs are a gold standard in the design of symptom diaries (i.e., structured, closed-ended questionnaires to collect symptoms) and provide a reliable tool for screening and identifying symptoms of diseases with a certain degree of confidence [22,23,24]. There is, however, a concern that symptom diaries may be subject to multiple biases, e.g., poor recall and timing bias, fatigue bias, collection mode-related bias, relevance bias, etc., and thus may be insensitive [25,26,27,28,29]. As a result, they are often the subject of careless reporting and may trigger a significant number of false positives and negatives [30,31,32]. Finally, symptom-based screening can induce additional stress and anxiety, especially in populations with high risk [33]. Personal digital diaries, on the other hand, may represent a more sensitive tool, especially for prescreening and assessing treatment responses [29,34]. Namely, because people provide frequent reports on the ”important” events and experiences of their daily lives, personal diaries offer a unique window on human phenomenology [35]. Furthermore, due to the subjective nature of symptoms, the semi-structured and open-ended designs of such diaries encourage individuals to identify what to record (report) [36,37]. The ability to use their own words and understanding may significantly decrease the perception of “wrong” questions and thus have a potential to significantly increase adherence and the quality of reporting.
Overall, digital diaries and digital screening tools (ePROMs) represent promising means of collecting real-time information and following people with complex (chronic) conditions [38,39,40]. However, to be truly valuable, weaknesses such as poor data utility and lack of in-depth information, patient burden and compliance, recall bias and diary fatigue, no continuous quantification, rigorous training in a challenging population, etc., need to be overcome [39,40,41,42,43,44]. The main motivation behind this paper is the efficient collection of high-quality, near real-time information, with significantly decreased complexity and burden of reporting. The framework simulates “doctor-patient” communication, i.e., creating an initial assessment and providing follow-up symptom-specific questions. We first deployed a pre-screening tool, which can collect an initial set of symptoms from speech. The response was analyzed and possible symptoms and causes extracted using a natural language processing tool. Using the symptoms and causes, we created a symptom-specific discourse to generate more in-depth information and insight. At the same time, the discourse ensures consistent quality of reporting. Finally, the information collected can then be used to trigger disease-specific (e)PROMs, when required, to further mitigate the self-reporting bias. Thus, the proposed tool also optimizes the use of PROMs and requests them from the patients on a necessity basis.
The paper is structured as follows. First, we represent the technology of the MRAST framework, which includes the mHealth patient and clinician app; the MSN, which includes the automatic speech recognition (ASR) SPREAD system [45]; and application of the Symptoma AI, emphasizing the utilization of Symptoma’s proprietary technology for extracting medical risk factors. The technology’s proficiency has been established in previous studies [46,47,48,49]. The MRAST framework also includes the FHIR server with implementation of the HAPI FHIR server [50] to store the patient data. Second, the paper provides insights into connectivity tests, elucidating the system’s robustness and scalability. Third, the patient evaluation section sheds light on the real-world usability of the integrated technologies, as reflected through general patient feedback and three rounds of a self-defined, patient-centered survivorship care plan after cancer treatments, based on big data and artificial intelligence technologies (PERSIST) block ABC surveys with questions related to the mHealth app, video diaries, and chatbot questionnaires. Section 4.3, Patient Evaluation, includes some of the results related to the MRAST framework. System usability scale (SUS) surveys were also provided to grade the mHealth app for patients and clinicians. Finally, the discussion section interprets and contextualizes the obtained results, exploring the implications of the MRAST framework. The paper concludes by summarizing the key findings and offering perspectives on the future implications and enhancements of the integrated technologies in healthcare.
This paper contributes multiple results, as presented in the results section. The MRAST framework was developed to provide a convenient and nonintrusive way of collecting, integrating, and representing patient-collected real-world data and outcomes in the form of structured questionnaires and video diaries. The paper presents the connectivity tests with the Symptoma AI, which provided the proposed symptoms and causes based on the ASR results, and we measured the resource’s consumption. Patients graded the use of the mHealth app, questionnaires, and video recordings, and this paper shows their responses.

2. Related Work

There are multiple technical implementations of solutions for collecting ePROMs being used in practice more and more often, ranging from web platforms [51,52,53,54,55] to dedicated mobile solutions [56,57,58,59]. Mobile solutions are observed as more practical, yield higher response rates, and result in fewer missing items [60,61]. With the digital PROMs offering many advantages over paper-based collection, they are preferred also from a practical perspective; e.g., they offer improved data quality, faster completion time, and decreased costs of data collection [62]. As already highlighted in the introduction, the main disadvantages of ePROMs from a patient’s perspective, and reasons for patients not using ePROMs, can be summarized as (i) ability to use (e.g., physical ability due to health issues), (ii) engagement (e.g., patients do not find them relevant because no symptoms exist), (iii) technical issues and usability (e.g., low technical proficiency), (iv) data security and trust [14,52]. Specifically tackling technical issues and usability, chatbots have been an efficient solution to improve usability and simplify the app functionalities and user experience [45,59,63,64,65]. Namely, chatbots exploit artificial intelligence and natural language processing to interact with patients without human intervention. At the same time, they can personalize the engagement, decrease the complexity compared with standard computer-based surveys, and overall offer a more “human-like” natural language collection of information [66]. The response rates are still relatively low [67]. The reasons for the low response rates are best summarized as disinterest, lack of time, inability to comprehend the questions, and anxiety [68].
Also from a practical perspective, the main barrier to a sustainable collection of PROs relates to engagement, which denotes disinterest and lack of time and comprehension. The digital diaries can improve the patient experience and decrease anxiety [69]. Using semi-structured and unstructured digital diaries with open-ended questions (i.e., interviews) represents an efficient alternative method of collecting patient experiences and outcomes [70]. The diaries with open questions represent collection of data that is unstructured in nature [71]. Overall, the nature of the information captured in such semi-structured interviews with open questions (i.e., in-depth interviews) offers the most reliable means of soliciting information from patients, from multiple perspectives [72,73]. However, when not conducted face-to-face, significant data loss could occur [74]. Namely, the method of in-depth interviewing is carried out as an interview guided by a flexible interview protocol and supplemented by follow-up questions, probes, and comments [75] that make the conversation focused and flexible and prevent poor or limited responses. Recently, there has been a growing interest in AI-enabled chatbot-based systems [76,77].
The chatbot technology could potentially mimic clinical interviews as specific activities such as health surveys, retrieving and analyzing health data, or the translation of diagnostic patterns considering behavioral indicators such as physical activity, sleep, nutrition, etc. Regarding data collection from patients, a plethora of research analyzed the use of chatbots in place of traditional form-based surveys and traditional (e)PROMs [78,79,80,81]. In fact, with closed-ended questions (which PROMs essentially are), the chatbots collect the same quality, if not higher, compared with digital surveys [33,82]. The solutions conducting interviews with open-ended questions generate more noise (less precise data); however, compared with the closed-ended solutions, their participants showed higher engagement and provided higher-quality responses when engaged with the chatbot [83,84,85]. With the recent breakthrough in large language models (LLMs), e.g., OpenAI’s GPT3, Google’s PALM, Gopher, Bing Chat, and Azure Health Bot, etc. [86,87,88,89,90,91,92], the technology has truly led to the development of powerful AI chatbots capable of engaging in natural and human-like conversations. In fact, LLMs are rapidly advancing to achieve near human-level performance on medical question-and-answering benchmarks [93,94]. As a main barrier to their facilitation in the health domain, the LLMs reflect the biases inherent to the systems they were trained on, i.e., the human interactions and internet data. This means their use can lead to manipulation, deception, and can even damage the users’ trust and negatively impact the users’ well-being [95,96]. Furthermore, the LLMs fail significantly when faced with a list of specific questions [97].
To sum up, it is well-acknowledged that incident reporting under-represents the actual frequency of events [98]. A significant part of this can be attributed to the effects of hindsight and outcome bias and the ability to discern “right” from “wrong” [99]. This is further emphasized by individuals’ ability to express themselves and articulate specific issues [19]. In this paper, we propose using a semi-structured, short questionnaire capable of capturing participants’ perspectives regarding an experience or an issue using their own words [100], eliminating the issue of comprehension. We describe a solution to collect patient self-reports using a storyline based (not LLM) speech-enabled chatbot, where the storyline is built dynamically based on symptoms expressed by the patient during each open-ended question. This means that the open-ended questions can be followed by a series of closed-ended questions, targeted to provide further context on the symptoms expressed by the patient. Compared with LLMs, the solution is limited in terms of discourse diversity; however, it does not introduce bias or responses potentially dangerous/negative to end users. Furthermore, it extends the traditional open-ended medical chatbot solutions with the capacity of extracting symptoms and causes from user responses and thus extending the interview with follow-up questions, and compared with traditional (e)PROMs, modeling the discourse to be relevant to actual issues experienced by the patient. Regarding the evaluation of user experience, we have found and compared some studies [101,102,103,104,105] that also include mHealth apps for cancer patients. The comparison of those studies is presented in Section 4.4, Feasibility of MRAST Framework in the Real World.

3. Methodology

The methodology section of this study presents a comprehensive approach known as the MRAST framework, designed to harness the power of decentralized architecture and advanced technologies in the context of healthcare. This framework comprises several key components, each serving a unique role in enhancing patient care and generating valuable insights. In this introduction, we delve into the fundamental aspects of the MRAST framework, with a focus on its primary elements, including the PERSIST mHealth application, a multimodal sensing network, and the big data platform based on Open Health Connect (OHC). Additionally, we explore the speech recognition engine, which plays a crucial role in extracting information from diary recordings. The methodology also highlights the transformation of diary recordings into updated patient profiles and the utilization of Symptoma’s AI to mine valuable information from the extracted data. Furthermore, it outlines the role of the big data platform and the HL7 FHIR server in managing and integrating patient information from various sources. Overall, the MRAST framework represents an innovative and holistic approach to healthcare, combining cutting-edge technologies and data-driven insights to improve patient well-being and disease management.

3.1. Environment

Figure 1 shows the MRAST framework as a whole. The MRAST framework consists of the mHealth patient and clinician apps, OHC platform, Symptoma AI, and MSN. The MSN’s architectural structure can be broken down into three components: Apache Camel, Apache ActiveMQ Artemis, and Apache Kafka. Apache Camel serves a dual role as both an external access point to the MSN and an internal link between Apache ActiveMQ Artemis and Apache Kafka. It operates as a Spring Boot application with an embedded Apache Tomcat server, facilitating HTTP requests. This Spring Boot application runs on a virtual machine and functions as a router, enabling the seamless exchange of data across various protocols.
Apache ActiveMQ Artemis operates as an MQTT broker, serving both internal and external connections to the mHealth app, which subscribes to specific topics. On the Apache Camel side, a REST API is implemented using REST DSL with Java, complemented by Swagger UI (OpenAPI) for documentation and REST endpoint testing. The mHealth app communicates with the Apache Camel REST API for managing questionnaires and user validation. The integration between Apache Camel and Apache ActiveMQ Artemis occurs through two methods: one via the Java Messaging Service (JMS), functioning similarly to MQTT with its support for topics and queues. Meanwhile, Apache Kafka, a distributed event streaming platform, plays a vital role in the MSN, managing data distribution through AI-based microservices.
MSN is tasked to enable communication between the components outside of the MSN, which include the patient and clinician mHealth apps, the Open Health Connect platform with FHIR server, and Symptoma’s AI.

3.2. The MRAST Framework

The MRAST framework in Figure 2 is built on a fully decentralized architecture that consists of four main components: (i) the PERSIST mHealth application, serving as the main interface for patients and clinicians; (ii) a multimodal sensing network, delivering software sensors to extract symptoms and causes; and (iii) the big data platform based on Open Health Connect (OHC) [106], a digital platform that provides the building blocks of connected health and care systems. The mHealth application consists of an application for patients and an application for clinicians. Within the scope of the MRAST framework (Figure 1), the role of the application for the patients is to enable and guide the diary recording process (1), deliver notifications (5), and display the disease-centric discourse (DCD) (6). Within the scope of the MRAST framework, the role of the application for clinicians is to display symptoms from the diary (2, 3), allow them to trigger the DCD (4), and finally, review the results and possible causes of the symptoms (3, 7). The process starts with the patient doing a video recording. In the mHealth application, the patient gets a notification to record a video diary. Once the patient taps on the notification patient, they will be led to a screen where the patient can start the recording. When the patient finishes with the recording, the video is stored to an OHC server and linked to the FHIR server resource as a link. Over the link, the MRAST framework can access it to process it further. From the video recording, the MRAST framework then extracts the symptoms. The second part is the collection of the questionnaire responses to obtain detailed specifications of symptoms. Over the mHealth application, each patient gets a notification over the MQTT protocol to fill in the questionnaire. Once the user taps on that notification, the questionnaire opens, and the user answers the questions until the end. For providing the new questions and retrieving the patient answers, we implement the REST protocol with a Rasa-based chatbot. Collected answers on questions are stored to the FHIR server when the user answers the last question.
The multimodal sensing network [107] represents the “brain” of the MRAST framework. It consists of components and end-to-end services to facilitate the symmetric interaction [45], including the speech recognition engine SPREAD, speech synthesis framework PLATOS, natural language services (including a Rasa-based chatbot), and conversational language generation services, i.e., the embodied virtual agent framework EVA [108]. Additionally, the framework integrates a symptoms extraction and tracking framework, which includes a depression classification pipeline and a risk assessment component built on top of Symptoma AI [47,109] for extracting clinical cues from free text, assessing risk factors, and returning risk scoring. MSN is a microservice-based component where the services are running as a virtual machine on an Ubuntu-based server that runs the Proxmox Virtual Environment. Some of the services, mostly the ones that need the use of GPUs, are running directly as services on Ubuntu-based servers with GPU cards (2x NVIDIA GeForce RTX 3050 Ti) used for inference. MSN is protected with the implementation of the VPN and firewall that allows specific ports to specific IP addresses with the use of SFPT and SSH to access the files and the commands terminal.
Finally, the OHC is a complete integration and streaming platform for large-scale distributed environments. Unlike a traditional messaging system or stream processing API, OHC enables all interfaces to be connected to and make decisions across disparate data sources in real time. OHC provides the framework and set of tools for the integration, ingestion, storage, indexing, and surfacing of patient information. The OHC platform is also a microservice-based platform where services are mostly running as Docker containers. The main component is the HL7 FHIR server, the Keycloak identity and access management service that generates the JWT token for safe access to the data that are stored on the FHIR server. OHC also offers a Kibana search that was used for stored data representation in the form of graphs. OHC is explained in more detail in Section 3.2.3, Big Data Platform and HL7 FHIR Server.

3.2.1. Speech Recognition Engine

To extract speech from diary recordings, we deploy an end-to-end multilingual automatic speech recognition (ASR) system, SPREAD [45]. SPREAD is built on an end-to-end connectionist temporal classification-based deep neural model. The acoustic model is based on the B × R Jasper model [110]. In SPREAD, we extend it with a natural language model and spell-checker. To facilitate the challenges of the “data in the wild” [111], the system includes a spell-checker model and a 6-gram KenLM [112]-based language model. The overall architecture is outlined in Figure 3.
In this end-to-end ASR model, acoustic and pronunciation models are replaced with a convolutional neural network (CNN). In the preprocessing phase, the mel filter bank features are calculated from 20 ms windows, and a 10 ms overlap is used. The ASR engine outputs a probability distribution over characters per frame. The engine has a block architecture; therefore, the B × R model has 10 blocks, each with 5 sub-blocks. Several operations are applied to each sub-block, such as a 1D-convolution, batch norm, ReLU, and dropout. Within each block, all sub-blocks have the same number of output channels. There is a residual connection between each block, which consists of a projection layer followed by batch normalization. The NovoGrad [113], an optimizer similar to Adam [114], is used to compute second moments per layer instead of per weight.
The decoder converts a probability distribution over characters into text. There are different types of decoders that are usually employed with CTC-based models: greedy decoder and beam search decoder, with or without spell-checker model, with or without language model re-scoring, etc. A greedy decoder outputs the most probable character at each time step. It is very fast, and it can produce transcripts that are very close to the original pronunciation. However, it may introduce many small misspelling errors. Due to the nature of the word-error-rate (WER) metric, even one character error makes a whole word incorrect. Thus, a beam search decoder with language model re-scoring and spell-checking allows for many possible decodings (beams) at once, assigning a higher score for more probable N-grams according to a given language model. The language model helps to correct misspelling errors. The downside is that it is slower than a greedy decoder.
The spelling correction (SC) models are used to explicitly correct acoustic ASR errors. In SPREAD, we are utilizing text-only data by training a supervised “spelling correction” model to explicitly correct the errors made by the acoustic model. Instead of predicting the likelihood of emitting a word based on the surrounding context, as in RNN-LM [115], the SC model in SPREAD only identifies likely errors in the acoustic model and proposes alternatives. We integrate a context-aware spell-checking library for automatic spelling correction. Correction and error detection targets to correct up to three edit distance errors and splits two merged words when needed. The SC uses a combination of CatBoost gradient-boosted decision trees, N-gram language models, and a static dictionary for error detection and candidates ranking. For each word, a set of features is generated, such as word length, prediction of 2-gram lm, 3-gram lm, 4 masked gram, absence or presence in the dictionary, and others. A fast classifier makes a prediction whether the word is correct or not. For “misspelled” words, a list of candidates is also generated.
Language modeling is the task of assigning probability to sentences in a given language. In addition to assigning a probability to each sequence of words, the language models (LMs) also assign a probability for the likelihood of a given word (or a sequence of words) to follow a sequence of words. N-gram language models are still among the most popular statistical language models today. During speech recognition decoding, candidates are evaluated using both acoustic scores and LM scores. As outlined in Figure 3, based on experiments, we propose to use a word-level N-gram language model after the spell-checking model in order to generate a candidate list using beam search with a specific width. Namely, an external LM model can re-score the final list of possible candidates. All LMs in SPREAD are trained on text datasets independently from the acoustic models. We use the scalable modified Kneser-Ney language model (KenLM) [112] estimation approach for training 6 g models for all languages. KenLM is a library that implements two data structures for efficient language model queries, reducing both time and memory costs.

3.2.2. From Diary Recording to Updated Insights on Patient Condition

The main objective of MRAST is to generate additional insights on symptoms and well-being in real-life settings. Namely, multiple studies have shown that symptoms extracted from conversation can greatly improve the accuracy of disease identification and disease progression [116]. The implementation and workflow are outlined in Figure 4.
The diary text, automatically extracted from diary recordings (see Section 3.2.1, Speech Recognition Engine) is first sent to the symptom extraction service based on Symptoma’s AI [117]. The approach aims at mining data features from free text that are medically relevant and can represent the documented content. This exploits Symptoma’s proprietary disease database built from proprietary disease concepts as well as its proprietary ontology, structuring disease—symptom—risk factor—etc. Symptoma’s proprietary ontology is developed by analyzing medical articles, case reports, and patient-generated data from 36 languages. Further, learning from billions of anonymized keywords entered by more than 10 million monthly users allows the addition of (lay) terminology variants, which is unparalleled, looking at the existing ontology landscape. The AI technology developed by Symptoma is a significant technological breakthrough, built on more than 15 years of research and development in the field. It is designed to understand the medical context of information units and to identify disease–symptom–risk factor relations, making it a valuable tool for medical professionals and patients alike. Symptoma AI boasts an accuracy rate exceeding 95% across a vast spectrum of over 20,000 diseases. The extracted concepts are stored as FHIR compositions [118] on the big data platform (See Section 3.2.3, Big Data Platform and HL7 FHIR Server). The extracted concepts are further transmitted to the “disease-centric discourse” to (1) trigger relevant PROs to be filled in or (2) to trigger a simple Q&A (with binary answers) to provide further context on the symptoms identified.
If (1) the Rasa framework-based DCD is activated to facilitate the completion of PROs, the outcomes provide valuable subjective information directly from the patients, contributing to a holistic understanding of their health status. The activation is initiated through a REST API called from the Open Health Connect (OHC) platform, prompting the patient to engage with the questionnaire. Simultaneously, the initially extracted concepts are supplied as input, enhancing the questionnaire’s relevance to the individual’s health concerns. The responses obtained from the patient are then relayed to the Symptoma endpoint through the UM REST API, facilitating a seamless integration of patient-reported data with the symptomatic information extracted from the diary recordings.
Moreover, (2) the DCD employs a tailored questionnaire designed for binary answers. This approach aims to further contextualize the concepts identified during the initial extraction process. The DCD, guided by Symptoma’s AI and considering the patient’s responses, refines its understanding of the symptoms, signs, and risk factors. This refined information is pivotal in generating an updated patient profile, enhancing the accuracy and relevance of the insights derived from the diary recordings.
As illustrated in Figure 3, the journey from diary recording to an updated patient profile embodies a dynamic feedback loop. Symptoma’s AI is not only extracting relevant data features, but also contextualizing their relations to diseases, suggesting further leading questions. This iterative process of data extraction, contextualization, and user interaction establishes a robust foundation for generating real-world insights into patient conditions and disease progression.

3.2.3. Big Data Platform and HL7 FHIR Server

The big data platform is based on Dedalus’s Open Health Connect (OHC), a predecessor of Digital Connect 4 Healthcare [119]. The OHC Digital Health Platform comprises sets of components that are orchestrated together in a holistic platform. The platform enables healthcare organizations to access, cleanse, integrate, ingest, and semantically “tag” their own data held across multiple clinical and operational systems. Unlike in traditional messaging systems or stream-processing APIs, Open Health Connect enables all interfaces to be connected to and make decisions across disparate data sources in real time. Open Health Connect provides the framework and set of tools for the integration, ingestion, storage, indexing, and surfacing of patient information. OHC facilitates innovation through near real-time access to longitudinal patient information by combining data in a defined FHIR format from a wide range of systems of record. OHC can adapt to unique business results owing to our design, which is open and flexible on purpose.
The results from the MRAST framework are stored on the UM’s FHIR server, using the CASIDE data model for cancer survivorship [120]. The UM FHIR server is based on the HAPI FHIR v2 [50] using the JSON to store the resources. FHIR defines a set of resources for representing and exchanging healthcare information, and it is designed to be easy to implement and support modern web technologies. In project PERSIST, we utilize the following resources to integrate real-world data collected from the patient diaries, i.e., the diagnostic report, observation, and composition. An FHIR diagnostic report resource is used to represent and communicate the results of diagnostic investigations, such as laboratory tests, imaging studies, or other diagnostic procedures. The diagnostic report resource provides a structured representation of the key information related to a diagnostic report, including the patient, the requestor, the service provider, the date and time of the report, and the actual diagnostic results. The FHIR Observation resource is used to represent measurements or simple assertions made about a patient or other subject. It is used for capturing clinical data related to various health parameters such as vital signs, laboratory results, and other observations. Observations can cover a wide range of clinical data, including numerical measurements, categorical assessments, and coded observations. As for the composition resource in FHIR, it is used to represent a set of healthcare-related information that is a coherent set and has clinical meaning. A composition resource typically includes metadata about the composition, such as the author, date, and context, and it may contain references to other resources, such as observations, conditions, or medications, to represent a comprehensive clinical document. Compositions are often used to create structured documents, such as discharge summaries or clinical notes, that capture a snapshot of a patient’s health status at a particular point in time. In the context of the PERSIST project, we store the patient symptoms together with the patient text transcription as results from the MRAST framework to the composition. In another composition resource that is linked to the patient resource, we also store the speech, text, and facial hand-crafted features. In the third composition resource are the extracted symptoms, signs, and causes as well as patient-provided answers to the DCD questions. In the PERSIST project, the diagnostic report was linked with the diary videos that were retrieved from the patients’ smartphones. Once the process of the MRAST framework finished, we added notes in the existing resource to mark that the linked video was annotated. The observation resource contained the final depression decision result.

3.3. Case Study with Full Patient Journey

A patient records their daily experiences, capturing not just words but the essence of their ailments in a video diary. The UM seamlessly extracts audio from these narratives with the help of the MRAST framework. This audio journey undergoes a transformation, owing to an automated speech recognition engine. It skillfully transcribes the patient’s spoken words into text, revealing valuable information. Consider a snippet from this textual tapestry: “I have a slight fever today, my head hurts, and my throat hurts. Since I’m not feeling well, all I want to do is sleep.” Five symptoms emerge—sore throat, low fever, headache, pain, and insensitivity to pain, offering a glimpse into the patient’s world.
The UM has put Symptoma’s concept extraction from video diaries as the MRAST framework presented in Figure 5 to use and evaluated it. The UM pulls audio from the patient video in the flow. The automated speech recognition engine is then used to extract the transcribed text from the audio. The following sample of text has five symptoms: “I have a slight fever today, my head hurts, and my throat hurts. Since I’m not feeling well, all I want to do is sleep.” From these results, at the end as the result we can see that the sore throat, low fever, headache, pain, and insensitivity to pain are recognized as possible symptoms. Those results are then stored to a UM local FHIR server to be validated.
The identification of medical domain-specific risk factors is accomplished by leveraging Symptoma AI, which includes an AI disease engine and a sophisticated disease database. This technology is designed to analyze symptoms, signs, and risk factors gathered from millions of medical articles and publications in 36 languages, to enable the recognition of terminology variants. Symptoma AI has been extensively validated for its performance through various studies and is also utilized in the Symptoma Digital Health Assistant, a Class I medical device.
Next, we describe the real-world implementation of DCD. This section describes the step-by-step implementation of the pipeline. This pipeline implements a disease-oriented dialogue system framework to provide additional context for symptoms detected within the PERSIST MSN by conversing with patients to collect additional symptoms beyond their self-reports or the content within diaries. Figure 6 depicts an updated communication flow for the disease-centric discourse. DH is activating the flow automatically via the REST API on OHC, requesting to offer the questionnaire to the specific patient and supplying the initial patient symptom or symptoms. This is submitted to the UM REST API as a JSON payload, which is then transferred to an MQTT message with the same JSON payload.
This MQTT payload is then transmitted as a notice to the mHealth application. On the user’s smartphone, a message arrives, requesting the user to complete the required questionnaire. When the user taps on the message, the UM REST API receives a REST request. The Symptoma endpoint is then notified by UM to provide the first question for the patient. Symptoma then delivers the inquiry to the UM endpoint, which is utilized for the text-to-speech (TTS) and embodied conversational agent (ECA) machine learning (ML) microservice. This way, the UM can create the video containing the ECA video together with the TTS audio result that is presenting the retrieved question. UM transmits the received queries to the mHealth application and receives the user’s response back in the form of “YES” or “NO” answers. Symptoma receives UM’s responses, and the discussion continues until the final relevant question/answer is resolved. EMO keeps track of conversation statistics, which it will send to the DH OHC FHIR server as an FHIR composition resource. The next sections depict how this workflow works for a specific problem and for a specific questionnaire.
In Figure 7 is shown an example of the FHIR composition resource that contains the extracted symptoms and original patient text received from the speech by ASR. In the specific resource we have information about the version, so we can see how many times this resource has been updated. There is a date with a timestamp of that latest update. In the document is the reference to the patient resource to which this composition belongs. In the Symptoma’s AI results part, we have the text extraction results, and at the end is the original text input provided by the text-to-speech part over the patient voice.

4. Results

In this section, we provide the outcomes of our study, revealing insights across key dimensions of our framework; i.e., the multimodal sensing network (MSN) architecture, the automatic speech recognition (ASR) system, SPREAD, and the scalable MRAST framework. A real-world case study illustrates the operational context. Lastly, the PERSIST clinical trial involving 166 patients unveils valuable insights, with the system usability survey showcasing evolving perspectives on the mHealth app usability.

4.1. ASR Results

The ASR system SPREAD is constructed around an end-to-end deep neural model based on connectionist temporal classification (CTC), similar to models like DeepSpeech. The term “end-to-end” denotes that it relies solely on speech samples and their associated transcripts, without the need for additional information. This approach enables the system to establish a correspondence between audio and text. The ASR model within SPREAD can be distilled into two significant components: training and inference. The training of the acoustic deep learning (DL) model for SPREAD is an offline procedure.
We had enough data for training and testing in Slovenian, Latvian, English, Russian, and French language datasets. The Slovenian training took 55 days, resulting in a highly accurate and low word error rate. Similarly, Latvian training for 55 days showed exceptional performance. For English, the model trained for 81 days, demonstrating outstanding accuracy and a low word error rate. Russian training, spanning 145 days, also yielded exceptional results. Finally, French training for 185 days showcased outstanding performance with high accuracy and minimal word error rate, as shown on Table 1.

4.2. FHIR Server and Connectivity Tests

The raw patient files are taken from the OHC FHIR server and processed by the MRAST framework, and the refined patient files are stored to the UM FHIR server, where they are checked for errors and then created on the OHC FHIR server. Connectivity load tests were made from UM REST API, executing the requests in batches from 50 to 1000 requests per batch. The requests included the dummy patient text with 7540 characters or 1455 words. Requests were formed and sent with the Python 3.8.12 script, wherein each request is a new thread. This way, we can see if the system would support the specific number of requests or system users, which in our case is the number of patients. Also, we can see the time needed for a specific request batch to be executed at the same time and make further decisions about future developments of the system. In the next figures, we present the results of the load tests. In Figure 8, we can see the response time in seconds based on the sent number of requests, between 50 and 1000. As we can see, the response time rises linearly from 50 requests, where we have a response time of 0.29 s, to 1000 requests, where the response time is 3.06 s.
In Figure 9, we can see the response time for requests in seconds based on the sent number of requests between 50 and 1000. Here, we are observing the total response time divided by the number of requests. As we can see, the response time drops logarithmically from 50 requests, where we have a response time of 5.94 milliseconds, to 1000 requests, where the response time is 3.07 milliseconds.
In Figure 10, we can see the RAM usage per request batch between 50 and 1000. RAM usage rises linearly but with very low or unnoticed rise of consumption from the 50 requests, where we have CPU requests of 5.81 GB, to 1000 requests, where RAM usage is 5.85 GB.
In Figure 11, we can see the CPU usage per request batch between 50 and 1000. CPU usage rises linearly, but with a low rise of consumption from the 50 requests, where we have CPU usage of 3.14%, to 1000 requests, where CPU usage is 5.81%.
In Figure 12, we track network traffic per request batch between 50 and 1000. Network traffic rises exponentially from the 50 requests, where we have network traffic of 2.49 k, to 1000 requests, where network traffic is 479.5 k.
Compared with Suresh et. al. [121], the proposed system can carry out a higher number of requests in less time and have better performance, but with a smaller size FHIR resource used while making tests. In their paper, the performance testing was made where they took a specific HL7 FHIR resource questionnaire for GAD-7. They executed 50 concurrent users in over 20 min and received an average response time between 0.3 and 0.5 s.

4.3. Patient Evaluation

A total of 166 patients took part in the PERSIST clinical trial across four hospitals, as detailed in Table 2. Among these participants, 85 individuals were diagnosed with breast cancer, while 81 were affected by colorectal cancer. The average age of the patients upon enrollment was 55 years. The study comprised 37 male and 129 female patients, resulting in an uneven gender distribution. This disparity can be attributed to the infrequency of breast cancer in men and a slightly higher representation of women in the group of patients with colorectal cancer. This gender imbalance in inclusion is further explained by the clinicians’ observations, who noted that men displayed less interest in participating in the study compared with women.
To collect feedback from patients, app-based questionnaires named PERSIST block ABC [122] were administered at three distinct time intervals to gain insights into their participation experience in the study and to highlight their key observations. The surveys from Table 3, Table 4, and Table 5 received responses from a total of twenty participants across three time points, with four participants from CHU, four from SERGAS, and twelve from UKCM. Notably, no participants from UL responded throughout all three survey instances. Those surveys were distributed at the commencement of the app-based questionnaire (First), after the introduction of the virtual agent (Middle), and at the conclusion of the study in October 2022 (Last). Participants rated their experiences on a scale of 1 (poor) to 10 (excellent). Analysis revealed no statistically significant differences between any two time points, as indicated by the Friedman one-way repeated measure analysis of variance by ranks with a p-value of 0.779. Furthermore, Conover’s post-hoc pairwise comparisons demonstrated no significant differences in the p-values among the initial-mid (p = 0.490), initial-final (p = 0.843), and mid-final (p = 0.622), affirming the consistency of responses across the survey periods. Table 3 shows descriptive statistics that summarize the answers to the question, “How do you rate your experience with questionnaires in the app?”
The information presented in Table 4 summarizes answers to the question, “How do you rate your experience with the mHealth app?” That information indicates that there were no statistically significant variations in participants’ assessments of the app’s ease of use between any two time points. The outcomes from the Friedman one-way repeated measure analysis of variance by ranks indicate a non-significant p-value of 0.279, suggesting that observed discrepancies in the ratings were likely due to random chance. Additionally, Conover’s post-hoc pairwise comparisons underscore that there were no notable distinctions between the initial and middle time points (p = 0.891). However, significant differences were observed between the initial and final time points (p = 0.138) and the middle and final time points (p = 0.176).
Table 5 represents the statistics that answer the question, “How do you rate your experience with diary recording?” Table 5 reveals that Friedman one-way repeated measure analysis of variance by ranks indicated no statistically significant differences between any two time points (p = 0.581). Conover’s post-hoc pairwise comparisons further confirm the absence of significant differences for the initial-mid (p = 0.304), initial-final (p = 0.512), and mid-final (p = 0.707) time points.
SUS questionnaires [122] were provided to patients with the same approach as the PERSIST block ABC questionnaires. Participants completed the survey three times: at the start, in the middle, and at the end of the clinical study. The 27 patients finished all questionnaires all three times. For each patient, the SUS score was calculated based on their responses to the 10 questions. According to the system usability levels, at the beginning of the study, most patients perceived the system as having “usability issues” (10 responses) and being “acceptable to good” (10 responses). This perception could be linked to patients’ prior experiences with technology in general, including different types of applications, and their ability to adapt to the mHealth app, which was still under development. Throughout the study, the proportion of participants who deemed the system to have “excellent usability” rose from 14% to 33%. This improvement can be attributed to the continuous enhancements made to the mHealth app in collaboration with technical partners. By the study’s conclusion, the predominant scoring category for the system was “Experiencing usability issues.” However, despite this, 44.44% of patients rated the usability as good or excellent, combining responses of “Acceptable to good” and “Excellent usability.”
Clinicians working with the mHealth app at the four participating hospitals also received SUS questionnaires. Two rounds of responses were collected: the first round involved only the mHealth app web version, and the second round included the mHealth app mobile version at the end of the study. The findings from the SUS questionnaire distributed to clinicians utilizing both the mHealth web and mobile app versions indicate that a majority of clinicians identified some usability issues (81.55% in the first round and 87.5% in the second round). However, it is noteworthy that the scores did not significantly differ between the two rounds, suggesting that the introduction of the app version did not bring about new usability issues.

4.4. Feasibility of MRAST Framework in the Real World

Table 6 represents a comparison of the project PERSIST study with similar studies that provided mHealth applications and similar systems for use to patients and then evaluated the overall user experience, usability, and general feedback of using the mHealth app. In the table we add the included studies and which questionnaires they used to collect the patient feedback. We present the number of patients included in each study and patient feedback in a range from strongly negative to strongly positive. All studies showed positive to strongly positive patient feedback and grades of the apps.
Short et. al. [117] performed an evaluation with 10 cancer patients. On a scale from 1 (strongly negative) to 5 (strongly positive) patients graded the app with a high grade of 4.4, which was a little bit higher than the grade of the patients in our study, but we used a bigger patient sample size. The apps that were suggested were typically perceived as user-friendly, with a grade of 4.1; the process of aligning apps with participant preferences was viewed as beneficial and was graded as 4.2. Nonetheless, all results show that the patient feedback was generally positive. Loh et. al. [102]’s primary outcome was usability assessed by the system usability scale (SUS). Their study included 18 cancer patients and 13 caregivers on a scale from 1 (strongly negative) to 5 (strongly positive). A significant portion of patients and caregivers expressed appreciation for and enjoyed the experience, recognizing the value of the proposed app. The overall satisfaction using the app for patients was graded 3.4, while the other question in the SUS varied from grade 3.0 to 3.6. Moorthy et. al. [103] had 133 cancer patients using their app and grading it in a range from 0 to 100. In that study, descriptive statistics were analyzed for the SUS and MAUQ (mobile app usability questionnaire) to evaluate the usability of their app. The participants demonstrated a notably high perceived usability, as evidenced by the SUS score with a mean of 88.3, surpassing the average score of 68. Similarly, the MAUQ produced a mean score of 85.89, providing additional confirmation of the positive perception of usability. In the study of Teckie et. al. [104], out of the 32 participants eligible for analysis, 53% (17) completed all scheduled sessions, 63% (20) completed 75% or more, and 78% (25) completed at least 50% by the study’s end. At the study’s conclusion, 53% (17/32) reconsented for SUS. The mean SUS score (95% CI) was 71.9 (64.3–79.5), indicating an “acceptable” rating. Subscale analysis revealed the learnability domain mean (95% CI) as 78.7 (71.2–86.1) and the usability domain mean (95% CI) as 70.2 (61.8–78.7). In the SUS responses, 88% found their app “easy to use,” 94% believed most could learn it quickly, and 82% felt very confident. Regarding usefulness, 76% agreed, with 59% and 71% agreeing with PRO frequency and length. Additionally, 76% would recommend their app, and 29% provided feedback, using words like “informative,” “helpful,” and “valuable” to express their positive experience. In the study by Paulissen et. al. [105], a total of 15 cancer patients returned SUS forms that were used in the analysis of this study which graded the mHealth app in range from 0 to 100. The outcomes from the computed SUS scores indicate a mean score of 86.8, categorizing the mHealth app as excellent. Participants found it helpful to respond to health questions before their appointments using the application. They mentioned that it allowed them to discuss all health issues more thoroughly during their visit, making it a more efficient and effective experience within the allotted time. This aligns with other compared studies, where the use of mHealth apps is generally acknowledged. Questionnaires from our study show that the patient general feedback about the mHealth app, questionnaires, and video diaries is strongly positive.

5. Discussion

In the realm of healthcare, the unparalleled significance of real-world data, particularly in the context of chronic diseases, cannot be overstated. Traditional measures such as patient-reported experience and outcomes measures (PREMs/PROMs) often grapple with limitations stemming from subjectivity and a lack of precision in targeting the nuanced experiences of individuals facing chronic conditions. The intricate and diverse array of symptoms experienced by these patients poses a challenge in selecting appropriate questionnaires to accurately capture their multifaceted realities. This limitation is where the potential of diary recordings emerges as a promising solution. Unlike conventional measures, diaries offer a comprehensive lens into psychological well-being by encompassing both physiological and psychological symptoms. Beyond symptomatology, these records illuminate non-symptomatic aspects and lifestyle choices, providing healthcare providers with a holistic understanding of a patient’s life. Recognizing the enormous utility of such data, not only does this facilitate a more nuanced measurement of everyday experiences, but it also paves the way for patient-centered care by offering insights into aspects previously overlooked. Moreover, the automated extraction of information from diaries represents a pivotal stride toward a patient-centered healthcare paradigm where decisions are rooted in real-world experiences and tailored to individual needs. Embracing this approach holds the potential to revolutionize healthcare practices and substantially enhance patient outcomes.
The good performance of the ASR system, SPREAD, across multiple languages is noteworthy. The substantial training durations for the models indicate the depth of learning, and the low word error rates (WERs) affirm the accuracy achieved. The results highlight the efficacy of end-to-end deep neural models based on CTC, emphasizing their potential for real-world applications in diverse linguistic environments. The integration of Symptoma’s AI showcases a significant advancement in medical domain-restricted risk factor extraction and opens the possibilities of predictive precision medicine. The utilization of AI for identifying and standardizing medically relevant concepts from free text makes unstructured data comparable, thus laying the foundation for building predictive models on top. The successful implementation of the FHIR server, coupled with the results of connectivity load tests, underscores the system’s robustness. Linear increases in response time and RAM usage, along with low CPU usage increase, indicate efficient handling of requests. The exponential rise in network traffic illustrates the scalability of the system, a crucial aspect for handling a substantial number of patients in real-world scenarios.
There are some drawbacks and limitations of the MRAST framework that should be considered if taking this approach in practice. The first one is the problem with the use of the diary recording functionality in the mHealth app. During the initial tests, some patients had difficulties with the use of some of the diary functionality. In order to improve the user experience of a mobile health (mHealth) app, co-creation sessions were conducted, resulting in the integration of specific elements such as a frame showcasing the position of the face, pause functionality, and an automatic end to the recording after 5 min of inactivity. The impact of these modifications and improvements was highlighted by the continuous increase in perceived experience with the app. Future research should focus on investigating the functional perspectives of mHealth apps, with a specific emphasis on introducing tasks into the everyday life of patients as nonintrusively as possible [123]. The second limitation relates to the errors in speech recognition. Namely, although with continuous retraining of the acoustic model the WER decreases, the WER might not be a realistic representation of what happens in the wild [124]. As also observed during the initial tests in real-world environments, the challenges posed by factors such as background noise, speaker variations, and the presence of multiple dialects have significant impacts on the actual accuracy of the model, e.g., the batch WER vs. test WER in Table 1. To correct misspellings and grammatical errors, we added a language model and spell-checking model to correct ASR errors. Overall, the final model achieves real accuracy over 92%, even for non-mainstream languages, which is comparable to the existing state of the art. However, future research should continue to enhance the accuracy and precision of speech recognition systems, particularly in the presence of background noise, various accents, and dialects [125]. Furthermore, as also highlighted in this paper, advancements in natural language understanding can enable more seamless and intuitive communication, as well as the integration of multimodal sensing for improved accuracy [126].
As observed in the results, the distributed architecture deployed in the experiments may cause certain delays in the overall execution. Depending on the availability of system resources and network traffic, our simulations already show that the responsiveness of the system may quickly traverse from real time, with execution times in milliseconds, to interactive time, with execution in several seconds. However, the benefits of distributed architecture with multiple networks carrying out specific tasks include overall performance and scalability, reliability and resilience, and efficiency (e.g., resource sharing, use of energy-efficient hardware, etc.). Finally, in the current use of the pipeline, the delays did not present a specific relevance or limitation. Namely, our approach was to represent the results to clinicians as background used during follow-up. If, however, similar systems should be used in concepts such as prescreening, research should focus on developing efficient communication protocols and network optimization techniques to minimize delays and ensure reliable data transmission within distributed systems [127].
The patient evaluation in the PERSIST clinical trial provides insights into the real-world usability and acceptance of the integrated technologies. Participants generally expressed a positive outlook toward the app-based questionnaires, yielding a mean score of 7.48 across all time points (initial, middle, and final). The median score of 8 indicates that most participants rated their experience as either “good” or “excellent.” The relatively low standard deviation implies minimal variability in participants’ ratings. Moreover, the absence of statistically significant differences between any two time points suggests that participants’ perceptions of their questionnaire experience remained relatively stable over time. This consistency implies that the app maintained a steady level of usability and effectiveness throughout the study duration. In summary, the data indicates a favorable participant experience with the app-based questionnaires. However, it is crucial to recognize that this assessment represents just one facet of the app’s performance. Further research may be necessary to comprehensively assess its effectiveness and user-friendliness. Respondents to this inquiry typically reported a favorable encounter with the diary recording in the application, giving it an average rating of 7 out of 10. The median rating remained consistent at 8 out of 10 for both the middle and final time assessments, suggesting a sustained positive experience for these individuals over time. Furthermore, there were no statistically significant differences observed between any two time points, indicating the stability of the positive experience with diary recording for these participants throughout the study.
In general, patients consistently evaluated their experience with the mHealth app positively, and these ratings exhibit a slight upward trend over time. Importantly, there are no statistically significant differences observed between any two time points, signifying the app’s consistently favorable reception among patients throughout the study. Notably, the middle test saw patients from all centers awarding higher scores, indicating the app’s consistent utility across diverse locations. Despite CHU patients providing the lowest ratings, the absence of statistically significant differences implies that the app was generally well received irrespective of the specific center.
The use of AI technologies in healthcare, such as Symptoma AI and chatbots, also presents some challenges and ethical dilemmas. These include privacy concerns and data bias. Patients may have concerns about the confidentiality of their information shared with the MRAST framework. AI language models, including chatbots, are susceptible to various biases, which can compromise the accuracy and fairness of medical information.
The utilization of an MRAST framework for the collection and storage of patient data raises specific privacy and security concerns. Patients may disclose sensitive health information during interactions with the MRAST framework, leading to the inclusion of this information in the FHIR database. The security and confidentiality of patient records, known as medical privacy or health privacy, are paramount and require robust safeguards to prevent unauthorized access or disclosure. It is also crucial to use as little identifying data as possible and to seek patient consent for data usage. In cases where consent is not feasible, special permissions and ethical approvals are required to use patient data for research or analysis. Personal data breaches, including unauthorized access or disclosure, must be addressed in accordance with legal requirements.
By leveraging Symptoma AI to extract and analyze symptoms from patient video diaries, the MRAST framework can offer numerous benefits. The MRAST framework can aid in the early recognition of potential symptoms by analyzing patient-provided data, enabling better treatment. It can contribute to more accurate symptom analysis, leading to improved and more tailored treatment plans. This solution was created for and tested by 166 oncology patients without looking at their socioeconomic backgrounds. By facilitating remote symptom analysis, the MRAST framework can improve access to care for patients, particularly those from diverse linguistic and socioeconomic backgrounds who may face barriers to traditional healthcare services. The framework’s ability to identify symptoms and potential causes using AI and chatbot technology can support the development of personalized care recommendations, thereby improving patient quality of life.

6. Conclusions

In conclusion, the work presented in this paper represents a significant step forward in enhancing patient evaluation and real-world data collection. The ASR system demonstrates robust performance across various languages, with the use of Symptoma’s AI extracting medical risk factors from free text and contextualizing them in relation to possible causes. The successful implementation of the FHIR server ensures seamless connectivity and scalability. The patient evaluation underscores the importance of continuous improvement in mHealth apps. The outcomes of the PERSIST clinical trial, combining advanced technologies and patient-centric approaches, provide a foundation for future developments in healthcare. The discussed technologies offer promising avenues for improving diagnostics, patient engagement, and overall healthcare outcomes. As technology continues to evolve, ongoing research and refinements will further enhance the integration of ASR and AI-driven chatbots in clinical settings, contributing to more effective and patient-friendly healthcare solutions. The article identifies some issues that were resolved during study and some with which we will deal in our future studies, such as problems with recording diaries in the mHealth app, errors in ASR results, slow execution time from diary input to symptom retrieval, and engagement-related issues. Improving the video diary recording through co-creation significantly reduced technical glitches and enhanced the user experience. The accuracy of ASR was significantly improved by coupling it with natural understanding modules (i.e., spell-checker and the language model). To sum up, the outcomes underscore the MRAST framework’s potential to enhance patient quality of life and provide clinicians with deeper insights into patients’ daily experiences with their illnesses. Our future research will focus on how technology is changing healthcare and affecting how patients become involved in their care. We aim to develop creative methods that improve patient care and give useful information to doctors. Using advanced technologies, patient-focused methods, and the MRAST framework is essential for shaping the future of healthcare research and practices.

Author Contributions

Conceptualization, V.Š., I.M., M.R., U.S.; data curation, S.L., M.C., M.H., D.B.; methodology, V.Š., I.M., S.L., M.R., U.S.; software, V.Š, I.M., M.R., U.A., S.L., J.N., A.G.M.; validation, N.P., M.H., D.B.; formal analysis, N.P., U.S., M.C., M.H., D.B; investigation, V.Š., I.M, S.L., M.C., M.H., D.B.; writing—original draft preparation, V.Š., I.M., S.L., D.B.; writing—review and editing, V.Š., I.M., M.R., U.S., S.L., M.C., M.H., D.B., N.P., M.R., J.N., A.G.M., U.A.; visualization, V.Š., I.M., S.L.; supervision, I.M., J.N., M.R.; funding acquisition, I.M., J.N., M.C., D.B., M.H., S.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Slovenian Research Agency (Research Core Funding) No. 0552-0796 P2-0069, Young Researcher Funding 0733/2022/P157/522-KZ, and by the European Union Horizon 2020 Research and Innovation Program, project PERSIST, supported under Grant Agreement no. 875406. The content of this paper does not reflect the official opinion of the European Union or any other institution. Responsibility for the information and views expressed therein lies entirely with the authors.

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki, and as part of protocol registered at ISRCTN: https://doi.org/10.1186/ISRCTN97617326, accessed on 19 June 2021. This study has been approved by relevant ethical committees in Belgium (Institutional Ethics Committee of CHU de Liege, approval ref. no: 2020/248), Latvia (Riga Eastern Clinical University Hospital Support Foundation Medical and Biomedical Research Ethics Committee, approval ref. no: 8-A/20), Slovenia (National Ethics Committee, approval ref. no. 0120-352/2020/5), and Spain (Regional Institutional Review Board, approval ref. no. 2020/394). A written informed consent was obtained from each participant.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data are not publicly available due to restrictions that apply to the availability of these data.

Conflicts of Interest

The authors declare no conflict of interest. The funders or Symptoma GmbH had no role in the design of the study; in the collection, analysis, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Salmond, S.W.; Echevarria, M. Healthcare Transformation and Changing Roles for Nursing. Orthop Nurs 2017, 36, 12–25. [Google Scholar] [CrossRef]
  2. Anderson, G. Chronic Conditions: Making the Case for Ongoing Care. Partnersh. Solut. Johns Hopkins Univ. 2002. Available online: https://www.policyarchive.org/handle/10207/21756 (accessed on 18 December 2023).
  3. Li, J.; Porock, D. Resident Outcomes of Person-Centered Care in Long-Term Care: A Narrative Review of Interventional Research. Int. J. Nurs. Stud. 2014, 51, 1395–1415. [Google Scholar] [CrossRef]
  4. Calvert, M.; Kyte, D.; Price, G.; Valderas, J.M.; Hjollund, N.H. Maximising the Impact of Patient Reported Outcome Assessment for Patients and Society. BMJ 2019, 364, k5267. [Google Scholar] [CrossRef]
  5. Nguyen, H.; Butow, P.; Dhillon, H.; Sundaresan, P. A Review of the Barriers to Using Patient-Reported Outcomes (PROs) and Patient-Reported Outcome Measures (PROMs) in Routine Cancer Care. J. Med. Radiat. Sci. 2021, 68, 186–195. [Google Scholar] [CrossRef] [PubMed]
  6. Black, N. Patient Reported Outcome Measures Could Help Transform Healthcare. BMJ 2013, 346, f167. [Google Scholar] [CrossRef] [PubMed]
  7. Greenhalgh, J. The Applications of PROs in Clinical Practice: What Are They, Do They Work, and Why? Qual. Life Res. 2009, 18, 115–123. [Google Scholar] [CrossRef] [PubMed]
  8. Snyder, C.F.; Aaronson, N.K.; Choucair, A.K.; Elliott, T.E.; Greenhalgh, J.; Halyard, M.Y.; Hess, R.; Miller, D.M.; Reeve, B.B.; Santana, M. Implementing Patient-Reported Outcomes Assessment in Clinical Practice: A Review of the Options and Considerations. Qual. Life Res. 2012, 21, 1305–1314. [Google Scholar] [CrossRef]
  9. Churruca, K.; Pomare, C.; Ellis, L.A.; Long, J.C.; Henderson, S.B.; Murphy, L.E.D.; Leahy, C.J.; Braithwaite, J. Patient-Reported Outcome Measures (PROMs): A Review of Generic and Condition-Specific Measures and a Discussion of Trends and Issues. Health Expect. 2021, 24, 1015–1024. [Google Scholar] [CrossRef] [PubMed]
  10. Foster, A.; Croot, L.; Brazier, J.; Harris, J.; O’Cathain, A. The Facilitators and Barriers to Implementing Patient Reported Outcome Measures in Organisations Delivering Health Related Services: A Systematic Review of Reviews. J. Patient-Rep. Outcomes 2018, 2, 46. [Google Scholar] [CrossRef] [PubMed]
  11. Ju, A.; Tong, A. Considerations and Challenges in Selecting Patient-Reported Outcome Measures for Clinical Trials in Nephrology. Clin. J. Am. Soc. Nephrol. 2017, 12, 1882–1884. [Google Scholar] [CrossRef]
  12. Montgomery, N.; Bartlett, S.J.; Brundage, M.D.; Bryant-Lukosius, D.; Howell, D.; Ismail, Z.; Krzyzanowska, M.K.; Moody, L.; Snyder, C.F.; Staley Liang, M.; et al. Defining a Patient-Reported Outcome Measure (PROM) Selection Process: What Criteria Should Be Considered When Choosing a PROM for Routine Symptom Assessment in Clinical Practice? J. Clin. Oncol. 2018, 36, 187. [Google Scholar] [CrossRef]
  13. Keller, S.; Dy, S.; Wilson, R.; Dukhanin, V.; Snyder, C.; Wu, A. Selecting Patient-Reported Outcome Measures to Contribute to Primary Care Performance Measurement: A Mixed Methods Approach. J. Gen. Intern. Med. 2020, 35, 2687–2697. [Google Scholar] [CrossRef] [PubMed]
  14. Patients’ Reasons for Non-Use of Digital Patient-Reported Outcome Concepts: A Scoping Review—Amalie Søgaard Nielsen, Kristian Kidholm, Lars Kayser, 2020. Available online: https://journals.sagepub.com/doi/full/10.1177/1460458220942649 (accessed on 26 October 2023).
  15. Calvert, M.J.; Cruz Rivera, S.; Retzer, A.; Hughes, S.E.; Campbell, L.; Molony-Oates, B.; Aiyegbusi, O.L.; Stover, A.M.; Wilson, R.; McMullan, C.; et al. Patient Reported Outcome Assessment Must Be Inclusive and Equitable. Nat. Med. 2022, 28, 1120–1124. [Google Scholar] [CrossRef] [PubMed]
  16. Garcia Farina, E.; Rowell, J.; Revette, A.; Haakenstad, E.K.; Cleveland, J.L.F.; Allende, R.; Hassett, M.; Schrag, D.; McCleary, N.J. Barriers to Electronic Patient-Reported Outcome Measurement Among Patients with Cancer and Limited English Proficiency. JAMA Netw. Open 2022, 5, e2223898. [Google Scholar] [CrossRef] [PubMed]
  17. Lavallee, D.C.; Chenok, K.E.; Love, R.M.; Petersen, C.; Holve, E.; Segal, C.D.; Franklin, P.D. Incorporating Patient-Reported Outcomes into Health Care to Engage Patients And Enhance Care. Health Aff. 2016, 35, 575–582. [Google Scholar] [CrossRef] [PubMed]
  18. Sanders, C.; Rogers, A.; Bowen, R.; Bower, P.; Hirani, S.; Cartwright, M.; Fitzpatrick, R.; Knapp, M.; Barlow, J.; Hendy, J.; et al. Exploring Barriers to Participation and Adoption of Telehealth and Telecare within the Whole System Demonstrator Trial: A Qualitative Study. BMC Health Serv. Res. 2012, 12, 220. [Google Scholar] [CrossRef] [PubMed]
  19. Long, C.; Beres, L.K.; Wu, A.W.; Giladi, A.M. Patient-Level Barriers and Facilitators to Completion of Patient-Reported Outcomes Measures. Qual. Life Res. 2022, 31, 1711–1718. [Google Scholar] [CrossRef] [PubMed]
  20. Miller, D.; Steele Gray, C.; Kuluski, K.; Cott, C. Patient-Centered Care and Patient-Reported Measures: Let’s Look Before We Leap. Patient 2015, 8, 293–299. [Google Scholar] [CrossRef]
  21. Fan, X.; Miller, B.C.; Park, K.-E.; Winward, B.W.; Christensen, M.; Grotevant, H.D.; Tai, R.H. An Exploratory Study about Inaccuracy and Invalidity in Adolescent Self-Report Surveys. Field Methods 2006, 18, 223–244. [Google Scholar] [CrossRef]
  22. Stuart, A.L.; Pasco, J.A.; Jacka, F.N.; Brennan, S.L.; Berk, M.; Williams, L.J. Comparison of Self-Report and Structured Clinical Interview in the Identification of Depression. Compr. Psychiatry 2014, 55, 866–869. [Google Scholar] [CrossRef]
  23. Dell’Osso, L.; Carmassi, C.; Rucci, P.; Conversano, C.; Shear, M.K.; Calugi, S.; Maser, J.D.; Endicott, J.; Fagiolini, A.; Cassano, G.B. A Multidimensional Spectrum Approach to Post-Traumatic Stress Disorder: Comparison between the Structured Clinical Interview for Trauma and Loss Spectrum (SCI-TALS) and the Self-Report Instrument (TALS-SR). Compr. Psychiatry 2009, 50, 485–490. [Google Scholar] [CrossRef]
  24. McColl, E. Best Practice in Symptom Assessment: A Review. Gut 2004, 53 (Suppl. S4), iv49–iv54. [Google Scholar] [CrossRef] [PubMed]
  25. Raskin, S.A. Memory for Intentions Screening Test: Psychometric Properties and Clinical Evidence. Brain Impair. 2009, 10, 23–33. [Google Scholar] [CrossRef]
  26. Sato, H.; Kawahara, J. Selective Bias in Retrospective Self-Reports of Negative Mood States. Anxiety Stress Coping 2011, 24, 359–367. [Google Scholar] [CrossRef] [PubMed]
  27. Robinson, M.A.; Boies, K. On the Quest for Quality Self-Report Data: HEXACO and Indicators of Careless Responding. Can. J. Behav. Sci. Rev. Can. Des Sci. Du Comport. 2021, 53, 377–380. [Google Scholar] [CrossRef]
  28. Maniaci, M.R.; Rogge, R.D. Caring about Carelessness: Participant Inattention and Its Effects on Research. J. Res. Personal. 2014, 48, 61–83. [Google Scholar] [CrossRef]
  29. Okupa, A.Y.; Sorkness, C.A.; Mauger, D.T.; Jackson, D.J.; Lemanske, R.F. Daily Diaries vs Retrospective Questionnaires to Assess Asthma Control and Therapeutic Responses in Asthma Clinical Trials: Is Participant Burden Worth the Effort? Chest 2013, 143, 993–999. [Google Scholar] [CrossRef]
  30. Jeong, H.; Yim, H.W.; Lee, S.-Y.; Lee, H.K.; Potenza, M.N.; Kwon, J.-H.; Koo, H.J.; Kweon, Y.-S.; Bhang, S.; Choi, J.-S. Discordance between Self-Report and Clinical Diagnosis of Internet Gaming Disorder in Adolescents. Sci. Rep. 2018, 8, 10084. [Google Scholar] [CrossRef] [PubMed]
  31. Siggeirsdottir, K.; Aspelund, T.; Sigurdsson, G.; Mogensen, B.; Chang, M.; Jonsdottir, B.; Eiriksdottir, G.; Launer, L.J.; Harris, T.B.; Jonsson, B.Y.; et al. Inaccuracy in Self-Report of Fractures May Underestimate Association with Health Outcomes When Compared with Medical Record Based Fracture Registry. Eur. J. Epidemiol. 2007, 22, 631–639. [Google Scholar] [CrossRef]
  32. Okura, Y.; Urban, L.H.; Mahoney, D.W.; Jacobsen, S.J.; Rodeheffer, R.J. Agreement between Self-Report Questionnaires and Medical Record Data Was Substantial for Diabetes, Hypertension, Myocardial Infarction and Stroke but Not for Heart Failure. J. Clin. Epidemiol. 2004, 57, 1096–1103. [Google Scholar] [CrossRef]
  33. Kim, A.; Chung, K.C.; Keir, C.; Patrick, D.L. Patient-Reported Outcomes Associated with Cancer Screening: A Systematic Review. BMC Cancer 2022, 22, 223. [Google Scholar] [CrossRef] [PubMed]
  34. Merlo, J.; Berglund, G.; Wirfält, E.; Gullberg, B.; Hedblad, B.; Manjer, J.; Hovelius, B.; Janzon, L.; Hanson, B.S.; Ostergren, P.O. Self-Administered Questionnaire Compared with a Personal Diary for Assessment of Current Use of Hormone Therapy: An Analysis of 16,060 Women. Am. J. Epidemiol. 2000, 152, 788–792. [Google Scholar] [CrossRef] [PubMed]
  35. Bolger, N.; Davis, A.; Rafaeli, E. Diary Methods: Capturing Life as It Is Lived. Annu. Rev. Psychol. 2003, 54, 579–616. [Google Scholar] [CrossRef] [PubMed]
  36. Janssens, K.A.M.; Bos, E.H.; Rosmalen, J.G.M.; Wichers, M.C.; Riese, H. A Qualitative Approach to Guide Choices for Designing a Diary Study. BMC Med. Res. Methodol. 2018, 18, 140. [Google Scholar] [CrossRef] [PubMed]
  37. Saeidzadeh, S.; Gilbertson-White, S.; Kwekkeboom, K.; Babaieasl, F.; Seaman, A. Using Online Self-Management Diaries for Qualitative Research. Int. J. Qual. Methods 2021, 20, 160940692110388. [Google Scholar] [CrossRef]
  38. Timmers, T.; Janssen, L.; Stohr, J.; Murk, J.L.; Berrevoets, M.a.H. Using eHealth to Support COVID-19 Education, Self-Assessment, and Symptom Monitoring in the Netherlands: Observational Study. JMIR Mhealth Uhealth 2020, 8, e19822. [Google Scholar] [CrossRef]
  39. Mendoza, J.; Seguin, M.L.; Lasco, G.; Palileo-Villanueva, L.M.; Amit, A.; Renedo, A.; McKee, M.; Palafox, B.; Balabanova, D. Strengths and Weaknesses of Digital Diaries as a Means to Study Patient Pathways: Experiences With a Study of Hypertension in the Philippines. Int. J. Qual. Methods 2021, 20, 16094069211002746. [Google Scholar] [CrossRef]
  40. Aiyegbusi, O.L.; Nair, D.; Peipert, J.D.; Schick-Makaroff, K.; Mucsi, I. A Narrative Review of Current Evidence Supporting the Implementation of Electronic Patient-Reported Outcome Measures in the Management of Chronic Diseases. Ther. Adv. Chronic. Dis. 2021, 12, 20406223211015958. [Google Scholar] [CrossRef]
  41. Papapetropoulos, S.S. Patient Diaries as a Clinical Endpoint in Parkinson’s Disease Clinical Trials. CNS Neurosci. Ther. 2012, 18, 380–387. [Google Scholar] [CrossRef]
  42. Broderick, J.E. Electronic Diaries: Appraisal and Current Status. Pharm. Med. 2008, 22, 69–74. [Google Scholar] [CrossRef]
  43. Piasecki, T.M.; Hufford, M.R.; Solhan, M.; Trull, T.J. Assessing Clients in Their Natural Environments with Electronic Diaries: Rationale, Benefits, Limitations, and Barriers. Psychol. Assess. 2007, 19, 25–43. [Google Scholar] [CrossRef]
  44. Lizée, T.; Basch, E.; Trémolières, P.; Voog, E.; Domont, J.; Peyraga, G.; Urban, T.; Bennouna, J.; Septans, A.-L.; Balavoine, M.; et al. Cost-Effectiveness of Web-Based Patient-Reported Outcome Surveillance in Patients with Lung Cancer. J. Thorac. Oncol. 2019, 14, 1012–1020. [Google Scholar] [CrossRef]
  45. Mlakar, I.; Šafran, V.; Hari, D.; Rojc, M.; Alankuş, G.; Pérez Luna, R.; Ariöz, U. Multilingual Conversational Systems to Drive the Collection of Patient-Reported Outcomes and Integration into Clinical Workflows. Symmetry 2021, 13, 1187. [Google Scholar] [CrossRef]
  46. Nateqi, J.; Lin, S.; Krobath, H.; Gruarin, S.; Lutz, T.; Dvorak, T.; Gruschina, A.; Ortner, R. From symptom to diagnosis-symptom checkers re-evaluated: Are symptom checkers finally sufficient and accurate to use? An update from the ENT perspective. HNO 2019, 67, 334–342. [Google Scholar] [CrossRef] [PubMed]
  47. Martin, A.; Nateqi, J.; Gruarin, S.; Munsch, N.; Abdarahmane, I.; Zobel, M.; Knapp, B. An Artificial Intelligence-Based First-Line Defence against COVID-19: Digitally Screening Citizens for Risks via a Chatbot. Sci. Rep. 2020, 10, 19012. [Google Scholar] [CrossRef] [PubMed]
  48. Munsch, N.; Martin, A.; Gruarin, S.; Nateqi, J.; Abdarahmane, I.; Weingartner-Ortner, R.; Knapp, B. Diagnostic Accuracy of Web-Based COVID-19 Symptom Checkers: Comparison Study. J. Med. Internet Res. 2020, 22, e21299. [Google Scholar] [CrossRef] [PubMed]
  49. Lin, S.; Nateqi, J.; Weingartner-Ortner, R.; Gruarin, S.; Marling, H.; Pilgram, V.; Lagler, F.B.; Aigner, E.; Martin, A.G. An Artificial Intelligence-Based Approach for Identifying Rare Disease Patients Using Retrospective Electronic Health Records Applied for Pompe Disease. Front. Neurol. 2023, 14, 1108222. [Google Scholar] [CrossRef]
  50. HAPI FHIR—The Open Source FHIR API for Java. Available online: https://hapifhir.io/ (accessed on 19 June 2021).
  51. Fox Insight Collects Online, Longitudinal Patient-Reported Outcomes and Genetic Data on Parkinson’s Disease|Scientific Data. Available online: https://www.nature.com/articles/s41597-020-0401-2 (accessed on 27 October 2023).
  52. Jones, J.B.; Snyder, C.F.; Wu, A.W. Issues in the Design of Internet-Based Systems for Collecting Patient-Reported Outcomes. Qual. Life Res. Int. J. Qual. Life Asp. Treat. Care Rehabil. 2007, 16, 1407–1417. [Google Scholar] [CrossRef] [PubMed]
  53. Frost, J.; Okun, S.; Vaughan, T.; Heywood, J.; Wicks, P. Patient-Reported Outcomes as a Source of Evidence in off-Label Prescribing: Analysis of Data from PatientsLikeMe. J. Med. Internet Res. 2011, 13, e6. [Google Scholar] [CrossRef] [PubMed]
  54. Haun, J.N.; Alman, A.C.; Melillo, C.; Standifer, M.; McMahon-Grenz, J.; Shin, M.; Lapcevic, W.A.; Patel, N.; Elwy, A.R. Using Electronic Data Collection Platforms to Assess Complementary and Integrative Health Patient-Reported Outcomes: Feasibility Project. JMIR Med. Inf. 2020, 8, e15609. [Google Scholar] [CrossRef]
  55. Penedo, F.J.; Oswald, L.B.; Kronenfeld, J.P.; Garcia, S.F.; Cella, D.; Yanez, B. The Increasing Value of eHealth in the Delivery of Patient-Centred Cancer Care. Lancet Oncol. 2020, 21, e240–e251. [Google Scholar] [CrossRef] [PubMed]
  56. Lu, D.J.; Girgis, M.; David, J.M.; Chung, E.M.; Atkins, K.M.; Kamrava, M. Evaluation of Mobile Health Applications to Track Patient-Reported Outcomes for Oncology Patients: A Systematic Review. Adv. Radiat. Oncol. 2021, 6, 100576. [Google Scholar] [CrossRef] [PubMed]
  57. Benze, G.; Nauck, F.; Alt-Epping, B.; Gianni, G.; Bauknecht, T.; Ettl, J.; Munte, A.; Kretzschmar, L.; Gaertner, J. PROutine: A Feasibility Study Assessing Surveillance of Electronic Patient Reported Outcomes and Adherence via Smartphone App in Advanced Cancer. Ann. Palliat. Med. 2019, 8, 104–111. [Google Scholar] [CrossRef] [PubMed]
  58. Werhahn, S.M.; Dathe, H.; Rottmann, T.; Franke, T.; Vahdat, D.; Hasenfuß, G.; Seidler, T. Designing Meaningful Outcome Parameters Using Mobile Technology: A New Mobile Application for Telemonitoring of Patients with Heart Failure. ESC Heart Fail 2019, 6, 516–525. [Google Scholar] [CrossRef]
  59. Juengst, S.B.; Terhorst, L.; Nabasny, A.; Wallace, T.; Weaver, J.A.; Osborne, C.L.; Burns, S.P.; Wright, B.; Wen, P.-S.; Kew, C.-L.N.; et al. Use of mHealth Technology for Patient-Reported Outcomes in Community-Dwelling Adults with Acquired Brain Injuries: A Scoping Review. Int. J. Environ. Res. Public Health 2021, 18, 2173. [Google Scholar] [CrossRef]
  60. Sato, Y.; Maruyama, T. Examining Difference between Paper- and Web-Based Self-Reported Departure/Arrival Time Using Smartphone-Based Survey. Transp. Res. Procedia 2020, 48, 1390–1400. [Google Scholar] [CrossRef]
  61. Descamps, J.; Le Hanneur, M.; Bouché, P.-A.; Boukebous, B.; Duranthon, L.-D.; Grimberg, J. Do Web-Based Follow-up Surveys Have a Better Response Rate than Traditional Paper-Based Questionnaires Following Outpatient Arthroscopic Rotator Cuff Repair? A Randomized Controlled Trial. Orthop. Traumatol. Surg. Res. 2023, 109, 103479. [Google Scholar] [CrossRef]
  62. Meirte, J.; Hellemans, N.; Anthonissen, M.; Denteneer, L.; Maertens, K.; Moortgat, P.; Van Daele, U. Benefits and Disadvantages of Electronic Patient-Reported Outcome Measures: Systematic Review. JMIR Perioper Med. 2020, 3, e15588. [Google Scholar] [CrossRef]
  63. Petracca, F.; Tempre, R.; Cucciniello, M.; Ciani, O.; Pompeo, E.; Sannino, L.; Lovato, V.; Castaman, G.; Ghirardini, A.; Tarricone, R. An Electronic Patient-Reported Outcome Mobile App for Data Collection in Type A Hemophilia: Design and Usability Study. JMIR Form Res 2021, 5, e25071. [Google Scholar] [CrossRef]
  64. Ma, D.; Orner, D.; Ghaly, M.M.; Parashar, B.; Ames, J.W.; Chen, W.C.; Potters, L.; Teckie, S. Automated Health Chats for Symptom Management of Head and Neck Cancer Patients Undergoing Radiation Therapy. Oral Oncol. 2021, 122, 105551. [Google Scholar] [CrossRef]
  65. Chaix, B.; Bibault, J.-E.; Romain, R.; Guillemassé, A.; Neeral, M.; Delamon, G.; Moussalli, J.; Brouard, B. Assessing the Performances of a Chatbot to Collect Real-Life Data of Patients Suffering from Primary Headache Disorders. Digit Health 2022, 8, 20552076221097783. [Google Scholar] [CrossRef]
  66. Bibault, J.-E.; Chaix, B.; Nectoux, P.; Pienkowski, A.; Guillemasé, A.; Brouard, B. Healthcare Ex Machina: Are Conversational Agents Ready for Prime Time in Oncology? Clin. Transl. Radiat. Oncol. 2019, 16, 55–59. [Google Scholar] [CrossRef]
  67. Te Pas, M.E.; Rutten, W.G.M.M.; Bouwman, R.A.; Buise, M.P. User Experience of a Chatbot Questionnaire Versus a Regular Computer Questionnaire: Prospective Comparative Study. JMIR Med. Inf. 2020, 8, e21982. [Google Scholar] [CrossRef]
  68. Schamber, E.M.; Takemoto, S.K.; Chenok, K.E.; Bozic, K.J. Barriers to Completion of Patient Reported Outcome Measures. J. Arthroplast. 2013, 28, 1449–1453. [Google Scholar] [CrossRef] [PubMed]
  69. Veloso Costa, A.; Padfield, O.; Elliott, S.; Hayden, P. Improving Patient Diary Use in Intensive Care: A Quality Improvement Report. J. Intensive Care Soc. 2021, 22, 27–33. [Google Scholar] [CrossRef] [PubMed]
  70. Knott, E.; Rao, A.H.; Summers, K.; Teeger, C. Interviews in the Social Sciences. Nat. Rev. Methods Primers 2022, 2, 1–15. [Google Scholar] [CrossRef]
  71. Thomas, J.A. Using Unstructured Diaries for Primary Data Collection. Nurse Res. 2015, 22, 25–29. [Google Scholar] [CrossRef] [PubMed]
  72. Ahlin, E. Semi-Structured Interviews with Expert Practitioners: Their Validity and Significant Contribution to Translational Research. In Sage Research Methods Cases Part 2; Sage Publications Ltd.: London, UK, 2019. [Google Scholar] [CrossRef]
  73. Henriksen, M.G.; Englander, M.; Nordgaard, J. Methods of Data Collection in Psychopathology: The Role of Semi-Structured, Phenomenological Interviews. Phenomenol. Cogn. Sci. 2022, 21, 9–30. [Google Scholar] [CrossRef]
  74. Kakilla, C. Strengths and Weaknesses of Semi-Structured Interviews in Qualitative Research: A Critical Essay. Preprints 2021, 2021060491. [Google Scholar] [CrossRef]
  75. DeJonckheere, M.; Vaughn, L.M. Semistructured Interviewing in Primary Care Research: A Balance of Relationship and Rigour. Fam. Med. Community Health 2019, 7, e000057. [Google Scholar] [CrossRef] [PubMed]
  76. You, Y.; Gui, X. Self-Diagnosis through AI-Enabled Chatbot-Based Symptom Checkers: User Experiences and Design Considerations. AMIA Annu. Symp. Proc. 2021, 2020, 1354–1363. [Google Scholar]
  77. Sharma, D.; Kaushal, S.; Kumar, H.; Gainder, S. Chatbots in Healthcare: Challenges, Technologies and Applications. In Proceedings of the 4th International Conference on Artificial Intelligence and Speech Technology (AIST), Delhi, India, 9–10 December 2022; pp. 1–6. [Google Scholar] [CrossRef]
  78. Bemmann, F.; Schoedel, R. Chatbots for Experience Sampling—Initial Opportunities and Challenges. In Proceedings of the Joint ACM Conference on Intelligent User Interfaces Workshops, ACMIUI-WS 2021, College Station, TX, USA, 13–17 April 2021. [Google Scholar]
  79. Xiao, Z.; Zhou, M.X.; Liao, Q.V.; Mark, G.; Chi, C.; Chen, W.; Yang, H. Tell Me About Yourself: Using an AI-Powered Chatbot to Conduct Conversational Surveys with Open-Ended Questions. ACM Trans. Comput.-Hum. Interact. 2020, 27, 1–37. [Google Scholar] [CrossRef]
  80. Jannach, D.; Manzoor, A.; Cai, W.; Chen, L. A Survey on Conversational Recommender Systems. ACM Comput. Surv. 2021, 54, 36. [Google Scholar] [CrossRef]
  81. Valtolina, S.; Barricelli, B.R.; Di Gaetano, S. Communicability of Traditional Interfaces VS Chatbots in Healthcare and Smart Home Domains. Behav. Inf. Technol. 2020, 39, 108–132. [Google Scholar] [CrossRef]
  82. Beam, E.A. Social Media as a Recruitment and Data Collection Tool: Experimental Evidence on the Relative Effectiveness of Web Surveys and Chatbots. J. Dev. Econ. 2023, 162, 103069. [Google Scholar] [CrossRef]
  83. Chaix, B.; Bibault, J.-E.; Pienkowski, A.; Delamon, G.; Guillemassé, A.; Nectoux, P.; Brouard, B. When Chatbots Meet Patients: One-Year Prospective Study of Conversations Between Patients With Breast Cancer and a Chatbot. JMIR Cancer 2019, 5, e12856. [Google Scholar] [CrossRef] [PubMed]
  84. Han, X.; Zhou, M.; Turner, M.; Yeh, T. Designing Effective Interview Chatbots: Automatic Chatbot Profiling and Design Suggestion Generation for Chatbot Debugging. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems, Yokohama, Japan, 6 May 2021; pp. 1–15. [Google Scholar]
  85. Wei, J.; Kim, S.; Jung, H.; Kim, Y.-H. Leveraging Large Language Models to Power Chatbots for Collecting User Self-Reported Data. arXiv 2023, arXiv:2301.05843. [Google Scholar] [CrossRef]
  86. Brown, T.B.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language Models Are Few-Shot Learners. Adv. Neural Inf. Process. Syst. 2020, 33, 1877–1901. [Google Scholar]
  87. Lin, B.; Cecchi, G.; Bouneffouf, D. Psychotherapy AI Companion with Reinforcement Learning Recommendations and Interpretable Policy Dynamics. In Proceedings of the Companion Proceedings of the ACM Web Conference 2023; Association for Computing Machinery: New York, NY, USA, 2023; pp. 932–939. [Google Scholar]
  88. Kasneci, E.; Seßler, K.; Küchemann, S.; Bannert, M.; Dementieva, D.; Fischer, F.; Gasser, U.; Groh, G.; Günnemann, S.; Hüllermeier, E.; et al. ChatGPT for Good? On Opportunities and Challenges of Large Language Models for Education. Learn. Individ. Differ. 2023, 103, 102274. [Google Scholar] [CrossRef]
  89. Floridi, L.; Chiriatti, M. GPT-3: Its Nature, Scope, Limits, and Consequences. Minds Mach. 2020, 30, 681–694. [Google Scholar] [CrossRef]
  90. Chowdhery, A.; Narang, S.; Devlin, J.; Bosma, M.; Mishra, G.; Roberts, A.; Barham, P.; Chung, H.W.; Sutton, C.; Gehrmann, S.; et al. PaLM: Scaling Language Modeling with Pathways. J. Mach. Learn. Res. 2023, 24, 1–113. [Google Scholar]
  91. Rae, J.W.; Borgeaud, S.; Cai, T.; Millican, K.; Hoffmann, J.; Song, F.; Aslanides, J.; Henderson, S.; Ring, R.; Young, S.; et al. Scaling Language Models: Methods, Analysis & Insights from Training Gopher. arXiv 2021, arXiv:2112.11446. [Google Scholar] [CrossRef]
  92. Teubner, T.; Flath, C.; Weinhardt, C.; Aalst, W.; Hinz, O. Welcome to the Era of ChatGPT et al.: The Prospects of Large Language Models. Bus. Inf. Syst. Eng. 2023, 65, 95–101. [Google Scholar] [CrossRef]
  93. BioMedLM: A Domain-Specific Large Language Model for Biomedical Text. Available online: https://www.mosaicml.com/blog/introducing-pubmed-gpt (accessed on 27 October 2023).
  94. Singhal, K.; Azizi, S.; Tu, T.; Mahdavi, S.S.; Wei, J.; Chung, H.W.; Scales, N.; Tanwani, A.; Cole-Lewis, H.; Pfohl, S.; et al. Large Language Models Encode Clinical Knowledge. Nature 2022, 620, 172–180. [Google Scholar] [CrossRef]
  95. Lin, B.; Bouneffouf, D.; Cecchi, G.; Varshney, K.R. Towards Healthy AI: Large Language Models Need Therapists Too. arXiv 2023, arXiv:2304.00416. [Google Scholar] [CrossRef]
  96. Yeung, J.A.; Kraljevic, Z.; Luintel, A.; Balston, A.; Idowu, E.; Dobson, R.J.; Teo, J.T. AI Chatbots Not yet Ready for Clinical Use. Front. Digit. Health 2023, 5, 60. [Google Scholar] [CrossRef] [PubMed]
  97. Omar, R.; Mangukiya, O.; Kalnis, P.; Mansour, E. ChatGPT versus Traditional Question Answering for Knowledge Graphs: Current Status and Future Directions Towards Knowledge Graph Chatbots. arXiv 2023, arXiv:2302.06466. [Google Scholar] [CrossRef]
  98. Shojania, K.G. The Frustrating Case of Incident-Reporting Systems. Qual. Saf. Health Care 2008, 17, 400–402. [Google Scholar] [CrossRef]
  99. O’Hara, J.K.; Reynolds, C.; Moore, S.; Armitage, G.; Sheard, L.; Marsh, C.; Watt, I.; Wright, J.; Lawton, R. What Can Patients Tell Us about the Quality and Safety of Hospital Care? Findings from a UK Multicentre Survey Study. BMJ Qual. Saf. 2018, 27, 673–682. [Google Scholar] [CrossRef]
  100. McIntosh, M.J.; Morse, J.M. Situating and Constructing Diversity in Semi-Structured Interviews. Glob. Qual. Nurs. Res. 2015, 2, 2333393615597674. [Google Scholar] [CrossRef]
  101. Short, C.E.; Finlay, A.; Sanders, I.; Maher, C. Development and Pilot Evaluation of a Clinic-Based mHealth App Referral Service to Support Adult Cancer Survivors Increase Their Participation in Physical Activity Using Publicly Available Mobile Apps. BMC Health Serv. Res. 2018, 18, 27. [Google Scholar] [CrossRef]
  102. Loh, K.P.; Ramsdale, E.; Culakova, E.; Mendler, J.H.; Liesveld, J.L.; O’Dwyer, K.M.; McHugh, C.; Gilles, M.; Lloyd, T.; Goodman, M.; et al. Novel mHealth App to Deliver Geriatric Assessment-Driven Interventions for Older Adults with Cancer: Pilot Feasibility and Usability Study. JMIR Cancer 2018, 4, e10296. [Google Scholar] [CrossRef] [PubMed]
  103. Moorthy, P.; Weinert, L.; Harms, B.C.; Anders, C.; Siegel, F. German Version of the mHealth App Usability Questionnaire in a Cohort of Patients with Cancer: Translation and Validation Study. JMIR Hum. Factors 2023, 10, e51090. [Google Scholar] [CrossRef]
  104. Teckie, S.; Solomon, J.; Kadapa, K.; Sanchez, K.; Orner, D.; Kraus, D.; Kamdar, D.P.; Pereira, L.; Frank, D.; Diefenbach, M. A Mobile Patient-Facing App for Tracking Patient-Reported Outcomes in Head and Neck Cancer Survivors: Single-Arm Feasibility Study. JMIR Form. Res. 2021, 5, e24667. [Google Scholar] [CrossRef] [PubMed]
  105. Paulissen, J.M.J.; Zegers, C.M.L.; Nijsten, I.R.; Reiters, P.H.C.M.; Houben, R.M.; Eekers, D.B.P.; Roelofs, E. Performance and Usability Evaluation of a Mobile Health Data Capture Application in Clinical Cancer Trials Follow-Up. Tech. Innov. Patient Support Radiat. Oncol. 2022, 24, 107–112. [Google Scholar] [CrossRef]
  106. Open Health Connect. Building Connected Health and Care Systems: Integrating Services with Actionable Data. Available online: https://www.dedalus.com/na/our-offer/products/open-health-connect/ (accessed on 10 December 2023).
  107. Šafran, V.; Hari, D.; Ariöz, U.; Mlakar, I. PERSIST Sensing Network: A Multimodal Sensing Network Architecture for Collection of Patient-Generated Health Data in The Clinical Workflow. In Proceedings of the 2021 International Conference on Electrical, Computer, Communications and Mechatronics Engineering (ICECCME), Mauritius, Mauritius, 7–8 October 2021; pp. 1–6. [Google Scholar]
  108. Rojc, M.; Mlakar, I.; Kačič, Z. The TTS-Driven Affective Embodied Conversational Agent EVA, Based on a Novel Conversational-Behavior Generation Algorithm. Eng. Appl. Artif. Intell. 2017, 57, 80–104. [Google Scholar] [CrossRef]
  109. Miner, A.S.; Laranjo, L.; Kocaballi, A.B. Chatbots in the Fight against the COVID-19 Pandemic. NPJ Digit. Med. 2020, 3, 65. [Google Scholar] [CrossRef]
  110. Li, J.; Lavrukhin, V.; Ginsburg, B.; Leary, R.; Kuchaiev, O.; Cohen, J.M.; Nguyen, H.; Gadde, R.T. Jasper: An End-to-End Convolutional Neural Acoustic Model. arXiv 2019, arXiv:1904.03288. [Google Scholar] [CrossRef]
  111. Del Rio, M.; Delworth, N.; Westerman, R.; Huang, M.; Bhandari, N.; Palakapilly, J.; McNamara, Q.; Dong, J.; Zelasko, P.; Jette, M. Earnings-21: A Practical Benchmark for ASR in the Wild. In Proceedings of the Interspeech 2021, Brno, Czech Republic, 30 August 2021; pp. 3465–3469. [Google Scholar]
  112. Heafield, K. KenLM: Faster and Smaller Language Model Queries. In Proceedings of the Sixth Workshop on Statistical Machine Translation; Association for Computational Linguistics: Edinburgh, Scotland, 2011; pp. 187–197. [Google Scholar]
  113. Ginsburg, B.; Castonguay, P.; Hrinchuk, O.; Kuchaiev, O.; Lavrukhin, V.; Leary, R.; Li, J.; Nguyen, H.; Zhang, Y.; Cohen, J.M. Stochastic Gradient Methods with Layer-Wise Adaptive Moments for Training of Deep Networks. arXiv 2019, arXiv:1905.11286. [Google Scholar] [CrossRef]
  114. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar] [CrossRef]
  115. Hori, T.; Watanabe, S.; Zhang, Y.; Chan, W. Advances in Joint CTC-Attention Based End-to-End Speech Recognition with a Deep CNN Encoder and RNN-LM. arXiv 2017, arXiv:1706.02737. [Google Scholar] [CrossRef]
  116. Picone, M.; Inoue, S.; DeFelice, C.; Naujokas, M.F.; Sinrod, J.; Cruz, V.A.; Stapleton, J.; Sinrod, E.; Diebel, S.E.; Wassman, E.R. Social Listening as a Rapid Approach to Collecting and Analyzing COVID-19 Symptoms and Disease Natural Histories Reported by Large Numbers of Individuals. Popul. Health Manag. 2020, 23, 350–360. [Google Scholar] [CrossRef]
  117. Eureka|Symptoma: The World’s Most Accurate AI Symptom Checker. Available online: https://www.eurekanetwork.org/blog/symptoma-the-world-s-most-accurate-ai-symptom-checker (accessed on 16 November 2023).
  118. Hong, N.; Wen, A.; Stone, D.J.; Tsuji, S.; Kingsbury, P.R.; Rasmussen, L.V.; Pacheco, J.A.; Adekkanattu, P.; Wang, F.; Luo, Y.; et al. Developing a FHIR-Based EHR Phenotyping Framework: A Case Study for Identification of Patients with Obesity and Multiple Comorbidities from Discharge Summaries. J. Biomed. Inform. 2019, 99, 103310. [Google Scholar] [CrossRef]
  119. The Dedalus Platform. Dedalus Ways to Digital Connect 4 Healthcare (DC4H). Available online: https://www.dedalus.com/global/en/our-offer/continuum-of-care/newdedalus-platform/ (accessed on 10 December 2023).
  120. González-Castro, L.; Cal-González, V.M.; Del Fiol, G.; López-Nores, M. CASIDE: A Data Model for Interoperable Cancer Survivorship Information Based on FHIR. J. Biomed. Inform. 2021, 124, 103953. [Google Scholar] [CrossRef]
  121. Mukhiya, S.K.; Rabbi, F.; I Pun, V.K.; Rutle, A.; Lamo, Y. A GraphQL Approach to Healthcare Information Exchange with HL7 FHIR. Procedia Comput. Sci. 2019, 160, 338–345. [Google Scholar] [CrossRef]
  122. Documents Download Module. Available online: https://ec.europa.eu/research/participants/documents/downloadPublic?documentIds=080166e5f99add65&appId=PPGMS (accessed on 24 January 2024).
  123. Alwakeel, L.; Lano, K. Functional and Technical Aspects of Self-Management mHealth Apps: Systematic App Search and Literature Review. JMIR Hum. Factors 2022, 9, e29767. [Google Scholar] [CrossRef]
  124. Cao, J.; Ganesh, A.; Cai, J.; Southwell, R.; Perkoff, E.M.; Regan, M.; Kann, K.; Martin, J.H.; Palmer, M.; D’Mello, S. A Comparative Analysis of Automatic Speech Recognition Errors in Small Group Classroom Discourse. In Proceedings of the 31st ACM Conference on User Modeling, Adaptation and Personalization; Association for Computing Machinery: New York, NY, USA, 2023; pp. 250–262. [Google Scholar]
  125. Denby, B.; Csapó, T.G.; Wand, M. Future Speech Interfaces with Sensors and Machine Intelligence. Sensors 2023, 23, 1971. [Google Scholar] [CrossRef] [PubMed]
  126. Wang, S.; Zhou, Z.; Li, B.; Li, Z.; Kan, Z. Multi-Modal Interaction with Transformers: Bridging Robots and Human with Natural Language. Robotica 2024, 42, 415–434. [Google Scholar] [CrossRef]
  127. Verbraeken, J.; Wolting, M.; Katzy, J.; Kloppenburg, J.; Verbelen, T.; Rellermeyer, J.S. A Survey on Distributed Machine Learning. ACM Comput. Surv. 2020, 53, 1–33. [Google Scholar] [CrossRef]
Figure 1. The architecture of the MRAST framework.
Figure 1. The architecture of the MRAST framework.
Sensors 24 01101 g001
Figure 2. Overall architecture of the MRAST framework.
Figure 2. Overall architecture of the MRAST framework.
Sensors 24 01101 g002
Figure 3. ASR SPREAD: an end-to-end architecture.
Figure 3. ASR SPREAD: an end-to-end architecture.
Sensors 24 01101 g003
Figure 4. From diary recording to updated patient profile.
Figure 4. From diary recording to updated patient profile.
Sensors 24 01101 g004
Figure 5. MRAST framework basic flow.
Figure 5. MRAST framework basic flow.
Sensors 24 01101 g005
Figure 6. DCD communication flow of real-world implementation.
Figure 6. DCD communication flow of real-world implementation.
Sensors 24 01101 g006
Figure 7. Refined FHIR composition resource including the extracted symptoms.
Figure 7. Refined FHIR composition resource including the extracted symptoms.
Sensors 24 01101 g007
Figure 8. Total response time of request batches between the UM REST API and SYM symptom extractor.
Figure 8. Total response time of request batches between the UM REST API and SYM symptom extractor.
Sensors 24 01101 g008
Figure 9. Response time for single request between UM REST API and SYM symptom extractor.
Figure 9. Response time for single request between UM REST API and SYM symptom extractor.
Sensors 24 01101 g009
Figure 10. RAM usage per request batch on UM REST API side.
Figure 10. RAM usage per request batch on UM REST API side.
Sensors 24 01101 g010
Figure 11. CPU usage per request batch on UM REST API side.
Figure 11. CPU usage per request batch on UM REST API side.
Sensors 24 01101 g011
Figure 12. Network traffic per request batch on UM REST API side.
Figure 12. Network traffic per request batch on UM REST API side.
Sensors 24 01101 g012
Table 1. Language parameters for testing and training.
Table 1. Language parameters for testing and training.
LanguageTraining DataTesting DataTraining TimeModel SizePlatformBatch WERTest WER
Slovenian336.74 h85.84 h55 days2.6 GBHPC GPU 2xRTX80000.0032%2.3%
Latvian782.65 h197.08 h93 days2.6 GBHPC GPU 2xRTX80002.03%0.35%
English1272.87 h319.97 h81 days2.6 GBHPC GPU 8xA1000.7%2.92%
Spanish1272.87 h319.97 h81 days2.6 GBHPC GPU 8xA1000.7%2.92%
Russian2796.00 h709.42 h145 days2.6 GBHPC GPU 6xA1009.1%2.7%
French1272.48 h335.49 h185 days2.6 GBHPC GPU 4xV1005.3%7.6%
Table 2. General description of patients.
Table 2. General description of patients.
Clinical PartnerRecruited PatientsMean AgeBreast CancerColorectal CancerMaleFemale
UL46542422739
UKCM405720201129
CHU41552120734
SERGAS395620191227
TOTAL16655858137129
Table 3. Descriptive statistics of the rate of user experience (1–10) with questionnaires.
Table 3. Descriptive statistics of the rate of user experience (1–10) with questionnaires.
FirstMiddleLast
Mean7.6007.2507.600
Median8.0008.0008.000
Std. deviation1.6352.0231.789
Minimum5.0002.0004.000
Maximum10.00010.00010.000
25th percentile6.0006.7506.000
50th percentile8.0008.0008.000
75th percentile8.2508.0009.000
Table 4. Descriptive statistics of the rate of user experience (1–10) with the mHealth app.
Table 4. Descriptive statistics of the rate of user experience (1–10) with the mHealth app.
FirstMiddleLast
Mean7.6007.3507.900
Median7.5008.0008.000
Std. deviation1.6671.8991.553
Minimum5.0003.0005.000
Maximum10.00010.00010.000
25th percentile6.0006.0007.000
50th percentile7.5008.0008.000
75th percentile9.0008.2509.000
Table 5. Descriptive statistics of the rate of user experience (1–10) with diary recordings.
Table 5. Descriptive statistics of the rate of user experience (1–10) with diary recordings.
FirstMiddleLast
Mean6.6507.0007.000
Median7.0008.0008.000
Std. deviation2.4552.7532.695
Minimum1.0001.0001.000
Maximum10.00010.00010.000
25th percentile5.7506.7506.000
50th percentile7.0008.0008.000
75th percentile8.0009.0009.000
Table 6. Comparison of PERSIST with other studies for cancer patients using mHealth apps.
Table 6. Comparison of PERSIST with other studies for cancer patients using mHealth apps.
StudyQuestionnairesPatient SamplePatient Feedback
Short et. al. [101]Self-defined10Strongly positive
Loh et. al. [102]SUS18Positive
Moorthy et. al. [103]SUS, MAUQ133Strongly positive
Teckie et. al. [104]SUS17Positive
Paulissen et. al. [105]SUS15Strongly positive
PERSIST [122] SUS, Self-defined166Strongly positive
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Šafran, V.; Lin, S.; Nateqi, J.; Martin, A.G.; Smrke, U.; Ariöz, U.; Plohl, N.; Rojc, M.; Bēma, D.; Chávez, M.; et al. Multilingual Framework for Risk Assessment and Symptom Tracking (MRAST). Sensors 2024, 24, 1101. https://doi.org/10.3390/s24041101

AMA Style

Šafran V, Lin S, Nateqi J, Martin AG, Smrke U, Ariöz U, Plohl N, Rojc M, Bēma D, Chávez M, et al. Multilingual Framework for Risk Assessment and Symptom Tracking (MRAST). Sensors. 2024; 24(4):1101. https://doi.org/10.3390/s24041101

Chicago/Turabian Style

Šafran, Valentino, Simon Lin, Jama Nateqi, Alistair G. Martin, Urška Smrke, Umut Ariöz, Nejc Plohl, Matej Rojc, Dina Bēma, Marcela Chávez, and et al. 2024. "Multilingual Framework for Risk Assessment and Symptom Tracking (MRAST)" Sensors 24, no. 4: 1101. https://doi.org/10.3390/s24041101

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop