Next Article in Journal
Blockchain-Driven Intelligent Scheme for IoT-Based Public Safety System beyond 5G Networks
Next Article in Special Issue
Research on Medical Security System Based on Zero Trust
Previous Article in Journal
Research on Orbital Angular Momentum Recognition Technology Based on a Convolutional Neural Network
Previous Article in Special Issue
Prioritization-Driven Congestion Control in Networks for the Internet of Medical Things: A Cross-Layer Proposal
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

FedSepsis: A Federated Multi-Modal Deep Learning-Based Internet of Medical Things Application for Early Detection of Sepsis from Electronic Health Records Using Raspberry Pi and Jetson Nano Devices

Department of Computer and Systems Sciences, Stockholm University, 16407 Stockholm, Sweden
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(2), 970; https://doi.org/10.3390/s23020970
Submission received: 14 December 2022 / Revised: 6 January 2023 / Accepted: 11 January 2023 / Published: 14 January 2023
(This article belongs to the Special Issue Internet of Medical Things and Smart Healthcare)

Abstract

:
The concept of the Internet of Medical Things brings a promising option to utilize various electronic health records stored in different medical devices and servers to create practical but secure clinical decision support systems. To achieve such a system, we need to focus on several aspects, most notably the usability aspect of deploying it using low-end devices. This study introduces one such application, namely FedSepsis, for the early detection of sepsis using electronic health records. We incorporate several cutting-edge deep learning techniques for the prediction and natural-language processing tasks. We also explore the multimodality aspect for the better use of electronic health records. A secure distributed machine learning mechanism is essential to building such a practical internet of medical things application. To address this, we analyze two federated learning techniques. Moreover, we use two different kinds of low-computational edge devices, namely Raspberry Pi and Jetson Nano, to address the challenges of using such a system in a practical setting and report the comparisons. We report several critical system-level information about the devices, namely CPU utilization, disk utilization, process CPU threads in use, process memory in use (non-swap), process memory available (non-swap), system memory utilization, temperature, and network traffic. We publish the prediction results with the evaluation metrics area under the receiver operating characteristic curve, the area under the precision–recall curve, and the earliness to predict sepsis in hours. Our results show that the performance is satisfactory, and with a moderate amount of devices, the federated learning setting results are similar to the single server-centric setting. Multimodality provides the best results compared to any single modality in the input features obtained from the electronic health records. Generative adversarial neural networks provide a clear superiority in handling the sparsity of electronic health records. Multimodality with the generative adversarial neural networks provides the best result: the area under the precision–recall curve is 96.55%, the area under the receiver operating characteristic curve is 99.35%, and earliness is 4.56 h. FedSepsis suggests that incorporating such a concept together with low-end computational devices could be beneficial for all the medical sector stakeholders and should be explored further.

1. Introduction

The traditional healthcare system was designed to ensure well-tested and intentioned efforts performed by expert professionals to improve or rehabilitate individuals’ physical and psychological health. Therefore, the most fundamental goal of the current medical approach is to give the resources to individuals to prevent or cure medically established diseases and disabilities [1]. The ever-expanding interconnectivity among different information generating electronic devices used by everyday people has provided us with the opportunity to help the healthcare system and medical facilities improve their overall aspects. We refer to this relatively new structural and usability paradigm shift as the Internet of Medical Things (IoMT). Healthcare professionals, patients, and other stakeholders can be interconnected through different electronic devices to exchange valuable related information with each other. Considering this, the structural (connection, communication, and information generation) and the usability (application) aspects are vital components in IoMT [2,3,4]. IoMT successfully utilizes the new wave of information explosion, the scale of inter-connectivity among electronic devices that we have never seen before, and the healthcare revolution transforming our notion of traditional healthcare systems. Effective prevention of diseases, patient-centric healthcare, real-time distance-based monitoring with automatic diagnosis support tools, improved collaboration among caregivers and patients, sustainable health and longevity, and low-cost healthcare for everyone are critical examples of IoMT applications [5]. We hope that this patient-centric approach will help reduce hospital costs, errors, and visit times. It will ensure effective distance-based management and remote monitoring of the patient’s health which will also help achieve sustainable development by reducing the urbanization trend as remote areas tend not to have better medical facilities. One of the keys to ensuring this is by employing effective clinical decision support systems to detect the disease using patient data early [6].
Clinical decision support systems (CDSSs) accommodate medical professionals in their decision-making process by providing intelligent and relevant information to diagnose, monitor, or manage patients’ health. A well-designed CDSS can enhance the overall performance by providing valid results, which should be convenient and confident in its prediction [7]. The notable examples of CDSS are disease detection/classification, sequential prediction of clinical events, concept embedding, data augmentation, and de-identification [8].
One way to ensure the excellent performance of a CDSS is by efficiently utilizing the medical data. IoMT can generate vast amounts of medical data. A crucial example of this type of data is electronic health records (EHRs). EHRs can be described as the digital records of all types of medical information of patients stored in different medical infrastructures. Therefore, it is ideal to utilize EHRs to create a robust CDSS using the IoMT setup. EHR can be extraordinarily sparse and heterogeneous, consisting of prescription notes about medicines; diagnosis history; clinical text from doctors and nurses; image-based, numeric, and categorical information from body vitals and laboratory tests, and demographic information about the patient [9,10]. This exceptional heterogeneity of the data and sensitivity of the information introduces some fascinating but challenging problems to consider before utilizing it in a CDSS. The effective utilization of EHRs should ensure improved predictability in terms of diagnosis. The heterogeneity aspect should also be handled carefully to obtain the optimum performance from the CDSS. The data should be explored with the utmost security and privacy to avoid information breaches.
Machine learning (ML) [11] is a solid choice to ensure improved predictability in CDSSs; which is a compendium of probabilistic computational methods to predict a decision using data for a particular task. Usually, several parameter values are fine-tuned to achieve such a high-accuracy prediction. These parameters are primarily specific to the data that have been used in an ML method. This effective fine-tuning of the parameters is known as ’learning the model’. Deep learning (DL) [12] is a section of ML. The input feature information is transformed in a hierarchical fashion so that a DL model can learn the representation by organising the lower-level information to a gradually higher-level one. The key idea is that this in-between transformation that bridges the map between input and output prediction should have minimal human feature engineering involvements from the input to the final prediction. It can be composed using deep neural networks, which can use a considerable amount of hyper-parameters to learn the useful features from the input data directly as a complex function of functions. It can be argued that it mimics the human learning process as we also gradually develop our knowledge over the iteration of different fundamental knowledge bases. Because of this, DL is an essential topic in the artificial intelligence domain.
Multimodality can help us deal with the heterogeneity aspect of EHRs. Multimodality investigates the data in all the associated modalities or forms. It can help us provide in-depth insights into the diagnosis of a CDSS. It can reveal new insights and discoveries. Multimodality enriches the EHR’s usage capability [13]. Typical medical diagnosis processes are multifaceted; therefore, a multi-modal approach provides a holistic method for diagnosis, resembling a real-world scenario instead of using only one modality [14]. Moreover, recent ML and DL methods are well-adjusted to working with multi-modal data to utilize this useful concept in a prediction model for CDSS [15,16].
The security and privacy issue can be handled using the federated learning (FL) concept. FL can be described as different distributed ML approaches where the data are distributed among several individual devices. The training process of these data is specific to the respective devices only. Usually, a central server is used to manage the whole training process, but it is strictly enforced that this server is never allowed to view or obtain the data. The challenging part is effectively aggregating different locally trained models to improve overall performance [17]. FL is practical, operational cost saving, secure as the ownership control can be fully employed, and private as the absolute minimum usage or exposure of the data is allowed [18,19].
Based on the discussion mentioned earlier, we need to investigate the IoMT setup to obtain improved CDSSs with the effective utilization of EHRs using the ML/DL, multimodality, and FL techniques. However, as described in Section 2, it can be observed that the empirical investigation and subsequent evaluation of such setup in the real-time low-end edge computational devices are in scarcity. The usability of the resource-constrained edge computational devices should be investigated more as this can expand the reach of IoMT to a much wider audience (different stakeholders) compared to the current high-end device setup we have. To address this issue, we have attempted to address the following research question in this study:
“What is the performance difference of the low-end edge devices in a federated IoMT-based CDSS application for the early detection of sepsis diagnosis prediction task?”
We choose early detection of sepsis as an example of an IoMT-based CDSS application because of its relevance to this setup mentioned above. We refer to this application as FedSepsis. Sepsis can be described as a potentially life-threatening condition. Sepsis is caused by the body’s overwhelming responses to infection. It is one of the major reasons for mortality and morbidity in the hospital [20]. The survival depends almost entirely on the earliest intervention by applying appropriate antimicrobial treatment [21]. The chance of mortality increases by 7.6% when the treatment is delayed for every hour [22], which means it is extremely crucial to detect it as early as possible.
In this study, we use two different kinds of low-end edge devices, namely Raspberry Pi [23] and Jetson Nano [24] to implement FedSepsis. We have used multi-modal EHR data and utilized different state of the art DL and natural-language processing (NLP) methods [25] to handle the sparsity and heterogeneity issue. Two different types of federated aggregation techniques are studied to investigate their applicability. Our previous work attempted to address the early sepsis detection in a single-server setting with single-modal data [26]. We also attempted to investigate the performance of the Raspberry Pi device in an FL setting with the basic aggregation technique [27]. FedSepsis shows that the results provide a clear superiority to using the multi-modal data over the single-modal data to detect sepsis. It also suggests that using different types of low-end edge devices can provide better performances in an FL setup. We also report the other system-level performance for different edge devices. Overall, our study indicates that IoMT setup using low-end edge computational devices for CDSSs can be a promising new prospect in the future.
The remainder of this article is organized as follows. Section 2 highlights the importance of exploring the usage of low-end edge devices by summarizing the current research on this. Section 3 provides an overview of the different methods and data processing we used for implementing FedSepsis. We also provide a detailed description of the sepsis-detection problem here. In Section 4, we discuss the experimental setup of FedSepsis application and explain the evaluation metrics used in different experiments. Section 5 reports the results of different experiments. Here, we also report critical system-level information about the Raspberry Pi and Jetson Nano devices. Finally, in Section 6, we discuss the results and their importance in a broader context.
The code implementation of FedSepsis can be obtained using the following link (accessed on 6 January 2023): https://github.com/anondo1969/FedSepsis.

2. Related Works

The primary focus of our study is to investigate the usability of the low-end edge devices in a federated IoMT-based CDSS application for CDSS tasks. In this section, we summarized the importance of low-end edge devices by analyzing the recent related works.
Kairouz et al. [17] provided an in-depth overview of the advances and open problems in FL. One key challenge identified in the platform development and deployment area is the difference in hardware among different devices. They mainly focused on the performance and device stability and mentioned that running computation must not affect it. Communication, storage, and computational capabilities of different devices are categorized as a core challenge in an FL setting by Li et al. [28]. Their primary concerns were the power (battery level) and hardware (CPU (central processing unit), memory). Scarce computational resources and relatively small storage capacity of the devices are mentioned by Bonawitz et al. [29], discussing the scaling issue of federated learning with the practical system design case. Gao et al. [30] used Raspberry Pi devices and investigated the performance of speech commands and electrocardiogram-related tasks in the FL setup. Orescanin et al. [31] investigated the fine-tuning performance in FL setting using Raspberry Pi devices using a pre-trained MobileNetV2 model on the CelebA dataset. The comparison of different edge devices was not performed in these works. The abovementioned discussions indicate the challenges to using low-end computational devices and highlight why it should be explored more in an FL setting to obtain the best performance.

3. Methods and Materials

In this section, we discuss the data and different methods used to conduct the experiments to address the research question described in Section 1 to implement FedSepsis.

3.1. Dataset

We used EHRs from the MIMIC-III (Multiparameter Intelligent Monitoring in Intensive Care) database, version 1.4 [32]. It comprises over 58,000 hospital admissions of over 45,000 patients between June 2001 and October 2012.

3.1.1. Data Selection Criteria

Patients older than 15 years were considered and the data were taken up until the first sepsis onset (see Section 3.1.2 for details), death or being discharged from the hospital. From this end point up until 48 h of continuous data were considered. We denote one such continuous event as a care episode where the total time (maximum of 48 h) is binned into one hour intervals. This interval is denoted as the time-window. Figure 1 illustrates one care episode with 48 one-hour time-windows. One should notice that the care episode length (total time) varies as can be seen in Figure 2.

3.1.2. Sepsis Definition

We used the Sepsis-3 clinical criteria [20,33] to define sepsis. With these criteria, the sepsis onset time is regarded as the latest time-window where both organ dysfunction and suspected infection criteria are satisfied. Suspected infection is fulfilled when at least two doses of antimicrobial treatment are given and any culture taking test is ordered. If the antimicrobial treatment is given first, then the culture test must be conducted within 24 h. On the other hand, if the culture test was conducted first then the antimicrobial treatment must be initiated within 72 h after the culture test. Organ dysfunction is measured by the sequential organ failure assessment (SOFA) score [34]. If the SOFA score is greater than or equal to 2 points from its baseline then the condition is denoted as organ dysfunction. Organ dysfunction is needed to be measured 48 h before to 24 h after the occurrence of the suspected infection. The baseline should be considered as the latest value measured before the 72 h time-window. If there is no known pre-existing organ dysfunction then the baseline should be regarded as zero. As the data are obtained from an intensive care unit (ICU)-based dataset, we can view the sepsis cases as a combination of community-acquired and hospital-acquired cases.

3.1.3. Feature Selection

Both the clinical-text and the non-text data were considered as features. Clinical-text data were collected as all the physician and nursing notes had no error mentioned in the database fields. The non-text data consisted of the demographic, vital, and the lab data from the EHRs. Some of the feature values can be extremely sparse as can be seen in Figure 2 (right). Therefore, we only considered the care episodes where the features have a less than 90% missingness in the non-text data. There is indeed quite a significant amount of missingness to be considered. Because of this, we researched extensively to tackle the missingness to obtain an improved result, as discussed in Section 4.1. Table 1 provides the name of all the features used in the experiments.

3.2. Federated Learning

Figure 3 illustrates the federated learning framework architecture we used. Initially the global model is created with some random weights W r from the server and then the server sends it to all the clients. It is trained and updated locally in the client devices. The trained model W r c is then sent back to the server. The server then aggregates all these models to create one global model W r + 1 = c S c S W r c . Here, c is each client, r is round, meaning the process or one cycle to create one aggregated global model. Usually there are multiple rounds to find the best trained and aggregated global model. S is the total sample size in the dataset. S c is the sample size in each client. We have used the following aggregation techniques.

3.2.1. Federated Averaging (Simple)

The aggregation technique is described as the FederatedAveraging [35] algorithm as shown in Figure 3. In this algorithm, the different locally trained models are aggregated using the weighted average of all the models.

3.2.2. Federated Optimization (Opt)

The aggregation technique is described as the FEDOPT algorithm [36], where the global model is updated in the server by applying the global optimizer (usually a gradient based one) using the average of the local models.

3.3. Machine Learning Methods

We used several state-of-the-art machine learning methods for the experiments. Recurrent neural network-based long short-term memory networks (RNN-LSTM) [37,38] were used for the sepsis prediction task. Generative adversarial imputation nets (GAINs) [39] were used for the missing non-text data imputation. Two different bidirectional encoder representations [40,41] from transformers (BERTs) [42] models were used to generate text embeddings from the clinical-text data.

3.3.1. Long Short-Term Memory Networks (RNN-LSTM)

Care episode information is time-series data which means it is sequential. To predict sepsis as early as possible from a given time-window, we are allowed to use only the information we had in the previous time-windows. A suitable DL model for this case can be recurrent neural networks (RNNs) [37] because of their capability of preserving and restoring information in its internal memory. A carefully constructed loop of all the care episodes is calculated by iterating all the time-windows in RNNs. Therefore, to make a prediction, it can use only the current and all the previous time-windows to detect sepsis. We need to be careful about two aspects of RNNs, namely, the exploding gradients problem and the vanishing gradients problem. The first problem occurs when RNNs assign an unusually high value to the model weights. The second problem can happen when this value is unreasonably low, which is relatively common in long sequences or, in our case, care episodes with a relatively high length.
Long short-term memory (LSTM) networks [38] provide solutions for these two critical problems. Its memory can be extended to learn essential and relevant information, even if it resides in long past time-window(s). Three gated cells (input, forget, and output) are used here to ensure the specific information to store or delete. The input gate is responsible for information gaining, the forget gate for detecting non-relevant information, and the output gate ensures the previous outputs are used as important ones. One could notice that it is similar to the medical diagnosis process as, in most cases, recent information is given more priority to diagnose an outcome.
Equation (1) represents the feedforward networks or multi-layer perceptron (fully connected neural networks). X i is the input.
h = f ( X i )
Usually, a hidden state is needed to preserve the previous information from a care episode. This state can be interpreted as additional features from the previous time-windows. If we want to predict whether the current stage is positive or not at a given time-window then we need to consider both the input X t and the hidden state from the previous time step h t 1 , as can be seen in the following equation:
h t = f ( x t , h t 1 )
h is the hidden states (1) composed from the previous time-windows. We can use it to (2) predict the subsequent outcome. Usually, LSTM sub-divides these two usages into two distinct variables h t and C. C symbolizes the LSTM hidden cell. Now we can control the information flow using the three gates we mentioned above,
g a t e f o r g e t = σ ( W f x X t + W f h h t 1 + b f )
g a t e i n p u t = σ ( W i x X t + W i h h t 1 + b i )
g a t e o u t = σ ( W o x X t + W o h h t 1 + b o )
If we need to update the cell state and hidden state, then we can use the following equations,
C * = tanh ( W c x X t + W c h h t 1 + b c )
C t = g a t e f o r g e t · C t 1 + g a t e i n p u t · C *
h t = g a t e f o r g e t · tanh ( C t )
σ ( x ) and tanh ( x ) are known as activation functions. σ ( x ) or sigmoid function can take any real-valued input and produces output in a range between 0 and 1,
σ ( x ) = 1 ( 1 + e x p ( x ) )
If the output sum of a sigmoid function is equal to one, then it is known as the soft-max function. Usually, a log form of soft-max function is used for prediction for multi-label classification because we can then consider the outputs as probability scores. Therefore it can be regarded as a probabilistic classification. tanh ( x ) also takes a real-valued input and provides the output in a range between [ 1 , 1 ] ,
tanh ( x ) = 2 σ ( 2 x ) 1

3.3.2. Generative Adversarial Imputation Nets (GAIN)

Generative adversarial imputation nets (GAINs) [39] are used for the imputation of the missing data. GAIN is a generalization of the GAN (generative adversarial networks) [43] architecture. Following the GAN architecture principle, in GAIN the generator attempts to impute the missing data whereas the discriminator tries to minimize the classification loss between the actual and observed imputations. In the GAIN architecture, a discriminator is assisted with additional ‘hints’ to ensure that the samples are being generated according to the real implicit data distribution.

3.3.3. Bidirectional Encoder Representations from Transformers (BERT)

BERT (bidirectional encoder representations from transformers) [42] is a deep learning model based on the transformer–encoder architecture [44] that pre-trains language representation with the use of a huge amount of data. It can be used to generate text features (text embedding) from a given text which can be used for a later classification task using different machine learning-based models (downstream tasks). BERT can also be used directly for a number of NLP classification tasks such as next sentence prediction, name–entity recognition, and question-answering. A self-attention mechanism is used in the transformer–encoder architecture where the model’s pre-training objective function is composed of mask language modeling and next sentence prediction unsupervised tasks. Stochastic optimization is used to estimate the text embeddings and model parameters. Further fine-tuning can be required to perform problem-specific downstream tasks.
We have used two such fine-tuned BERT modes for generating the text embeddings from clinical-text data. These text embeddings are later used to predict the early detection of sepsis. In this paper, we refer to these models as ClinicalBERT-Alsentzer [41] and ClinicalBERT-Huang [40]. Both ClinicalBERT-Alsentzer and ClinicalBERT-Huang used the clinical texts from the MIMIC-III v1.4 database [32].
ClinicalBERT-Alsentzer aggregates the 15 types of clinical texts into discharge summary type or non-discharge summary type data and then tokenized them using the ScispaCy [45]. These tokenized sentences were used to fine-tune the original BERT base model. ClinicalBERT-Huang additionally fine-tuned the model for the hospital-readmission prediction task. In BERT, the maximum token length is 512 which can be limited for long clinical texts. ClinicalBERT-Huang tackles it by splitting the long text and combining these multiple predictions for one global prediction in an efficient way.

4. Experimental Setup

In this section, we discuss several aspects of the experimental setup to address the research question described in Section 1.

4.1. Missingness Representation in the Care Episode

A time-window in a care episode (Figure 1) can contain multiple input-feature values or no values at all. We need to represent this missingness in an efficient way for a better prediction as our data are quite sparse (Figure 2). We have devised two methods for this missing-data imputation task based on the randomness notion. We used two strategies for the missing not-at-random case. One is using GAIN (discussed in Section 3.3.2) to generate the imputation data (GAIN-imputation). For the other case we did not impute but used an integer value which is entirely absent in the whole dataset (distinct-value). The goal is that this integer should indicate that the value is missing for a particular reason or missing not-at-random. For the missing at-random case we performed the carry-forward of the present data to all the next windows as long as we do not see another value in the forthcoming window. The later value is then again carry-forwarded if any later time-window value is missing. If the missing value for a particular input feature is presented in the whole care episode then we use the global average of that particular value across the whole data (carry-forward-mean-imputation).

4.2. Training–Testing–Tuning Data Generation

The total data are subdivided into three portions for the training (80%), testing (10%), and tuning (10%) of the prediction model. We kept both the sepsis class-based and care episode length-based distribution similar in these three datasets. Positive class is considered when the Sepsis-3 criteria is met. Because of this distribution similarity ensurity, we omitted 50 care episodes where a particular care episode length contains five or fewer cases. Table 2 provides the overview of the data. We can see that the negative episodes are almost 90% which provides us with highly-imbalanced datasets and thus makes the sepsis (positive class) prediction quite challenging.

4.3. Text Embeddings Generation from Clinical Texts

We used both ClinicalBERT-Alsentzer and ClinicalBERT-Huang (Section 3.3.3) to generate the text embeddings from the clinical text. All the clinical-texts in each care episode were tokenized first (up to 512 characters for each text). We then used it for the inference in the ClinicalBERT models. The values from the last four hidden layers of the ClinicalBERT(s) were considered to generate the feature vectors. Two approaches were taken to create the final text embeddings. First, the feature vector was generated from the average of all the layers (with a total length of (768). This is referred to as short-text-embedding. Second, the feature vector was generated from the concatenation of all the layers (with a total length of 768 ∗ 4 = 3072). This is referred to as long-text-embedding.

4.4. RNN-LSTM and GAIN Hyper-Parameter Tuning

We tuned the hyper-parameters for RNN-LSTM and GAIN models as described in Table 3. We randomly selected 100 data points among these parameters and thus conducted 100 experiments with the training data to find the best tuning parameter values. We oversampled the distribution in the training data to make the data balanced (50% positive) in each mini-batch. We used tuning data for evaluation and used AUPRC (area under the precision recall curve, described in Section 4.7) to select the best-tuned model. Based on these fine-tuned parameters we finally evaluated the trained model using the testing data.

4.5. Configuration of Server and Client Devices

For the Experiment A (described in Section 4.6.1), which is performed in a single machine we used an Ubuntu 20.04.3 LTS (GNU/Linux 5.13.0-28-generic x86_64) based server with GeForce GTX 1080 Ti graphics processing unit (GPU). For the Experiment B (described in Section 4.6.2), an Ubuntu 18.04.5 LTS (GNU/Linux 4.15.0-142-generic x86_64) based server with GeForce RTX 2080 Ti graphics processing unit (GPU) was used as the server for the federated learning experiments (Experiment B, described in Section 4.6.2). For clients, we used two different kinds of edge devices. 10 Raspberry Pi [23] 4b devices (Ubuntu 20.10 as GNU/Linux 5.8.0-1024-raspi aarch64) and three Jetson Nano [24] Developer Kit devices (Linux-4.9.253-tegra-aarch64-with-Ubuntu-18.04-bionic) were used for different experiments. Figure 4 illustrates the client edge devices used in different experiments.
Socket programming [46] was used to establish the federated learning (discussed in Section 3.2) setup. In socket programming we can connect two nodes on a network to establish a communication channel. A server usually creates the listener socket where multiple clients can connect. PyTorch [47] framework was used to develop the deep learning models.

4.6. Experiments

We conducted several experiments to evaluate the performance of early detection of sepsis using multi-modal data. We compared the performance in a federated learning based setting using two different kinds of edge devices. The experiments are classified into following categories.

4.6.1. Experiment A: Single-Server Setting

We generated the text embeddings using the process described in Section 4.3. The non-text input features (showed in Table 1) were processed in terms of missingness, as described in Section 4.1. If there were multiple values for a particular feature in a time window, then we used the average value. These clinical features were then used as the input for the RNN-LSTM model (discussed in Section 3.3.1). The hyper-parameter tuning process of the deep learning models is discussed in Section 4.4. The RNN-LSTM model predicts the output as a probability score based on the current and previous features presented in a current time window. The threshold of >0.5 was used to determine the sepsis or non-sepsis case for one time-window and once it obtains a positive prediction then the later time window prediction scores are ignored. Three different RNN-LSTM models were considered based on the difference of the input features, namely multi-modal, non-text, and clinical-text. We concatenated the text embeddings and the non-text processed features for the multi-modal representation of the input features. The mini-batch size used for training non-text, and clinical-text models was 1000, and for multi-modal model it was 500. Figure 5 illustrates the schematic diagram of the proposed model for the single-server setting.

4.6.2. Experiment B: Federated Learning Setting

The training data were split equally among all the client devices. We then performed Experiment A (Section 4.6.1) in a federated learning setting. Due to the computational limitation constrain compared to the server, the mini-batch size used for training multi-modal and clinical-text models was 500, and for non-text model it was 1000. The details about the devices and federated learning setting is discussed in Section 4.5. Both simple and opt (described in Section 3.2) aggregation techniques were used to create the global RNN-LSTM model. Due to time-constraint issue we only conducted the experiments compared to the best three models (based on AUPRC score) obtained from Experiment A (Section 4.6.1). As the training dataset is relatively small, therefore to overcome the overfitting issue we used a total of five rounds for each experiment. Along with the evaluation results we also report the system information of the client devices.

4.7. Evaluation Metrics

There are four possible outcomes for binary classification, with a positive and a negative class. If the model assigns the positive class to a positive example then it is a true positive ( T P ). If the model assigns the positive class to a negative example then it is a false positive ( F P ). If the model assigns the negative class to a negative example then it is a true negative ( T N ). If the model assigns the negative class to a positive example then it is a false negative ( F N ).
Now we can calculate the precision and recall in the following ways,
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
Area under the receiver operating characteristic curve (AUROC) and area under the precision–recall curve (AUPRC) were used to evaluate the model performance.
For a classification task with a positive and a negative class, a model will estimate the probability for an example belonging to the positive class, and if this value surpasses a threshold, the example will be classified as positive. The AUROC score is based on the true positive rate (the proportion of correct classifications of positive examples) and the false positive rate (the proportion of negative examples classified as belonging to the positive class). By varying the threshold for when an example is considered as belonging to the positive class, pairs of true positive rates and false positives rate will be obtained. These pairs can be plotted as a curve, with the false positive rate on the x-axis, and the true positive rate on they y-axis. This curve is known as the receiver operating characteristic and the AUROC value corresponds to the area under this curve.
Reviewing both precision and recall is useful in cases where there is an imbalance in the observations between the two classes. Specifically, there are many examples of no event (class 0) and only a few examples of an event (class 1). The reason for this is that typically the large number of class 0 examples means we are less interested in the skill of the model at predicting class 0 correctly, e.g., high true negatives. The key to the calculation of precision and recall is that the calculations do not make use of the true negatives. It is only concerned with the correct prediction of the minority class, class 1. A precision–recall curve is a plot of the precision (y-axis) and the recall (x-axis) for different thresholds.
The earliness in predicting sepsis is calculated based on the median prediction time from the sepsis onset in hours for all the true positive cases. A prediction score threshold of >0.5 is used to determine the correct prediction, and we retained the first positive time-window value of the care episode to calculate earliness. Therefore, the larger value of earliness represents better prediction in general. As earliness is calculated based on true positives only, we should consider it together with AUROC and AUPRC values to obtain a more balanced overview of the model’s performance.
Table 4 shows the client device system information topics that we reported in Section 5.3.

5. Results

In this section, we present the results obtained by conducting different experiments discussed in Section 4.6.

5.1. Experiment A Results: Single-Server Setting

Table 5 shows the results obtained using only the non-text data. We can see that GAIN-imputation-based missingness representation provides the best result and carry-forward-mean-imputation provides the worst result. The best evaluation scores obtained as AUPRC: 87.30%, AUROC: 98.31%, and earliness: 4.45 h.
Table 6 shows the results obtained using only the clinical-text data. We can see that the result is comparatively worse and it is quite challenging to select a better model as different evaluation metrics provides best results in different rounds. The best evaluation scores obtained as AUPRC: 14.06%, AUROC: 59.99%, and earliness: 7.28 h.
Table 7 shows the results obtained using both the clinical-text and non-text data. It can be seen that using multi-modal setting provides the best results. The best evaluation scores obtained as AUPRC: 96.55%, AUROC: 99.35%, and earliness: 4.58 h.

5.2. Experiment B Results: Federated Learning Setting

We have used two different types of edge device to conduct the experiments in federated learning setting. The results are reported here.

5.2.1. Raspberry Pi

Table 8 shows the results obtained using non-text data in a federated learning setting using 10 Raspberry Pi devices. We can see that federated aggregation technique simple performed better over opt. The best evaluation scores obtained as AUPRC: 71.39%, AUROC: 97.60%, and earliness: 4.54 h.
Table 9 shows the results obtained using clinical-text data in a federated learning setting using 10 Raspberry Pi devices. Although the result is comparatively worse, but similar to before the result improves after each round and aggregation technique simple performed better over opt. The best evaluation scores obtained as AUPRC: 14.62%, AUROC: 58.91%, and earliness: 7.73 h.
Table 10 shows the results obtained using multi-modal data in a federated learning setting using 10 Raspberry Pi devices. multi-modal model provides the best results. It is worth noting that the best AUPRC score was obtained after the round 4. The best evaluation scores obtained as AUPRC: 84.80%, AUROC: 98.49%, and earliness: 4.54 h.
Figure 6 provides the loss comparison for different federated aggregation techniques using 10 Raspberry Pi devices for the training of the multi-modal model. We can see that the loss value gradually decreased after each round and it is much smoother in the case of the aggregation technique simple over opt.
Table 11 shows the results obtained using multi-modal data in a federated learning setting using different number of Raspberry Pi devices. It provides insights into the empirical settings regarding data and the number of devices. We can see that the best results were obtained using only one or two devices, and after increasing it to six and more devices the AUPRC score decreased. This suggests that using a relatively small amount of data can increase the chance of overfitting, and we have to be careful in that case. Choosing less number of devices may be a walk around in that case. The best evaluation scores obtained using two devices as AUPRC: 99.20%, and AUROC: 99.91%, and using eight devices the earliness score is 4.53 h.

5.2.2. Jetson Nano

Table 12 shows the results obtained using non-text data in a federated learning setting using three Jetson Nano devices. Similar to the Raspberry Pi case, federated Aggregation Technique simple provided the best results. We can also see that fewer devices provide better results than using more. The best evaluation scores obtained as AUPRC: 94.43%, AUROC: 99.51%, and earliness: 4.53 h.
Table 13 shows the results obtained using clinical-text data in a federated learning setting using three Jetson Nano devices. We can observe that the performance is significantly worse, similar to the Raspberry Pi cases. It suggests that it may be possible that the text modality alone sometimes may not be a wise choice as input features. The best evaluation scores obtained as AUPRC: 15.12%, AUROC: 63.58%, and earliness: 7.73 h.
Table 14 shows the results obtained using multi-modal data in a federated learning setting using three Jetson Nano devices. Once again, the multi-modal representation of the input features provided the best results. Compared to Table 11 the results with three devices provides a better AUPRC score. The best evaluation scores obtained as AUPRC: 98.99%, AUROC: 99.89%, and earliness: 4.56 h.
Figure 7 provides the loss comparison for different federated aggregation techniques using three Jetson Nano devices for the training of multi-modal model. We can see that the loss value was gradually decreased after each round and it is much smoother in the case of the aggregation technique simple over opt.
Table 15 shows the results obtained using multi-modal data in a federated learning setting using different Jetson Nano devices. As it can be observed, with two devices, the best AUPRC score was obtained. The best evaluation scores obtained as AUPRC: 99.36%, AUROC: 99.93% with two devices, and earliness: 4.51 h with three devices. The best performance here is slightly worse than the best performance obtained with the similar experiments using Raspberry Pi devices, as can be compared with Table 11.

5.3. Edge Device System Information

In this section, we report different edge device system information as mentioned in Table 4. For the Raspberry Pi, the experiment was conducted with 10 devices (see Table 10), and for the Jetson Nano, three devices (see Table 14) were used. For both cases, multi-modal input features and simple federated aggregation method were used.
Figure 8 provides the CPU utilization comparison for different Raspberry Pi and Jetson Nano devices. We can see the precise segmentation of each round (of a total of five). Raspberry Pi’s CPU utilization is much higher, and it is consistent across all 10 devices. This is understandable as Jetson Nano uses GPU. Therefore, the majority of computational works are performed there. This indicates that we need to be careful if we plan to run other resource-heavy computations on both devices, mainly Raspberry Pi.
Figure 9 provides the disk utilization comparison for different Raspberry Pi and Jetson Nano devices. The total storage memory capacity for 8 of the 10 Raspberry Pi devices was 40 Gigabytes (GB). The rest had a capacity of 128 GB. All three Jetson Nano devices had a storage memory capacity of 128 GB. We can see that the disk utilization is also consistent in all the relatively low-storage memory devices. Even with a 40 GB total capacity, the disk utilization is below 60% for the Raspberry Pi devices. This suggests that we need to optimize the disk storage given the dataset size.
Figure 10 provides the comparison regarding the number of ’CPU process threads in use’ for different Raspberry Pi and Jetson Nano devices. We can see that it is entirely consistent across all the devices, and in almost all the cases, it is close to 12. It suggests that if we can effectively utilize all the cores of the CPU, then we may obtain a faster and better parallel performance, as can be seen in these graphs.
Figure 11 provides the comparison of process memory in use for different Raspberry Pi and Jetson Nano devices. Process memory is the collection of processes responsible for obtaining, preserving, and retrieving information. We can see that process memory utilization is relatively low in Raspberry Pi devices, and it is even much less (with a bit of irregularity) in Jetson Nano devices.
Figure 12 provides the process memory available comparison for different Raspberry Pi and Jetson Nano devices. If we observe this figure jointly with Figure 11 then we can see that Raspberry Pi uses significantly less process memory compared to Jetson Nano devices. Even with the small dataset and relatively small DL models used in FedSepsis experiments, it takes almost all the process memory in Jetson Nano. Therefore, Raspberry Pi is suited for more data and more complex training models.
Figure 13 provides the system memory utilization comparison for different Raspberry Pi and Jetson Nano devices. System memory or random access memory (RAM) stores data and instructions temporarily so that the CPU can use it later by accessing it randomly. Similar to process memory, system memory utilization is also much higher in Jetson Nano devices than Raspberry Pi devices. We need to be very cautious about the dataset size (mini-batch size) and the complexity of the DL training model while utilizing Jetson Nano devices.
Figure 14 provides the network traffic comparison for different Raspberry Pi and Jetson Nano devices. The data-size pattern is very similar as we exchanged the same types of information for every round. If the model is more complex, the data size will be bigger. Therefore, network usage should be carefully constructed and monitored.
Figure 15 provides the temperature comparison for different Raspberry Pi and Jetson Nano devices. We can see that the temperature obtains significantly high for Raspberry Pi devices. As we used a relatively small dataset and simple DL training model there, on average, it took 40 to 50 min to complete one multi-modal training. Therefore, we could provide sufficient time-gap to cool down the devices. In a more practical scenario, using a cooling fan is advisable due to this high temperature during the training process.

6. Discussion

In this study, we introduced FedSepsis, an FL-based IoMT application for the early prediction of sepsis using DL and EHRs. We utilized two recent federated aggregation algorithms for the FL setup. We incorporated several cutting-edge concepts and methods, such as multimodality, ClinicalBERT for the NLP, RNN-LSTM neural networks, and GAN to implement FedSepsis. Moreover, to investigate the aspects of practical usability, we deployed two popular low-end edge devices, namely Raspberry Pi and Jetson Nano. To provide a more balanced and multifaceted evaluation, we presented the results using AUPRC, AUROC, and earliness in terms of hours. We also reported system-level information on the low-end devices to provide insights into using such devices in a practical setting. Instead of focusing on the model improvement only, we also emphasized the applicability aspect by integrating cutting-edge techniques with a keen interest in low-edge devices. Therefore, we heavily invested in analyzing the system-level aspects of these low-edge devices. We expect that by taking this combined approach, our FedSepsis application will be more effective and ensure far-reaching benefits.
We used genuine (de-identified) EHRs to extract the input features. EHRs are relatively sparse, or the missingness in EHR is generally quite high, which was in our case also. To tackle this issue, we used a GAN-based technique which proved to be successful as the best results were obtained based on this technique to handle missingness.
Representing the clinical text from the EHRs for prediction tasks is one of the most challenging problems in the NLP domain. To investigate this issue in-depth, we used two latest ClinicalBERT models and two different representations from each. The comparison in terms of best performance is inconclusive as both models performed relatively well along with different representations. This study concatenated the text representations (text embeddings) with the non-text input features as a combined multi-modal input feature. In the future, we will explore different training combinations to attach it to different layers of RNN-LSTM.
Incorporating multimodality provided a precise and superior performance over any form of single-modal input features. It also supports the real-time diagnosis procedure and suggests that in future, we should investigate further in utilizing this concept more efficiently. Effective utilization of the image data is one possible option here. The overall optimization in the training model should also be fine-tuned according to different modalities, as we can see that different modalities may carry different degrees of importance in terms of the overall prediction.
Compared to the FL setting we used, the single-server setting’s storage (one terabyte) and computational capacity were considerably superior to the FedSepsis application performance. Therefore, a bit of degrading performance in the FL setup should be taken for granted compared to the single-server setup. We also showed that the performance is close to the single-server setup for this task with a smaller number of devices (five, for example). It suggests that we need to be careful about the device choice and the data distribution in an FL setting not to overfit the model. The model weight aggregation techniques also need to be explored further so that the effective loss optimization can be fine-tuned according to the model or the data we have for a particular task.
For the early sepsis prediction task, we reported AUROC, AUPRC, and earliness prediction in hours. AUROC and AUPRC provide a more balanced and clinically relevant evaluation of the prediction as we should be more careful about the true positives. Overall the AUPRC score is satisfactory. However, it should be considered together with earliness to understand the effectiveness of the prediction. On average, the earliness is close to 4.5 h which is satisfactory given the sparsity of the data and the variance in the care episode lengths.
The key aim of this study is to evaluate the performance difference of the low-end edge devices as the research question mentioned in Section 1. We can categorize this performance into model-prediction-level and system-level perspectives.
There is a clear performance difference in model prediction while incorporating GAIN for handling missingness. On average, the AUPRC improved 23.17%, and the AUROC improved 6.89% compared to the next imputation strategy across all non-text, clinical-text, and multi-modal cases in the single-server setting. The combination of clinical-text with non-text or the multi-modal case always provided superior performance in all three settings (single-server, Raspberry Pi, and Jetson Nano). In the single-server setting, for the multi-modal case, the AUPRC improvement is 9.25%, and the AUROC improvement is 1.04%, compared to the non-text modality. Using Raspberry Pi, the AUPRC improvement is 11.06%, and the AUROC improvement is 0.89%. Using Jetson Nano the AUPRC improvement is 4.56%, and the AUROC improvement is 0.39%. The federated aggregation technique (discussed in Section 3.2) simple performed slightly better than the technique opt. For the multi-modal case, compared to opt technique, using Raspberry Pi the AUPRC improvement is 1.86%, and the AUROC improvement is 0.43% in using the simple technique. Using Jetson Nano, the AUPRC improvement is 3.99%, and the AUROC improvement is 0.54%. Overall, Jetson Nano provided better performance than Raspberry Pi. For the multi-modal case, compared to Raspberry Pi, using Jetson Nano, the AUPRC improvement is 16.55%, and the AUROC improvement is 1.40%. Using fewer devices (three) compared to Raspberry Pi (10) and the GPU could be one reason for this improvement. It is worth noting that different combinations of ClinicalBERT-based text embedding representation (discussed in Section 4.3) did not provide significant performance differences. It should be investigated further in the future.
Section 5.3 provides an in-depth overview of the performance difference from a system-level perspective. For a faster implementation, Jetson Nano can be a viable choice; on the other hand, Raspberry Pi is less memory exhaustive, given that the temperature issue is effectively handled. Effective CPU utilization can also be achieved by exploring the feature importance issue instead of using all the features. Reducing unimportant features could enhance the speed and efficient CPU utilization. The device system information provides valuable insights regarding the practical usability of FedSepsis. CPU utilization, process memory, and system memory are crucial to ensure optimum performance in these low-computational devices, and, therefore, the training model or the data to create input features should be fine-tuned accordingly. Device temperature is also a critical issue to consider to ensure that it runs effectively. Network traffic could be vital, and effective allocation needs to be ensured, especially when we have low power and lossy network (LLN), typical for such IoMT infrastructures.
EHRs are either stored or generated in different medical equipment and servers across different hospital physical locations. We should be very cautious about combining or storing these precious and copious amounts of data in a single place from cost–benefit and security–privacy aspects. As CDSSs can be built with these EHRs, which can substantially positively impact the healthcare system, effective interconnection among the medical equipment and servers should be paramount. IoMT can be one excellent solution for this distributed platform. It can be expanded further with a minimal additional interconnection cost to connect it with other hospitals, even across different countries, while ensuring the utmost security and privacy to create a global impact. This has the potential to revolutionize the healthcare industry with a long-lasting impact to help every stakeholder.
We must be cautious about the security and privacy issue as the EHRs are extraordinarily vulnerable, and breaching the information could bring disastrous after-effects. Therefore, the IoMT platform and applications should ensure the reliability, security, and safety aspects as much as possible [48]. Advanced level data intelligence, supported by cutting-edge DL and ML techniques, should be deployed to enable the IoMT platform and applications to be interactive, stateful, iterative, adaptive, and contextual. Along with NLP, we should also focus on other cross-disciplinary domains to make it more effective, such as speech recognition [49], computer vision [50], dialogue and narrative generation, and human–computer interaction.
Effective data aggregation with the help of metadata aggregation [51], and medical sensors [52] can provide us with new insights. As the inter-relations of different medical conditions or diseases are ubiquitous phenomena among patients (comorbidity), incorporating this aspect (for example, general infection detection from various input sources and outcomes) could be an exciting option to explore further. The optimum use of the low-end edge devices will ensure that these new generation computer systems are cost-effective, patient-centric, robust, and are here to create a long-lasting impact, which will help all parties involved, including the patients and the healthcare professionals.
One potential limitation regarding the acceptance of FedSepsis as a CDSS to the medical community is the lack of interpretability of the black box model. The medical significance should be investigated [53] as well while predicting the model to make it more acceptable. We can denote the medical significance of a symptom or disease process as a type of clinical manifestation affecting the regular functionalities of organ systems, tissues, living cells, or organs, for example, in the chest area [54]. We will explore this issue in the future.

7. Conclusions

In this study, we presented an Internet of Medical Things-based application, namely FedSepsis, for early detection of sepsis as an example of a promising clinical decision support system. We used cutting-edge deep learning methods and relevant concepts, such as multimodality, natural-language processing and federated learning to make a robust and practical application. Moreover, we utilized low-end edge devices, such as Raspberry Pi and Jetson Nano, to show the usability aspects of such applications. Multimodality incorporating clinical-text data provided improved performance. Jetson Nano yielded better results from both the prediction-level and system-level perspectives. The results suggest that FedSepsis is viable to use and can be an ideal choice given that the interpretability aspect is also ensured to a degree as the prediction scores are entirely satisfactory and the possibility of relatively low-cost device usage can benefit all the stakeholders. In the future, sensor integration, interpreting the medical significance of the prediction, and enhancing the security aspects will be explored.

Author Contributions

Conceptualization, M.U.A.; data curation, M.U.A.; investigation, M.U.A.; methodology, M.U.A.; resources, R.R.; supervision, R.R.; writing—original draft, M.U.A.; writing—review & editing, M.U.A. and R.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

MIMIC-III (Multiparameter Intelligent Monitoring in Intensive Care) database, version 1.4 [32].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Plsek, P.E.; Greenhalgh, T. The challenge of complexity in health care. Bmj 2001, 323, 625–628. [Google Scholar] [CrossRef] [PubMed]
  2. Irfan, M.; Ahmad, N. Internet of Medical Things: Architectural model, motivational factors and impediments. In Proceedings of the 2018 15th Learning and Technology Conference (L&T), Jeddah, Saudi Arabia, 25–26 February 2018; pp. 6–13. [Google Scholar]
  3. Joyia, G.J.; Liaqat, R.M.; Farooq, A.; Rehman, S. Internet of Medical Things (IoMT): Applications, benefits and future challenges in healthcare domain. J. Commun. 2017, 12, 240–247. [Google Scholar] [CrossRef] [Green Version]
  4. Manogaran, G.; Chilamkurti, N.; Hsu, C.H. Emerging trends, issues, and challenges in Internet of Medical Things and wireless networks. Pers. Ubiquitous Comput. 2018, 22, 879–882. [Google Scholar] [CrossRef] [Green Version]
  5. Goyal, S.; Sharma, N.; Bhushan, B.; Shankar, A.; Sagayam, M. Iot enabled technology in secured healthcare: Applications, challenges and future directions. In Cognitive Internet of Medical Things for Smart Healthcare; Springer International Publishing: Cham, Switzerland, 2021; pp. 25–48. [Google Scholar]
  6. Kukhtevich, I.; Goryunova, V.; Goryunova, T.; Zhilyaev, P. Medical Decision Support Systems and Semantic Technologies in Healthcare. In Proceedings of the Russian Conference on Digital Economy and Knowledge Management (RuDEcK 2020), Voronezh, Russia, 27–29 February 2020; Atlantis Press: Amsterdam, The Netherlands, 2020; pp. 370–375. [Google Scholar] [CrossRef]
  7. Souza-Pereira, L.; Ouhbi, S.; Pombo, N. Quality-in-use characteristics for clinical decision support system assessment. Comput. Methods Programs Biomed. 2021, 207, 106169. [Google Scholar] [CrossRef] [PubMed]
  8. Xiao, C.; Choi, E.; Sun, J. Opportunities and challenges in developing deep learning models using electronic health records data: A systematic review. J. Am. Med. Inform. Assoc. 2018, 25, 1419–1428. [Google Scholar] [CrossRef] [Green Version]
  9. Müller, H.; Unay, D. Retrieval From and Understanding of Large-Scale Multi-modal Medical Datasets: A Review. IEEE Trans. Multimed. 2017, 19, 2093–2104. [Google Scholar] [CrossRef] [Green Version]
  10. Itani, S.; Lecron, F.; Fortemps, P. Specifics of medical data mining for diagnosis aid: A survey. Expert Syst. Appl. 2019, 118, 300–314. [Google Scholar] [CrossRef]
  11. Molnar, C. Interpretable Machine Learning: A Guide for Making Black Box Models Explainable. 2022. Available online: https://christophm.github.io/interpretable-ml-book/ (accessed on 6 January 2023).
  12. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
  13. Cai, Q.; Wang, H.; Li, Z.; Liu, X. A survey on multimodal data-driven smart healthcare systems: Approaches and applications. IEEE Access 2019, 7, 133583–133599. [Google Scholar] [CrossRef]
  14. Huang, S.C.; Pareek, A.; Seyyedi, S.; Banerjee, I.; Lungren, M.P. Fusion of medical imaging and electronic health records using deep learning: A systematic review and implementation guidelines. NPJ Digit. Med. 2020, 3, 1–9. [Google Scholar] [CrossRef]
  15. Baltrušaitis, T.; Ahuja, C.; Morency, L.P. Multimodal machine learning: A survey and taxonomy. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 423–443. [Google Scholar] [CrossRef] [Green Version]
  16. Ramachandram, D.; Taylor, G.W. Deep multimodal learning: A survey on recent advances and trends. IEEE Signal Process. Mag. 2017, 34, 96–108. [Google Scholar] [CrossRef]
  17. Kairouz, P.; McMahan, H.B.; Avent, B.; Bellet, A.; Bennis, M.; Bhagoji, A.N.; Bonawitz, K.; Charles, Z.; Cormode, G.; Cummings, R.; et al. Advances and open problems in federated learning. arXiv 2019, arXiv:1912.04977. [Google Scholar]
  18. Briggs, C.; Fan, Z.; Andras, P. A Review of Privacy Preserving Federated Learning for Private IoT Analytics. arXiv 2020, arXiv:2004.11794. [Google Scholar]
  19. Xu, J.; Glicksberg, B.S.; Su, C.; Walker, P.; Bian, J.; Wang, F. Federated learning for healthcare informatics. J. Healthc. Inform. Res. 2021, 5, 1–19. [Google Scholar] [CrossRef]
  20. Singer, M.; Deutschman, C.S.; Seymour, C.W.; Shankar-Hari, M.; Annane, D.; Bauer, M.; Bellomo, R.; Bernard, G.R.; Chiche, J.D.; Coopersmith, C.M.; et al. The third international consensus definitions for sepsis and septic shock (Sepsis-3). JAMA 2016, 315, 801–810. [Google Scholar] [CrossRef]
  21. Ferrer, R.; Martin-Loeches, I.; Phillips, G.; Osborn, T.M.; Townsend, S.; Dellinger, R.P.; Artigas, A.; Schorr, C.; Levy, M.M. Empiric antibiotic treatment reduces mortality in severe sepsis and septic shock from the first hour: Results from a guideline-based performance improvement program. Crit. Care Med. 2014, 42, 1749–1755. [Google Scholar] [CrossRef] [Green Version]
  22. Kumar, A.; Roberts, D.; Wood, K.E.; Light, B.; Parrillo, J.E.; Sharma, S.; Suppes, R.; Feinstein, D.; Zanotti, S.; Taiberg, L.; et al. Duration of hypotension before initiation of effective antimicrobial therapy is the critical determinant of survival in human septic shock. Crit. Care Med. 2006, 34, 1589–1596. [Google Scholar] [CrossRef]
  23. Johnston, S.J.; Cox, S.J. The Raspberry Pi: A Technology Disrupter, and the Enabler of Dreams. Electronics 2017, 6, 51. [Google Scholar] [CrossRef] [Green Version]
  24. Cass, S. Nvidia makes it easy to embed AI: The Jetson nano packs a lot of machine-learning power into DIY projects-[Hands on]. IEEE Spectr. 2020, 57, 14–16. [Google Scholar] [CrossRef]
  25. Otter, D.W.; Medina, J.R.; Kalita, J.K. A survey of the usages of deep learning for natural language processing. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 604–624. [Google Scholar] [CrossRef] [Green Version]
  26. Alam, M.U.; Rahmani, R. Cognitive Internet of Medical Things Architecture for Decision Support Tool to Detect Early Sepsis Using Deep Learning. In Proceedings of the Biomedical Engineering Systems and Technologies, BIOSTEC 2020, Valletta, Malta, 24–26 February 2020; Revised Selected Papers. Ye, X., Soares, F., De Maria, E., Gómez Vilda, P., Cabitza, F., Fred, A., Gamboa, H., Eds.; Springer International Publishing: Cham, Switzerland, 2021; pp. 366–384. [Google Scholar]
  27. Alam, M.U.; Rahmani, R. Federated Semi-Supervised Multi-Task Learning to Detect COVID-19 and Lungs Segmentation Marking Using Chest Radiography Images and Raspberry Pi Devices: An Internet of Medical Things Application. Sensors 2021, 21, 5025. [Google Scholar] [CrossRef] [PubMed]
  28. Li, T.; Sahu, A.K.; Talwalkar, A.; Smith, V. Federated Learning: Challenges, Methods, and Future Directions. IEEE Signal Process. Mag. 2020, 37, 50–60. [Google Scholar] [CrossRef]
  29. Bonawitz, K.; Eichner, H.; Grieskamp, W.; Huba, D.; Ingerman, A.; Ivanov, V.; Kiddon, C.; Konečný, J.; Mazzocchi, S.; McMahan, H.B.; et al. Towards Federated Learning at Scale: System Design. Proc. Mach. Learn. Syst. 2019, 1, 374–388. [Google Scholar]
  30. Gao, Y.; Kim, M.; Abuadbba, S.; Kim, Y.; Thapa, C.; Kim, K.; Camtepe, S.A.; Kim, H.; Nepal, S. End-to-End Evaluation of Federated Learning and Split Learning for Internet of Things. arXiv 2020, arXiv:2003.13376. [Google Scholar]
  31. Orescanin, M.; Ergezer, M.; Singh, G.; Baxter, M. Federated Fine-Tuning Performance on Edge Devices. In Proceedings of the 2021 20th IEEE International Conference on Machine Learning and Applications (ICMLA), Online, 13–15 December 2021; pp. 1174–1181. [Google Scholar] [CrossRef]
  32. Johnson, A.E.; Stone, D.J.; Celi, L.A.; Pollard, T.J. The MIMIC Code Repository: Enabling reproducibility in critical care research. J. Am. Med. Inform. Assoc. 2018, 25, 32–39. [Google Scholar] [CrossRef] [Green Version]
  33. Seymour, C.W.; Liu, V.X.; Iwashyna, T.J.; Brunkhorst, F.M.; Rea, T.D.; Scherag, A.; Rubenfeld, G.; Kahn, J.M.; Shankar-Hari, M.; Singer, M.; et al. Assessment of clinical criteria for sepsis: For the Third International Consensus Definitions for Sepsis and Septic Shock (Sepsis-3). JAMA 2016, 315, 762–774. [Google Scholar] [CrossRef] [Green Version]
  34. Vincent, J.L.; Moreno, R.; Takala, J.; Willatts, S.; De Mendonça, A.; Bruining, H.; Reinhart, C.; Suter, P.; Thijs, L. The SOFA (Sepsis-related Organ Failure Assessment) score to describe organ dysfunction/failure. Intensive Care Med. 1996, 22, 707–710. [Google Scholar] [CrossRef]
  35. McMahan, B.; Moore, E.; Ramage, D.; Hampson, S.; y Arcas, B.A. Communication-efficient learning of deep networks from decentralized data. In Proceedings of the Artificial Intelligence and Statistics, PMLR, Fort Lauderdale, FL, USA, 20–22 April 2017; pp. 1273–1282. [Google Scholar]
  36. Reddi, S.; Charles, Z.; Zaheer, M.; Garrett, Z.; Rush, K.; Konečnỳ, J.; Kumar, S.; McMahan, H.B. Adaptive federated optimization. arXiv 2020, arXiv:2003.00295. [Google Scholar]
  37. Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning representations by back-propagating errors. Nature 1986, 323, 533–536. [Google Scholar] [CrossRef]
  38. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  39. Yoon, J.; Jordon, J.; Schaar, M. Gain: Missing data imputation using generative adversarial nets. In Proceedings of the International Conference on Machine Learning, PMLR, Baltimore, MD, USA, 17–23 July 2018; pp. 5689–5698. [Google Scholar]
  40. Huang, K.; Altosaar, J.; Ranganath, R. ClinicalBERT: Modeling Clinical Notes and Predicting Hospital Readmission. arXiv 2019, arXiv:1904.05342. [Google Scholar]
  41. Alsentzer, E.; Murphy, J.; Boag, W.; Weng, W.H.; Jin, D.; Naumann, T.; McDermott, M. Publicly Available Clinical BERT Embeddings. In Proceedings of the 2nd Clinical Natural Language Processing Workshop, Minneapolis, MN, USA, 7–10 June 2019; pp. 72–78. [Google Scholar] [CrossRef] [Green Version]
  42. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  43. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. Adv. Neural Inf. Process. Syst. 2014, 27, 139–144. [Google Scholar]
  44. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 5998–6008. [Google Scholar]
  45. Neumann, M.; King, D.; Beltagy, I.; Ammar, W. ScispaCy: Fast and robust models for biomedical natural language processing. arXiv 2019, arXiv:1902.07669. [Google Scholar]
  46. Goerzen, J. Foundations of Python Network Programming; Apress: New York, NY, USA, 2004. [Google Scholar]
  47. Paszke, A.; Gross, S.; Chintala, S.; Chanan, G.; Yang, E.; DeVito, Z.; Lin, Z.; Desmaison, A.; Antiga, L.; Lerer, A. Automatic differentiation in PyTorch. In Proceedings of the NIPS-W 2017, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  48. Hatzivasilis, G.; Soultatos, O.; Ioannidis, S.; Verikoukis, C.; Demetriou, G.; Tsatsoulis, C. Review of security and privacy for the Internet of Medical Things (IoMT). In Proceedings of the 2019 15th International Conference on Distributed Computing in Sensor Systems (DCOSS), Santorini, Greece, 29–31 May 2019; pp. 457–464. [Google Scholar]
  49. Malik, M.; Malik, M.K.; Mehmood, K.; Makhdoom, I. Automatic speech recognition: A survey. Multimed. Tools Appl. 2021, 80, 9411–9457. [Google Scholar] [CrossRef]
  50. Goel, A.; Tung, C.; Lu, Y.H.; Thiruvathukal, G.K. A survey of methods for low-power deep learning and computer vision. In Proceedings of the 2020 IEEE 6th World Forum on Internet of Things (WF-IoT), New Orleans, LA, USA, 2–16 June 2020; pp. 1–6. [Google Scholar]
  51. Alam, M.U.; Rahmani, R. Intelligent context-based healthcare metadata aggregator in internet of medical things platform. Procedia Comput. Sci. 2020, 175, 411–418. [Google Scholar] [CrossRef]
  52. Ray, P.P.; Dash, D.; Kumar, N. Sensors for Internet of Medical Things: State-of-the-art, security and privacy issues, challenges and future directions. Comput. Commun. 2020, 160, 111–131. [Google Scholar] [CrossRef]
  53. Alam, M.U.; Baldvinsson, J.R.; Wang, Y. Exploring LRP and Grad-CAM visualization to interpret multi-label-multi-class pathology prediction using chest radiography. In Proceedings of the 2022 IEEE 35th International Symposium on Computer-Based Medical Systems (CBMS), Shenzhen, China, 21–22 July 2022; pp. 258–263. [Google Scholar] [CrossRef]
  54. Goodman, L.R. Felson’s Principles of Chest Roentgenology, a Programmed Text; Elsevier Health Sciences: Amsterdam, The Netherlands, 2014. [Google Scholar]
Figure 1. A care episode having with 48 time-windows (care episode length = 48).
Figure 1. A care episode having with 48 time-windows (care episode length = 48).
Sensors 23 00970 g001
Figure 2. Histogram of different care episode lengths (left) and missing rates (right).
Figure 2. Histogram of different care episode lengths (left) and missing rates (right).
Sensors 23 00970 g002
Figure 3. Federated learning architecture.
Figure 3. Federated learning architecture.
Sensors 23 00970 g003
Figure 4. Raspberry Pi (left) and Jetson Nano (right) edge devices used in the experiments.
Figure 4. Raspberry Pi (left) and Jetson Nano (right) edge devices used in the experiments.
Sensors 23 00970 g004
Figure 5. Schematic diagram of the proposed model for the single-server setting. For details please see Section 4.6.1.
Figure 5. Schematic diagram of the proposed model for the single-server setting. For details please see Section 4.6.1.
Sensors 23 00970 g005
Figure 6. Loss comparison for different federated aggregation techniques using 10 Raspberry Pi devices. The x-axis represents the incremental number of training mini-batches, the y-axis represents loss-values.
Figure 6. Loss comparison for different federated aggregation techniques using 10 Raspberry Pi devices. The x-axis represents the incremental number of training mini-batches, the y-axis represents loss-values.
Sensors 23 00970 g006
Figure 7. Loss comparison for different federated aggregation methods using Jetson Nano devices. x-axis represents the incremental number of training mini-batches, y-axis represents loss-values.
Figure 7. Loss comparison for different federated aggregation methods using Jetson Nano devices. x-axis represents the incremental number of training mini-batches, y-axis represents loss-values.
Sensors 23 00970 g007
Figure 8. CPU utilization comparison for different Raspberry Pi and Jetson Nano devices. The x-axis represents the time in minutes, the y-axis represents the CPU utilization percentage (%).
Figure 8. CPU utilization comparison for different Raspberry Pi and Jetson Nano devices. The x-axis represents the time in minutes, the y-axis represents the CPU utilization percentage (%).
Sensors 23 00970 g008
Figure 9. Disk utilization comparison for different Raspberry Pi and Jetson Nano devices. The x-axis represents the time in minutes, the y-axis represents the disk utilization percentage (%).
Figure 9. Disk utilization comparison for different Raspberry Pi and Jetson Nano devices. The x-axis represents the time in minutes, the y-axis represents the disk utilization percentage (%).
Sensors 23 00970 g009
Figure 10. Process CPU threads in use comparison for different Raspberry Pi and Jetson Nano devices. The x-axis represents the time in minutes, the y-axis represents the number of CPU threads in use.
Figure 10. Process CPU threads in use comparison for different Raspberry Pi and Jetson Nano devices. The x-axis represents the time in minutes, the y-axis represents the number of CPU threads in use.
Sensors 23 00970 g010
Figure 11. Process memory in use comparison for different Raspberry Pi and Jetson Nano devices. The x-axis represents the time in minutes, the y-axis represents the used process memory in megabytes (MB).
Figure 11. Process memory in use comparison for different Raspberry Pi and Jetson Nano devices. The x-axis represents the time in minutes, the y-axis represents the used process memory in megabytes (MB).
Sensors 23 00970 g011
Figure 12. Process memory available comparison for different Raspberry Pi and Jetson Nano devices. The x-axis represents the time in minutes, the y-axis represents the available process memory in megabytes (MB).
Figure 12. Process memory available comparison for different Raspberry Pi and Jetson Nano devices. The x-axis represents the time in minutes, the y-axis represents the available process memory in megabytes (MB).
Sensors 23 00970 g012
Figure 13. System memory utilization comparison for different Raspberry Pi and Jetson Nano devices. The x-axis represents the time in minutes, the y-axis represents the system memory utilization in percentage (%).
Figure 13. System memory utilization comparison for different Raspberry Pi and Jetson Nano devices. The x-axis represents the time in minutes, the y-axis represents the system memory utilization in percentage (%).
Sensors 23 00970 g013
Figure 14. Network traffic comparison for different Raspberry Pi and Jetson Nano devices. The x-axis represents the time in minutes, the y-axis represents network traffic in bytes.
Figure 14. Network traffic comparison for different Raspberry Pi and Jetson Nano devices. The x-axis represents the time in minutes, the y-axis represents network traffic in bytes.
Sensors 23 00970 g014
Figure 15. Temperature comparison for different Raspberry Pi and Jetson Nano devices. The x-axis represents the incremental number of training mini-batches, the y-axis represents temperature in Celsius ( C).
Figure 15. Temperature comparison for different Raspberry Pi and Jetson Nano devices. The x-axis represents the incremental number of training mini-batches, the y-axis represents temperature in Celsius ( C).
Sensors 23 00970 g015
Table 1. List of the data features.
Table 1. List of the data features.
Non-TextClinical Text
Vital DataDemographic DataLaboratory Data
Systolic Blood PressureGenderAlbuminNursing Notes
Diastolic Blood PressureAdmission AgeBands (Immature Neutrophils)Physician Notes
Mean Blood PressureEthnicityBicarbonate
Respiratory RateAdmission TypeBilirubin
Heart RateAdmission LocationCreatinine
SpO2 (Pulsoxymetry) Chloride
Temperature Celsius Sodium
Cardiac Output Potassium
Tidal Volume Set Lactate
Tidal Volume Observed Hematocrit
Tidal Volume Spontaneous Hemoglobin
Peak Inspiratory Pressure Platelet Count
Total Peep Level Partial Thromboplastin Time
O2 flow INR (Standardized Quick)
FiO2 (Fraction of Inspired Oxygen) Blood Urea Nitrogen
White Blood Cells
Creatine Kinase
Creatine Kinase MB
Fibrinogen
Lactate Dehydrogenase
Magnesium
Calcium (free)
pO2 Bloodgas
pH Bloodgas
pCO2 Bloodgas
SO2 Bloodgas
Glucose
Troponin T
Prothrombin Time (Quick)
Table 2. Dataset distribution.
Table 2. Dataset distribution.
DataPositiveNegative% Positive% Negative
Training135212,9829.4390.57
Testing18116439.9290.08
Tuning17216249.5890.42
Total170516,2499.5090.50
Table 3. Neural Network Parameters.
Table 3. Neural Network Parameters.
NameValues
hidden layers1, 2, 3
neurons512, 800, 1024, 2048
drop out0, 10, 20, 30
epochs1, 2
classification functionlog-softmax
optimizerAdam optimizer
GAIN alpha0.1, 1, 10, 100
GAIN mini-batch512
GAIN hint rate0.9
GAIN total iterations15,000
Table 4. Topics of client device system information.
Table 4. Topics of client device system information.
TopicMeasurement Unit
CPU Utilization%
Disk Utilization%
Process CPU Threads In UseNumber of threads
Process Memory In Use (non-swap)Megabytes
Process Memory Available (non-swap)Megabytes
System Memory Utilization%
TemperatureCelsius
Network TrafficBytes
Table 5. Evaluation results using non-text input features in single-server setting.
Table 5. Evaluation results using non-text input features in single-server setting.
Missingness RepresentationAUPRC
(%)
AUROC
(%)
Earliness
(h)
GAIN-imputation87.3098.314.45
distinct-value18.3575.624.01
carry-forward-mean-imputation14.2266.474.15
Table 6. Evaluation results using clinical-text input features in single-server setting.
Table 6. Evaluation results using clinical-text input features in single-server setting.
clinicalBERTText
Embeddings
Representation
Missingness RepresentationAUPRC
(%)
AUROC
(%)
Earliness
(h)
HuangshortGAIN-imputation14.06 57.516.35
Huanglongdistinct-value13.9758.506.53
Alsentzershortcarry-forward-mean-imputation13.7059.487.15
Alsentzershortdistinct-value13.4659.996.49
Alsentzerlongcarry-forward-mean-imputation13.4456.315.77
Alsentzerlongdistinct-value13.3556.807.28
AlsentzershortGAIN-imputation13.0657.116.70
Huangshortcarry-forward-mean-imputation12.9156.936.68
AlsentzerlongGAIN-imputation12.3655.726.26
HuanglongGAIN-imputation12.3255.176.43
Huangshortdistinct-value12.3054.917.00
Huanglongcarry-forward-mean-imputation11.6953.076.62
Table 7. Evaluation results using multi-modal (both clinical-text and non-text) input features in single-server setting.
Table 7. Evaluation results using multi-modal (both clinical-text and non-text) input features in single-server setting.
clinicalBERTText
Embeddings
Representation
Missingness RepresentationAUPRC
(%)
AUROC
(%)
Earliness
(h)
HuangshortGAIN-imputation 96.55 99.35 4.56
HuanglongGAIN-imputation96.0998.904.53
AlsentzerlongGAIN-imputation94.0497.904.45
AlsentzershortGAIN-imputation93.9598.61 4.58
Huanglongdistinct-value31.9582.903.71
Huangshortdistinct-value31.3181.743.86
Alsentzershortdistinct-value31.0682.683.94
Huanglongcarry-forward-mean-imputation28.9881.253.87
Huangshortcarry-forward-mean-imputation28.6180.573.71
Alsentzerlongcarry-forward-mean-imputation28.3781.853.58
Alsentzerlongdistinct-value24.3379.603.90
Alsentzershortcarry-forward-mean-imputation24.2377.713.98
Table 8. Evaluation results using non-text input features in a federated learning setting using 10 Raspberry Pi devices.
Table 8. Evaluation results using non-text input features in a federated learning setting using 10 Raspberry Pi devices.
After RoundFederated Aggregation: SimpleFederated Aggregation: Opt
AUPRC
(%)
AUROC
(%)
Earliness
(h)
AUPRC
(%)
AUROC
(%)
Earliness
(h)
130.9786.734.0022.5482.074.01
239.7788.454.0619.7879.403.94
38.4192.804.4418.6677.863.98
455.2095.694.5323.0881.524.04
571.3997.604.5434.8386.994.14
Table 9. Evaluation results using clinical-text input features in a federated learning setting using 10 Raspberry Pi devices.
Table 9. Evaluation results using clinical-text input features in a federated learning setting using 10 Raspberry Pi devices.
After RoundFederated Aggregation: SimpleFederated Aggregation: Opt
AUPRC
(%)
AUROC
(%)
Earliness
(h)
AUPRC
(%)
AUROC
(%)
Earliness
(h)
113.2055.604.1212.9952.067.73
213.2354.055.6313.2251.846.63
313.7956.536.4812.8552.156.98
413.8457.526.2913.1152.936.45
514.6258.916.3013.4654.586.85
Table 10. Evaluation results using multi-modal (both clinical-text and non-text) input features in a federated learning setting using 10 Raspberry Pi devices.
Table 10. Evaluation results using multi-modal (both clinical-text and non-text) input features in a federated learning setting using 10 Raspberry Pi devices.
After RoundFederated Aggregation: SimpleFederated Aggregation: Opt
AUPRC
(%)
AUROC
(%)
Earliness
(h)
AUPRC
(%)
AUROC
(%)
Earliness
(h)
145.3093.104.4746.1793.474.43
269.2697.324.5354.6695.434.49
360.4095.734.5142.0691.894.49
484.8098.324.5450.0294.664.52
582.4498.494.5180.5898.064.51
Table 11. Evaluation results using multi-modal (both clinical-text and non-text) input features in a federated learning setting using a different number of Raspberry Pi devices.
Table 11. Evaluation results using multi-modal (both clinical-text and non-text) input features in a federated learning setting using a different number of Raspberry Pi devices.
Total
Devices
AUPRC
(%)
AUROC
(%)
Earliness
(h)
Federated Aggregation: Simple
1082.4498.494.51
994.5499.60 4.53
896.1399.53 4.53
797.0399.634.51
696.5599.634.51
598.7999.874.50
498.5699.764.52
397.5999.774.50
299.2099.914.50
199.4399.944.49
Table 12. Evaluation results using non-text input features in a federated learning setting using three Jetson Nano devices.
Table 12. Evaluation results using non-text input features in a federated learning setting using three Jetson Nano devices.
After RoundFederated Aggregation: SimpleFederated Aggregation: Opt
AUPRC
(%)
AUROC
(%)
Earliness
(h)
AUPRC
(%)
AUROC
(%)
Earliness
(h)
140.4886.103.8438.3985.463.90
255.9595.554.5042.8787.054.04
380.4298.554.5244.3189.004.18
492.0799.304.5248.5392.034.26
594.4399.514.5352.2093.914.30
Table 13. Evaluation results using clinical-text input features in a federated learning setting using three Jetson Nano devices.
Table 13. Evaluation results using clinical-text input features in a federated learning setting using three Jetson Nano devices.
After RoundFederated Aggregation: SimpleFederated Aggregation: Opt
AUPRC
(%)
AUROC
(%)
Earliness
(h)
AUPRC
(%)
AUROC
(%)
Earliness
(h)
113.3155.695.8412.7954.187.27
215.0761.735.7913.1255.656.90
315.1263.586.4313.8857.427.06
414.3360.886.9014.1458.497.02
514.7362.247.7314.2159.727.06
Table 14. Evaluation results using multi-modal (both clinical-text and non-text) input features in a federated learning setting using three Jetson Nano devices.
Table 14. Evaluation results using multi-modal (both clinical-text and non-text) input features in a federated learning setting using three Jetson Nano devices.
After RoundFederated Aggregation: SimpleFederated Aggregation: Opt
AUPRC
(%)
AUROC
(%)
Earliness
(h)
AUPRC
(%)
AUROC
(%)
Earliness
(h)
194.7799.344.5694.7199.444.55
295.0399.374.5494.8099.394.56
398.0799.784.5393.5899.184.52
498.8299.874.5195.8699.484.51
598.9999.894.5194.9999.354.52
Table 15. Evaluation results using multi-modal (both clinical-text and non-text) input features in a federated learning setting using different number of Jetson Nano devices.
Table 15. Evaluation results using multi-modal (both clinical-text and non-text) input features in a federated learning setting using different number of Jetson Nano devices.
Total
Devices
AUPRC
(%)
AUROC
(%)
Earliness
(h)
Federated Aggregation: Simple
398.9999.894.51
299.3699.934.50
199.3199.924.52
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Alam, M.U.; Rahmani, R. FedSepsis: A Federated Multi-Modal Deep Learning-Based Internet of Medical Things Application for Early Detection of Sepsis from Electronic Health Records Using Raspberry Pi and Jetson Nano Devices. Sensors 2023, 23, 970. https://doi.org/10.3390/s23020970

AMA Style

Alam MU, Rahmani R. FedSepsis: A Federated Multi-Modal Deep Learning-Based Internet of Medical Things Application for Early Detection of Sepsis from Electronic Health Records Using Raspberry Pi and Jetson Nano Devices. Sensors. 2023; 23(2):970. https://doi.org/10.3390/s23020970

Chicago/Turabian Style

Alam, Mahbub Ul, and Rahim Rahmani. 2023. "FedSepsis: A Federated Multi-Modal Deep Learning-Based Internet of Medical Things Application for Early Detection of Sepsis from Electronic Health Records Using Raspberry Pi and Jetson Nano Devices" Sensors 23, no. 2: 970. https://doi.org/10.3390/s23020970

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop