Next Article in Journal
Teaching Essential EMG Theory to Kinesiologists and Physical Therapists Using Analogies Visual Descriptions, and Qualitative Analysis of Biophysical Concepts
Next Article in Special Issue
A Systematic Review of Internet of Things in Clinical Laboratories: Opportunities, Advantages, and Challenges
Previous Article in Journal
SpecSeg Network for Specular Highlight Detection and Segmentation in Real-World Images
Previous Article in Special Issue
Automated Chicago Classification for Esophageal Motility Disorder Diagnosis Using Machine Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Classification of Tetanus Severity in Intensive-Care Settings for Low-Income Countries Using Wearable Sensing

1
Department of Engineering Science, University of Oxford, Oxford OX1 3PJ, UK
2
Oxford University Clinical Research Unit, Ho Chi Minh City 700000, Vietnam
3
Hospital of Tropical Diseases, Ho Chi Minh City 700000, Vietnam
4
Hthe Oxford Suzhou Centre for Advanced Research, University of Oxford, Suzhou Dushu Lake Science and Education Innovation District, Suzhou 215123, China
*
Author to whom correspondence should be addressed.
The members of VITAL Consortium are listed in Acknowledgments.
Sensors 2022, 22(17), 6554; https://doi.org/10.3390/s22176554
Submission received: 25 July 2022 / Revised: 19 August 2022 / Accepted: 22 August 2022 / Published: 30 August 2022
(This article belongs to the Special Issue AI and IoT Enabled Solutions for Healthcare)

Abstract

:
Infectious diseases remain a common problem in low- and middle-income countries, including in Vietnam. Tetanus is a severe infectious disease characterized by muscle spasms and complicated by autonomic nervous system dysfunction in severe cases. Patients require careful monitoring using electrocardiograms (ECGs) to detect deterioration and the onset of autonomic nervous system dysfunction as early as possible. Machine learning analysis of ECG has been shown of extra value in predicting tetanus severity, however any additional ECG signal analysis places a high demand on time-limited hospital staff and requires specialist equipment. Therefore, we present a novel approach to tetanus monitoring from low-cost wearable sensors combined with a deep-learning-based automatic severity detection. This approach can automatically triage tetanus patients and reduce the burden on hospital staff. In this study, we propose a two-dimensional (2D) convolutional neural network with a channel-wise attention mechanism for the binary classification of ECG signals. According to the Ablett classification of tetanus severity, we define grades 1 and 2 as mild tetanus and grades 3 and 4 as severe tetanus. The one-dimensional ECG time series signals are transformed into 2D spectrograms. The 2D attention-based network is designed to extract the features from the input spectrograms. Experiments demonstrate a promising performance for the proposed method in tetanus classification with an F1 score of 0.79 ± 0.03, precision of 0.78 ± 0.08, recall of 0.82 ± 0.05, specificity of 0.85 ± 0.08, accuracy of 0.84 ± 0.04 and AUC of 0.84 ± 0.03.

1. Introduction

Infectious diseases remain a common cause of morbidity and mortality among people living in low- and middle-income countries [1,2,3]. Tetanus disease is an infection caused by a toxin produced by the Clostridium tetani bacteria [4]. This powerful neurotoxin inhibits transmission at central synapses resulting in muscle stiffness and spasms and in severe cases, cardiovascular system instability. These symptoms generally progress over a period of 2–5 days. Approximately half of all patients progress to severe disease where spasm control necessitates paralysis and mechanical ventilation. The most common cause of death in settings with access to mechanical ventilation is autonomic nervous system (ANS) dysfunction, occurring in approximately 25% of all patients. Therefore, the early detection of those likely to have severe disease requiring mechanical ventilation or ANS dysfunction is highly valuable as it enables timely intervention and allows appropriate resource allocation [5,6].
The most widely used system for tetanus severity classification is the Ablett score, ranging from 1 to 4. While grades 1 and 2 describe mild or moderate disease progressions where a mechanical ventilation is typically not required, grades 3 and 4 represent severe disease requiring mechanical ventilation. Grade 4 is the most severe form, in which patients have signs of ANS dysfunction [2,7]. Similar to other infectious diseases, early and accurate diagnosis of tetanus severity is extremely important to improve both short- and long-term patient outcomes [8,9,10]. However, in the low- and middle-income countries where tetanus most commonly occurs, facilities and equipment for treatment are often limited. Experienced doctors and nurses have limited time to frequently monitor patients with tetanus. Heart rate variability (HRV) has been shown to be valuable in ANS detection [11,12].
To improve the clinical outcomes and disease incidence of tetanus, we aim to develop a severity warning tool. This tool will use the patient’s electrocardiogram (ECG) data to classify disease severity, aiming to function as a screening tool and guide to clinicians. Such a tool will predict the severity of the disease to help clinicians determine whether the patient must undergo close monitoring and start planning for admission to an intensive care unit (ICU). If the predicted symptoms are mild, the patient can be monitored less frequently in a normal ward. Such a tool would be of particular value for inexperienced or overloaded staff, prevent unnecessary ICU admissions and reduce treatment delays.
In this paper, we design scientific steps as follows: Firstly, the physiological data - electrocardiogram (ECG) data - are collected from the low-cost wearable monitors. Secondly, we propose a warning tool with a deep learning approach for the diagnosis of infectious disease (e.g., tetanus) patients. The aim of this tool is to classify the tetanus severity level, represented as the Ablett score, based on this low-cost ECG data. The contributions of this work are as follows:
  • To the best of our knowledge, this is the first attempt to exploit deep learning with a channel-wise attention mechanism in tetanus diseases detection, which models the channel relationship and boosts the performance of a network. Since the method is completely data-driven, this concept could be transferable to similar infectious diseases.
  • We demonstrate the effectiveness of the proposed method on the low-cost ECG data. We show that our novel method outperforms the sequential techniques. The sequential techniques, including the time-dependent versions of the attention-based network, do not work on low-cost ECG data because the noise of the low quality data disturbs time series analysis.
  • We explore the robustness of the proposed method for the minimal window length of the log-spectrogram.
The paper is structured as follows: Section 2 introduces related work in the diagnosis of tetanus diseases in low- and middle-income countries and deep learning approaches in imaging fields related to our work. Section 3 describes the proposed approach for the tetanus diseases classification. Section 4 provides the details of the tetanus dataset, implementation details, a comparison of baseline methods and the evaluations of the classification results with several performance metrics. Section 5 presents and discusses the experimental results. Finally, Section 6 provides the conclusion of our work.

2. Related Work

The early diagnosis of lethal infectious diseases plays an important role in patient treatment. Heart rate is controlled by the ANS and heart rate variability (HRV), i.e., the beat-to-beat changes in RR intervals, is linked to changes in ANS activity [11,12]. In tetanus, disease severity is associated with ANS activity, and changes in conventional HRV parameters measured from ECG have been shown to correlate with disease severity. To evaluate autonomic nervous system disturbance (ANSD), the HRV-based methods need an extra preprocessing step and then require features such as RR intervals and QRS complex extraction [6,13,14]. Lin et al. [15] use HRV as an index to detect the disease progression which is caused by enterovirus infection. However, it is still a challenge to robustly extract RR intervals.
Recently, machine learning approaches have been used to diagnose and classify severity of infectious diseases. Traditional machine learning approaches require the feature engineering process for manually selecting and transforming features from the dataset. Tadesse et al. [16] applied support vector machines (SVM) to automatically detect the ANS dysfunction level for hand, foot and mouth disease (HFMD) and tetanus. They also demonstrated SVM outperforms HRV on infectious diseases detection. Tadesse et al. [17] use spectrograms of ECG and PPG with transfer learning to classify severity of two infectious diseases, tetanus and HFMD, and prove deep learning methods outperform traditional machine learning methods (e.g., SVM). Kiyasseh et al. [18] suggest generating pathological photoplethysmogram (PPG) signals to boost diagnosis performance (e.g., tetanus and HFMD). The previous works from [16,17,18] study a small dataset of tetanus; consequently, their results are limited.
One prominent advantage of convolutional neural networks (CNNs) is their capability to implicitly learn to extract relevant features. The one-dimensional (1D) CNNs have been widely used in signal processing applications, such as biomedical data classification and early diagnosis [19,20]. In order to use the 2D CNN, the common approach transforms the 1D signal to a 2D representation by time–frequency analysis, such as spectrogram, log spectrogram, mel spectrogram and scalogram [17,21,22,23,24], and the 2D representation obtained is considered an image. The spectrogram provides a visual presentation of dynamic information which can be composed of low-level features such as lines and edges [17]. Based on the recent literature work, we know the 2D CNN performs well in image classification. Using 2D spectrograms, an image-based ECG signal classification structure achieves a better performance than the 1D CNN [25].
Computer vision and image analysis have been revolutionized by the attention mechanism [26]. The benefits of the attention mechanism range are across different topics, from image classification [27,28] to action recognition [29,30], for improving representation power of networks. There are two families of attention mechanisms in deep learning, soft attention and hard attention. In soft attention, the features of the image are multiplied with a mask of values between zero and one. In hard attention, the deep learning model focuses on the input information from a small portion of the whole image, with a mask of values of zero or one. In general, attention is implemented by a combination with a gating function, such as a sigmoid or softmax. Attention can also be implemented via combinating with sequential techniques, such as long short-term memory (LSTM) [31]. So far, the attention mechanism has not been implemented in infectious diseases for improving diagnostic accuracy. Therefore, it is a novel application of the attention mechanism to an ECG dataset acquired from patients with infectious diseases. In previous studies, spectrograms of tetanus with transfer learning [17] do not consider which part of the feature maps is more important. Our work will use soft self-attention to weight the channel-wise responses in the convolutional layers for modelling inter-dependencies among the channel-wise features.

3. Method

The proposed framework includes the following steps:
  • Data Preprocessing: ECG noise removal;
  • Spectrogram analysis of single-lead ECG signal: Generated 2D log-spectrograms as inputs of the proposed method;
  • Feature extraction with CNN: Feed the log-spectrograms into convolutional layers to extract features;
  • Attention Mechanism: Model the inter-dependencies among the channel features of the convolutional layers.
The proposed method is named 2D-CNN + Channel-wise Attention. To understand how the proposed method makes a decision and what the network sees in each layer in the method, we explore the visual explanation algorithm-gradient-weighted class activation mapping (Grad-CAM) [32]-in the proposed method. Figure 1 shows the overview of the proposed method and the visual explanations of features in the last layers of 6 convolutional blocks of the method.

3.1. Data Preprocessing

There are mainly two types of noise that influence the ECG signal. Firstly, patient muscle movement causes low band frequency noises [33]. Secondly, the electrical source which operates the ECG monitor leads to high band frequency noise [33]. Given low-cost ECG signal, we use a one-lead ECG from an ECG signal and perform preprocessing to clean the data and remove the background noise from the input ECG signal using a Butterworth filter. We set a cutoff point of 0.05 Hz for the high-pass filter and 100 Hz for the low-pass filter. The implementation is performed utilizing the SciPy package [34].

3.2. Logarithmic Spectrogram Generation

A 1D ECG is not able to use 2D CNNs. If the ECG is represented in 2D, such as an image, we can use the successful approaches in image classification to deal with the signal. Hence, we transform the preprocessed ECG into spectrograms. The spectrogram is a 2D time–frequency representation based on the consecutive Fourier transform. The logarithmic spectrogram is a log-scaled spectrogram based on the consecutive Fourier transform, and it pays more attention to lower frequencies. Next, we normalise the spectrograms by their maximum value and scale the value in the range 0 to 255 and logarithmic scale of the normalised spectrograms (see Equation (1)).
V ˜ = log V m a x ( V ) 255
Figure 2 shows examples of spectrograms and normalised logarithmic spectrograms of ECG with mild and severe symptoms. The visible image patterns in spectrograms are hard to see. By using a log scale, the low frequencies are easier to see as log-spectrograms. Hence, the use of log-spectrogram image patterns enhances understanding of 2D CNNs.

3.3. Attention-Based Network

3.3.1. Convolutional Layers

The convolutional layers explore the spatial information in each 2D spectrogram (intra-slice information). The architecture of each block was inspired by Zihlmann et al. [22] and consists of the convolutional blocks of the 2D convolutional layers ( 3 × 3 kernel size), ELU and 2D batch normalization. The second, fourth and sixth convolutional blocks are followed by a 2D max pooling layer ( 2 × 2 window), respectively (See Figure 1). A logarithmic spectrogram is input to the convolutional layers.

3.3.2. Channel-Wise Attention

Changing the weight of the different channels in the feature maps, the proposed model can emphasise the most important features and suppress less useful features. Hence, the channel-wise attention mechanism can explore the relationships of features among different channels and add weights as soft attention for each channel. Inspired by the squeeze-and-excitation networks [28], we add the channel-wise attention at the end of each convolutional block. The illustration in Figure 3 shows the structure of the channel-wise attention mechanism.
Let U = [ u 1 , u 2 , …, u n ], where u i R W × H denotes the feature map on the i -th channel, n is the number of channels, and the W and H are width and height of u i . For the squeeze operation, we aim to squeeze global spatial information into a channel feature. We apply a 2D adaptive average pooling to obtain a single value for each channel feature (see Equation (2)). The channel features m can be represented as M = [ m 1 , m 2 , , m n ] , where m i R n denotes the i -th channel features, and m i is the average of the vector u i . The channel features can be calculated by
m i = 1 W × H p = 1 W q = 1 H u i ( p , q ) .
Next, we perform an excitation operation on the single value for obtaining the channel weights that represent channel-wise dependencies. The excitation operation uses a gating function with a sigmoid activation, which can be represented as
s i = σ ( W 2 ϕ ( W 1 m i ) ) ,
where σ and ϕ refer to the sigmoid and ReLU function, respectively. W 1 and W 2 represent the learnable parameter matrices.
The output of the excitation operation - channel weights - are element-wise multiplied on the output features of each convolutional block. The final output of the channel-wise attention block can be represented as
x i ˜ = u i m i ,
where ⨂ is the channel-wise multiplication between the feature map u i and the weight vector m i .

3.3.3. Loss Function

The binary cross-entropy (BCE) loss function [35] is used in the proposed method, which is defined as
L B C E = 1 N i = 1 N ( y i · l o g y i ^ + ( 1 y i ) · l o g ( 1 y i ^ ) ,
where y i is the ith target label, y ^ is the prediction of the ith label, and N is the batch size. We combine a sigmoid layer and a BCELoss in one single class. This combined loss function is more numerically stable than using a plain sigmoid followed by a BCELoss; by combining the operations into one layer, we take advantage of the log-sum-exp trick for numerical stability.
After the attention layer of the last convolutional block, we choose 3 fully connected layers that output our 2 labels. The output of the fully connected layer is fed as inputs to the sigmoid layer, and the output of the sigmoid layer is turned into the probability of the tetanus mild and severe classes.

4. Experiments

4.1. ECG Acquisition for Tetanus Patients

To acquire ECG data from tetanus patients, we use the low-cost wearable monitor ePatch [36] (see Figure 4). The ePatch (ePatch. https://www.myheartmonitor.com/device/epatch/ (accessed on 21 August 2022)) sensor includes all the electronic components: a rechargeable battery, a signal processing component, a data storage component and wireless data transmission equipment [37]. The doctor attaches the lightweight cardiac monitor firmly to the patient’s chest. The ePatch will automatically record the ECG once the system is installed. Figure 2 shows examples of ECG waveforms collected from tetanus patients.
The study data collection has been approved by the relevant ethical committees. This dataset has been published previously [38] and is collected from 110 patients at the Hospital for Tropical Diseases, Ho Chi Minh City, Vietnam. The ECG waveforms from the tetanus patients are collected with a sampling rate of 256 Hz. The first 24 h ECG data is recorded on day 1 when a patient is admitted to the infectious disease department. The second 24 h ECG data is recorded on day 5 of hospitalization.
The dataset used in this study consists of 178 ECG example files from 110 patients on days 1 and 5. We use trimmed ECG data [38] which removes most of the noise from the ECG recording. Then we split our data into the training/validation/test datasets with a 141/19/18 ratio.
The ePatch can collect two-lead ECG signals. We extract the one-lead ECG signal for our experiments. We perform spectrogram analysis of this single-lead ECG signal. We transform the ECG signal to a 2D image, extract features from a 2D CNN and then model inter-dependencies among the channel-wise features. Our deep learning model will classify the ECG signal into two categories; label 0 represents mild tetanus (Ablett grade 1 and 2), and label 1 represents severe tetanus (Ablett grade 3 and 4).

4.2. Implementation Details

4.2.1. Data Preprocessing

The time series ECG waveform is divided into a sequence of ECG samples without overlapped windows. We set the duration of the window length as 60 s. We choose 30 60-s ECG samples from each ECG example file. There are 4230 (141 × 30) ECG log-spectrograms in the training set, including 2370 samples of the mild disease and 1860 samples of the severe disease; 540 (19 × 30) ECG log-spectrograms in the validation set, including 270 samples of the mild disease and 270 samples of the severe disease; 570 (18 × 30) ECG log-spectrograms in the test set, including 360 samples of the mild disease and 210 samples of the severe disease.
Spectrograms are computed by scipy.signal.spectrogram in SciPy [34]. We choose the Tukey window width to be 25 % of a window’s length overlap. We set the nperseg-length of each segment as 64, and the noverlap numbers of points to overlap between segments as 32. There are 15,360 = 256 Hz × 60 s sampling points in a window of length which are used to compute a spectrogram; these are based on 60 s at the sampling rate of 256 Hz of the ECG data. We then perform normalization and logarithmic scale on the spectrogram (see Figure 2). The spectrogram is saved as a PNG format image with the default ’viridis’ colormap. Finally, the rectangular picture of the spectrogram ( 479 × 33 pixels of the log-spectrograms on every 60 s ECG) is ready for the proposed deep learning approach.

4.2.2. Training

The model is trained over 100 epochs using the Adam optimizer with a learning rate 0.001 and a batch size of 32. The mean squared error (MSE) is chosen as the evaluation metric. We choose torch.nn.BCEWithLogitsLoss for the loss function. The proposed network was implemented using Python 3.7 with Pytorch. Experiments are run with computational hardware NVidia GeForce GTX 1080 Ti GPU 10 GB and NVidia GeForce RTX 3060 12 GB.

4.3. Baseline Methods

Because the attention mechanism can be implemented by combining it with either the gating function or sequential techniques, we aim to compare these two attention combination style networks. In our work, we compare the proposed 2D-CNN + Channel-wise Attention method with six different methods. Figure 5 shows six 2D deep learning methods: 2D-CNN, 2D-CNN + Channel-wise Attention + ConvLSTM, 2D-CNN + Channel-wise Attention + LSTM, 2D-CNN + LSTM, 2D-CNN + ConvLSTM and 2D-CNN + Dual Attention. Here Channel-wise Attention and Dual Attention belong to the gating function; ConvLSTM and LSTM belong to sequential techniques.
To investigate how the attention mechanism works in the proposed method, we compare the methods with and without attention layers. We also compare the 2D-CNN with the 1D-CNN for testing the image-based ECG signal classification method.

4.4. Evaluation Metrics

In the binary classification, the terms true positive (TP), true negative (TN), false positive (FP) and false negative (FN) are used to calculate accuracy, precision, specificity, recall and F1-score [17].
A c c u r a c y = T P + T N T P + T N + F P + F N
P r e c i s i o n = T P T P + F P
S p e c i f i c i t y = T N T N + F P
R e c a l l = T P T P + F N
F 1 = 2 p r e c i s i o n r e c a l l p r e c i s i o n + r e c a l l
We run each model five times and calculate the mean and the standard deviation of the performance metrics on the test dataset.

5. Results and Discussion

In this section, we evaluate the proposed method and show how it works. Then, we compare it to the vanilla 2D-CNN as a benchmark. We also present results on the longitudinal data, which are all essentially time-dependent versions of the previously used ones. Moreover, we analyse the method’s parameters regarding more efficient computation. In addition, we compare the proposed method to the traditional machine learning method of random forest. In our experiments, we run each model five times with the same split training/validation/test datasets. We perform the splitting of the dataset into training, validation and test based on unique ECG samples. After splitting, we apply windowing on ECG time series to split each signal into 60 s time series. Therefore, we made sure that ECG samples in each split of the dataset are unique.

5.1. Attention Layers

We have investigated different attention mechanisms including spatial attention and channel-wise attention. Due to GPU memory capacity limitations, the self-attention model cannot be tested in our experiments. However, we are able to compare the dual attention model (position attention and channel attention modules) [39] to the proposed 2D-CNN + channel-wise attention model. According to the experimental results in Table 1, we found that the channel-wise attention outperforms the dual attention mechanism. Figure 6 shows the examples of Grad-CAM visual explanations of the features for label 0 - mild tetanus - in all different layers of the baseline 2D-CNN method and the proposed 2D-CNN + Channel-wise Attention method. The different important locations of features are visualised by colours. The red colours emphasise the most important parts where the model focuses on different layers for classification. Compared to the last layer in block six, there are more red areas in the proposed method, meaning that this area influences the final decision for label 0 - mild tetanus.

5.2. Sequential Techniques

The different longitudinal models are all essentially time-dependent versions of the previously used ones: 2D-CNN and 2D-CNN + Channel-wise Attention.

5.2.1. Recurrent Neural Network Layers

As shown in Table 1, the 2D-CNN + Channel-wise Attention performs better than the 2D-CNN + Channel-wise Attention + LSTM, and the 2D-CNN performs slightly better than the 2D-CNN + LSTM. The low-cost ECG signal quality is too low to perform longitudinal data analysis with recurrent neural networks. There is a great deal of background noise in the ECG data. Although we performed preprocessing to filter out the noise, the cleaned ECG data still contains noise, which influences the results of the recurrent neural network.

5.2.2. Convolutional LSTMs Model

Inspired by Lu et al. [40], we use a single layer convolutional LSTM (ConvLSTM) [41] to explore the temporal relationships among the three log-spectrograms. As shown in Figure 5, the 2D-CNN + ConvLSTM and 2D-CNN + Channel-wise Attention + ConvLSTM methods are explored in our experiment. The output of the convolutional layers will be the input of the ConvLSTM layer. We set T = 3 , 60 s for a 20-s window duration. The ConvLSTM makes decisions on the features of three log-spectrogram samples. Table 1 shows that the ConvLSTM models do not perform well in 1-lead ECG data, suggesting that the ECG signal quality in resource-limited settings is too low for temporal information analysis.

5.3. 1D Convolutional Model

The 1D-CNN has the same architecture as the 2D-CNN (shown in Figure 5). However, the 1D-CNN model uses 1D convolution instead of 2D convolution at its convolutional layers. As shown in Table 1, the performance of the 1D-CNN is slightly lower than the 2D-CNN using a 60-s window length log-spectrogram without downsampling. Compared to the 2D-CNN + Channel-wise Attention, the 1D-CNN has lower performance metrics. The results show that the image-base method works better, and the channel-wise information boosts the performance of diagnosing tetanus.

5.4. Downsample Spectrogram

Due to the computational limits of the GPUs, we aim to develop a deep learning pipeline with low computational cost. Therefore, we perform experiments on downsampled spectrograms using scipy.signal.decimate. We downsample spectrograms four times in the horizontal axis and the vertical axis, respectively. As shown in Table 2, spectrograms without downsampling produce better F1 scores, specificities and accuracies than those with downsampling for 60 s window length spectrograms. Because the downsampled spectrograms are too small, the convolutional LSTM methods - 2D-CNN + ConvLSTM and 2D-CNN + Channel-wise Attention + ConvLSTM - fail. Hence, we suggest using spectrograms without downsampling as inputs in the proposed model.

5.5. Misclassification

In the training phase, we run each method five times, which gives five trained models in each method. Next, we obtain five different confusion matrices using the test dataset. Here we average confusion matrix numbers over the five different runs. The confusion matrices in Figure 7 show a holistic view of how well each method in our experiments performs and what kind of misclassification they make between the mild and severe levels. As shown in Figure 7a, the true successful detection of the severe tetanus diagnosis increases from 122 to 171 after employing channel-wise attention layers. It also increases to 129 after employing dual attention layers. Figure 7c shows the 1D-CNN better predicts severe tetanus than mild tetanus, with 162 correct severe tetanus diagnoses out of 210 samples and 253 correct severe mild diagnoses out of 360 samples. Compared to the same method from Figure 7a,b, the correct mild and severe tetanus diagnosis numbers are higher in (a) than (b). This demonstrates the 60 s window length log-spectrograms without downsampling as inputs work better than the downsampled log-spectrograms.

5.6. Window Length of Spectrogram

In order to evaluate the robustness of the proposed method (spectrograms without downsampling), we perform experiments on the window duration of the ECG and check the minimal window length of the spectrogram that can still diagnose tetanus. We have investigated the 60 s, 50 s, 40 s, 30 s, 20 s, 10 s and 5 s window lengths of raw ECG data for spectrogram generation. For the experiments, the size of the training/validation/test dataset does not change. As shown in Table 3, the 10 s and 5 s window lengths of the raw ECG data are too short to generate a useful spectrogram for deep learning approaches. Comparisons with the 60 s window length show that the 50 s, 40 s, 30 s and 20 s window length spectrograms can still maintain an accurate tetanus diagnosis.
Figure 8 shows the examples of Grad-CAM visual explanations of the features for mild tetanus in the different layers in the proposed method. The log-spectrograms without downsampling are the inputs of the proposed method. From the visual explanations of features, we can see that the channel-wise attention layer emphasises some parts of the feature image compared to the previous layer in the convolutional block, particularly in the green rectangle area of Figure 8.

5.7. Traditional Machine Learning

We compare the proposed 2D-CNN + Channel-wise Attention with the traditional machine learning method random forest [42,43]. The details of extracted features are shown in Table 4, including eight HRV time domain features. There are several open-source toolboxes that compute HRV based on raw ECG signal [44,45,46,47]. In our work, we detect r peaks of ECG using the open-source packages py-ecg-detectors 1.3.2 [48] and extract features using the open-source packages hrv-analysis 1.0.4 [47].
The comparisons are shown in Table 5. The F1 score is higher for random forest with HRV time domain features compared to the proposed 2D-CNN + Channel-wise Attention. The F1 score is the harmonic mean of precision and recall. Precision evaluates how precisely a method predicts severe tetanus (TP). Recall measures the percentage of the correctly predicted severe tetanus (TP) that a method detects. Random forest with HRV time domain features yields a better prediction of severe tetanus.

6. Conclusions

We proposed a deep learning method, 2D-CNN with a Channel-wise Attention mechanism, to classify the severity of tetanus using wearable monitors in a resource-limited setting. We cleaned the background noise, and we were able to classify tetanus symptoms as mild or severe tetanus. Despite this, there are limitations to this method. Firstly, the ECG data from the wearable monitors have a much lower signal-to-noise ratio. This makes reducing the large amount of noise from the wearable monitors ECG data a significant challenge. However, the low-cost ECG data are affordable in low- and middle-income countries, and we are able to reliably use this low-quality data. Secondly, due to the small dataset used, to make a classification of tetanus severity requires ECG data recorded on day 1 and day 5. In the future, we will extend the dataset in order to predict the severity of tetanus on day 5 using the ECG data from only day 1.
We investigate the window length of the spectrogram and investigate the range of window lengths that can maintain an accurate tetanus diagnosis. In our experiments, a 50 s window has a relatively higher value of performance metrics than other window lengths. We will explore time series imaging further in future work, which will aim to find the optimal range of time windows.
In future work, we will consider sequence learning via transformers. A combination of CNN and transformer networks are used to extract both local features and global dependencies [49,50,51]. Moreover, we will also apply knowledge distillation techniques to this combination network for further improving the accuracy of tetanus diagnoses. The rationale for using the knowledge distillation is the low processing and sensor costs. Knowledge distillation extracts the knowledge from the large complex teacher model and passes it on to the small simple student model [52,53]. This distilled procedure will not require the training of a large number of tetanus data.

Author Contributions

Conceptualization, P.L., H.B.H., N.V.H., P.N.Q.K., L.D.V.K. and D.A.C.; data curation, P.L. and H.B.H.; formal analysis, P.L., T.Z. and D.A.C.; funding acquisition, L.T. and D.A.C.; investigation, D.A.C.; methodology, P.L., S.G., J.H., L.T. and D.A.C.; resources, P.N.Q.K., N.V.H. and D.A.C.; supervision, T.Z., L.T. and D.A.C.; writing—original draft, P.L.; writing—review and editing, S.G., J.H., H.B.H., N.V.H., L.D.V.K., L.T., T.Z. and D.A.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Wellcome Trust under grant 217650/Z/19/Z. This work was supported in part by the National Institute for Health Research (NIHR) Oxford Biomedical Research Centre (BRC), and in part by InnoHK Project on Project 1.1—Wearable Intelligent Sensing Engineering (WISE) at Hong Kong Centre for Cerebro-cardiovascular Health Engineering (COCHE). DAC is an Investigator in the Pandemic Sciences Institute, University of Oxford, Oxford, UK. The views expressed are those of the authors and not necessarily those of the NHS, the NIHR, the Department of Health, the University of Oxford, or InnoHK—ITC.

Institutional Review Board Statement

Our study was approved by the Scientific and Ethical Com- mittee of the HTD, Ho Chi Minh City in Vietnam with the protocol number 1009/BVBND-HDDD and Oxford Tropical Research Ethics Committee with the protocol number 522-20.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Data not publicly available due to ethical restrictions.

Acknowledgments

Vietnam ICU Translational Applications Laboratory (VITAL) investigators. OUCRU inclusive authorship list in Vietnam (alphabetic order by surname): Sayem Ahmed, Dang Phuong Thao, Dang Trung Kien, Doan Bui Xuan Thy, Dong Huu Khanh Trinh, Du Hong Duc, Ronald Geskus, Ho Bich Hai, Ho Quang Chanh, Ho Van Hien, Huynh Trung Trieu, Evelyne Kestelyn, Lam Minh Yen, Le Dinh Van Khoa, Le Thanh Phuong, Le Thuy Thuy Khanh, Luu Hoai Bao Tran, Luu Phuoc An, Angela Mcbride, Nguyen Lam Vuong, Nguyen Quang Huy, Nguyen Than Ha Quyen, Nguyen Thanh Ngoc, Nguyen Thi Giang, Nguyen Thi Diem Trinh, Nguyen Thi Le Thanh, Nguyen Thi Phuong Dung, Nguyen Thi Phuong Thao, Ninh Thi Thanh Van, Pham Tieu Kieu, Phan Nguyen Quoc Khanh, Phung Khanh Lam, Phung Tran Huy Nhat, Guy Thwaites, Louise Thwaites, Tran Minh Duc, Trinh Manh Hung, Hugo Turner, Jennifer Ilo Van Nuil, Vo Tan Hoang, Vu Ngo Thanh Huyen, Sophie Yacoub. Hospital for Tropical Diseases, Ho Chi Minh City (alphabetic order by surname): Cao Thi Tam, Duong Bich Thuy, Ha Thi Hai Duong, Ho Dang Trung Nghia, Le Buu Chau, Le Mau Toan, Le Ngoc Minh Thu, Le Thi Mai Thao, Luong Thi Hue Tai, Nguyen Hoan Phu, Nguyen Quoc Viet, Nguyen Thanh Nguyen, Nguyen Thanh Phong, Nguyen Thi Kim Anh, Nguyen Van Hao, Nguyen Van Thanh Duoc, Nguyen Van Vinh Chau, Pham Kieu Nguyet Oanh, Phan Thi Hong Van, Phan Tu Qui, Phan Vinh Tho, Truong Thi Phuong Thao. University of Oxford (alphabetic order by surname): Natasha Ali, David Clifton, Mike English, Shadi Ghiasi, Heloise Greeff, Jannis Hagenah, Ping Lu, Jacob McKnight, Chris Paton, Tingting Zhu. Imperial College London (alphabetic order by surname): Pantelis Georgiou, Bernard Hernandez Perez, Kerri Hill-Cawthorne, Alison Holmes, Stefan Karolcik, Damien Ming, Nicolas Moser, Jesus Rodriguez Manzano. King’s College London (alphabetic order by surname): Liane Canas, Alberto Gomez, Hamideh Kerdegari, Marc Modat, Reza Razavi, Miguel Xochicale. University of Ulm (alphabetic order by surname): Walter Karlen. The University of Melbourne (alphabetic order by surname): Linda Denehy, Thomas Rollinson. Mahidol Oxford Tropical Medicine Research Unit (MORU) (alphabetic order by surname): Luigi Pisani, Marcus Schultz.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Thuy, D.B.; Campbell, J.I.; Thanh, T.T.; Thuy, C.T.; Loan, H.T.; Van Hao, N.; Minh, Y.L.; Van Tan, L.; Boni, M.F.; Thwaites, C.L. Tetanus in southern Vietnam: Current situation. Am. J. Trop. Med. Hyg. 2017, 96, 93. [Google Scholar] [CrossRef]
  2. Thwaites, C.; Yen, L.; Glover, C.; Tuan, P.; Nga, N.; Parry, J.; Loan, H.; Bethell, D.; Day, N.; White, N.; et al. Predicting the clinical outcome of tetanus: The tetanus severity score. Trop. Med. Int. Health 2006, 11, 279–287. [Google Scholar] [CrossRef]
  3. Van Hao, N.; Yen, L.M.; Davies-Foote, R.; Trung, T.N.; Duoc, N.V.T.; Trang, V.T.N.; Nhat, P.T.H.; Anh, N.T.K.; Lieu, P.T.; Thuy, T.T.D.; et al. The management of tetanus in adults in an intensive care unit in Southern Vietnam. Wellcome Open Res. 2021, 6, 107. [Google Scholar]
  4. Thwaites, C. Botulism and tetanus. Medicine 2017, 45, 739–742. [Google Scholar] [CrossRef]
  5. What are the Symptoms of Tetanus? Available online: https://healthclinics.superdrug.com/tetanus-symptoms/ (accessed on 21 August 2022).
  6. Duong, H.T.H.; Tadesse, G.A.; Nhat, P.T.H.; Van Hao, N.; Prince, J.; Duong, T.D.; Kien, T.T.; Nhat, L.T.H.; Van Tan, L.; Pugh, C.; et al. Heart rate variability as an indicator of autonomic nervous system disturbance in tetanus. Am. J. Trop. Med. Hyg. 2020, 102, 403. [Google Scholar] [CrossRef]
  7. Afshar, M.; Raju, M.; Ansell, D.; Bleck, T.P. Narrative review: Tetanus—A health threat after natural disasters in developing countries. Ann. Intern. Med. 2011, 154, 329–335. [Google Scholar] [CrossRef]
  8. The Importance of Diagnostic Tests in Fighting Infectious Diseases. Available online: https://www.lifechanginginnovation.org/medtech-facts/importance-diagnostic-tests-fighting-infectious-diseases.html (accessed on 21 August 2022).
  9. Trieu, H.T.; Anh, N.T.K.; Vuong, H.N.T.; Dao, T.; Hoa, N.T.X.; Tuong, V.N.C.; Dinh, P.T.; Wills, B.; Qui, P.T.; Van Tan, L.; et al. Long-term outcome in survivors of neonatal tetanus following specialist intensive care in Vietnam. BMC Infect. Dis. 2017, 17, 646. [Google Scholar] [CrossRef]
  10. Lam, P.K.; Trieu, H.T.; Lubis, I.N.D.; Loan, H.T.; Thuy, T.T.D.; Wills, B.; Parry, C.M.; Day, N.P.; Qui, P.T.; Yen, L.M.; et al. Prognosis of neonatal tetanus in the modern management era: An observational study in 107 Vietnamese infants. Int. J. Infect. Dis. 2015, 33, 7–11. [Google Scholar] [CrossRef]
  11. Cygankiewicz, I.; Zareba, W. Heart rate variability. Handb. Clin. Neurol. 2013, 117, 379–393. [Google Scholar]
  12. Task Force of the European Society of Cardiology the North American Society of Pacing Electrophysiology. Heart rate variability: Standards of measurement, physiological interpretation, and clinical use. Circulation 1996, 93, 1043–1065. [Google Scholar]
  13. Bolanos, M.; Nazeran, H.; Haltiwanger, E. Comparison of heart rate variability signal features derived from electrocardiography and photoplethysmography in healthy individuals. In Proceedings of the 2006 International Conference of the IEEE Engineering in Medicine and Biology Society, New York, NY, USA, 30 August–3 September 2006; pp. 4289–4294. [Google Scholar]
  14. Ghiasi, S.; Zhu, T.; Lu, P.; Hagenah, J.; Khanh, P.N.Q.; Hao, N.V.; Thwaites, L.; Clifton, D.A.; Consortium, V. Sepsis Mortality Prediction Using Wearable Monitoring in Low–Middle Income Countries. Sensors 2022, 22, 3866. [Google Scholar] [CrossRef] [PubMed]
  15. Lin, M.T.; Wang, J.K.; Lu, F.L.; Wu, E.T.; Yeh, S.J.; Lee, W.L.; Wu, J.M.; Wu, M.H. Heart rate variability monitoring in the detection of central nervous system complications in children with enterovirus infection. J. Crit. Care 2006, 21, 280–286. [Google Scholar] [CrossRef] [PubMed]
  16. Tadesse, G.A.; Zhu, T.; Thanh, N.L.N.; Hung, N.T.; Duong, H.T.H.; Khanh, T.H.; Van Quang, P.; Tran, D.D.; Yen, L.M.; Van Doorn, R.; et al. Severity detection tool for patients with infectious disease. Healthc. Technol. Lett. 2020, 7, 45–50. [Google Scholar] [CrossRef] [PubMed]
  17. Tadesse, G.A.; Javed, H.; Thanh, N.L.N.; Thi, H.D.H.; Thwaites, L.; Clifton, D.A.; Zhu, T. Multi-modal diagnosis of infectious diseases in the developing world. IEEE J. Biomed. Health Inform. 2020, 24, 2131–2141. [Google Scholar] [CrossRef]
  18. Kiyasseh, D.; Tadesse, G.A.; Thwaites, L.; Zhu, T.; Clifton, D. Plethaugment: Gan-based ppg augmentation for medical diagnosis in low-resource settings. IEEE J. Biomed. Health Inform. 2020, 24, 3226–3235. [Google Scholar] [CrossRef]
  19. Tutuko, B.; Nurmaini, S.; Tondas, A.E.; Rachmatullah, M.N.; Darmawahyuni, A.; Esafri, R.; Firdaus, F.; Sapitri, A.I. AFibNet: An Implementation of Atrial Fibrillation Detection with Convolutional Neural Network. BMC Med. Inform. Decis. Mak. 2021, 21, 216. [Google Scholar] [CrossRef]
  20. Kiranyaz, S.; Avci, O.; Abdeljaber, O.; Ince, T.; Gabbouj, M.; Inman, D.J. 1D convolutional neural networks and applications: A survey. Mech. Syst. Signal Process. 2021, 151, 107398. [Google Scholar] [CrossRef]
  21. Ullah, A.; Anwar, S.M.; Bilal, M.; Mehmood, R.M. Classification of arrhythmia by using deep learning with 2-D ECG spectral image representation. Remote Sens. 2020, 12, 1685. [Google Scholar] [CrossRef]
  22. Zihlmann, M.; Perekrestenko, D.; Tschannen, M. Convolutional recurrent neural networks for electrocardiogram classification. In Proceedings of the 2017 Computing in Cardiology (CinC), Rennes, France, 24–27 September 2017; pp. 1–4. [Google Scholar]
  23. Diker, A.; Cömert, Z.; Avcı, E.; Toğaçar, M.; Ergen, B. A novel application based on spectrogram and convolutional neural network for ecg classification. In Proceedings of the 2019 1st International Informatics and Software Engineering Conference (UBMYK), Ankara, Turkey, 6–7 November 2019; pp. 1–6. [Google Scholar]
  24. Liu, G.; Han, X.; Tian, L.; Zhou, W.; Liu, H. ECG quality assessment based on hand-crafted statistics and deep-learned S-transform spectrogram features. Comput. Methods Programs Biomed. 2021, 208, 106269. [Google Scholar] [CrossRef]
  25. Wu, Y.; Yang, F.; Liu, Y.; Zha, X.; Yuan, S. A comparison of 1-D and 2-D deep convolutional neural networks in ECG classification. arXiv 2018, arXiv:1810.07088. [Google Scholar]
  26. Guo, M.H.; Xu, T.X.; Liu, J.J.; Liu, Z.N.; Jiang, P.T.; Mu, T.J.; Zhang, S.H.; Martin, R.R.; Cheng, M.M.; Hu, S.M. Attention Mechanisms in Computer Vision: A Survey. arXiv 2021, arXiv:2111.07624. [Google Scholar] [CrossRef]
  27. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  28. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  29. Wang, X.; Girshick, R.; Gupta, A.; He, K. Non-local neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7794–7803. [Google Scholar]
  30. Du, W.; Wang, Y.; Qiao, Y. Recurrent spatial-temporal attention network for action recognition in videos. IEEE Trans. Image Process. 2017, 27, 1347–1360. [Google Scholar] [CrossRef] [PubMed]
  31. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [PubMed]
  32. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
  33. Byeon, Y.H.; Kwak, K.C. Pre-configured deep convolutional neural networks with various time-frequency representations for biometrics from ECG signals. Appl. Sci. 2019, 9, 4810. [Google Scholar]
  34. Virtanen, P.; Gommers, R.; Oliphant, T.E.; Haberland, M.; Reddy, T.; Cournapeau, D.; Burovski, E.; Peterson, P.; Weckesser, W.; Bright, J.; et al. SciPy 1.0: Fundamental Algorithms for Scientific Computing in Python. Nat. Methods 2020, 17, 261–272. [Google Scholar] [CrossRef] [Green Version]
  35. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT press: Cambridge, MA, USA, 2016. [Google Scholar]
  36. ePatch™. The World’s Most Wearable Holter Monitor. Available online: https://www.gobio.com/clinical-research/cardiac-safety/epatch/ (accessed on 21 August 2022).
  37. Dorthe Bodholt, S.; Helge Bjarup Dissing, S.; Ingeborg Helbech, H.; Kenneth, E.; Poul, J.; Karsten, H. Available online: https://backend.orbit.dtu.dk/ws/portalfiles/portal/102966188/ePatch_A_Clinical_Overview_DTU_Technical_Report.pdf (accessed on 21 August 2022).
  38. Van, H.M.T.; Van Hao, N.; Quoc, K.P.N.; Hai, H.B.; Yen, L.M.; Nhat, P.T.H.; Duong, H.T.H.; Thuy, D.B.; Zhu, T.; Greeff, H.; et al. Vital sign monitoring using wearable devices in a Vietnamese intensive care unit. BMJ Innov. 2021, 7. [Google Scholar] [CrossRef]
  39. Fu, J.; Liu, J.; Tian, H.; Li, Y.; Bao, Y.; Fang, Z.; Lu, H. Dual attention network for scene segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3146–3154. [Google Scholar]
  40. Lu, P.; Qiu, H.; Qin, C.; Bai, W.; Rueckert, D.; Noble, J.A. Going deeper into cardiac motion analysis to model fine spatio-temporal features. In Proceedings of the Annual Conference on Medical Image Understanding and Analysis; Springer: Berlin/Heidelberg, Germany, 2020; pp. 294–306. [Google Scholar]
  41. Xingjian, S.; Chen, Z.; Wang, H.; Yeung, D.Y.; Wong, W.K.; Woo, W.c. Convolutional LSTM network: A machine learning approach for precipitation nowcasting. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; pp. 802–810. [Google Scholar]
  42. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  43. Lu, P.; Barazzetti, L.; Chandran, V.; Gavaghan, K.; Weber, S.; Gerber, N.; Reyes, M. Highly accurate facial nerve segmentation refinement from CBCT/CT imaging using a super-resolution classification approach. IEEE Trans. Biomed. Eng. 2017, 65, 178–188. [Google Scholar] [CrossRef]
  44. Vest, A.N.; Da Poian, G.; Li, Q.; Liu, C.; Nemati, S.; Shah, A.J.; Clifford, G.D. An open source benchmarked toolbox for cardiovascular waveform and interval analysis. Physiol. Meas. 2018, 39, 105004. [Google Scholar]
  45. Vollmer, M. HRVTool—An Open-Source Matlab Toolbox for Analyzing Heart Rate Variability. In Proceedings of the Computing in Cardiology 2019, Singapore, 8–11 September 2019; Volume 46. [Google Scholar] [CrossRef]
  46. Vollmer, M. A robust, simple and reliable measure of heart rate variability using relative RR intervals. In Proceedings of the 2015 Computing in Cardiology Conference (CinC), Nice, France, 6–9 September 2015; pp. 609–612. [Google Scholar] [CrossRef]
  47. Heart Rate Variability Analysis. Available online: https://pypi.org/project/hrv-analysis/ (accessed on 21 August 2022).
  48. Seven ECG Heartbeat Detection Algorithms and Heartrate Variability Analysis. Available online: https://www.ecdc.europa.eu/en/tetanus/facts (accessed on 21 August 2022).
  49. Touvron, H.; Cord, M.; Douze, M.; Massa, F.; Sablayrolles, A.; Jégou, H. Training data-efficient image transformers & distillation through attention. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual, 18–24 July 2021; pp. 10347–10357. Available online: https://proceedings.mlr.press/v139/touvron21a.html (accessed on 21 August 2022).
  50. Khan, S.; Naseer, M.; Hayat, M.; Zamir, S.W.; Khan, F.S.; Shah, M. Transformers in vision: A survey. ACM Comput. Surv. CSUR 2021. [Google Scholar] [CrossRef]
  51. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  52. Lopes, R.G.; Fenu, S.; Starner, T. Data-free knowledge distillation for deep neural networks. arXiv 2017, arXiv:1710.07535. [Google Scholar]
  53. Gou, J.; Yu, B.; Maybank, S.J.; Tao, D. Knowledge distillation: A survey. Int. J. Comput. Vis. 2021, 129, 1789–1819. [Google Scholar]
Figure 1. Overview of the proposed method and the visual explanations of features for the label 0-mild tetanus in the last layers of 6 convolutional blocks of the method. The 60-s window length log-spectrogram of raw ECG data is the input of the proposed method called 2D-CNN + Channel-wise Attention. The output of the proposed method is the label prediction, label 0 (mild tetanus) and label 1 (severe tetanus).
Figure 1. Overview of the proposed method and the visual explanations of features for the label 0-mild tetanus in the last layers of 6 convolutional blocks of the method. The 60-s window length log-spectrogram of raw ECG data is the input of the proposed method called 2D-CNN + Channel-wise Attention. The output of the proposed method is the label prediction, label 0 (mild tetanus) and label 1 (severe tetanus).
Sensors 22 06554 g001
Figure 2. Examples of ECG waveforms collected from tetanus patients, spectrograms and normalised log-spectrograms generated for each tetanus classification: scores 1 and 2 refer to mild symptoms and scores 3 and 4 refer to severe symptoms.
Figure 2. Examples of ECG waveforms collected from tetanus patients, spectrograms and normalised log-spectrograms generated for each tetanus classification: scores 1 and 2 refer to mild symptoms and scores 3 and 4 refer to severe symptoms.
Sensors 22 06554 g002
Figure 3. Illustration of the channel-wise attention mechanism structure. The channel-wise attention is added at the end of each convolutional block, which models the interdependencies among the channel features of the convolutional layers.
Figure 3. Illustration of the channel-wise attention mechanism structure. The channel-wise attention is added at the end of each convolutional block, which models the interdependencies among the channel features of the convolutional layers.
Sensors 22 06554 g003
Figure 4. Wearable monitor for ECG data acquisition; ePatch (left) and example of ePatch placement on the chest (right).
Figure 4. Wearable monitor for ECG data acquisition; ePatch (left) and example of ePatch placement on the chest (right).
Sensors 22 06554 g004
Figure 5. The architecture of deep learning methods which are used to compare with the proposed method. The methods from left to right, top: 2D-CNN; 2D-CNN + Channel-wise Attention + ConvLSTM; 2D-CNN + Channel-wise Attention + LSTM; bottom: 2D-CNN + LSTM; 2D-CNN + ConvLSTM; 2D-CNN + Dual Attention.
Figure 5. The architecture of deep learning methods which are used to compare with the proposed method. The methods from left to right, top: 2D-CNN; 2D-CNN + Channel-wise Attention + ConvLSTM; 2D-CNN + Channel-wise Attention + LSTM; bottom: 2D-CNN + LSTM; 2D-CNN + ConvLSTM; 2D-CNN + Dual Attention.
Sensors 22 06554 g005
Figure 6. Examples of visual explanations of the features in all different layers of the baseline 2D-CNN method and the proposed 2D-CNN + Channel-wise Attention method. The 60-s window length log-spectrogram of raw ECG data is the input of these two methods. The green rectangle highlights the huge visual difference between the adjacent layers in the proposed method.
Figure 6. Examples of visual explanations of the features in all different layers of the baseline 2D-CNN method and the proposed 2D-CNN + Channel-wise Attention method. The 60-s window length log-spectrogram of raw ECG data is the input of these two methods. The green rectangle highlights the huge visual difference between the adjacent layers in the proposed method.
Sensors 22 06554 g006
Figure 7. Confusion matrices of the tetanus severity level diagnosis using different deep learning methods: (a) tetanus confusion matrices for the methods using 60 s window length log-spectrograms without downsampling as inputs; (b) tetanus confusion matrices for the methods using 60 s window length log-spectrograms with downsampling as inputs; (c) tetanus confusion matrices for the 1D-CNN method (no time series images as inputs).
Figure 7. Confusion matrices of the tetanus severity level diagnosis using different deep learning methods: (a) tetanus confusion matrices for the methods using 60 s window length log-spectrograms without downsampling as inputs; (b) tetanus confusion matrices for the methods using 60 s window length log-spectrograms with downsampling as inputs; (c) tetanus confusion matrices for the 1D-CNN method (no time series images as inputs).
Sensors 22 06554 g007
Figure 8. Examples of visual explanations of the features in the last layer of each convolutional block and each attention layer of the proposed method. The log-spectrograms without downsampling are the inputs of the proposed method. These log-spectrograms are generated from the 50 s, 40 s, 30 s, 20 s, 10 s and 5 s window lengths of raw ECG data. The green rectangle highlights the huge visual difference between the adjacent layers.
Figure 8. Examples of visual explanations of the features in the last layer of each convolutional block and each attention layer of the proposed method. The log-spectrograms without downsampling are the inputs of the proposed method. These log-spectrograms are generated from the 50 s, 40 s, 30 s, 20 s, 10 s and 5 s window lengths of raw ECG data. The green rectangle highlights the huge visual difference between the adjacent layers.
Sensors 22 06554 g008
Table 1. Quantitative comparison on the proposed 2D-CNN + Channel-wise Attention method and the baseline methods. The results are presented as mean ± standard deviation. The best performance is indicated in bold.
Table 1. Quantitative comparison on the proposed 2D-CNN + Channel-wise Attention method and the baseline methods. The results are presented as mean ± standard deviation. The best performance is indicated in bold.
60 s Window Length Log-Spectrogram without Downsampling
MethodF1 ScorePrecisionRecallSpecificityAccuracyAUC
2D-CNN0.61 ± 0.140.68 ± 0.070.57 ± 0.190.85 ± 0.020.75 ± 0.070.72 ± 0.09
2D-CNN + Dual Attention0.65 ± 0.190.71 ± 0.170.61 ± 0.210.86 ± 0.090.76 ± 0.110.74 ± 0.13
2D-CNN + Channel-wise Attention0.79 ± 0.030.78 ± 0.080.82 ± 0.050.85 ± 0.080.84 ± 0.040.84 ± 0.03
2D-CNN + LSTM0.61 ± 0.150.71 ± 0.160.59 ± 0.200.83 ± 0.170.74 ± 0.100.71 ± 0.10
2D-CNN + ConvLSTM0.52 ± 0.320.77 ± 0.230.46 ± 0.330.95 ± 0.040.77 ± 0.110.71 ± 0.15
2D-CNN + Channel-wise Attention + ConvLSTM0.38 ± 0.170.67 ± 0.100.29 ± 0.160.92 ± 0.060.68 ± 0.050.60 ± 0.06
2D-CNN + Channel-wise Attention + LSTM0.59 ± 0.320.70 ± 0.340.56 ± 0.340.92 ± 0.920.79 ± 0.120.74 ± 0.16
No Time Series Images
MethodF1 ScorePrecisionRecallSpecificityAccuracyAUC
1D-CNN0.65 ± 0.140.61 ± 0.050.77 ± 0.250.70 ± 0.130.73 ± 0.050.74 ± 0.08
Table 2. Quantitative comparison on the proposed 2D-CNN + Channel-wise Attention method and the baseline method using downsampled log-spectrogram. The results are presented as mean ± standard deviation. The best performance is indicated in bold.
Table 2. Quantitative comparison on the proposed 2D-CNN + Channel-wise Attention method and the baseline method using downsampled log-spectrogram. The results are presented as mean ± standard deviation. The best performance is indicated in bold.
60 s Window Length Log-Spectrogram with Downsampling
MethodF1 ScorePrecisionRecallSpecificityAccuracyAUC
2D-CNN0.58 ± 0.160.68 ± 0.050.53 ± 0.190.85 ± 0.060.74 ± 0.050.69 ± 0.07
2D-CNN + Dual Attention0.54 ± 0.080.57 ± 0.170.57 ± 0.210.69 ± 0.230.65 ± 0.090.63 ± 0.06
2D-CNN + Channel-wise Attention0.60 ± 0.100.82 ± 0.100.51 ± 0.160.92 ± 0.080.77 ± 0.300.71 ± 0.05
2D-CNN + LSTM0.52 ± 0.120.67 ± 0.030.43 ± 0.140.88 ± 0.030.71 ± 0.040.66 ± 0.06
2D-CNN + Channel-wise Attention + LSTM0.63 ± 0.130.75 ± 0.050.56 ± 0.190.89 ± 0.040.77 ± 0.050.73 ± 0.08
Table 3. Quantitative comparison on window length of log-spectrograms as the inputs of the proposed 2D-CNN + Channel-wise Attention method. The results are presented as mean ± standard deviation. The best performance is indicated in bold.
Table 3. Quantitative comparison on window length of log-spectrograms as the inputs of the proposed 2D-CNN + Channel-wise Attention method. The results are presented as mean ± standard deviation. The best performance is indicated in bold.
The Proposed Method (Spectrograms without Downsampling)
Window DurationF1 ScorePrecisionRecallSpecificityAccuracyAUC
50 s0.81 ± 0.050.81 ± 0.060.82 ± 0.040.88 ± 0.040.86 ± 0.040.85 ± 0.04
40 s0.80 ± 0.040.84 ± 0.080.77 ± 0.070.91 ± 0.050.86 ± 0.030.84 ± 0.03
30 s0.74 ± 0.050.79 ± 0.070.79 ± 0.070.87 ± 0.060.84 ± 0.040.83 ± 0.04
20 s0.79 ± 0.050.80 ± 0.080.78 ± 0.070.88 ± 0.070.84 ± 0.040.83 ± 0.04
10 s0.55 ± 0.330.74 ± 0.160.45 ± 0.380.90 ± 0.060.77 ± 0.120.72 ± 0.17
5 s0.43 ± 0.320.98 ± 0.020.34 ± 0.290.99 ± 0.010.75 ± 0.100.67 ± 0.14
Table 4. List of extracted heart rate variability (HRV) features in traditional machine learning.
Table 4. List of extracted heart rate variability (HRV) features in traditional machine learning.
Parameters
HRV time domain features
mean_nnimean of RR-intervals
sdnnstandard deviation of RR-intervals
sdsdstandard deviation of differences between adjacent RR-intervals
rmssdsquare root of the mean of the sum of the squares of differences between adjacent NN-intervals
mean_hrmean Heart Rate
max_hrmax heart rate
min_hrmin heart rate
std_hrstandard deviation of heart rate
Table 5. Quantitative comparison of the proposed method (2D-CNN + Channel-wise Attention) and the baseline methods (traditional machine learning), using the original 60 s window length ECG as input. The results are presented as mean ± standard deviation. The best performance is indicated in bold.
Table 5. Quantitative comparison of the proposed method (2D-CNN + Channel-wise Attention) and the baseline methods (traditional machine learning), using the original 60 s window length ECG as input. The results are presented as mean ± standard deviation. The best performance is indicated in bold.
60 s Window Length Log-Spectrogram
MethodF1 ScorePrecisionRecallSpecificityAccuracyAUC
2D-CNN + Channel-wise Attention0.79 ± 0.030.78 ± 0.080.82 ± 0.050.85 ± 0.080.84 ± 0.040.84 ± 0.03
No Time Series Images
MethodF1 ScorePrecisionRecallSpecificityAccuracyAUC
Random Forest (HRV time domain features)0.81 ± 0.000.77 ± 0.000.85 ± 0.010.85 ± 0.000.85 ± 0.000.80 ± 0.00
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lu, P.; Ghiasi, S.; Hagenah, J.; Hai, H.B.; Hao, N.V.; Khanh, P.N.Q.; Khoa, L.D.V.; VITAL Consortium; Thwaites, L.; Clifton, D.A.; et al. Classification of Tetanus Severity in Intensive-Care Settings for Low-Income Countries Using Wearable Sensing. Sensors 2022, 22, 6554. https://doi.org/10.3390/s22176554

AMA Style

Lu P, Ghiasi S, Hagenah J, Hai HB, Hao NV, Khanh PNQ, Khoa LDV, VITAL Consortium, Thwaites L, Clifton DA, et al. Classification of Tetanus Severity in Intensive-Care Settings for Low-Income Countries Using Wearable Sensing. Sensors. 2022; 22(17):6554. https://doi.org/10.3390/s22176554

Chicago/Turabian Style

Lu, Ping, Shadi Ghiasi, Jannis Hagenah, Ho Bich Hai, Nguyen Van Hao, Phan Nguyen Quoc Khanh, Le Dinh Van Khoa, VITAL Consortium, Louise Thwaites, David A. Clifton, and et al. 2022. "Classification of Tetanus Severity in Intensive-Care Settings for Low-Income Countries Using Wearable Sensing" Sensors 22, no. 17: 6554. https://doi.org/10.3390/s22176554

APA Style

Lu, P., Ghiasi, S., Hagenah, J., Hai, H. B., Hao, N. V., Khanh, P. N. Q., Khoa, L. D. V., VITAL Consortium, Thwaites, L., Clifton, D. A., & Zhu, T. (2022). Classification of Tetanus Severity in Intensive-Care Settings for Low-Income Countries Using Wearable Sensing. Sensors, 22(17), 6554. https://doi.org/10.3390/s22176554

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop