Next Article in Journal
An Inductively Powered Implantable System to Study the Gastrointestinal Electrophysiology in Freely Behaving Rodents
Previous Article in Journal
VEGF Polymorphism rs3025039 and Human T-Cell Leukemia Virus 1 (HTLV-1) Infection among Older Japanese Individuals: A Cross-Sectional Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Predicting Analyte Concentrations from Electrochemical Aptasensor Signals Using LSTM Recurrent Networks

by
Fatemeh Esmaeili
1,
Erica Cassie
2,3,
Hong Phan T. Nguyen
2,3,
Natalie O. V. Plank
2,3,
Charles P. Unsworth
1,3 and
Alan Wang
4,5,6,*
1
Department of Engineering Science, University of Auckland, Auckland 1010, New Zealand
2
School of Chemical and Physical Sciences, Victoria University of Wellington, Wellington 6021, New Zealand
3
The MacDiarmid Institute for Advanced Materials and Nanotechnology, Victoria University of Wellington, Wellington 6021, New Zealand
4
Auckland Bioengineering Institute, University of Auckland, Auckland 1010, New Zealand
5
Faculty of Medical and Health Sciences, University of Auckland, Auckland 1010, New Zealand
6
Centre for Brain Research, University of Auckland, Auckland 1010, New Zealand
*
Author to whom correspondence should be addressed.
Bioengineering 2022, 9(10), 529; https://doi.org/10.3390/bioengineering9100529
Submission received: 5 September 2022 / Revised: 29 September 2022 / Accepted: 2 October 2022 / Published: 6 October 2022

Abstract

:
Nanomaterial-based aptasensors are useful devices capable of detecting small biological species. Determining suitable signal processing methods can improve the identification and quantification of target analytes detected by the biosensor and consequently improve the biosensor’s performance. In this work, we propose a data augmentation method to overcome the insufficient amount of available original data and long short-term memory (LSTM) to automatically predict the analyte concentration from part of a signal registered by three electrochemical aptasensors, with differences in bioreceptors, analytes, and the signals’ lengths for specific concentrations. To find the optimal network, we altered the following variables: the LSTM layer structure (unidirectional LSTM (LSTM) and bidirectional LSTM (BLSTM)), optimizers (Adam, RMSPROP, SGDM), number of hidden units, and amount of augmented data. Then, the evaluation of the networks revealed that the highest original data accuracy increased from 50% to 92% by exploiting the data augmentation method. In addition, the SGDM optimizer showed a lower performance prediction than that of the ADAM and RMSPROP algorithms, and the number of hidden units was ineffective in improving the networks’ performances. Moreover, the BLSTM nets showed more accurate predictions than those of the ULSTM nets on lengthier signals. These results demonstrate that this method can automatically detect the analyte concentration from the sensor signals.

Graphical Abstract

1. Introduction

Aptamer-based biosensors have been widely used in various fields, such as environmental monitoring [1], food quality and safety [2], and medical diagnostics and therapy [3], due to the physical and chemical features of aptamers for detection and small binding substances [4]. Moreover, the advancement in nanostructured materials has attracted much attention in recent years due to their potential applications and unique properties, including high reactivity, high functionalization, large surface-area-to-volume ratio, and small size [5]. Thus, advanced nanostructured materials have been utilized to improve the sensing capacities of aptasensors [6], lower the limits of detection of analytes [7], and amplify the sensors’ signals [8]. Nanomaterial-based aptasensors have been used as effective instruments for recognizing small analytes in clinical health diagnostics [6,7], medical therapy [9], and disease biomarker detection [10]. In addition, using and improving these analytical devices for identifying and quantifying a target analyte is beneficial due to their having higher specificity and selectivity and their elimination of labor-intensive and time-consuming procedures, expensive instruments, and multiple analytical steps [11].
Applications of machine learning (ML) algorithms have been widely used in healthcare as powerful tools for creating prediction models and making precise decisions [12]. Consequently, machine learning has significantly improved biosensors, such as through the analysis of sensing data for anomaly detection, noise reduction, classification, and pattern recognition [13]. Identifying ultra-low levels of biological species is a critical objective in improving biosensors in medical diagnostics and therapy [14].
Deep learning (DL) algorithms, as a subcategory of machine learning algorithms, have progressed remarkably on broad datasets with distinctive modalities, including time series, images, and natural languages [15]. Significant progress in deep learning has been beneficial for solving problems in many domains, including the medical and healthcare fields, and it has also defeated the conventional machine learning models [15]. For example, convolutional neural networks (CNNs) are a better fit for problems dealing with image processing, and recurrent neural networks (RNNs) are suitable for modeling problems that require the processing of time series or sequential data [16].
There are different RNN-based networks, and their main distinguishing feature is the difference in how they remember the input data [16]. For example, an original RNN is incapable of remembering past data, while a long short-term memory (LSTM) network is a modified version of RNN-based networks capable of remembering and learning from past data [16]. This means that LSTMs are suitable for making a prediction model when the datasets are in the form of time series due to their ability to learn temporal dependencies by employing a gating mechanism for data analysis [17]. Moreover, LSTM solves the vanishing gradient problem, unlike the original RNN [18].
LSTM networks have a broad range of applications in data processing and decision-making in healthcare [19,20]. Saltepe et al. [21] utilized two LSTM networks to detect and classify gold ion concentrations ranging from 0 to 100 μ M to decrease the time needed for the detection of gold ions, which ranged from 30 min to 3 h. The first network was a binary classification network and was designed to detect the existence of gold ions in the sample. This network detected the gold ions’ presence in the sample with prediction accuracies of 78% and about 98% from the 30-min and 3-h time series, respectively. The second LSTM network that was designed for classifying the gold concentrations showed a prediction accuracy of 82% from the 3-h time series. Klosowski et al. [22] applied two LSTM networks to classify six types of heart dysfunctions based on ECG time-series signals. The LSTM networks used the raw and the double spectral ECG data, and the accuracies of their predictions were 70.8% and 100%, respectively.
Some research has been conducted regarding the classification of biosensors’ signals [21,23]. The similarity among these studies is that the analyte concentration remained constant from the start to the end of the signal registration. This means that signals of different analyte concentration levels were recorded separately and independently. To the best of our knowledge, there has not been enough research on the detection of concentration from electrochemical biosensor signals in conditions in which signals are registered while the concentration of the analyte increases after a specific time and in which the sensing platform does not show that it is sensing the analyte anymore. Finding a suitable workflow for classifying these biosensor signals motivates this research, since it might ease the development of biosensor. Thus, this study presents a deep learning model composed of recurrent LSTM layers that are capable of classifying analyte concentrations from aptasensor signals.
In this paper, we utilized LSTM networks to analyze the transient signals from three nanomaterial-based aptasensors in order to detect the presence and concentrations of target analytes and automate signal classification regarding the analyte concentration. In addition, all computations and deep learning algorithms were implemented in MATLAB 2021b. The following describes the workflow followed in this paper. First, a preprocessing technique was applied to the sensor signals. Then, the preprocessed original signals were split into two categories: original data (OD) network data and control group data. In the next step, the OD network was used to make prediction models, and the control group was used to assess the prediction models. Then, a data augmentation method was developed to increase the data size. Then, both original and augmented data were used to make and train the LSTM models, and the test set and control group were responsible for assessing the performance of the prediction models.

2. Materials and Methods

This section describes the methods applied in this study, including those for data collection, data preprocessing, data augmentation, LSTM architecture and optimization, and evaluation of the prediction models.

2.1. Dataset Descriptions

The datasets used in this study were (1) 35-mer adenosine, (2) 31-mer oestradiol, and (3) 35-mer oestradiol. The datasets contained several time-series signals representing the drain current of three different aptasensors. Table 1 describes three key features of the sensors used for data collection to provide a quick and brief comparison of the datasets. The distinguishing features of these datasets were two main components of their sensors: their target analytes and their bioreceptors. However, their transducers, another main component used for these aptasensors, were carbon nanotube (CNT) field-effect transistors (FETs). Explaining all of details of the functionalization of these sensors is beyond the scope of this paper. However, detailed information on the 35-mer adenosine sensor, including transistor fabrication and aptamer functionalization, can be found in [24].
As the sensing protocols for the drain current measurements might provide a clear insight into the registered signals, the following explains the method of measuring the signals. The sensing protocols for measuring the 31-mer and 35-mer oestradiol sensors’ responses were similar, but they were different from those of the adenosine sensors. Table 2 summarizes and compares the sensing protocols of the adenosine and oestradiol datasets.
The sensing responses for adenosine and oestradiol were measured in time intervals of 1 and 1.081 s with a standard deviation of 5 × 10 3 , respectively, with similar gate and drain voltages, i.e., V G = 0 V and V D = 100 mV. The buffer selected for the adenosine sensor was 2 mM Tris-HCI, and that for the oestradiol sensors was 0.05 × PBS (phosphate-buffered saline) with 5% ethanol (EtOH).
Regarding the adenosine sensor, the initial load for each measurement was 110 μ M of 2 mM Tris-HCI in a polydimethylsiloxane (PDMS) well, which lasted for 1000 s. Then, the adenosine solution was added to the PDMS well every 500 s in successively greater concentrations, considering the adenosine concentration in the PDMS well before each addition. The process of adding the adenosine solution increased the adenosine concentration, which varied from 1 pM to 10 μ M in the PDMS well.
Regarding the oestradiol sensors, the initial load for each measurement was 100 μ L of 0.05 × PBS 5% EtOH in the well, which lasted for 300 s. Then, in the next 300 s, 20 μ L of 0.05 × PBS 5% EtOH was added, while the oestradiol concentration did not increase. Then, the oestradiol solution was added to the well every 300 s in successively greater concentrations, considering the oestradiol concentration in the well before each addition. In addition, each time the oestradiol concentration was increased, a solution of 20 μ L of 0.05 × PBS 5% EtOH was added to the well. The process of adding the oestradiol solution increased the oestradiol concentration, which varied from 1 nM to 10 μ M in the well.

2.2. Contextual Outlier Detection

A contextual outlier, also known as a conditional anomaly, is defined as a data instance whose pattern does not conform to that of other well-defined data instances with similar contextual information [27,28]. Regarding the experiments related to this study, factors that could cause contextual outliers were background noises in the lab, the use of broken transistors, issues in immobilizing the aptamers on carbon nanotube surfaces, issues in fabricating the sensing interface, and so on.
The patterns of signals affected by these factors deviate from the patterns of well-defined and normal signals. The purpose of removing outliers is to eliminate non-informative signals or segments. As there were a few signals in the datasets, removing the outliers was performed with prior knowledge of the biosensors’ behaviors and through data visualization.
In this paper, the signals were preprocessed with data normalization before being fed into the DL models. It needs to be mentioned that data normalization was applied to the entire signal. Data normalization or feature scaling puts all of the signals in a dataset on the same scale and prevents a feature from dominating and controlling the others. The data normalization applied in this paper was a Z-score scaling that used the mean ( μ ) and standard deviation ( σ ) of a signal.
Suppose that X = [ x 1 , x 2 , , x i , , x n ] is an entire signal, where the n is the number of data points within the given signal or the length of the signal. Then, Equation (1) shows the new signal X ^ created by Z-score scaling.
X ^ = [ x ^ 1 , x ^ 2 , , x ^ i = x i μ σ , , x ^ n ] .

Segmentation and Labeling

After rescaling the signal, it was split into different segments. Each segment was a part of the signal for which the concentration of analyte remained constant from its beginning to its end. Then, each segment was labeled with its corresponding analyte concentration. This means that the labels for the three datasets were: No Analyte, 1 nM, 10 nM, 100 nM, 1 μ M, and 10 μ M. As shown in Table 3, these six labels and concentrations were considered as the six different classes, in the same order.

2.3. Data Split

The data fed into the DL model needed to be split into three subsets, namely, the training, validation, and test set, with the proportions of 60%, 20%, and 20%, respectively. The training set was used to extract meaningful information and find the optimal parameters, the validation set was used to tune the parameters, and the test set was used to assess the model’s performance [13].
In this paper, the original data were split into two sets—network and test sets—with proportions of approximately 70% and 30%, respectively. These sets were named the original data (OD) network and OD test set, respectively. The former set was used to make DL models based on original data and for data augmentation. The latter, the OD test set, assessed the DL models and acted as a control group. The reasons for considering this OD test set were to prove the functionality of the data augmentation method in making the prediction models and to avoid biased results. In order to complete the information related to the data split, it must be mentioned that the augmented data and OD networks were randomly shuffled and separated again into the training set (60%), validation set (20%), and network test set (20%).

2.4. Data Augmentation

In machine learning, small amounts of training data might cause overfitting and might not be enough for training models [29]. The need for data augmentation is more critical for real-world data, since acquiring large enough real-world datasets has not always been possible due to cost or time limitations. Generating synthetic data, which is also known as data augmentation, is a solution for overcoming the problem of insufficient data samples [30] or compensating for datasets with imbalanced classes [31]. Data augmentation helps to increase the generalization capability of an ML prediction model and improve the model’s performance by increasing the variability of the training and validation data samples [32,33].
In this paper, we utilized a data augmentation method to increase the size of the available datasets. Suppose the S ^ i and S ^ j are two preprocessed segments from an identical dataset with similar analyte concentrations. Then, S a u g is an augmented segment generated with the following Equation (2):
S a u g = w × S ^ i + ( 1 w ) × S ^ j ,
where w ( 0 , 1 ) and is a normally distributed random number generated by the r a n d n function in MATLAB R2021b.

2.5. Background of LSTM

This subsection explains long short-term memory (LSTM) and its architecture. Then, the unidirectional and bidirectional LSTM structures, as well as their similarities and differences, are discussed.
The advantage of using an LSTM network over a recurrent neural network (RNN) is that LSTM can capture the temporal dependency of input sequences during the training process [21,34]. An LSTM network is an RNN that prevents the long-term dependency problem by utilizing gates and calculating a hidden state with an enhanced function [17]. The building blocks of LSTM networks are LSTM cells, which means that an LSTM layer consists of recurrently connected cells [16,17,34]. An LSTM cell, or an LSTM hidden unit, consists of four parts: the forget gate, input gate, output gate, and a cell candidate. Figure 1 presents the structure of a cell. This cell decides to ignore or remember something in its memory by using a gating mechanism. The role of the three gates is to selectively control and transfer needed information into and out of the cell [34]. This figure can also be considered an LSTM layer consisting of only one memory cell or hidden unit, where X t and h t , respectively, are the input and output of the LSTM layer.
Consider X t as a sequence input into the memory block at time step t; the forget gate selects which data to erase and which data to remember. As shown in Equation (3), these decisions are made by the s i g m o i d layer:
f t = σ ( W f x X t + W f h h t 1 + b f ) .
The input gate is responsible for controlling the level at which the cell state is updated by using another s i g m o i d layer. As shown in Equation (4), the s i g m o i d layer of the input gate decides which data need to be updated. In the next step, as shown in Equation (5), the cell candidate ( c ˜ t ) is responsible for adding information to the cell state by using the t a n h layer. Now, the cell state is ready to be updated with the combination of the forget and input gates and new candidate values of c ˜ t . Equation (6) describes the mathematical formula for calculating the cell state:
i t = σ ( W i x X t + W i h h t 1 + b i ) ,
c ˜ t = t a n h ( W c x X t + W c h h t 1 + b c ) ,
c t = f t × c t 1 + i t × c ˜ t .
The output gate, which is shown in Equation (7), utilizes a s i g m o i d layer to decide which part of the cell state contributes to the output. Now, the hidden state or output of the memory cell is ready to be calculated. The output gate and the cell state are contributors to the hidden state. Equation (8) presents its mathematical formula:
o t = σ ( W o x X t + W o h h t 1 + b o ) ,
h t = o t × t a n h ( c t ) .
Note that W i x , W f x , W o x , and W c x refer to the input weight matrices for the input gate, forget gate, output gate, and cell value, respectively, and W i h , W f h , W o h , and W c h are the recurrent weights for the gates and the cell value in the same order. Their corresponding bias vectors are b i , b f , b o , and b c .
Moreover, it can be seen that the cell state and gate activation functions (AFs) are, respectively, t a n h (Equation (9)) and s i g m o i d (Equation (10)); these map the nonlinearity and make decisions:
t a n h ( z ) = e 2 z 1 e 2 z + 1 ,
σ ( z ) = 1 1 + e z .
An LSTM layer in a deep neural network consists of a set of LSTM cells. LSTM layers can be categorized into unidirectional LSTM (ULSTM) and bidirectional LSTM (BLSTM) layers. Figure 2 represents a ULSTM structure. It can be said that a ULSTM structure is an RNN that uses LSTM cells instead. The unfolded figure of the ULSTM shows that the output of each cell is the input for the next cell in the same layer. It should be mentioned that an LSTM block refers to several LSTM cells or hidden units.
Figure 3 depicts a BLSTM structure consisting of forward and backward layers. The unfolded figure shows that the forward layer moves in a positive temporal direction, while the backward movement is in a negative temporal direction. In addition, the outputs from both the forward and backward LSTM cells are joined and concatenated as the layer’s output.
Figure 4 presents the flow of information in an LSTM layer during different time steps. In this figure, N is the length of the sequential input for the LSTM layer, L is the number of hidden units in the LSTM layer, and T is the length of the training set. Note that in this figure, h t can be considered as just the forward movement ( h t ) in the ULSTM layer or as a concatenation of both forward ( h t ) and backward ( h t ) movements in the BLSTM layer.

2.6. LSTM Architecture

In this work, two LSTM networks were employed to classify the analyte concentrations. The objective was to classify the input data into six different concentration classes: 0 M, 1 nM, 10 nM, 100 nM, 1 μ M , and 10 μ M. The target outputs of each class were labeled in a binary vector format, where the desired class was labeled with “1” and the others were labeled with “0”. Recall that the input data were the corresponding concentration segments of the signals, as well as the original and/or augmented segments.
Figure 5 visualizes the architectures of both networks. The networks comprised five successive layers: a sequential input layer, an LSTM layer, a fully connected layer, a softmax layer, and a classification layer. The only difference between the networks was in their LSTM layers. The LSTM layer in the first network was a unidirectional LSTM layer, while this was a bidirectional LSTM layer in the second network. It should be taken into consideration that the fully connected (FC) layers were affected by the previous LSTM layers and the number of output classes in the classification layer.
Table 4 describes and compares the layers and the properties of the unidirectional and bidirectional LSTM networks depicted in Figure 5. Recall that the size of the input layer entering the networks was equal to the length of the segments and was considered as one sequence, and the output size of the networks (m) was identical to the number of classes in the data. It should be mentioned that all of the input weights, the recurrent weight, and the bias matrices were concatenated together to form the input weights ( W x = [ W i x ; W f x ; W c x ; W o x ] ), recurrent weights ( W h = [ W i h ; W f h ; W c h ; W o h ] ), and bias ( b = [ b i ; b f ; b c ; b o ] ).

2.7. LSTM Optimization

We adopted the following variables to find the optimal prediction model: the number of hidden units in the LSTM layers, the optimizers (Adam, RMSPROP, and SGDM) [35], and the number of segments fed into the networks. Initially, the original datasets were used to train the networks with the total size of a segment. The altered variables were: the number of hidden units in the LSTM layers—starting from 50 and increasing to 500 in increments of 50—and the use of the three optimizers mentioned. In the next step, we used both the original and augmented signals to train the networks. The sums of the original and augmented segments per class were 50 and 100.
It should be noted that all deep learning algorithms were implemented with the MATLAB R2021b Deep Learning Toolbox.

2.8. Evaluation Metrics

After training the prediction models, the classification performance of the neural networks needed to be assessed with relevant metrics. In this paper, the LSTM networks’ performances were assessed using two standard metrics: overall accuracy (ACC) and macro F1-score (MF1) [21,36]. In this work, we created a confusion matrix using the predictions from the test data, and then the overall accuracy and the macro F1-score were calculated with the confusion matrix.
The overall accuracy was calculated from the sum of the diagonal numbers of the confusion matrix divided by the total number of elements. In other words, the overall accuracy (Equation (11)) was the proportion of correctly classified elements among all elements in the test set.
ACC = Number of true testing outputs Total number of elements in the test set .
Before mentioning the method for calculating the macro F1-score, its building blocks—recall, precision, and F1-score—need to be defined. The recall, which is the true positive rate (TPR) and is presented in Equation (12), is the number of correctly classified positive elements among all positive elements. Precision, which is the positive predicted value (PPV) and is presented in Equation (13), is the number of correctly classified positive elements among the elements classified as positives by the model. Then, the harmonic mean of the recall and precision is called the F1-score (Equation (14)). The macro F1-score (Equation (15)) is the mean of the class-wise F1-score of each concentration.
Recall = TPR = TP TP + FN ,
Precision = PPV = TN TP + FP ,
F 1 - score = 2 1 / precision + 1 / recall ,
Macro F 1 - score = MF 1 = 1 m i = 1 m { F 1 - score } i ,
where m is the number of classes in a given dataset.

3. Results

3.1. Datasets

Figure 6 presents the typical raw signals of the available datasets. Vertical dashed lines separate the different analyte concentrations (ACs). Note that the initial ACs for the 35-mer adenosine experiments were not necessarily identical. Figure 6A,B represent the drain current in two experiments from the adenosine dataset in which the initial AC for the former experiment was 1 nM and that for the latter one was 1 μ M. However, the initial ACs for all of the experiments of the 31-mer and 35-mer oestradiol datasets were completely similar; these are depicted in Figure 6C,D, respectively.
In addition, we need to clarify the notions of an entire signal and a segment, as these notions will be repeatedly used in the rest of this study. An entire signal refers to all of the data points registered from the beginning to the end of an experimental measurement. For example, in Figure 6A, the entire signal comprises the data points at t [ 1 , 3500 ] . On the other hand, a segment refers to a part of a signal that represents the sensor response regarding a specific analyte concentration. For example, Figure 6B contains three segments: the No Analyte segment for t [ 1 , 1000 ] , the 1 μ M segment for t [ 1001 , 1500 ] , and the 10 μ M segment for t [ 1501 , 2000 ] .

3.2. Contextual Outlier Detection

Figure 7 compares the normal signals with the contextual outliers. Figure 7A shows the normal pattern of the 35-mer oestradiol signals that were registered with a well-fabricated device. It can be seen that these signals had similar trends when the analyte concentration increased. On the other hand, Figure 7B shows that the signals that did not conform to normal signals. The red and blue lines were registered with non-sensing and broken transistor devices. All of the signals that were registered by the non-sensing and broken transistor devices were completely removed from the datasets.
In addition, Figure 7C shows two instances of signals that showed temporal abnormalities although the sensors were normal and other parts of the signals showed reasonable sensing. In this figure, the drain currents registered between 300–750 s were considered the contextual outliers. Thus, there was partial removal of the signals that were related to the abnormal segments. Consequently, the other segments of the signals remained in their relevant datasets for utilization in the deep learning models.
Figure 7. The plots compare the normal behavior in the 35-mer oestradiol signals with the behavior of the contextual outliers: (A) signals with normal patterns; (B) signals with contextual outliers (the red and blue lines in the plot represent signals generated by non-sensing and broken transistor devices, respectively); (C) abnormal temporal signals that were registered with normal sensors but showed unusual behavior between 300 and 750 s.
Figure 7. The plots compare the normal behavior in the 35-mer oestradiol signals with the behavior of the contextual outliers: (A) signals with normal patterns; (B) signals with contextual outliers (the red and blue lines in the plot represent signals generated by non-sensing and broken transistor devices, respectively); (C) abnormal temporal signals that were registered with normal sensors but showed unusual behavior between 300 and 750 s.
Bioengineering 09 00529 g007
It should be noted that removing the segments that showed temporal abnormalities was performed after data normalization, since the mean and standard deviation used for data normalization depended on the information of an entire signal.
Table 5 shows the number of available normal segments in each dataset that were considered normal and suitable for utilization in the DL networks. However, as there were insufficient segments of adenosine with a concentration of 100 pM, these segments were excluded from the adenosine dataset.

3.3. Data Preprocessing

Figure 8 presents and compares the instances of preprocessed signals according to the Z-score scaling. Recall that this preprocessing method was applied to the entire signal. Figure 8A shows a raw signal from the 35-mer adenosine dataset, and Figure 8B shows the preprocessed signal according to Equation (1).

3.4. Data Split

Table 6 presents the numbers of segments used in the OD network and the OD test sets in each dataset after the original data split. The network sets were used for feeding into the networks, and the OD test sets, which consisted of original data, were used to assess the networks. The data split in the oestradiol datasets was done by considering three and four entire signals for the 31-mer and 35-mer as their test sets, respectively. However, the data split in the 35-mer adenosine dataset was not as straightforward as that for the oestradiol datasets. The OD test set was selected segment-wise.

3.5. Data Augmentation

Figure 9 presents the augmented segments generated from the normalized drain currents.

3.6. LSTM Optimization

Table 7 fully describes the model hyperparameters for optimizing the ULSTM and BLSTM. Moreover, the MATLAB Deep Learning Toolbox set the other hyperparameters and functions that are not mentioned in this table to their default values. For example, the function used to initialize the input weights was the default function, i.e., the g l o r o t weight initialization function.

3.7. Evaluation

It was mentioned that we used the overall accuracy (ACC) and macro F1-score (MF1) as the performance metrics. Moreover, all of the networks were assessed with two datasets: the network test set and the OD test dataset. This means that there were two tables of evaluation results for each dataset after altering the effective parameters for optimization. Thus, these tables were organized according to the network and OD test datasets.
Table 8 and Table 9 present the performance metrics of the test and control data, respectively, from the networks trained with the 35-mer adenosine dataset. Table 10 and Table 11 present the performance metrics for the network and control sets from the networks trained with the 31-mer oestradiol dataset, respectively. Table 12 and Table 13 present the evaluation metrics for the test data when trained with the 35-mer oestradiol signals.
For a better understanding of the tables, the boldface numbers indicate the highest performance of a network based on the optimizers and the LSTM layer structure for 50 and 100 augmented segments. Moreover, we ignored the best classifiers when the numerical metrics for almost all hidden units were similar. For example, in Table 9, the best classifiers were not selected for accuracy when using 50 segments in the ULSTM and BLSTM structures and the SGDM optimizer. The reason for this was that the use of more than four hidden units resulted in the same accuracy.
Figure 10 shows the overall accuracy of the OD test set from the network trained with the original segments from the 35-mer adenosine dataset. The accuracies of the prediction models trained with the original data varied approximately from 30% to 55% for the three datasets.
Figure 11 depicts the effect of increasing the number of augmented segments on the prediction models. This figure shows the accuracy of the test data from the networks trained with the ADAM optimizer and the 35-mer adenosine dataset. In general, it can be seen that the performance of the prediction models significantly improved by utilizing the data augmentation method. However, the prediction models were slightly improved with the increase in the amount of augmented data.
Figure 12 shows the overall accuracy from the control data of the ULSTM and BLSTM networks trained with the 35-mer adenosine dataset when 100 augmented segments per class were used.
In more detail, regarding the 35-mer adenosine dataset, as shown in Table 8 and Table 9, the MF1 for the original data changed approximately from 50% to 72%. Then, the ACC and MF1 reached 85% and 94%, respectively, by utilizing the augmented segments for the control data, as shown in Table 9. For the 31-mer oestradiol dataset, Table 10 and Table 11 show that the accuracy reached 90% and the MF1 increased approximately from 50–83% to 54–93%. For the 35-mer oestradiol dataset, Table 12 and Table 13 depict that the accuracy range of 25–50% reached 70% by using the augmented data, and the MF1 increased from the range of 30–63% to 76% after data augmentation.

4. Discussion

Nanomaterial-based aptasensors are useful biosensors that are capable of detecting small chemicals and species. A vital goal in the advancement of biosensors is the identification and measurement of low levels of target analytes. Deep learning methods are attractive tools for the advancement of biosensor technology and the analysis biosensing data. RNNs exploit temporal information in time-series inputs to make prediction models for classification and regression problems.
In our work, we successfully employed LSTM networks to automatically predict analyte concentrations from parts of drain current signals registered by three different electrochemical aptasensors. The differences in these sensors were their bioreceptors, their analytes, and the lengths of the signals for specific concentrations. Among RNN-based models, LSTM networks, which contain a gating schema for data analysis, are suitable models due to their ability to learn temporal dependencies. Thus, we utilized ULSTM and BLSTM networks with different optimizers and hidden units to identify the optimal classification models for various concentrations.
Moreover, similarly to most real-world problems, the available signals registered by these sensors were insufficient for training the networks. To overcome this limitation, we proposed a data augmentation method in order to increase the size of the available datasets and improve the prediction model’s generalization ability and overall performance. The augmentation method improved the model’s performance.
By comparing the evaluation results obtained with the original and augmented data, it can be seen that the applied data augmentation method significantly improved the classification performance. In addition, the results showed that the number of hidden units might not be very effective in enhancing LSTM models. Considering the optimizers, the LSTM networks that used the SGDM optimization algorithm showed a lower prediction performance than that obtained with the ADAM and RMSPROP algorithms [35].
According to the results showing the networks’ performance on the test data, the BLSTM networks used for the three datasets presented more accurate classification than that of the ULSTM networks [37]. However, this result was not identical to that obtained with the OD test data on the 31-mer and 35-mer oestradiol datasets, i.e., the prediction performance of the BLSTM models on both the test and control segments of the 35-mer adenosine dataset was higher than that of the ULSTM structures, but not for the oestradiol datasets. The reason for this failure to improve the performance of the BLSTM over that of the ULSTM for the oestradiol datasets might be the shorter length of the oestradiol segments compared to that of the adenosine segments.
Regarding the datasets, we observed that the least accurate models belonged to the 35-mer oestradiol dataset, and the most accurate networks were trained with the 35-mer adenosine dataset. This low accuracy might have occurred for one of two reasons: the aptamer length used to detect the analyte or the criteria chosen for the detection of contextual outliers. It was shown that successful detection of the analyte depended on the aptamer length, and the analyte–aptamer binding must occur within the Debye length to affect a CNT FET’s drain current [38]. Thus, the 35-mer oestradiol signals might not have shown obvious evidence of sensing. In the case that this low-accuracy model might be the result of the detection of contextual outliers, it can be assessed in a future study.
In future work, we can evaluate the effects of the segment length and automatic detection of contextual outliers on the prediction models. In order to address the unequal size of segments in future studies, data augmentation methods based on artificial neural networks might be a solution. In fact, by defining a regression model and employing deep learning networks [39], we can extend the length of available time-series data. Consequently, this technique can produce new datasets containing segments of similar lengths. Regarding the detection of contextual outliers, neural networks such as autoencoder-based [40] or LSTM-based models [41] can be utilized to assess and compare the effects of automatic anomaly detection on the prediction models.
The insufficiency of the data exposed a limitation in the preprocessing method applied in this work. Typical data preprocessing methods apply identical changes to the training and test sets. However, we used Z-score scaling for each signal based on its mean and standard deviation. It was impossible to estimate the mean and standard deviation with the available statistical methods, such as with simulations and hypothesis tests.

5. Conclusions

In this work, we exploited a data augmentation method and LSTM networks to analyze the drain current signals of three similar CNT FET aptasensors in the 35-mer adenosine dataset and 31-mer and 35-mer oestradiol datasets. The drain current signals reflected the sensing responses of the aptasensors, while the concentrations of target analytes successively increased from 1 nM to 10 μ M. The ultimate purpose of analyzing the signals was to automatically detect and classify the analyte concentration according to the corresponding sensing response.
The scaling-based data augmentation method was proven to be highly effective for improving and increasing the generalization ability of LSTM-based classification models. In addition, the results suggest that the applied data augmentation method might be more effective and useful in capturing the features from lengthier signals, e.g., 35-mer adenosine signals.
Moreover, the LSTM networks were successful in sensing response classification and in predicting the analyte concentration. In addition, the results suggest that using BLSTM networks does not necessarily result in making more accurate prediction models when using augmented data. Higher scores for evaluation metrics, namely, accuracy and MF1, of the BLSTM over the ULSTM might have resulted due to overfitting. Thus, using a control group when evaluating prediction models seems to be vital in order to obtain robust results. In addition, changing the number of hidden units was not effective with respect to the performance of the prediction models, and the SGDM algorithm was not a suitable optimisation algorithm for them.

Author Contributions

Conceptualization, F.E., N.O.V.P., C.P.U. and A.W.; methodology, F.E.; software, F.E.; validation, F.E.; formal analysis, F.E.; investigation, F.E. and A.W.; resources—computing resources and analysis tools, F.E.; resources—laboratory samples, E.C. and H.P.T.N.; resources, N.O.V.P.; data curation, E.C.; writing—original draft preparation, F.E.; writing—review and editing, F.E., N.O.V.P., C.P.U. and A.W.; visualization, F.E.; supervision, A.W.; funding acquisition, N.O.V.P. and C.P.U. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Marsden Fund Council using government funding, managed by the Royal Society of New Zealand, grant number 5000357.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study might be available on request from the corresponding author. There are restrictions on data availability due to their necessity for our future work.

Acknowledgments

The authors would like to thank Erica Happe for her support. The authors acknowledge that Erica Happe determined the contextual outliers of the 35-mer adenosine dataset.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
MLMachine Learning
DLDeep Learning
CNNConvolution Neural Network
RNNRecurrent Neural Networks
LSTMLong Short-Term Memory
ODOriginal Data
CNTCarbon Nanotube
FETField-Effect Transistor
PDMSPolydimethylsiloxane
AFActivation Function
ULSTMUnidirectional Long Short-Term Memory
BLSTMBidirectional Long Short-Term Memory
FCFully Connected
ACAnalyte Concentrations

References

  1. Satish, S.; Dey, A.; Tharmavaram, M.; Khatri, N.; Rawtani, D. Risk assessment of selected pharmaceuticals on wildlife with nanomaterials based aptasensors. Sci. Total. Environ. 2022, 836, 155622. [Google Scholar] [CrossRef] [PubMed]
  2. Nodoushan, S.M.; Nasirizadeh, N.; Sedighian, H.; Kachuei, R.; Azimzadeh-Taft, M.; Fooladi, A.A.I. Detection of Staphylococcal Enterotoxin A (SEA) using a sensitive nanomaterial-based electrochemical aptasensor. Diam. Relat. Mater. 2022, 127, 109042. [Google Scholar] [CrossRef]
  3. Omage, J.I.; Easterday, E.; Rumph, J.T.; Brula, I.; Hill, B.; Kristensen, J.; Ha, D.T.; Galindo, C.L.; Danquah, M.K.; Sims, N.; et al. Cancer Diagnostics and Early Detection Using Electrochemical Aptasensors. Micromachines 2022, 13, 522. [Google Scholar] [CrossRef] [PubMed]
  4. Lan, Y.; He, B.; Tan, C.S.; Ming, D. Applications of Smartphone-Based Aptasensor for Diverse Targets Detection. Biosensors 2022, 12, 477. [Google Scholar] [CrossRef] [PubMed]
  5. Chen, Z.; Xie, M.; Zhao, F.; Han, S. Application of Nanomaterial Modified Aptamer-Based Electrochemical Sensor in Detection of Heavy Metal Ions. Foods 2022, 11, 1404. [Google Scholar] [CrossRef] [PubMed]
  6. Yousef, H.; Liu, Y.; Zheng, L. Nanomaterial-Based Label-Free Electrochemical Aptasensors for the Detection of Thrombin. Biosensors 2022, 12, 253. [Google Scholar] [CrossRef] [PubMed]
  7. Sadeghi, M.; Kashanian, S.; Naghib, S.M.; Arkan, E. A high-performance electrochemical aptasensor based on graphene-decorated rhodium nanoparticles to detect HER2-ECD oncomarker in liquid biopsy. Sci. Rep. 2022, 12, 3299. [Google Scholar] [CrossRef] [PubMed]
  8. Park, J.A.; Kwon, N.; Park, E.; Kim, Y.; Jang, H.; Min, J.; Lee, T. Electrochemical biosensor with aptamer/porous platinum nanoparticle on round-type micro-gap electrode for saxitoxin detection in fresh water. Biosens. Bioelectron. 2022, 210, 114300. [Google Scholar] [CrossRef]
  9. Ayodele, O.O.; Adesina, A.O.; Pourianejad, S.; Averitt, J.; Ignatova, T. Recent Advances in Nanomaterial-Based Aptasensors in Medical Diagnosis and Therapy. Nanomaterials 2021, 11, 932. [Google Scholar] [CrossRef]
  10. Ma, R.; Gopinath, S.C.; Lakshmipriya, T.; Chen, Y. Carbon Material Hybrid Construction on an Aptasensor for Monitoring Surgical Tumors. J. Anal. Methods Chem. 2022, 2022. [Google Scholar] [CrossRef]
  11. Chen, Y.; Xiang, J.; Liu, B.; Chen, Z.; Zuo, X. Gold nanoparticle-engineered electrochemical aptamer biosensor for ultrasensitive detection of thrombin. Anal. Methods 2020, 12, 3729–3733. [Google Scholar] [CrossRef]
  12. Alanazi, A. Using machine learning for healthcare challenges and opportunities. Inform. Med. Unlocked 2022, 2022, 100924. [Google Scholar] [CrossRef]
  13. Cui, F.; Yue, Y.; Zhang, Y.; Zhang, Z.; Zhou, H.S. Advancing biosensors with machine learning. ACS Sens. 2020, 5, 3346–3364. [Google Scholar] [CrossRef]
  14. Zhou, K.; Liu, Y. Early-Stage Gas Identification Using Convolutional Long Short-Term Neural Network with Sensor Array Time Series Data. Sensors 2021, 21, 4826. [Google Scholar] [CrossRef]
  15. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  16. Siami-Namini, S.; Tavakoli, N.; Namin, A.S. The performance of LSTM and BiLSTM in forecasting time series. In Proceedings of the 2019 IEEE International Conference on Big Data (Big Data), Los Angeles, CA, USA, 9–12 December 2019; IEEE: Pisacatway, NJ, USA, 2019; pp. 3285–3292. [Google Scholar]
  17. Alawneh, L.; Mohsen, B.; Al-Zinati, M.; Shatnawi, A.; Al-Ayyoub, M. A comparison of unidirectional and bidirectional lstm networks for human activity recognition. In Proceedings of the 2020 IEEE International Conference on Pervasive Computing and Communications Workshops (PerCom Workshops), Austin, TX, USA, 23–27 March 2020; IEEE: Pisacatway, NJ, USA, 2020; pp. 1–6. [Google Scholar]
  18. Venugopalan, S.; Xu, H.; Donahue, J.; Rohrbach, M.; Mooney, R.; Saenko, K. Translating videos to natural language using deep recurrent neural networks. arXiv 2014, arXiv:1412.4729. [Google Scholar]
  19. Ali, O.M.A.; Kareem, S.W.; Mohammed, A.S. Evaluation of Electrocardiogram Signals Classification Using CNN, SVM, and LSTM Algorithm: A review. In Proceedings of the 2022 8th International Engineering Conference on Sustainable Technology and Development (IEC), Erbil, Iraq, 23–24 February 2022; IEEE: Pisacatway, NJ, USA, 2022; pp. 185–191. [Google Scholar]
  20. Saadatnejad, S.; Oveisi, M.; Hashemi, M. LSTM-based ECG classification for continuous monitoring on personal wearable devices. IEEE J. Biomed. Health Inform. 2019, 24, 515–523. [Google Scholar] [CrossRef] [Green Version]
  21. Saltepe, B.; Bozkurt, E.U.; Güngen, M.A.; Çiçek, A.E.; Şeker, U.Ö.Ş. Genetic circuits combined with machine learning provides fast responding living sensors. Biosens. Bioelectron. 2021, 178, 113028. [Google Scholar] [CrossRef]
  22. Kłosowski, G.; Rymarczyk, T.; Wójcik, D.; Skowron, S.; Cieplak, T.; Adamkiewicz, P. The use of time-frequency moments as inputs of lstm network for ecg signal classification. Electronics 2020, 9, 1452. [Google Scholar] [CrossRef]
  23. Banerjee, M.B.; Roy, R.B.; Tudu, B.; Bandyopadhyay, R.; Bhattacharyya, N. Black tea classification employing feature fusion of E-Nose and E-Tongue responses. J. Food Eng. 2019, 244, 55–63. [Google Scholar] [CrossRef]
  24. Nguyen, H.P.T.; Murugathas, T.; Plank, N.O. Comparison of Duplex and Quadruplex Folding Structure Adenosine Aptamers for Carbon Nanotube Field Effect Transistor Aptasensors. Nanomaterials 2021, 11, 2280. [Google Scholar] [CrossRef]
  25. Jauset-Rubio, M.; Botero, M.L.; Skouridou, V.; Aktas, G.B.; Svobodova, M.; Bashammakh, A.S.; El-Shahawi, M.S.; Alyoubi, A.O.; O’Sullivan, C.K. One-pot SELEX: Identification of specific aptamers against diverse steroid targets in one selection. ACS Omega 2019, 4, 20188–20196. [Google Scholar] [CrossRef]
  26. Alsager, O.A.; Kumar, S.; Zhu, B.; Travas-Sejdic, J.; McNatty, K.P.; Hodgkiss, J.M. Ultrasensitive colorimetric detection of 17β-estradiol: The effect of shortening DNA aptamer sequences. Anal. Chem. 2015, 87, 4201–4209. [Google Scholar] [CrossRef]
  27. Chandola, V.; Banerjee, A.; Kumar, V. Anomaly detection: A survey. ACM Comput. Surv. (CSUR) 2009, 41, 1–58. [Google Scholar] [CrossRef]
  28. Liang, J.; Parthasarathy, S. Robust contextual outlier detection: Where context meets sparsity. In Proceedings of the 25th ACM International on Conference on Information and Knowledge Management, Online Conference, 24 October 2016; pp. 2167–2172. [Google Scholar]
  29. Oh, C.; Han, S.; Jeong, J. Time-series data augmentation based on interpolation. Procedia Comput. Sci. 2020, 175, 64–71. [Google Scholar] [CrossRef]
  30. Olson, M.; Wyner, A.; Berk, R. Modern neural networks generalize on small data sets. Adv. Neural Inf. Process. Syst. 2018, 31. [Google Scholar]
  31. Hasibi, R.; Shokri, M.; Dehghan, M. Augmentation scheme for dealing with imbalanced network traffic classification using deep learning. arXiv 2019, arXiv:1901.00204. [Google Scholar]
  32. Goubeaud, M.; Joußen, P.; Gmyrek, N.; Ghorban, F.; Kummert, A. White Noise Windows: Data Augmentation for Time Series. In Proceedings of the 2021 7th International Conference on Optimization and Applications (ICOA), Wolfenbuttel, Germany, 19–20 May 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–5. [Google Scholar]
  33. Iwana, B.K.; Uchida, S. An empirical survey of data augmentation for time series classification with neural networks. PLoS ONE 2021, 16, e0254841. [Google Scholar] [CrossRef]
  34. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  35. Tuncer, E.; Bolat, E.D. Classification of epileptic seizures from electroencephalogram (EEG) data using bidirectional short-term memory (Bi-LSTM) network architecture. Biomed. Signal Process. Control. 2022, 73, 103462. [Google Scholar] [CrossRef]
  36. Dong, H.; Supratak, A.; Pan, W.; Wu, C.; Matthews, P.M.; Guo, Y. Mixed neural network approach for temporal sleep stage classification. IEEE Trans. Neural Syst. Rehabil. Eng. 2017, 26, 324–333. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Schuster, M.; Paliwal, K.K. Bidirectional recurrent neural networks. IEEE Trans. Signal Process. 1997, 45, 2673–2681. [Google Scholar] [CrossRef] [Green Version]
  38. Zheng, H.Y.; Alsager, O.A.; Wood, C.S.; Hodgkiss, J.M.; Plank, N.O. Carbon nanotube field effect transistor aptasensors for estrogen detection in liquids. J. Vac. Sci. Technol. Nanotechnol. Microelectron. Mater. Process. Meas. Phenom. 2015, 33, 06F904. [Google Scholar] [CrossRef]
  39. Demir, S.; Mincev, K.; Kok, K.; Paterakis, N.G. Data augmentation for time series regression: Applying transformations, autoencoders and adversarial networks to electricity price forecasting. Appl. Energy 2021, 304, 117695. [Google Scholar] [CrossRef]
  40. Kieu, T.; Yang, B.; Guo, C.; Jensen, C.S.; Zhao, Y.; Huang, F.; Zheng, K. Robust and explainable autoencoders for unsupervised time series outlier detection. In Proceedings of the 38th IEEE International Conference on Data Engineering, ICDE 2022, Online Conference, 9–11 May 2022. [Google Scholar]
  41. Provotar, O.I.; Linder, Y.M.; Veres, M.M. Unsupervised anomaly detection in time series using lstm-based autoencoders. In Proceedings of the 2019 IEEE International Conference on Advanced Trends in Information Theory (ATIT), Kyiv, Ukraine, 18–20 December 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 513–517. [Google Scholar]
Figure 1. Illustration of an LSTM cell. Note that ⊙ and ⊕ refer to element-wise multiplication and addition, respectively.
Figure 1. Illustration of an LSTM cell. Note that ⊙ and ⊕ refer to element-wise multiplication and addition, respectively.
Bioengineering 09 00529 g001
Figure 2. Flow of information in a unidirectional LSTM layer during different time steps with forward states, where X, Y, and h refer to the input, output, and forward states in the LSTM layer, respectively.
Figure 2. Flow of information in a unidirectional LSTM layer during different time steps with forward states, where X, Y, and h refer to the input, output, and forward states in the LSTM layer, respectively.
Bioengineering 09 00529 g002
Figure 3. Flow of information in a bidirectional LSTM layer during different time steps with forward and backward states, where X, Y, h , and h refer to the input, output, forward state, and backward state in the LSTM layer, respectively.
Figure 3. Flow of information in a bidirectional LSTM layer during different time steps with forward and backward states, where X, Y, h , and h refer to the input, output, forward state, and backward state in the LSTM layer, respectively.
Bioengineering 09 00529 g003
Figure 4. An unfolded LSTM layer that presents the flow of information during different time steps.
Figure 4. An unfolded LSTM layer that presents the flow of information during different time steps.
Bioengineering 09 00529 g004
Figure 5. A schematic of the network architecture. Note that the only difference between the two networks was in their LSTM layer, which used either a unidirectional or a bidirectional LSTM layer.
Figure 5. A schematic of the network architecture. Note that the only difference between the two networks was in their LSTM layer, which used either a unidirectional or a bidirectional LSTM layer.
Bioengineering 09 00529 g005
Figure 6. The typical raw signals of the datasets, (A) The 35-mer adenosine dataset with an initial concentration of 1 nM, (B) the 35-mer adenosine signal with an initial concentration of 1 μ M, (C) the 31-mer oestradiol dataset, (D) the 35-mer oestradiol dataset. Note that in these plots, AC refers to the analyte concentration.
Figure 6. The typical raw signals of the datasets, (A) The 35-mer adenosine dataset with an initial concentration of 1 nM, (B) the 35-mer adenosine signal with an initial concentration of 1 μ M, (C) the 31-mer oestradiol dataset, (D) the 35-mer oestradiol dataset. Note that in these plots, AC refers to the analyte concentration.
Bioengineering 09 00529 g006aBioengineering 09 00529 g006b
Figure 8. Preprocessed signal for the 35-mer adenosine dataset according to the Z-score scaling: (A) the raw signal; (B) the preprocessed signal.
Figure 8. Preprocessed signal for the 35-mer adenosine dataset according to the Z-score scaling: (A) the raw signal; (B) the preprocessed signal.
Bioengineering 09 00529 g008
Figure 9. Samples of augmented normalized segments from the available datasets: (A) augmented segments from the 35-mer adenosine segments with a concentration of 100 nM, (B) augmented segments from the 31-mer oestradiol segments with a concentration of 1 μ M, and (C) augmented segments from the 35-mer oestradiol segments with a concentration of 10 μ M.
Figure 9. Samples of augmented normalized segments from the available datasets: (A) augmented segments from the 35-mer adenosine segments with a concentration of 100 nM, (B) augmented segments from the 31-mer oestradiol segments with a concentration of 1 μ M, and (C) augmented segments from the 35-mer oestradiol segments with a concentration of 10 μ M.
Bioengineering 09 00529 g009aBioengineering 09 00529 g009b
Figure 10. Accuracy of the networks trained with the original 35-mer adenosine dataset: (A) ULSTM network; (B) BLSTM network.
Figure 10. Accuracy of the networks trained with the original 35-mer adenosine dataset: (A) ULSTM network; (B) BLSTM network.
Bioengineering 09 00529 g010
Figure 11. Accuracy of the test data from the networks trained with the 35-mer adenosine dataset and the ADAM optimizer: (A) ULSTM network; (B) BLSTM network.
Figure 11. Accuracy of the test data from the networks trained with the 35-mer adenosine dataset and the ADAM optimizer: (A) ULSTM network; (B) BLSTM network.
Bioengineering 09 00529 g011
Figure 12. Accuracy of the control data from the networks trained with the 35-mer adenosine dataset with 100 augmented and original data points per class: (A) ULSTM network; (B) BLSTM network.
Figure 12. Accuracy of the control data from the networks trained with the 35-mer adenosine dataset with 100 augmented and original data points per class: (A) ULSTM network; (B) BLSTM network.
Bioengineering 09 00529 g012aBioengineering 09 00529 g012b
Table 1. The datasets and the main components of the sensors that recorded the signals in each dataset.
Table 1. The datasets and the main components of the sensors that recorded the signals in each dataset.
Dataset NameAnalyteTransducerBioreceptorRef
35-mer AdenosineAdenosineCNT FET 5 -NH2-AAAAAAAAAACCTGGGGGAGTATTGCGGAGGAAGG- 3 [24]
31-mer OestradiolOestradiolCNT FET 5 -GGTCCTGACCGGAGGCTGACCGGAGTGGGAA- 3 [25] 1
35-mer OestradiolOestradiolCNT FET 5 -AAGGGATGCCGTTTGGGCCCAAGTTCGGCATAGTG- 3 [26]
1 The bioreceptor of this dataset was created by Erica S. Cassie and was a modification of that in the mentioned reference. This sequence took the common part of the best three oestradiol aptamers, and then some extra mers were added on either side.
Table 2. Comparison of the sensing protocols of the adenosine and oestradiol aptasensors. Note that the sensing protocols for both the 31-mer and 35-mer oestradiol aptasensors were completely the same. Thus, their relevant information is merged into one column.
Table 2. Comparison of the sensing protocols of the adenosine and oestradiol aptasensors. Note that the sensing protocols for both the 31-mer and 35-mer oestradiol aptasensors were completely the same. Thus, their relevant information is merged into one column.
CharacteristicsAdenosine AptasensorOestradil Aptasensors
Time interval of easurement1 s1.081 s with std 5 × 10 3
Gate voltage ( V G )0 V0 V
Drain voltage ( V D )100 mV100 mV
Buffer2 mM Tris-HCI0.05 × PBS with 5% EtOH
Initial step load chemical110 μ M of 2 mM Tris-HCI100 μ L of 0.05 × PBS 5% EtOH
Next steps load chemical-20 μ L of 0.05 × PBS 5% EtOH
Initial load time1000 s300 s
Time interval of adding analyte500 s300 s
Time interval of adding chemical-300 s
Analyte concentration range1 pM–10 μ M1 nM–10 μ M
Table 3. The available segments’ labels and their corresponding classes.
Table 3. The available segments’ labels and their corresponding classes.
LabelNo Analyte1 nM10 nM100 nM1 μ M10 μ M
Classclass 1class 2class 3class 4class 5class 6
Table 4. Layer description of the LSTM deep learning model.
Table 4. Layer description of the LSTM deep learning model.
Layer IDLayer TypeHyperparametersLearnable ParametersState Parameters
1Sequential inputOutput size:    1--
2ULSTMInput size:    1
Hidden units:    n W x : 4 n × 1 Hidden state: n × 1
Output size:    n W h : 4 n × n
State AF:    tanh b : 4 n × 1 Cell state: n × 1
Gate AF:    sigmoid
BLSTMInput size:    1
Hidden units:    n W x : 8 n × 1 Hidden state: 2 n × 1
Output size:     2 n W h : 8 n × n
State AF :    tanh b : 8 n × 1 Cell state: 2 n × 1
Gate AF:    sigmoid
3FC (ULSTM) 1Input size: nWeights: m × n
Output size: mBias: m × 1 -
FC (BLSTM) 2Input size: 2 n Weights: m × 2 n
Output size: mBias: m × 1 -
4Softmax---
5Output classification---
1 Fully connected layer that follows a ULSTM layer. 2 Fully connected layer that follows a BLSTM layer.
Table 5. The number of available segments for each dataset after removing the contextual outlier signals and segments.
Table 5. The number of available segments for each dataset after removing the contextual outlier signals and segments.
Analyte Concentration35-mer Adenosine31-mer Oestradil35-mer Oestradil
No Analyte996
100 pM1--
1 nM496
10 nM5912
100 nM7912
1 μ M9912
10 μ M9912
Table 6. The number of available segments for the OD network and the OD test sets that were relevant to each dataset.
Table 6. The number of available segments for the OD network and the OD test sets that were relevant to each dataset.
Analyte
Concentration
35-mer Adenosine31-mer Oestradil35-mer Oestradil
OD NetworkOD TestOD NetworkOD TestOD NetworkOD Test
No Analyte636342
1 nM316342
10 nM416384
100 nM526384
1 μ M636384
10 μ M636384
Table 7. The model hyperparameters for both the ULSTM and BLSTM neural networks with three optimizers.
Table 7. The model hyperparameters for both the ULSTM and BLSTM neural networks with three optimizers.
OptimiserADAMRMSPROPSGDM
Gradient decay factor0.9--
Squared gradient decay factor0.90.9-
Momentum--0.9
Initial learning rate0.0050.0050.005
Learning rate schedulepiecewisepiecewisepiecewise
Learning rate drop factor0.20.20.2
Learning rate drop period555
L 2 Regularization0.010.010.01
Maximum epochs505050
Minimum batch size202020
Shuffleevery epochevery epochevery epoch
Table 8. Performance metrics of the test data from networks trained with the 35-mer adenosine dataset.
Table 8. Performance metrics of the test data from networks trained with the 35-mer adenosine dataset.
ACCMF1
DatasetHidden UnitsAdamRMSPROPSGDMAdamRMSPROPSGDM
ULSTMBLSTMULSTMBLSTMULSTMBLSTMULSTMBLSTMULSTMBLSTMULSTMBLSTM
Original Data500.3330.3330.1670.5000.3330.3330.5000.5000.5000.7220.5830.500
1000.3330.3330.3330.3330.3330.3330.5000.5000.5000.5000.5330.583
1500.3330.5000.5000.3330.3330.3330.5330.7220.7220.5000.5330.583
2000.3330.3330.3330.3330.3330.3330.5830.5000.5330.5000.5000.500
2500.3330.3330.3330.3330.5000.3330.5000.5000.5330.5000.7220.500
3000.3330.3330.3330.3330.3330.5000.5000.5830.5330.5000.5330.722
3500.3330.3330.3330.5000.5000.3330.5000.5000.5000.7220.7220.500
4000.3330.3330.3330.5000.3330.3330.5330.5000.5330.7220.5330.500
4500.3330.3330.3330.3330.3330.3330.5330.5000.5330.5000.5330.500
5000.3330.3330.3330.3330.3330.3330.5330.5000.6670.5000.5330.500
50 Segments500.7830.8000.8830.8670.4500.4000.8640.7570.8860.8660.4980.523
1000.7500.8830.8500.8170.3500.3830.7070.8610.8030.7730.4190.393
1500.7330.9000.6670.8500.3330.3500.6830.8960.8330.8440.5440.430
2000.7830.8330.6330.8000.3500.3500.7690.8040.6810.7940.4110.431
2500.8500.8170.7670.8000.3500.3500.8420.8140.8510.7910.4150.426
3000.8500.8000.7670.8500.3500.3500.8420.7630.7360.8370.4370.432
3500.7500.9000.7500.8000.3500.3670.8200.9000.7040.7890.4320.356
4000.8500.8500.8170.5500.3500.3670.8030.8440.8080.6730.4180.362
4500.5330.7670.7830.7000.3500.3500.6420.7550.7640.6500.4280.419
5000.8830.8000.3330.8170.3670.3500.8850.7940.5490.9110.3670.419
100 Segments500.8500.8250.8330.8750.5420.4580.8440.8170.8280.8690.5300.450
1000.8170.8670.8330.8750.6420.4080.8100.8680.8260.8720.5870.440
1500.8250.9170.8580.9170.6420.5250.8170.9130.8530.9140.5910.453
2000.5670.8580.8170.9250.6250.5330.5230.8590.8060.9240.5640.553
2500.8500.8830.8420.8420.6420.6250.8430.8810.8360.8430.6020.566
3000.8500.9000.6420.8250.6750.6170.8480.8980.7010.8150.6410.652
3500.8170.8500.4920.8170.6920.6500.8090.8450.7130.8040.6700.597
4000.8500.9330.8500.9080.7000.6420.8460.9320.8430.9090.6770.593
4500.8250.9080.7500.7920.7000.6250.8130.9060.7080.7840.6800.568
5000.7000.8170.8420.7580.7000.6250.7590.8070.8350.7290.6750.567
Table 9. Performance metrics of the OD test (control) data from the networks trained with the 35-mer adenosine dataset.
Table 9. Performance metrics of the OD test (control) data from the networks trained with the 35-mer adenosine dataset.
ACCMF1
DatasetHidden UnitsAdamRMSPROPSGDMAdamRMSPROPSGDM
ULSTMBLSTMULSTMBLSTMULSTMBLSTMULSTMBLSTMULSTMBLSTMULSTMBLSTM
Original Data500.4620.4620.3850.4620.3080.3850.6790.6790.4710.7010.4310.610
1000.4620.4620.4620.4620.3080.4620.6790.6790.6790.6790.4310.729
1500.3850.4620.4620.4620.3080.6150.5640.7010.7010.6790.4140.726
2000.4620.4620.3080.4620.3850.4620.7010.6790.4310.6790.5640.679
2500.3850.4620.3080.5380.3080.4620.5640.6790.4140.6340.4730.679
3000.4620.4620.3080.4620.3080.5380.6790.6790.4140.6790.4140.619
3500.4620.4620.3850.6150.3850.4620.6790.6790.5830.7380.4390.679
4000.3080.4620.3850.6150.3080.4620.4140.6790.5640.6290.4140.679
4500.3080.4620.3080.4620.3080.4620.4140.6790.4140.6790.4140.679
5000.3080.4620.5380.4620.3080.4620.4140.6790.6340.6790.4140.679
50 Segments500.4620.7690.5380.6920.3850.4620.6060.8860.6130.7790.6100.679
1000.5380.5380.5380.6920.4620.4620.7310.7070.7070.8170.6790.679
1500.5380.4620.4620.6920.4620.4620.7380.6210.6100.7140.6790.679
2000.6150.7690.4620.8460.4620.4620.7110.8050.5630.9030.6790.679
2500.5380.7690.3850.6920.4620.4620.5850.8030.5370.6980.6790.679
3000.6150.6920.3850.6150.4620.4620.6730.8210.5370.6970.7010.679
3500.5380.6920.4620.6920.4620.4620.6060.8170.7560.8690.7010.679
4000.6150.7690.5380.6150.4620.4620.7040.7760.6310.7380.6790.679
4500.3080.6920.6920.4620.4620.4620.4500.6650.7810.7290.7010.679
5000.6920.6150.3850.5380.4620.4620.8110.7710.5640.6170.7010.679
100 Segments500.6150.5380.4620.6920.3080.4620.6050.6380.5130.7790.4500.729
1000.5380.3850.6150.6920.3080.4620.6430.4940.6510.6980.4730.701
1500.6920.6150.7690.7690.3080.4620.7250.7300.7540.7850.5000.701
2000.3850.8460.6150.6920.3850.4620.4710.8910.6890.7300.5000.701
2500.6150.7690.6920.5380.3850.4620.7140.8100.8110.6070.5000.729
3000.5380.7690.3850.4620.3080.4620.6130.7850.5150.6060.5000.729
3500.6920.6920.3080.8460.3850.4620.8000.8170.4670.8710.5000.729
4000.5380.9230.6150.6150.3850.4620.6060.9430.7140.7140.5000.729
4500.5380.9230.6150.5380.3850.4620.6480.9430.6050.6170.5000.729
5000.3850.8460.7690.5380.3850.4620.5150.8910.7760.5790.5000.729
Table 10. Performance metrics of the test data from the networks trained with the 31-mer Oestrdiol dataset.
Table 10. Performance metrics of the test data from the networks trained with the 31-mer Oestrdiol dataset.
ACCMF1
DatasetHidden UnitsAdamRMSPROPSGDMAdamRMSPROPSGDM
ULSTMBLSTMULSTMBLSTMULSTMBLSTMULSTMBLSTMULSTMBLSTMULSTMBLSTM
Original Data500.3330.3330.5000.5000.3330.3330.5330.5330.7780.8330.6670.533
1000.1670.3330.3330.5000.3330.3330.6670.5330.5000.8330.7500.533
1500.5000.5000.3330.5000.3330.1670.7220.8330.5330.7220.5330.400
2000.3330.3330.3330.5000.3330.3330.6670.5830.5830.8000.7000.583
2500.5000.5000.3330.5000.3330.3330.7220.7220.6670.7220.7000.533
3000.3330.5000.3330.3330.5000.3330.5000.8330.6670.5330.8000.533
3500.3330.3330.3330.3330.3330.3330.7500.5830.7000.5330.7000.533
4000.3330.5000.5000.5000.1670.3330.6670.6670.7780.8000.3330.533
4500.3330.5000.3330.3330.1670.5000.5000.7220.7000.5330.5000.833
5000.3330.5000.5000.3330.3330.3330.5330.7220.8000.5330.6670.533
50 Segments500.6170.8330.6000.8330.3330.3500.8220.9330.6400.8390.5330.457
1000.6170.9000.7170.8000.3830.3670.6650.8870.6850.8880.5180.496
1500.8330.8330.5500.8000.3330.4000.9330.9330.6630.8880.5330.442
2000.6170.8330.6000.8000.3330.3500.7540.9330.6080.8930.5370.421
2500.7330.8830.7500.8330.3330.3830.8180.8640.8150.9330.5330.415
3000.5830.9170.5170.7170.3330.3500.7480.9090.6300.7860.5330.432
3500.5500.8670.7170.7670.3500.3330.6670.8410.7940.7560.4250.530
4000.5000.8330.7500.8330.3500.3670.7220.9330.8220.9200.4250.364
4500.5500.8330.3330.7170.3330.3670.6670.9330.5700.8260.5330.478
5000.7000.4830.1670.3330.3500.3330.7770.5980.2900.5510.4250.530
100 Segments500.7080.8670.6670.8670.6330.5580.7900.8350.7460.8330.8330.683
1000.8000.8830.7830.8330.6830.5330.8760.8670.8590.9210.7600.564
1500.7080.8670.7080.8420.7000.5250.7800.8390.7940.7980.7830.553
2000.6580.8670.8000.8670.6420.5750.7110.8350.8800.8340.6980.625
2500.7000.8830.7920.8670.6080.6500.7790.8640.7470.8360.6540.610
3000.6420.8330.7170.8580.6330.5920.6330.8010.8000.8200.6850.539
3500.7580.9000.6580.8670.6000.6170.8340.8930.7300.8350.6410.578
4000.6420.9000.3330.8080.6580.6420.6900.8900.5420.7520.7270.604
4500.7080.8670.7920.8580.6080.6080.7900.8420.8700.8200.6540.556
5000.7670.8330.7920.8580.6420.6420.8480.9210.7530.8210.7060.607
Table 11. Performance metrics of the OD test (control) data from the networks trained with the 31-mer Oestrdiol dataset.
Table 11. Performance metrics of the OD test (control) data from the networks trained with the 31-mer Oestrdiol dataset.
ACCMF1
DatasetHidden UnitsAdamRMSPROPSGDMAdamRMSPROPSGDM
ULSTMBLSTMULSTMBLSTMULSTMBLSTMULSTMBLSTMULSTMBLSTMULSTMBLSTM
Original Data500.3330.3330.4440.3890.3330.3330.5330.5330.6940.5830.6050.533
1000.3890.3330.3330.5560.3330.3330.5190.5330.5030.7080.7500.533
1500.3330.5560.3330.4440.2220.2780.5030.7080.6050.6190.3190.417
2000.3330.4440.2780.6110.3890.3330.6050.6560.5000.8150.6250.583
2500.5000.5560.3330.5000.3330.3330.7220.6560.6760.6900.6290.533
3000.3330.3890.3330.3330.3330.3890.5230.5280.6050.5330.6050.550
3500.3890.3330.3330.3890.1670.3330.6670.6060.6050.5220.3530.548
4000.3330.4440.5560.5000.1670.3330.6050.5950.6930.8000.3530.533
4500.2780.5000.3330.3330.1670.5000.4390.6820.7000.5330.5000.833
5000.2220.5000.3330.3330.3330.3330.3190.6900.6050.5330.6050.533
50 Segments500.6670.7220.6110.6110.3330.3330.8750.7960.7650.5970.5330.564
1000.6110.7780.7780.6110.3890.3890.7890.8640.8550.6780.5320.543
1500.7780.7780.5000.5560.3330.3890.8710.8640.6900.5970.5480.532
2000.7220.7780.5560.7220.3330.3330.7710.8710.7000.8000.5330.564
2500.7220.7780.6670.7780.3330.3890.8000.8710.7310.8640.5330.543
3000.5000.8330.6110.5000.3330.3330.7020.8190.6640.5730.5330.548
3500.5560.7220.7220.6670.3330.3890.6780.8100.7640.7520.5480.543
4000.6110.7220.6670.7780.3330.3890.7890.8000.7360.8640.5480.532
4500.5000.6670.3330.6670.3330.3330.6900.7140.6050.7800.5330.564
5000.6670.5000.1670.3330.3890.3890.7310.8210.2860.5640.5320.543
100 Segments500.6110.7780.7220.7780.6670.4440.8360.8640.8000.8640.8750.656
1000.8330.7220.8330.7220.6670.5000.9330.8170.9210.7960.7310.649
1500.8330.7780.7220.7220.6670.5560.9210.9000.7880.7960.8750.599
2000.6670.7780.7780.7780.6110.6110.8750.8640.8710.8640.8010.681
2500.6670.7780.7780.7780.6110.6110.8510.7580.8860.8640.7890.681
3000.6670.7780.7220.7780.6110.6110.7670.7700.8000.8710.7890.681
3500.7780.7220.7220.7780.6110.6110.8530.7170.7880.8710.7890.681
4000.6670.7780.2780.7220.6670.6110.8750.8640.4860.7150.7390.681
4500.7220.7780.7780.7780.6110.6110.8090.8640.8710.8710.7890.681
5000.7220.7780.7780.7780.6110.6110.8000.9000.7660.8640.7890.681
Table 12. Performance metrics of the test data from the networks trained with the 35-mer Oestrdiol dataset.
Table 12. Performance metrics of the test data from the networks trained with the 35-mer Oestrdiol dataset.
ACCMF1
DatasetHidden UnitsAdamRMSPROPSGDMAdamRMSPROPSGDM
ULSTMBLSTMULSTMBLSTMULSTMBLSTMULSTMBLSTMULSTMBLSTMULSTMBLSTM
Original Data500.3000.4000.4000.5000.2000.5000.4500.5460.5790.6670.5000.656
1000.2000.4000.2000.4000.4000.3000.3430.5670.3430.5790.5830.395
1500.3000.4000.3000.4000.3000.3000.4860.5330.4860.5460.5000.429
2000.4000.4000.3000.4000.2000.3000.5460.5330.4860.5460.5000.429
2500.3000.5000.3000.3000.3000.3000.5360.6890.4500.6190.5000.429
3000.4000.5000.3000.5000.2000.3000.5560.5920.4330.5920.5000.429
3500.4000.5000.2000.4000.2000.5000.5790.5920.3670.5460.5000.635
4000.5000.4000.3000.5000.4000.3000.5470.5460.4500.6000.5830.429
4500.4000.5000.3000.4000.4000.4000.5220.5920.4500.5460.5830.544
5000.3000.4000.3000.4000.4000.3000.4500.5460.4500.5460.5830.429
50 Segments500.6170.6170.6170.6170.3670.4330.6460.6600.6650.6680.4590.511
1000.6670.6500.6670.6500.3330.4830.7290.7060.7250.6290.5080.547
1500.6830.7670.6670.6670.3330.4830.7470.7470.7290.6400.5130.553
2000.6830.7000.5830.6170.3330.4170.7470.6850.6360.6400.5130.480
2500.6830.6670.6170.6670.3830.5000.6470.6240.5790.7220.5030.600
3000.7500.7330.6500.6670.3330.4170.7480.7240.6900.7070.5180.508
3500.6170.7500.6000.6670.3330.4500.6770.7310.6650.7220.5800.509
4000.7000.7330.4170.5170.3330.4500.6960.7230.5590.5450.5300.656
4500.6330.5670.3330.6330.3330.5000.5880.7330.5120.6280.5080.564
5000.5670.8330.6330.5500.3330.4500.6140.8360.6900.5170.5420.627
100 Segments500.6330.7420.6830.7500.6000.5000.6320.7180.6210.7330.6110.480
1000.7420.9000.6920.7830.4920.5250.7260.8990.6300.7790.4890.638
1500.7670.8670.6830.7830.5250.5330.7530.8620.6050.7830.5260.530
2000.7750.8420.7250.8920.5920.5500.7370.8410.6950.8920.6300.557
2500.7920.8670.7250.8580.4920.5250.7740.8650.6990.8570.6120.541
3000.8000.8500.7500.8080.6170.6000.7840.8450.6940.7970.6570.644
3500.7420.9330.5830.8000.5330.5670.7220.9330.7090.7780.5490.600
4000.5750.9170.6000.6830.6420.6000.5730.9160.6040.7350.6980.640
4500.8250.7830.6000.6500.6080.6330.8010.7770.6430.6990.6350.580
5000.6170.8580.6000.7830.5920.5920.6140.8580.5930.7750.6210.532
Table 13. Performance metrics of the OD test (control) data from the networks trained with the 35-mer Oestrdiol dataset.
Table 13. Performance metrics of the OD test (control) data from the networks trained with the 35-mer Oestrdiol dataset.
ACCMF1
DatasetHidden UnitsAdamRMSPROPSGDMAdamRMSPROPSGDM
ULSTMBLSTMULSTMBLSTMULSTMBLSTMULSTMBLSTMULSTMBLSTMULSTMBLSTM
Original Data500.3500.4500.3500.4500.2000.3000.5170.6130.4690.5960.2670.462
1000.3000.4000.2000.3500.3000.3000.4720.5480.3930.4680.3330.445
1500.3500.4000.3500.4500.2000.3500.4720.5220.5110.6130.5000.522
2000.4500.4000.3500.4000.1500.3500.6070.5430.4560.5380.3530.500
2500.4000.4000.3000.3500.2500.3500.5370.5430.6000.5310.3470.522
3000.3000.4500.3500.4000.1500.3500.4000.6320.4200.6040.3330.500
3500.2500.4500.3500.4500.2000.3500.4220.5390.5320.6130.2710.508
4000.2000.4500.2500.4500.1500.3500.3010.6380.4260.6270.3000.508
4500.3000.4500.2500.4500.2500.2500.4890.6710.5500.6390.3220.298
5000.2500.4500.2500.4500.3000.3500.5880.6390.4530.6270.3830.500
50 Segments500.6500.5500.6000.5500.3000.4000.7560.6440.6870.6270.4310.598
1000.6000.5000.6500.6500.3000.4000.7050.5680.7400.6350.4720.561
1500.6500.5500.6500.6500.3000.4000.7500.5760.7450.6350.5170.598
2000.6500.6000.7000.5500.3500.4000.7500.6220.7850.7170.4440.578
2500.5500.5500.5000.5500.3500.4000.6070.6440.5500.6270.4760.482
3000.5500.5000.6000.5500.3000.4000.6080.5580.6730.6440.5000.598
3500.5500.5500.5500.5000.3000.4000.6070.6440.6430.6430.5580.561
4000.7500.5500.3500.6000.3000.4000.7490.5930.4470.6900.5170.561
4500.5500.5500.2500.6000.3500.4000.6340.6270.4830.6250.4440.561
5000.6000.6000.6000.3500.3000.4000.7000.6380.6880.5390.5000.561
100 Segments500.6000.5500.5500.5500.6000.4500.6650.5760.6530.5760.6750.594
1000.6000.5500.5500.6000.5500.4500.6140.5760.5450.6050.6480.613
1500.6000.6000.5500.5500.5000.4500.6130.6380.5420.6700.5770.613
2000.6000.5500.6000.6000.5500.4500.6060.5900.5870.6410.6480.613
2500.6500.6000.6000.6500.5000.5500.6980.6380.5930.7010.7130.644
3000.6000.5006000.6000.6000.4000.6310.5030.6360.6050.6070.522
3500.5500.5500.5500.5500.5000.4500.5930.5380.7170.6300.6970.613
4000.5000.6000.6000.6000.6000.4000.6080.6200.5610.6410.6070.522
4500.6500.5000.7000.6000.6000.4000.6550.6570.6910.6650.6950.522
5000.6000.5500.6000.5000.5500.4500.6670.5900.6650.6100.6560.613
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Esmaeili, F.; Cassie, E.; Nguyen, H.P.T.; Plank, N.O.V.; Unsworth, C.P.; Wang, A. Predicting Analyte Concentrations from Electrochemical Aptasensor Signals Using LSTM Recurrent Networks. Bioengineering 2022, 9, 529. https://doi.org/10.3390/bioengineering9100529

AMA Style

Esmaeili F, Cassie E, Nguyen HPT, Plank NOV, Unsworth CP, Wang A. Predicting Analyte Concentrations from Electrochemical Aptasensor Signals Using LSTM Recurrent Networks. Bioengineering. 2022; 9(10):529. https://doi.org/10.3390/bioengineering9100529

Chicago/Turabian Style

Esmaeili, Fatemeh, Erica Cassie, Hong Phan T. Nguyen, Natalie O. V. Plank, Charles P. Unsworth, and Alan Wang. 2022. "Predicting Analyte Concentrations from Electrochemical Aptasensor Signals Using LSTM Recurrent Networks" Bioengineering 9, no. 10: 529. https://doi.org/10.3390/bioengineering9100529

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop