Next Article in Journal
Evaluation of the Effect of Supervised Group Exercise on Self-Reported Sleep Quality in Pregnant Women with or at High Risk of Depression: A Secondary Analysis of a Randomized Controlled Trial
Next Article in Special Issue
Epidemiology of Adverse Outcomes in Teenage Pregnancy—A Northeastern Romanian Tertiary Referral Center
Previous Article in Journal
Mental Health and Well-Being Needs among Non-Health Essential Workers during Recent Epidemics and Pandemics
Previous Article in Special Issue
Systemic Steroids in Preventing Bronchopulmonary Dysplasia (BPD): Neurodevelopmental Outcome According to the Risk of BPD in the EPICE Cohort
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Deep Learning Approach to Estimate the Incidence of Infectious Disease Cases for Routinely Collected Ambulatory Records: The Example of Varicella-Zoster

1
Unit of Biostatistics, Epidemiology and Public Health, Department of Cardiac, Thoracic, Vascular Sciences and Public Health, University of Padova, Via Loredan, 18, 35131 Padova, Italy
2
Division of Pediatric Infectious Diseases, Department of Women’s and Children’s Health, University of Padova, 35131 Padova, Italy
3
Consorzio Arsenàl.IT, 35131 Padova, Italy
4
Società Servizi Telematici–Pedianet, 35138 Padova, Italy
*
Author to whom correspondence should be addressed.
Int. J. Environ. Res. Public Health 2022, 19(10), 5959; https://doi.org/10.3390/ijerph19105959
Submission received: 22 March 2022 / Revised: 3 May 2022 / Accepted: 10 May 2022 / Published: 13 May 2022
(This article belongs to the Special Issue Real World Data for Population-Based Pediatric Studies)

Abstract

:
The burden of infectious diseases is crucial for both epidemiological surveillance and prompt public health response. A variety of data, including textual sources, can be fruitfully exploited. Dealing with unstructured data necessitates the use of methods for automatic data-driven variable construction and machine learning techniques (MLT) show promising results. In this framework, varicella-zoster virus (VZV) infection was chosen to perform an automatic case identification with MLT. Pedianet, an Italian pediatric primary care database, was used to train a series of models to identify whether a child was diagnosed with VZV infection between 2004 and 2014 in the Veneto region, starting from free text fields. Given the nature of the task, a recurrent neural network (RNN) with bidirectional gated recurrent units (GRUs) was chosen; the same models were then used to predict the children’s status for the following years. A gold standard produced by manual extraction for the same interval was available for comparison. RNN-GRU improved its performance over time, reaching the maximum value of area under the ROC curve (AUC-ROC) of 95.30% at the end of the period. The absolute bias in estimates of VZV infection was below 1.5% in the last five years analyzed. The findings in this study could assist the large-scale use of EHRs for clinical outcome predictive modeling and help establish high-performance systems in other medical domains.

1. Introduction

Disease burden estimates for infectious diseases are crucial for public health resource allocation [1].
Active and passive surveillance systems are essential to evaluate the epidemiological impact of infectious diseases: they can provide background data to implement effective control strategies, such as vaccination campaigns, and monitor the trend [2].
In this framework, the detection of varicella-zoster virus (VZV) has recently raised interest due to the extensive debate about the need and cost–benefit profile of introducing a mass vaccination program for young people [3,4]. The national routine notification system adopted in Italy—which is mandatory—helps describe the epidemiology of VZV and allows one to evaluate historical temporal trends but is undoubtedly affected by under-notification and under-diagnosis [5].
The broad implementation of the Electronic Health Record (EHR) in primary care offers new opportunities for population-based studies through data generated by accesses in any care setting, making it accessible promptly [6,7]. In addition, automatic methods for data extraction from free-text reports are increasingly replacing manual extraction [8,9]. In particular, machine learning (ML) is gaining popularity in healthcare because of its power to extract and filter available information from big data and accurately solve demanding learning tasks such as classification, clustering, and numerical prediction [10,11].
In recent years, enhanced generalized linear model (GLM) techniques have been used for text mining from EHRs for the case detection of VZV [12]. However, these techniques can produce highly variable resulting models depending on the analyst’s choices, especially in a class imbalance [13].
Standard shallow ML algorithms cannot directly model data sequences from text streams: they rely on the independence assumptions between tokens, i.e., a single unit of textual information, which generally corresponds to “words,” covariates. Deep-recurrent neural network architectures can overcome these limitations: the reduction of input preprocessing and manipulation, the possibility to process the text as a data sequence, and the automatical learning of the correlations between features without superimposed structures meet the mentioned needs [14,15].
Applying the mentioned approach could, in turn, be used to achieve two critical goals: to estimate the incidence of VZV infection by automatic case identification and to make timely and consistent predictions over time.

Highlights

  • Estimates of infectious disease incidence can be time-consuming and tedious;
  • Deep learning, in particular RNN-GRU, for automatic data extraction from the free text could be a feasible and timely option;
  • Results obtained with MLT were promising, yet, in future development, this text-mining tool should be readily usable by non-technical users as well.

2. Materials and Methods

2.1. Electronic Medical Record Database

The Pedianet database (http://www.pedianet.it/en/, accessed on 9 May 2022) is a pediatric primary care database that contains clinical, demographic, prescription, and outcome data of children aged 0–14 years. Data are generated during daily clinical practice by about 150 family pediatricians who use the same software JuniorBit® (various versions); in the recent past, it has been exploited for infectious disease research studies [12,16,17]. In addition, Pedianet gathers details about specialist referrals, procedures, hospitalizations, medical examinations, and health status (according to the Guidelines of Health Supervision of the American Academy of Pediatrics).
Pedianet was the starting point for producing the gold-standard diagnosis of VZV infection for each record, according to the literature [18].
The study population includes all the children in the Veneto region (Northeast Italy) who were visited at least once between 2004 and 2014: data about 1,227,578 visits and 7631 children were collected. The baseline characteristics are shown in Table 1. stratified for outcome class, i.e., negative or positive case of VZV in the corresponding year. Sex, a categorical measure, is reported as a percentage and absolute frequency. Age, a continuous measure, is reported with I/II (median)/III quartiles.

2.2. Main Strategy

VZV infection exhibits unique characteristics. First, it can be contracted once in a lifetime (with notable exceptions). Furthermore, using a specific diagnostic code to identify VZV infection is not mandatory for a pediatrician. Consequently, it can be critical to ascertain cases and estimate the yearly incidence rate of VZV infection.
Three Pedianet researchers spent two years manually reviewing the medical records to obtain a gold standard for VZV infection; due to this mentioned period, a delay of two years was set as the starting point to develop our model. The main idea was to train the model to detect—through natural language processing (NLP)—the status of a child, i.e., infected or not by VZV, based on text fields of the Pedianet, to compare the predicted incidence with the gold standard and to use the newly acquired data to improve predictive performances and continuously update the existing models.
For illustrative purposes, let us assume that data for a given year are available, e.g., 2006. Under a “two-year delay,” the gold standard is established for data up to 2004. A first version of the model is trained on data up to 2004 and applied to classify data from 2005 to 2006. The following year, i.e., 2007, the model can be updated with the data from 2005 (only in 2007, the gold standard for data from 2005 is supposed to be available) and used to classify all the cases collected in the last two years, i.e., 2006 and 2007. This process went on year by year up to the last model trained on data from 2004 to 2013 and used to classify 2014 data (i.e., the last year for which we have the gold standard). Overall, 10 models were trained, i.e., with training data from 2004 alone, 2004–2005, 2004–2006, 2004–2007, 2004–2008, 2004–2009, 2004–2010, 2004–2011, 2004–2012, 2004–2013, each of them used to classify and to be tested on the next two years, i.e., 2005–2006, 2006–2007, 2007–2008, 2008–2009, 2009–2010, 2010–2011, 2011–2012, 2012–2013, 2013–2014, and 2014 alone, respectively. The general strategy is described in Figure 1.
The entire dataset was initially divided into ten groups, indexed by the last year of training data, from 2004 to 2013. Each group has two components. The first component contains the training data, divided into 80% and 20% for training and validation/fine-tuning, respectively. The second component includes the training data (from 2004 to the reference year) and the next two years of data for testing purposes. Models were tuned on the first component of each group, and hyperparameters were selected for optimizing each validation set. Once the tuning parameters were selected, the new models were (re-) trained on the whole training set using that selection of parameters, and tested on the testing one, i.e., the next two years respect the last one used in the corresponding training set. See Table 2 for the sets’ sizes.

2.3. Model Choice

Given the sequential nature of the input for the task, a deep learning architecture involving a recurrent component to process the text was set up. Compared to a traditional neural network, a recurrent neural network (RNN) is specifically designed to grasp the complex relations between sequential data, such as text reported in natural language.
Within the RNN base components, the gated recurrent unit (GRU) was chosen to meet the needs. A GRU can be viewed as an optimized Long Short-Term Memory (LSTM) module, i.e., the principal reference for advanced RNN networks [19]. LSTM, as well as its recent customized versions, would have also represented a valid option to explore [20]. On the other hand, we preferred GRU to LSTM and its variations because GRU has lower complexity and required resources [21]. In particular, GRU meets the needs of a fast-converging learner who can make an accurate prediction within a reasonable amount of computational resources and time.

2.4. Language Model

The text must be converted into numbers to serve as input in an ML model. One of the most effective strategies to represent the text as a vector of numbers is to resort to embeddings. Vectors are “trained” to convert syntactically and semantically similar words into geometrically close points of a vector space. Due to the highly specialized field and the high amount of data disposal, a self-trained embedding representation was chosen. The embeddings were created using the FastText algorithm with the SkipGram strategy and 300-dimension output vectors [22]. The entire Pedianet database was used, with an overall of 6,903,035 visits of 216,976 children collected by 144 family pediatricians starting from 1 January 2004 to 23 August 2017.

2.5. Implementation

First, the records were collapsed and arranged in chronological order to allow the GRU to work correctly. In the Pedianet, each row represented all available data for a specific visit of a child. Next, all the records for a given year were put in a single text cell. Each row was finally composed of a single cell reporting all the sequentially ordered history of textual notes for a child in a specific year.
Furthermore, it was necessary to have a fixed dimension for the input sequence to allow bidirectional RNN: a hard limit of 10,000 words was chosen. Records with more than 10,000 words were truncated (loss of information), while those with less than 10,000 words were filled with a fictitious word (__PAD__).
After these operations, each input was a three-dimensional matrix with rows corresponding to the corresponding amount of record (see Table 2), 10,000 columns (words in each record), and 300 deep-wise elements (embedding representation for each word).
A bidirectional RNN-GRU was added to the network considering 256 nodes in each direction (512 in total).
All the processed information was analyzed by two fully connected layers of 128 nodes each (256). The last 128 nodes are next connected to a single node with a logistic activation to output the probability of the possible status: VZV positive versus VZV negative for every child.
To summarize, the overall network was composed of the following main layers (see Figure 2):
  • Embeddings: representation of words converted into a syntactic and semantically coherent 300-dimensional structure—input N × 10.000/output N × 10.000 × 300, where N is the number of cases (i.e., one case is the collection of all the HERs of a single child for a given year) in the dataset/minibatch considered.
  • Two synchronized layers of bidirectional RNN-GRU modules are composed of 256 nodes each (512 nodes of each layer process the information as a pure sequence summarizing its “meaning”—input N × 10.000 × 300/output N × 512 (output for the first synchronized layer is equal to the input of the second: N × 10.000 × 512))
  • Two fully connected layers of 128 nodes each to process the “meaning” vector of information from each record—input N × 512/output N × 128 (output for the first synchronized layer is equal to the input of the second: N × 128).
  • Logistic output node: to produce a probability measure for the children being affected by VZV in the corresponding year—input N × 128/output N × 1.
Figure 2. Flowchart of the trained network. The boxes report shapes and shape interpretation of the data between each computation step, i.e., between layers of the network. Layers are reported as linking connections between the boxes. N represents the size of the record passed in input; for our minibatch training, N is 16; for overall records, N depends on the set reported in Table 2.
Figure 2. Flowchart of the trained network. The boxes report shapes and shape interpretation of the data between each computation step, i.e., between layers of the network. Layers are reported as linking connections between the boxes. N represents the size of the record passed in input; for our minibatch training, N is 16; for overall records, N depends on the set reported in Table 2.
Ijerph 19 05959 g002

2.6. Training

The training of the network was based on the optimization/minimization of the weighted binary cross-entropy log-loss (wbcell) measure, i.e., (minus) the mean of the logarithm of the probability of being classified in the correct class (i.e., p i for positive cases, ( 1 p i ) for the negative ones). For taking into account the issue of class imbalance, the contributions to the loss estimation were weighted by the inverse of the class relative frequency (considering f P the absolute frequency of positive cases, and f N the absolute frequency of the negative ones). With the notation adopted the optimization measure takes the following expression:
w b c e l l = 1 N i = 1 N ( N f P * y i * log p i + N f N * 1 y i * log 1 p i )
For network optimization, the Adam strategy was adopted [23]. The network was fed with a mini-batch of size 16, i.e., 16 records at a time [24]. An “epoch” was passed once the network was trained by all the records (batch by batch). The number of training loops was capped at 15 epochs, imposing an early stop if no improvement was seen within 5 epochs and choosing the best epoch [25].
A small nodal dropout was performed across the whole network to minimize the over-fitting [26].
The output model is a probabilistic one, i.e., for any given case, the output is its estimated probability of being a positive one. Deciding if a case is positive or not, i.e., defining a classifier, is a matter of selecting a cut-off for those probabilities. Selecting the cut-off is mainly a matter of utility/preference in preferring false negative or false positive as a potential error [27]. To evaluate a model independently from the cut-off, we can draw the receiver operating characteristic curve, i.e., the curve made up of all the points of coordinate ( s e n s i t i v i t y c , 1 s p e c i f i c i t y c ), at any possible cut-off c, from 0 to 1. The more area under the ROC curve, the better the model is. The area under the ROC curve (AUC-ROC) is used as the evaluation metric for the model trained. AUC-ROC ranges from 0 to 1, with a 0.5 value representing uninformative models, i.e., equivalent to random guesses.

3. Results

In total, 1,227,578 records recorded on Pedianet databases from 7631 children between 2004 and 2014 were merged by patient and year. They were divided into ten cumulative training sets, starting from data from 2004 and adding a year each time. Correspondingly, 10 test sets containing the two years’ data following the last one in the training set were built up. A dictionary of 122,607 words was trained as 300-dimension embedding vectors from Pedianet textual entries.
Across the ten model years, a median AUC-ROC of 97.97% (IQR 97.25%–98.07%) within the training sets and 71.36% (IQR 60.35%–82.47%) for the test sets were reached. A maximum test AUC-ROC of 95.30% was reached for the 2013 model, i.e., the last and the one with the most data in the training set. The minimum test AUC-ROC of 38.46% was reached for the model with data up to 2005. Progression of AUC-ROC performances in the training and test set are reported in Figure 3.
The performance of the training sets increases quickly in the initial models and then stabilizes. On the contrary, the performance of the test sets continuously improves over time, suggesting that the underlying model is suitable for the specific task (optimal training performances) and benefits from the update with new data to generalize its usability (increasing testing performances).
The outcomes of the trained models convey a probability for each case to be a positive one in the corresponding year; before obtaining a classifier, a threshold cut-off for the probabilities had to be selected to distinguish between positivity and negativity. Thus, an optimal cut-off was chosen for maximizing the product of precision (i.e., the portion of records correctly classified as positives) and recall (i.e., the portion of positive records that are correctly classified) [28]. The performances corresponding to each test set for all the models are reported in Table 3.
Each model, equipped with the corresponding cut-off, is then used to estimate the incidence of VZV infections in the two following years of the test. Results are reported in Table 4.
ROCs curves, highlighting the cut-off point and the corresponding difference in the estimation of VZV infections, are reported in Figure 4. The years of the training models are reported on the facets’ headers. Testing years are the following two up to 2014. Color variations in the curves represent the variation of the error in the incidence estimation. The optimal cut-off maximizing the product of precision and recall is reported (red dot) on the side of the corresponding error produced by classifying records using it.

Computational Environment

All the computations were performed using the R v4.1.2 [29] programming language, powered by TensorFlow and compiled for GPU usage, and its Keras interface for the deep learning modeling and training, the {targets} R-package for the global pipeline control, and the {tidyverse} R suite of packages for data management and plots. All the computations run on a Linux Ubuntu 20.04 operating system installed on a machine with a 16-Core Xenon processor equipped with 128 GB RAM and a CUDA NVIDIA Quadro RTX5000 graphic card with 16 GB of dedicated memory.

4. Discussion

Many medical organizations and networks still use manual processes to extract data from unstructured EHRs. Especially for infectious diseases, delays in producing and disseminating the results may hinder a prompt public health response. Thus, it is necessary to create and continuously update a data-based infectious disease prediction model to handle situations in real time [21].
While automatization, in general, represents a hot topic nowadays, the potential of these approaches still has to be fully explored, and the state of the art is hard to identified [30]. On the other hand, some recent works have already explored the field, finding potential best practices to address the issues concerning the classification of natural free-text fields from EHRs. Accordingly to our first step, one of the strategies appearing promising is the adoption of pre-trained language models as a starting macrolayer of the network [31].
This work proposes an alternative strategy to set up, maintain, and update a data-based VZV surveillance model to supplement existing systems. It could be thought of as the starting point to solving the issues within the medical database. It would be interesting to know whether NLP—a very well-known topic for other application fields—could be successfully used in this context [32]. It opens new options and perspectives to analyze medical data in an integrated environment.
Real-incidence rates would be at disposal when manual extraction is performed. Conversely, the automated extraction represented by the model purposed could provide instant estimates for the previous years: e.g., on the 1st of January of a specific year, when all the data about the previous year were collected, it could be possible to estimate the yearly incidence for a specific pathology. Moreover, it could also be possible to have an updated estimation of the number of cases detected up to now live. Finally, thanks to the NLP models, it is possible to calculate incidence for the upcoming years as well—i.e., by making predictions—with better precision than simply relying on recent incidence trends.
Several studies have used techniques from deep learning to predict infectious diseases. Our findings align with those of studies reporting that deep learning yields satisfactory results when used to perform tasks that are difficult for conventional analysis methods [33]. For example, through an indirect comparison of the same Pedianet data [12], RNN-GRU yielded better VZV prediction performance than enhanced GLM-based ML models. As such, methods for predicting infectious diseases, such as VZV that uses deep learning, help design effective models.
In future scenarios, the framework mentioned above could be expanded and used in similar tasks of predicting other infections than VZV ones. None of the methodologies are specific for VZV, and the whole procedure can be specialized and fine-tuned for other targets. One of the main advantages of (recurrent) neural networks is their ability to keep themselves up to date with low effort, thanks to the pre-trained scenario for transfer learning. Much of the ability learned by the network is retained, and only the terminal nodes used for the final classification are fine-tuned for the specific target of interest [34]. Due to this reason, the strategy is getting popular even for the general-purpose classification of biomedical text [35].
Furthermore, in the present work, the models learn from free-text fields only, without any additional input. Any evidence or suggestions from the literature were not given to our model; it is well known that some factors influence the VZV incidence. For example, the peak of infections is in the preschooler age and decreases afterwards. Given these considerations, our work could be thought of as part of a more complex data integration system, where essential variables of different nature and importance are put together to help maximize the desired result [36]. Deep learning is especially suitable and able to manage by design this mixed-type of information, from structured (e.g., tabular information) to unstructured types (e.g., free-text). Both the scenarios are promising: the former, where structured clinical data were used as additional metadata to improve NLP tasks, and the latter when NLP produces small, structured summary information to integrate classical statistical models.

Limitations

Some Limitations Must Be Acknowledged

The amount of textual-based information stored electronically is rapidly increasing. Accumulating information is easy; however, finding relevant information on demand can be difficult as the size of the collection continues to escalate. This article presents a general framework for text mining consisting of a large body of EHRs with text data that are inherently unstructured and fuzzy and cutting-edge advances in deep learning. However, barriers to adoption are one of the main challenges. Trusting that the current system extracts high-quality information (i.e., all actual VZV cases) is also likely to cause concern that an approach that deviates from the current standard (i.e., manual review) might not be of equal quality. In addition, this system is designed for trained knowledge specialists. In future developments, this text-mining tool should be readily usable by non-technical users.
It was recently shown that the Bidirectional Encoder Representations from Transformers (BERT)-based model outperforms other deep learning alternatives in classification tasks from EHRs [37]. BERT is based on a bi-directional representation of tokens incorporating attention layers. Pre-train BERT language models require a massive amount of computational resources (in contrast with FastText, i.e., our choice). In contrast, using a pre-trained one is significantly lighter. On the other hand, in the present work, we preferred to start exploring potential solutions based on a personalized pre-trained language model. For that reason, we did not adopt a BERT-first approach, leaving that for future exploration.
While the architecture does not theoretically require GPUs to be trained or evaluated, it is complex and inefficient for a general-purpose production environment. Exploration of optimization strategies for lowering the computational complexity of the network evaluation can be considered for real-world usage.
While the AUC-ROC on the training set goes up quite quickly (basically after the first year of data), AUC-ROC on the test set needs more time. It obtains satisfying results only when data were available for nine years and passing the 0.9 area only when ten years were considered. Of course, this variability reflects the dimension of the starting dataset.
It is also worth considering the actual nature of the task in supporting human efforts and decisions. In the present work, we have “mathematically” optimized the classifiers derived from the probabilistic models selecting a cut-off and optimizing the product of precision and recall. Given the high level of AUC-ROC reached by the final optimized models, the cut-off can be selected to improve the recall up to a trade-off that allows a human reviewer to only inspect a (small amount of) positive predicted record to exclude false-positive detection. That way, a dual benefit can be obtained: a more precise estimation of the incidence with a low additional human effort and a selection of highly relevant wrongly classified records that can be specifically used to further refine the model in the next update.
It is essential to understand that our model is not temporal; thus, it aims to obtain the case identification and benefit from additional data, year after year, rather than find a correlation in a specified temporal space. Its results are likely to help integrate temporal models on the side of other structured data.
There is also a lack of external validation. Other pediatricians in different regions with different habits can challenge the model. That is reflected in the high performance obtained and maintained on the training set since the early phases, while more time and data are needed to reach a similar level of performance on the testing sets. For training a deep learning NLP classifier, a large amount of different data are needed to explore the variability of language used in the field. On the other hand, as our study shows, with a sufficient amount of data, the model can learn enough to be used on entirely new data with similar performances shown during its training phase.

5. Conclusions

The proposed method achieves promising results which outperform other state-of-the-art algorithms, showing that the NPL approach can effectively predict the status of a child with reasonable accuracy. The deep-learning-based system built in this study could be applied to facilitate the large-scale use of family pediatrician notes for clinical outcome predictive modeling. The findings in this study could also assist in establishing high-performance systems in other medical domains integrating structured and unstructured data.

Author Contributions

Conceptualization, C.L., I.B. and D.G.; methodology, I.B. and C.L.; formal analysis, C.L.; data curation, E.B., L.T., A.S., L.C., C.G.; writing—original draft preparation, C.L. and A.F.; writing—review and editing, C.L., A.F., I.B. and D.G.; supervision, I.B., D.G. and C.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data available on reasonable request from Pedianet. Code developed is publicly available at www.github.com/UBESP-DCTV/varicella.due (accessed on 9 May 2022).

Acknowledgments

The authors thank all the family pediatricians collaborating in Pedianet. Alongi Angelo, Angelini Roberta, Avarello Giovanni, Azzoni Lucia, Balliana Franco, Barbazza Maria Carolina, Barberi Frandanisa Maria, Barbieri Patrizia, Belluzzi Gabriele, Benetti Eleonora, Bezzi Roberto, Bit Junior, Boe Franca, Bollettini Stefano, Bonfigli Emanuela, Bruna Andrea, Brusaterra Ivana, Budassi Roberto, Caccini Massimo, Cantalupi Laura, Cantarutti Luigi, Caprio Luigia, Castaldo Massimo, Castelli Stefano, Castronuovo Serenella, Cavedagni Monica, Censini Stefania, Cera Giuseppe Egidio, Ciscato Carla, Clerici Schoeller Mariangela, Collacciani Giuseppe, Comaita Fabrizio, Conte Ugo Alfredo, Costanzo Nicola, Cozzani Sandra, Cuboni Giancarlo, Curti Valentino, D’Amanti Vito Francesco, De Angelis Rita, De Clara Roberto, De Marchi Annamaria, De Nicolò Emanuele, Del Bono Gian Piero, Del Ponte Gigliola, Dell’Antonia Fabio, Di Giampietro Tiziana, Di Mauro Giuseppe, Di Renzo Anna Paola, Di Santo Giuseppe, Dolci Marco, Doria Mattia, Drago Stefano, Falco Pietro, Fama Mario, Faraci Marco, Favilli Tania, Federico Mariagrazia, Felice Michele, Ferrara Enrico, Ferrarese Marta, Ferretti Michele, Forcina Paolo, Frattini Claudio Paolo, Frison Ezio, Fusco Fabrizio, Gallo Giovanni, Galvagno Andrea, Gentili Alberta, Gentilucci Pierfrancesco, Giampaolo Giuliana, Giancola Giuseppe, Giaretta Letizia, Girotto Silvia, Gobbi Costantino, Grelloni Mauro, Grugnetti Mirco, Lagrasta Urania Elisabetta, Landi Massimo, Lasalvia Paola, Letta Maria Rosaria, Lietti Giuseppe, Lista Cinzia, Lucantonio Ricciardo, Luise Francesco, Luotti Diego, Macropodio Nadia, Marine Francesca, Mariniello Lorenzo, Marostica Gabriella, Masotti Sergio, Meneghetti Stefano, Milani Massimo, Milone Stella Vittoria, Monteleone Angela Maria, Mussinu Pierangela, Muzzolini Carmen, Nicoloso Flavia, Olimpi Laura Maria, Palma Maria Maddalena, Pandolfini Vittorio, Pasinato Angela, Passarella Andrea, Pazzola Pasquale, Perri Danilo, Pescosolido Silvana Rosa, Petrazzuoli Giovanni, Petrotto Giuseppe, Picco Patrizia, Pirola Ambrogina, Pisanello Lorena, Pittarello Daniele, Porro Elena, Profumo Elisabetta, Puma Antonino, Ragazzon Ferdinando, Rosas Paolo, Rosignoli Rino, Rossitto Mariella, Ruffato Bruno, Ruggieri Lucia, Ruscitti Annamaria, Russo Annarita, Salamone Pietro, Sambugaro Daniela, Saretta Luigi, Sarno Vittoria, Sciolla Nico Maria, Semenzato Flavio, Senesi Paolo, Silvan Carla, Spanevello Valter, Speciale Sergio Maria, Speranza Francesco, Sticco Maura, Storelli Francesco, Tamassia Gianni, Tambaro Paolo, Toffol Giacomo, Tonelli Gabriele, Tummarello Angelo, Ulliana Antonella, Venditti Sergio, Volpe Concetta, Volpe Francescopaolo and Vozzi Aldo. The authors would like to thank the PENTA Foundation VII.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kretzschmar, M.; Mangen, M.J.; Pinheiro, P.; Jahn, B.; Fevre, E.M.; Longhi, S.; Lai, T.; Havelaar, A.H.; Stein, C.; Cassini, A.; et al. New Methodology for Estimating the Burden of Infectious Diseases in Europe. PLoS Med. 2012, 9, e1001205. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Magill, S.S.; Dumyati, G.; Ray, S.M.; Fridkin, S.K. Evaluating Epidemiology and Improving Surveillance of Infections Associated with Health Care, United States. Emerg. Infect. Dis. 2015, 21, 1537–1542. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Baracco, G.J.; Eisert, S.; Saavedra, S.; Hirsch, P.; Marin, M.; Ortega-Sanchez, I.R. Clinical and Economic Impact of Various Strategies for Varicella Immunity Screening and Vaccination of Health Care Personnel. Am. J. Infect. Control 2015, 43, 1053–1060. [Google Scholar] [CrossRef] [PubMed]
  4. Damm, O.; Ultsch, B.; Horn, J.; Mikolajczyk, R.T.; Greiner, W.; Wichmann, O. Systematic Review of Models Assessing the Economic Value of Routine Varicella and Herpes Zoster Vaccination in High-Income Countries. BMC Public Health 2015, 15, 533. [Google Scholar] [CrossRef] [Green Version]
  5. Gabutti, G.; Rota, M.C.; Guido, M.; De Donno, A.; Bella, A.; Ciofi degli Atti, M.L.; Crovari, P. The Epidemiology of Varicella Zoster Virus Infection in Italy. BMC Public Health 2008, 8, 372. [Google Scholar] [CrossRef] [Green Version]
  6. De Bie, S.; Coloma, P.M.; Ferrajolo, C.; Verhamme, K.M.; Trifiro, G.; Schuemie, M.J.; Straus, S.M.; Gini, R.; Herings, R.; Mazzaglia, G.; et al. The Role of Electronic Healthcare Record Databases in Paediatric Drug Safety Surveillance: A Retrospective Cohort Study. Br. J. Clin. Pharmacol. 2015, 80, 304–314. [Google Scholar] [CrossRef] [Green Version]
  7. Sutherland, S.M.; Kaelber, D.C.; Downing, N.L.; Goel, V.V.; Longhurst, C.A. Electronic Health Record-Enabled Research in Children Using the Electronic Health Record for Clinical Discovery. Pediatr. Clin. 2016, 63, 251–268. [Google Scholar] [CrossRef]
  8. Rosier, A.; Burgun, A.; Mabo, P. Using Regular Expressions to Extract Information on Pacemaker Implantation Procedures from Clinical Reports. In Proceedings of the AMIA Annual Symposium Proceedings, Washington, DC, USA, 8–12 November 2008; pp. 81–85. [Google Scholar]
  9. Nassif, H.; Woods, R.; Burnside, E.; Ayvaci, M.; Shavlik, J.; Page, D. Information Extraction for Clinical Data Mining: A Mammography Case Study. In Proceedings of the 2009 IEEE International Conference on Data Mining Workshops, Miami, FL, USA, 6 December 2009; pp. 37–42. [Google Scholar]
  10. Obermeyer, Z.; Emanuel, E.J. Predicting the Future—Big Data, Machine Learning, and Clinical Medicine. N. Engl. J. Med. 2016, 375, 1216–1219. [Google Scholar] [CrossRef] [Green Version]
  11. Witten, I.H.; Frank, E.; Hall, M.A. Data Mining: Practical Machine Learning Tools and Techniques; Morgan Kaufmann Publishers Inc.: San Francisco, CA, USA, 2011. [Google Scholar]
  12. Lanera, C.; Berchialla, P.; Baldi, I.; Lorenzoni, G.; Tramontan, L.; Scamarcia, A.; Cantarutti, L.; Giaquinto, C.; Gregori, D. Use of Machine Learning Techniques for Case-Detection of Varicella Zoster Using Routinely Collected Textual Ambulatory Records: Pilot Observational Study. JMIR Med. Inform. 2020, 8, e14330. [Google Scholar] [CrossRef]
  13. Lanera, C.; Berchialla, P.; Sharma, A.; Minto, C.; Gregori, D.; Baldi, I. Screening PubMed Abstracts: Is Class Imbalance Always a Challenge to Machine Learning? Syst. Rev. 2019, 8, 317. [Google Scholar] [CrossRef]
  14. Hahn, U.; Oleynik, M. Medical Information Extraction in the Age of Deep Learning. Yearb. Med. Inform. 2020, 29, 208–220. [Google Scholar] [CrossRef] [PubMed]
  15. Jozefowicz, R.; Zaremba, W.; Sutskever, I. An Empirical Exploration of Recurrent Network Architectures. In Proceedings of the 32nd International Conference on Machine Learning, Lille, France, 7–9 July 2015; pp. 2342–2350. [Google Scholar]
  16. Dona, D.; Mozzo, E.; Scamarcia, A.; Picelli, G.; Villa, M.; Cantarutti, L.; Giaquinto, C. Community-Acquired Rotavirus Gastroenteritis Compared with Adenovirus and Norovirus Gastroenteritis in Italian Children: A Pedianet Study. Int. J. Pediatr. 2016, 2016, 5236243. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Barbieri, E.; Porcu, G.; Hu, T.; Petigara, T.; Senese, F.; Prandi, G.M.; Scamarcia, A.; Cantarutti, L.; Cantarutti, A.; Giaquinto, C. A Retrospective Database Analysis to Estimate the Burden of Acute Otitis Media in Children Aged <15 Years in the Veneto Region (Italy). Children 2022, 9, 436. [Google Scholar] [CrossRef] [PubMed]
  18. Nicolosi, A.; Sturkenboom, M.; Mannino, S.; Arpinelli, F.; Cantarutti, L.; Giaquinto, C. The Incidence of Varicella: Correction of a Common Error. Epidemiology 2003, 14, 99–102. [Google Scholar] [CrossRef] [PubMed]
  19. Cho, K.; van Merrienboer, B.; Bahdanau, D.; Bengio, Y. On the Properties of Neural Machine Translation: Encoder–Decoder Approaches. In Proceedings of the Proceedings of SSST-8, Eighth Workshop on Syntax, Semantics and Structure in Statistical Translation, Doha, Qatar, 25 October 2014; Association for Computational Linguistics: Doha, Qatar, 2014; pp. 103–111. [Google Scholar]
  20. Shi, J.; Gao, X.; Kinsman, W.C.; Ha, C.; Gao, G.G.; Chen, Y. DI++: A Deep Learning System for Patient Condition Identification in Clinical Notes. Artif. Intell. Med. 2022, 123, 102224. [Google Scholar] [CrossRef]
  21. Chae, S.; Kwon, S.; Lee, D. Predicting Infectious Disease Using Deep Learning and Big Data. Int. J. Environ. Res. Public Health 2018, 15, 1596. [Google Scholar] [CrossRef] [Green Version]
  22. Bojanowski, P.; Grave, E.; Joulin, A.; Mikolov, T. Enriching Word Vectors with Subword Information. Trans. Assoc. Comput. Linguist. 2017, 5, 135–146. [Google Scholar] [CrossRef] [Green Version]
  23. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar] [CrossRef]
  24. Robbins, H.; Monro, S. A Stochastic Approximation Method. Ann. Math. Stat. 1951, 22, 400–407. [Google Scholar] [CrossRef]
  25. Yao, Y.; Rosasco, L.; Caponnetto, A. On Early Stopping in Gradient Descent Learning. Constr. Approx. 2007, 26, 289–315. [Google Scholar] [CrossRef]
  26. Dropout: A Simple Way to Prevent Neural Networks from Overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958.
  27. Habibzadeh, F.; Habibzadeh, P.; Yadollahie, M. On Determining the Most Appropriate Test Cut-off Value: The Case of Tests with Continuous Results. Biochem Med. 2016, 26, 297–307. [Google Scholar] [CrossRef] [PubMed]
  28. Liu, X. Classification Accuracy and Cut Point Selection. Statist. Med. 2012, 31, 2676–2686. [Google Scholar] [CrossRef] [PubMed]
  29. Team, R.C. R: A Language and Environment for Statistical Computing; R Foundation for Statistical Computing: Vienna, Austria, 2022. [Google Scholar]
  30. Holub, K.; Hardy, N.; Kallmes, K. Toward Automated Data Extraction according to Tabular Data Structure: Cross-Sectional Pilot Survey of the Comparative Clinical Literature. JMIR Form. Res. 2021, 5, e33124. [Google Scholar] [CrossRef]
  31. Dai, H.-J.; Su, C.-H.; Lee, Y.-Q.; Zhang, Y.-C.; Wang, C.-K.; Kuo, C.-J.; Wu, C.-S. Deep Learning-Based Natural Language Processing for Screening Psychiatric Patients. Front. Psychiatry 2021, 11, 533949. [Google Scholar] [CrossRef]
  32. Ding, Y.; Ma, J.; Luo, X. Applications of Natural Language Processing in Construction. Autom. Constr. 2022, 136, 104169. [Google Scholar] [CrossRef]
  33. Xu, Q.; Gel, Y.R.; Ramirez Ramirez, L.L.; Nezafati, K.; Zhang, Q.; Tsui, K.-L. Forecasting Influenza in Hong Kong with Google Search Queries and Statistical Model Fusion. PLoS ONE 2017, 12, e0176690. [Google Scholar] [CrossRef] [Green Version]
  34. Howard, J.; Ruder, S. Universal Language Model Fine-Tuning for Text Classification. arXiv 2018, arXiv:1801.06146. [Google Scholar] [CrossRef]
  35. Lee, J.; Yoon, W.; Kim, S.; Kim, D.; Kim, S.; So, C.H.; Kang, J. BioBERT: A Pre-Trained Biomedical Language Representation Model for Biomedical Text Mining. Bioinformatics 2019, 36, 1234–1240. [Google Scholar] [CrossRef]
  36. Shi, J.; Liu, S.; Pruitt, L.C.C.; Luppens, C.L.; Ferraro, J.P.; Gundlapalli, A.V.; Chapman, W.W.; Bucher, B.T. Using Natural Language Processing to Improve EHR Structured Data-Based Surgical Site Infection Surveillance. In Proceedings of the AMIA Annual Symposium Proceedings, Washington, DC, USA, 16–20 November 2019; Volume 2019, pp. 794–803. [Google Scholar]
  37. Mitra, A.; Rawat, B.P.S.; McManus, D.D.; Yu, H. Relation Classification for Bleeding Events from Electronic Health Records Using Deep Learning Systems: An Empirical Study. JMIR Med. Inform. 2021, 9, e27527. [Google Scholar] [CrossRef]
Figure 1. Chart for the general strategy for model development and test. Top-line on the chart: each new year X (right-most position on the x-axis), an updated model can be trained on the already ready gold-standard data, i.e., up to the previous two years (blue), and used to predict the following two years X − 1 and X (white). Middle-line on the chart: the following year X + 1, a second, updated prediction (yellow) can be made on one of the years of tested data (X − 1) with the previous model. Bottom-line on the chart: in the second next year X + 2 the gold standard is supposed to be ready for that year X − 1, becoming a new training data (blue). The model can provide an updated prediction for the year X and a new prediction for the years X + 1 and X + 2 (i.e., the current one “just ended”).
Figure 1. Chart for the general strategy for model development and test. Top-line on the chart: each new year X (right-most position on the x-axis), an updated model can be trained on the already ready gold-standard data, i.e., up to the previous two years (blue), and used to predict the following two years X − 1 and X (white). Middle-line on the chart: the following year X + 1, a second, updated prediction (yellow) can be made on one of the years of tested data (X − 1) with the previous model. Bottom-line on the chart: in the second next year X + 2 the gold standard is supposed to be ready for that year X − 1, becoming a new training data (blue). The model can provide an updated prediction for the year X and a new prediction for the years X + 1 and X + 2 (i.e., the current one “just ended”).
Ijerph 19 05959 g001
Figure 3. AUC-ROC (y-axis) performance progression across epochs of training (x-axes) and model years(panels from 2004 to 2004–2013, from left to right) for both the train (green) and test (red). In total, 95% CI are reported as shadows.
Figure 3. AUC-ROC (y-axis) performance progression across epochs of training (x-axes) and model years(panels from 2004 to 2004–2013, from left to right) for both the train (green) and test (red). In total, 95% CI are reported as shadows.
Ijerph 19 05959 g003
Figure 4. Receiver operator curves (ROCs) of the model trained to classify VZV infections. The years of the training models are reported on the facets’ headers. Testing years are the following two up to 2014. Color variations in the curves represent the variation of the error in the incidence estimation. The optimal cut-off maximizing the product of precision and recall is reported (red dot) on the side of the corresponding error produced by classifying records using it.
Figure 4. Receiver operator curves (ROCs) of the model trained to classify VZV infections. The years of the training models are reported on the facets’ headers. Testing years are the following two up to 2014. Color variations in the curves represent the variation of the error in the incidence estimation. The optimal cut-off maximizing the product of precision and recall is reported (red dot) on the side of the corresponding error produced by classifying records using it.
Ijerph 19 05959 g004
Table 1. Characteristics of units analyzed. Descriptive statistics are reported and stratified for outcome class, i.e., negative or positive case of VZV in the corresponding year.
Table 1. Characteristics of units analyzed. Descriptive statistics are reported and stratified for outcome class, i.e., negative or positive case of VZV in the corresponding year.
NVZV Negative
(N = 58,334)
VZV Positive
(N = 2325)
Sex60,659
 Female 47% (27,340)46% (1068)
 Male 53% (30,994)54% (1257)
Age [days]60,3420.7/2.2/4.28 *0.6/1.4/3.1
* I/II (median)/III quartile.
Table 2. Cases in each set of the models trained. All child records for a given year represent a case, i.e., the same child in distinct years represents distinct and independent cases. Each row reports datasets for the training, validation, and test of a model.
Table 2. Cases in each set of the models trained. All child records for a given year represent a case, i.e., the same child in distinct years represents distinct and independent cases. Each row reports datasets for the training, validation, and test of a model.
YearsTraining PhaseTesting Phase
TrainTestTrain (#)Validation (#)Train (#)Test (#)
20042005–2006158839619847854
2004–20052006–20074405109955049454
2004–20062007–200878731965983810,852
2004–20072008–200911,969238914,95812,020
2004–20082009–201016,555413520,69013,062
2004–20092010–201121,586539226,98713,848
2004–20102011–201227,006674633,75214,139
2004–20112012–201332,666816040,82614,017
2004–20122013–201438,319957247,89112,768
2004–2013201443,88210,96154,8435816
Table 3. Number of positives, negatives, area under the receiver operating characteristic curve (AUC-ROC), predicted true-positives (tp) and true-negatives (tn), false-positives (fp), and false-negatives (fn). Precision or positive predictive value (prec) and recall or sensitivity (rec) for each model year (by row, indexed by the column year) related to their corresponding test sets. Bold face is used to highlight the best performance column wise.
Table 3. Number of positives, negatives, area under the receiver operating characteristic curve (AUC-ROC), predicted true-positives (tp) and true-negatives (tn), false-positives (fp), and false-negatives (fn). Precision or positive predictive value (prec) and recall or sensitivity (rec) for each model year (by row, indexed by the column year) related to their corresponding test sets. Bold face is used to highlight the best performance column wise.
Model YearPositivesNegativesAUC
ROC
tptnfpfnprecrec
2004–20046371.9540.8045405.1802.037970.2100.848
2004–20051723.3480.3851888.474357570.8430.199
2004–20064653.8690.58819410.024415930.8260.247
2004–20074804.6400.64913011.403334540.7980.223
2004–20083075.4250.58210212.470514390.6670.189
2004–20092776.0110.6529813.386463180.6810.236
2004–20102646.5100.7753713.870401920.4810.162
2004–20111526.9220.8354313.848191070.6940.287
2004–2012776.9880.8324512.64522560.6720.446
2004–2013736.8790.953175.69890110.1590.607
Table 4. Incidences of VZV infections observed in Pedianet and estimated by the model trained.
Table 4. Incidences of VZV infections observed in Pedianet and estimated by the model trained.
Model YearYears
Estimated
PositivesNegativesObserved
Incidence
(%)
Estimated
Incidence
(%)
Estimated
Incidence
Error (%)
20042005–20066371.9548.1132.824.7
2004–20052006–20071723.348102.36−7.64
2004–20062007–20084653.8697.252.17−5.09
2004–20072008–20094804.6404.861.36−3.5
2004–20082009–20103075.4254.141.17−2.97
2004–20092010–20112776.01131.04−1.96
2004–20102011–20122646.5101.620.54−1.08
2004–20112012–20131526.9221.070.44−0.63
2004–20122013–2014776.9880.790.52−0.27
2004–20132014736.8790.481.841.36
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lanera, C.; Baldi, I.; Francavilla, A.; Barbieri, E.; Tramontan, L.; Scamarcia, A.; Cantarutti, L.; Giaquinto, C.; Gregori, D. A Deep Learning Approach to Estimate the Incidence of Infectious Disease Cases for Routinely Collected Ambulatory Records: The Example of Varicella-Zoster. Int. J. Environ. Res. Public Health 2022, 19, 5959. https://doi.org/10.3390/ijerph19105959

AMA Style

Lanera C, Baldi I, Francavilla A, Barbieri E, Tramontan L, Scamarcia A, Cantarutti L, Giaquinto C, Gregori D. A Deep Learning Approach to Estimate the Incidence of Infectious Disease Cases for Routinely Collected Ambulatory Records: The Example of Varicella-Zoster. International Journal of Environmental Research and Public Health. 2022; 19(10):5959. https://doi.org/10.3390/ijerph19105959

Chicago/Turabian Style

Lanera, Corrado, Ileana Baldi, Andrea Francavilla, Elisa Barbieri, Lara Tramontan, Antonio Scamarcia, Luigi Cantarutti, Carlo Giaquinto, and Dario Gregori. 2022. "A Deep Learning Approach to Estimate the Incidence of Infectious Disease Cases for Routinely Collected Ambulatory Records: The Example of Varicella-Zoster" International Journal of Environmental Research and Public Health 19, no. 10: 5959. https://doi.org/10.3390/ijerph19105959

APA Style

Lanera, C., Baldi, I., Francavilla, A., Barbieri, E., Tramontan, L., Scamarcia, A., Cantarutti, L., Giaquinto, C., & Gregori, D. (2022). A Deep Learning Approach to Estimate the Incidence of Infectious Disease Cases for Routinely Collected Ambulatory Records: The Example of Varicella-Zoster. International Journal of Environmental Research and Public Health, 19(10), 5959. https://doi.org/10.3390/ijerph19105959

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop