Next Article in Journal
Bronchoalveolar Lavage Cell Count and Lymphocytosis Are the Important Discriminators between Fibrotic Hypersensitivity Pneumonitis and Idiopathic Pulmonary Fibrosis
Next Article in Special Issue
Digital Tomosynthesis as a Problem-Solving Technique to Confirm or Exclude Pulmonary Lesions in Hidden Areas of the Chest
Previous Article in Journal
Thyroglossal Duct Lipoma: A Case Report and a Systematic Review of the Literature for Its Management
Previous Article in Special Issue
Chest X-ray Interpretation: Detecting Devices and Device-Related Complications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Web-Based Platform for the Automatic Stratification of ARDS Severity

by
Mohammad Yahyatabar
1,
Philippe Jouvet
2,*,
Donatien Fily
2,
Jérome Rambaud
2,
Michaël Levy
2,
Robinder G. Khemani
3 and
Farida Cheriet
1,† on behalf of the Pediatric Acute Respiratory Distress Syndrome Incidence and Epidemiology (PARDIE) V3 Investigators and PALISI Network
1
Department of Computer and Software Engineering, Polytechnique Montréal, Montréal, QC H3T 1J4, Canada
2
Department of Pediatrics, Faculty of Medicine, University of Montréal, Montréal, QC H3C 3J7, Canada
3
Department of Anesthesiology and Critical Care Medicine, Children’s Hospital of Los Angeles, Los Angeles, CA 90027, USA
*
Author to whom correspondence should be addressed.
Membership of the Pediatric Acute Respiratory Distress Syndrome Incidence and Epidemiology (PARDIE) V3 Investigators and PALISI Network is provided in the Acknowledgments.
Diagnostics 2023, 13(5), 933; https://doi.org/10.3390/diagnostics13050933
Submission received: 15 January 2023 / Revised: 23 February 2023 / Accepted: 24 February 2023 / Published: 1 March 2023
(This article belongs to the Special Issue Chest X-ray Detection and Classification of Chest Abnormalities)

Abstract

:
Acute respiratory distress syndrome (ARDS), including severe pulmonary COVID infection, is associated with a high mortality rate. It is crucial to detect ARDS early, as a late diagnosis may lead to serious complications in treatment. One of the challenges in ARDS diagnosis is chest X-ray (CXR) interpretation. ARDS causes diffuse infiltrates through the lungs that must be identified using chest radiography. In this paper, we present a web-based platform leveraging artificial intelligence (AI) to automatically assess pediatric ARDS (PARDS) using CXR images. Our system computes a severity score to identify and grade ARDS in CXR images. Moreover, the platform provides an image highlighting the lung fields, which can be utilized for prospective AI-based systems. A deep learning (DL) approach is employed to analyze the input data. A novel DL model, named Dense-Ynet, is trained using a CXR dataset in which clinical specialists previously labelled the two halves (upper and lower) of each lung. The assessment results show that our platform achieves a recall rate of  95.25 %  and a precision of  88.02 % . The web platform, named PARDS-CxR, assigns severity scores to input CXR images that are compatible with current definitions of ARDS and PARDS. Once it has undergone external validation, PARDS-CxR will serve as an essential component in a clinical AI framework for diagnosing ARDS.

1. Introduction

Acute respiratory distress syndrome (ARDS) is a severe, even life-threatening condition, associated with respiratory failure, i.e., the inability of the lungs to fulfill their basic function of exchanging gases in the body. ARDS occurs in children and adults; its main causes include respiratory infection, aspiration, or trauma. The first description of ARDS as a separate disease was provided in 1967. Variability in the ability to identify ARDS causes difficulty in clinical trials. The Berlin definition introduced diagnostic criteria, such as acute onset, severe hypoxemia (lack of oxygen in the blood), bilateral diffuse infiltrates visible in chest radiography, and absence of any evidence of cardiac failure or fluid overload [1]. Despite intensive studies investigating ARDS (60,000+ articles found in PubMed), its mortality rate is still as high as  43 %  [2]. Among the survivors of ARDS, a significant portion experienced lasting damage to the lungs, especially in older patients. The Berlin definition grades the severity of ARDS as being mild, moderate, or severe. Table 1 illustrates the oxygenation criteria and mortality rates associated with these severity levels.
As seen in Table 1, considering the high mortality rate of ARDS and its rapid progression, early diagnosis of ARDS is vital. Furthermore, the mortality rate is directly associated with the severity of the syndrome. The risk benefit profile of therapies depends on ARDS severity, making early stratification of ARDS severity crucial for management. The Pediatric Acute Lung Injury Consensus Conferences (PALICC) [3,4,5] were organized to address pediatric ARDS (PARDS) specifications and give treatment and diagnosis recommendations. According to the most recent definition of PARDS, PALICC-2 [5], the criteria allow for new infiltrates in chest radiography, even if only a region within a single lung is affected. One of the main reasons for this change in diagnostic criteria was the lack of agreement in the interpretation of chest images between radiologists or between radiologists and intensive care practitioners on the presence of bilateral infiltrates, which are required in the Berlin standard. López-Fernández et al. showed that interobserver agreement for bilateral infiltrates and quadrants of consolidation in PARDS was “slight” (kappa 0.31 and 0.33) [6]. Sjoding et al. reported similar results, with interobserver reliability of ARDS diagnosis being “moderate” (kappa = 0.50; 95CI, 0.40–0.59). Hence, there is an urgent need to improve the reliability of Chest X-ray (CXR) intepretation in ARDS and PARDS to allow earlier diagnosis of the syndrome [7].
Several studies have applied machine learning (ML) and artificial intelligence (AI) approaches to analyze CXR images. One of the most common tasks reported in the literature is diagnosing pulmonary pathologies using chest radiography. Thanks to massive publicly available datasets, deep learning (DL) approaches have been broadly applied in medical pathology detection. However, there is as yet no dataset annotated with ARDS labels. Thus, few studies are found in the literature addressing the diagnosis of the syndrome.
To our knowledge, two papers present ML-based systems to identify ARDS in CXR images. The first one [8] proposed a method for detecting ARDS using a traditional ML approach based on hand-crafted features. The texture of intercostal image regions is considered as a discriminative feature for classifying samples. To highlight intercostal areas, a semi-automatic approach proposed by Plourde is utilized [9]. They succeed in reducing the inter-observer variability between clinicians in diagnosing PARDS. However, their approach is not automatic, and the rib segmentation step requires operator intervention. In the second work, an automatic ARDS detection and grading approach was proposed using a state-of-the-art DL model (Densenet) [10]. The authors first pretrained the model on public datasets (not containing ARDS samples) and then refined the model with a custom dataset consisting of ARDS-labeled images. Their approach performs well in diagnosing ARDS, but the model provides no evidence for the support system’s decisions. Thus, although it works well in analyzing ARDS cases, the model lacks interpretability, which is essential for a ML system to be used in clinical settings.
Recently, due to the COVID-19 outbreak, the research community has gotten more involved in computer-based analysis of chest X-ray images as one of the easiest and fastest ways to check for signs of the disease. Mobile Chest X-ray Analysis [11] and Chester [12] are prototype systems for CXR assessment developed using the aforementioned Densenet model, trained on the public Chest-Xray14 dataset [13]. Both systems provide evidence for the detected pathologies by means of saliency maps obtained using GradCam [14]. However, this can reveal areas that are irrelevant to the pathology being detected [15,16]. Thus, although these systems provide activation maps pointing out the references for the decisions, they are not sufficiently reliable to be used in clinics.
The main contributions of this paper are to create a tool for stratifying the severity of ARDS in CXR images and to build a web-based platform for external validation. The platform uses local information to classify X-rays based on the distribution of infiltrates in the different lung quadrants, and it provides a global severity score for the image that is applicable in both children and adults. The web-based platform, PARDS-CxR, can be used as a standalone tool, or it can be integrated with other ARDS analysis tools to offer a comprehensive approach for clinical use.
The following section first explains the details of the data collection used to train our DL model. Then, we describe the proposed DL model and its evaluation process, and we present the development of the Web platform. Section 3 presents the results of testing the ARDS assessment tool, and in Section 4, the strengths and drawbacks of our platform are discussed. Section 5 provides concluding remarks for this paper.

2. Materials and Methods

2.1. Methodology

This study contains four main phases, as illustrated in Figure 1. The end product is PARDS-CxR, the web-based application to detect ARDS. First, a substantial set of data is required to train the model. Existing public datasets do not include ARDS-labeled CXR images, so we created a new one. This data collection process is summarized in Section 2.2. Then, the proposed model must be trained on the CXR images. The model has two outputs associated with lung segmentation and ARDS classification, as explained in Section 2.3. The trained model is then tested on unseen data to be evaluated. Section 2.4, Section 2.5 and Section 2.6 detail the validation process. Finally, the model is uploaded to a server, and an interface is designed so the user can easily access it. The web application is addressed in Section 2.7.

2.2. ARDS Dataset

Collectively, the main publicly available CXR datasets provide around a million images with pathology labels [17]. This data motivated many researchers to employ AI techniques in this domain. However, no such datasets assign ARDS-specific labels to images. As our first step, we collected and annotated a dataset at Sainte-Justine Hospital, Montreal, Canada (CHUSJ), to address the lack of appropriate data. Our dataset comprises three data sources containing 373 CXR images. Ninety images and their corresponding labels came from a previous study by our team [8]. A further 100 images were taken from the Chest X-ray14 dataset [13] and relabeled by clinical experts (JR, ML) in the hospital. Another 183 images were provided by the PARDIE study, a multi-national study that prospectively gathered chest X-ray images of children with ARDS [6]. For each image, labels were associated with the four lung quadrants obtained by splitting each lung into upper and lower portions. We refer to each quadrant by its position: left upper (LR), left lower (LL), right upper (RU), and right lower (RL). According to the Berlin definition [1], visible bilateral infiltrates are a mandatory criterion for a case to be categorized as ARDS. Two intensivists from CHUSJ assessed the presence of infiltrates in each quadrant. A sample was included in the dataset only if the clinical observers reached a consensus on the labels. In addition, 138 CXR images were taken from the Montgomery dataset to represent the normal class [18]. These samples were labeled as non-ARDS if agreed by the clinical experts. By dropping images with disagreements in their labeling, our final ARDS dataset consisted of 356 images, of which 134 meet the bilateral infiltrates criteria in the Berlin definition [1].

2.3. Joint Segmentation and Classification Model

In computer-based diagnosis approaches, it is common to use segmentation ahead of classification to determine the region of interest. Lung segmentation separates the lung areas from the thoracic tissues surrounding them and is the primary image analysis step in many clinical decision support systems. Generalization to new datasets is a difficult challenge in the analysis of chest radiography. In that respect, segmentation is considered a strategy to limit the impact of specific imaging devices and settings, since it restricts the feature extraction to the lung fields and removes the effect of the image background [19,20]. However, serial usage of segmentation and classification propagates the segmentation error into the classification network. Dense-Ynet is a convolutional network that takes advantage of Densenet, Y-net, and U-net models to do both tasks simultaneously in a joint segmentation–classification model. The backbone of the network used in this study is our previously developed Dense-Unet [21]. Dense-Unet is a segmentation model in which dense connections between the feature maps in various layers facilitate the information flow throughout the model, letting designers choose a configuration with a small number of training parameters. Our proposed Dense-Ynet takes advantage of automatic feature extraction from both the original and segmented images (Figure 2). The model has two outputs and is trained using two loss functions: the lung segmentation loss and quadrant classification loss. The model works based on the convolution operation. A convolution is a mathematical operation that filters the information of its input and creates feature maps. An inevitable effect of the convolution operation is to change the dimensions of the feature maps. To tackle this issue, upsampling and strided convolution operations are used to ensure that feature maps coming from different layers can be concatenated. Squeeze and excitation (SE) blocks [22] are also used after each convolution layer to improve the representational power of the blocks by recalibrating the features. The key strengths of Dense-Ynet are use of lung segmentation in its architecture, specialized connectivity, which enable better generalization, and prediction of local labels for each image.
To reach the final decision based on the Berlin definition, we must test for existing bilateral infiltrates. To that end, a simple logical operation in Equation (1) combines the predictions of each quadrant to check this condition:
P A R D S = ( P R L P R U ) ( P L L P L U )
P R L P R U P L L , and  P L U  are the prediction labels for the right lower, right upper, left lower, and left upper quadrants, respectively.  P A R D S  is the inferred ARDS label, and ∨ and ∧ are logical or and and operations. The equation states that, if at least one quadrant is involved on each side, the case is recognized as (P)ARDS.

2.4. Experimental Design

In this work, 267 images of the ARDS dataset are used to train the Dense-Ynet model. In addition, 35 images are used to validate the training process. For the testing stage, 54 images previously unseen by the network are used. The algorithm is evaluated with the five-fold cross-validation strategy. Cross-validation is a method that tries various training and testing data combinations to confirm the reported results’ reliability. Data augmentation is a technique to enrich the training data by generating new images from the current training set. For this purpose, we use basic image processing techniques, such as random rotation, cropping, shifting, horizontal flipping, and intensity changing. The rectified linear unit ( R e L u ) activation function introduces non-linearity to network blocks. The  S i g m o i d  function provides valid labels between zero and one in both the segmentation and classification output layers.  A d a m  is the optimizer used for updating the model weights during training. To reach the optimal configuration, a set of hyperparameters must be explored to find the best model structure and training policy. The Web platform (see Section 2.7) employs six Dense-Ynet instances, corresponding to the best hyper-parameters sets. Using an ensemble approach, the final result presented to the user combines the values received from the individual models.
The PARDS-CxR application detects lung quadrants consolidation, and the final ARDS label is derived from the quadrant predictions using Equation (1).

2.5. Scoring Scheme

To analyze the severity of ARDS in CXR images, a scoring scheme is proposed based on the number and the position of affected lung quadrants (see Table 2). The scheme is compatible with the Berlin definition, in which existing bilateral infiltrates are an essential criterion for ARDS diagnosis in chest radiography.
Giving scores is important from two points of view. First, the score represents the severity of the diffused infiltrates throughout the lungs. Second, reporting disease severity helps clinicians follow appropriate treatment protocols or triaging. This type of system has been proposed for the Murray Lung Injury Score, as well as as part of the recently proposed RALE score in adult patients with ARDS.

2.6. Evaluation Metrics

Evaluation metrics are measured from the algorithm’s performance on unseen test data to assess the approach. There is no metric representing the total capacity of the PARDS-CxR platform. However, we use a set of performance metrics to provide a complete overview of the model’s operation. A confusion matrix quantifies the ability of the classifier to detect each class separately. It gives detailed measures comparing the actual and predicted labels, as shown in Figure 3.
The elements of the confusion matrix, namely, the true positive (TP), true negative (TN), false positive (FP), and false negative (FN) values, serve to calculate several assessment metrics as follows:
A c c u r a c y = T P + T N F P + F N + T P + T N
P r e c i s i o n = T P F P + T P
R e c a l l = T P F N + T P
F 1 = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l .
The  A c c u r a c y  metric represents the overall correctness of a classification algorithm. It cannot fully express the model performance, however, especially in the case of unbalanced testing data.  P r e c i s i o n  and  R e c a l l  reveal the model’s performance in discriminating between the different classes.  P r e c i s i o n  represents how precise the model is in identifying the target (positive) class. Specifically, it points out what portion of cases predicted as positive are really ARDS cases. On the other hand, the  R e c a l l  value shows what proportion of predicted ARDS cases are actually labeled as ARDS. These two metrics have a complementary role in describing the model’s behavior. The  F 1  score, derived from  P r e c i s i o n  and  R e c a l l  values, is a single metric to quantify the algorithm’s performance.
The receiver operating characteristic (ROC) curve illustrates the diagnostic capacity of a system by comparing true positive and false positive rates as the discrimination threshold (applied at the network’s output layer to decide between the two classes) varies. The area under the ROC curve (AUROC) represents the discriminatory power of the classifier.

2.7. Web-Based Platform

We designed a web-based platform to facilitate the diagnosis of ARDS in CXR images by medical professionals. The platform is intended as a tool to provide a second opinion to clinicians, but no direct medical use is recommended until medical professionals validate the tool using external data. The PARDS-CxR platform takes advantage of six Dense-Ynet instances to provide scores for each input image. The scores are given based on the number and combination of affected lung quadrants as explained in Section 2.5. A global score is assigned by combining the outputs from the model instances. In addition, the application provides accurate lung segmentation maps, which are helpful in AI-based analysis of CXR images.
The web application utilizes the React library to create a user-friendly and interactive user interface (UI) for delivering the specified services. The library enables efficient code writing and makes it easier to manage, refine, and integrate the application with other tools. The platform supports both English and French languages and has two main modes for ARDS definitions for adults (Berlin) and children (PALICC-2). The difference between the modes is that, when using PALICC-2 mode, the platform requires two input images. The application response includes segmentation maps, severity scores (local and global), and an interpretation based on the definition.
Although the deep models are trained using graphical processing units (GPUs), the evaluation model does not require a GPU and can process the results in 2–3 s. Thus, the running bottleneck could be the network connection speed. The application is capable of storing data and providing log files, but this feature is currently disabled and will be activated when the validation protocol is approved. The PARDS-CxR platform is detailed further in Section 3.3.

3. Results

3.1. Quadrant-Based Classification

The PARDS-CxR web-based platform uses Dense-Ynet as the joint segmentation-classification model. In classification, the model predicts four labels associated with lung quadrants, as explained in Section 2.3. The platform uses an ensemble of six Dense-Ynet model instances with different training and model structure configurations. Regarding model structures, we experimented with different channel depths in convolution blocks, loss functions, weights for merging loss functions, activation functions, and initial network weights. For the training configurations, we varied several hyperparameters, namely, the learning rate, training batch size, augmentation probability, and stopping criterion.
Figure 4 shows the confusion matrix of the ensemble of models. To merge the results from the model instances, a hard voting strategy is employed based on the labels predicted independently by the models. To be precise, each model is trained separately with its specific configuration. The testing is also done independently, and if at least three models decide that an image is an ARDS case, the combined result is positive. By combining models with various configurations, the intrinsic biases of each one to accept or reject an image as ARDS are balanced in the ensemble output. Thus, the final performance improves compared to any individual model.
Table 3 compares the classification performances of the Dense-Ynet instances in terms of the four assessment metrics seen previously. Some of the listed models achieve higher precision, while others reach better recall values. By combining the predicted labels provided by these models, the ensemble algorithm achieves the highest  F 1  score, representing the best compromise between precision and recall. Indeed, ensembling the models does not outperform every one in terms of  P r e c i s i o n  and  R e c a l l , but the final  F 1  and accuracy values improve.
In this paper, the problem of ARDS diagnosis is based on the classification of lung quadrants. Thus, the task can also be considered as a multi-label classification problem. Figure 5 shows the ROC curves of all quadrants’ predictions for the Dense-Ynet instances, i.e., the ROC curves associated with the binary classification of the lung quadrants, regardless of their positions. The AUROC metric is not directly related to the system’s performance in ARDS diagnosis, but the misclassification of one lung quadrant may cause an error in classifying the image as a whole.

3.2. ARDS Severity Prediction

As seen in Table 2, the application determines the severity of ARDS in CXR images based on the number and combination of affected lung quadrants. The platform provides a global score for each input image by taking the average of the scores from each model. CXR images are then categorized into one of three severity grades based on the predicted scores: non-ARDS, mild ARDS, and severe ARDS. The platform’s effectiveness in determining ARDS severity is illustrated in Figure 6. The three-class confusion matrix shows that the approach can detect ARDS and discriminate between mild and severe states of the syndrome.

3.3. PARDS-CxR, the Web-Based Platform

Our web application is currently loaded on a web server at CHUSJ and is accessible at the address (https://chestxray-reader.chusj-sip-ia.ca, accessed on 15 January 2023). The process of training and testing the deep model was programmed in Python using the PyTorch library [23]. The training process and hyperparameter search were executed on GPU, as they required intensive parallel computing. The trained model was then transferred to CPU to evaluate new images; thus, no graphical processor is necessary on the server to run the application. The graphical user interface was written in JavaScript and is compatible with various internet browsers on the client side. No data are kept on the server side, and the application output image is available to store in the user’s local storage. The user interface works in English and French, and CXR images can be uploaded using the menu option or drag-and-drop (see Figure 7).
The application bases itself on the most accepted definitions for ARDS and PARDS. Based on the Berlin definition, the presence of bilateral infiltrates in chest radiography is a criterion manifesting the existence of ARDS [1]. The platform processes the image and displays its decision by providing a percentage associated with the level of infiltration in each quadrant (Figure 7). A global percentage is also given based on infiltrate levels of infiltrates in quadrants and their combination as in Table 2. This value represents the severity of ARDS in the input image. An image with a global percentage above  60 %  is interpreted as an ARDS case, since, based on the proposed severity scoring system, infiltrates should be diffused through both lungs. Reporting each quadrant’s involvement is necessary, since it gives the rationale behind the global severity measure. As seen in Figure 7, a segmentation map highlighting the lung segments is also provided.
Identifying progression of ARDS is also possible, as two images taken at different times can be compared by the system. Additionally, an example of CXR image comparison is displayed in Figure 8.

4. Discussion

The proposed DenseY-net is a joint segmentation–classification model that diagnoses (P)ARDS based on lung quadrant-level classification. The results show that the model can accurately classify quadrants and, consequently, the entire input image. This labeling strategy offers a reasoning framework for decision-making and incorporates an interpretability feature into the platform. Ensemble modeling is used to combine the outcomes from six model instances. PARDS-CxR can also do lung field segmentation, which is a necessary element in many decision support systems. Our approach performs well in detecting the severity of ARDS by giving a score to each input determined by the number and position of affected lung quadrants. This makes the model compatible with both ARDS and PARDS definitions.
A few large chest radiography datasets are publicly available for the research community [13,24,25]. A key benefit of deep learning is its capacity to analyze and learn features from a substantial amount of data. Therefore, it is unsurprising that several ML researchers have investigated CXR image analysis in various contexts. However, important limitations of these datasets make them unsuitable for developing dependable systems for the hospital setting. Indeed, most of the data are annotated using clinicians’ notes processed by natural language processing (NLP) techniques [26]. This leads to erroneous labeling of a portion of the images. For example, a  10 %  error is reported for Chest X-ray 14 [13], even though it is one of the most frequently used CXR datasets. The clinical review in [27] reveals an even higher rate of data labeling errors in that dataset.
Although adding some level of noise to the training inputs can improve a deep model’s performance, biases and extensive labeling errors will decrease the model’s accuracy. This could be a reason for the relatively poor generalization ability of deep models when confronting new samples from other data sources. Furthermore, available samples are annotated for a limited number of pathologies. Public CXR datasets cover between 14 and 18 chest pathologies, but these do not include ARDS or PARDS. To address this constraint, we collected our own CXR dataset from three different sources and annotated it for PARDS at CHUSJ. This dataset was labeled at the lung quadrant level, and the lung fields were manually identified in each image to establish a segmentation ground truth. The resulting dataset contains 356 CXR images, including 134 that meet the bilateral criteria for ARDS. Annotating data is costly in the clinical field, even more so considering that the DenseY-net model needs lung maps and quadrant-level ground-truth labels. Consequently, our ARDS dataset is relatively small. Nonetheless, our model is designed in such a way as to train adequately on small datasets. The specialized connectivity within the model allows for the creation of a lighter model with shallower intermediate feature maps, resulting in a smaller number of training parameters. A model with fewer parameters is more appropriate for training with small datasets. The algorithm was assessed on our own dataset, as explained in Section 2.4. A bigger dataset could increase the generalization capacity of the model ensemble. Moreover, external validation of the platform using data from various health centers will make it more reliable as a tool for prospective clinical research. Thus, as next steps in the web application’s development, external validation and improving interpretability are two major points, since both are necessary to turn the platform into a practical tool in clinics.
Moreover, according to the (P)ARDS definition, co-occurrence of detectable infiltrates in CXR and hypoxemia is necessary when no evidence of cardiogenic pulmonary edema is observed. Thus, although the presence of infiltrates in chest radiography is known as the most limiting factor for diagnosing ARDS, meeting other criteria is a requisite. The Clinical Decision Support System (CDSS) lab at CHUSJ has the capacity to investigate other ARDS diagnosis criteria, including cardiac failure and hypoxemia. Le et al. have employed NLP techniques and ML algorithms to detect cardiac failure in children [28]. Sauthier et al. have developed a method to accurately estimate Pao2 levels using noninvasive data [29]. Integrating the tool proposed in this study with other works will lead to a system giving comprehensive ARDS diagnoses. Sufficient electronic medical infrastructure is available in the PICU of CHUSJ to facilitate the flow of data from various sources [30]. By accessing data from clinical narrative analysis, measuring oxygenation indices, and detecting infiltrates in CXR images, it will be possible to make clinical decisions in real time. Therefore, an important objective for our team will be to implement an ARDS diagnosis package at CHUSJ, integrating all these criteria and data sources.
The strength of this study lies in the development of an algorithm that, in comparison to existing approaches, is more interpretable and automated and is compatible with existing ARDS definitions. Unlike an earlier ARDS diagnosis method proposed by our research team [8], the DL-based approach used in this application does not need any interaction from clinicians or operators to guide the algorithm. The novel model provides an end-to-end process that is simple for the user and provides the diagnotic outputs instantaneously. Recently, Sjoding et al. [10] proposed annother automatic algorithm for detecting ARDS in CXR images. However, their approach lacks explainability, i.e., the system’s decisions are not supported by further information. By contract, since PARDS-CxR detects infiltrates in each lung quadrant, the basis for the decision is integral to our method. This strengthens the platform’s reliability, since the user can reject or accept the decision by observing the delivered explanation. In addition, the proposed approach is compatible with both PARDS and ARDS definitions [1,3], as the scoring scheme used translates to a disease severity level. At present, the main limitation of our algorithm is its lack of external validation. Indeed, its development relied on a limited number of CXR images with a single team annotating them. For this reason, we have implemented the algorithm on a web platform to allow researchers to conduct validation studies.

5. Conclusions

This work has described a deep learning method and web-based platform for diagnosing acute respiratory distress syndrome (ARDS) from chest X-ray (CXR) images. The platform uses an ensemble of novel Dense-Ynet networks that can accurately detect lung infiltrates in different quadrants and combine this information to detect ARDS and grade its severity. This approach ensures that our tool is compatible with various ARDS definitions in both adults and children. Following feedback from clinical researchers during a validation phase, the platform will be integrated into a complete clinical decision system for ARDS. The tool presented here will serve as the CXR analysis component within an AI-based framework that will monitor other factors, such as hypoxemia and occurence of cardiac arrest.

Author Contributions

P.J., F.C., M.Y. and R.G.K. conceptualized and designed the study. M.Y., P.J., F.C., D.F., M.L. and J.R. developed the study protocol. M.Y., D.F., M.L. and J.R. conducted the algorithm development. M.Y., F.C. and P.J. drafted the initial manuscript. All authors approved the final manuscript as submitted. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by grants from IVADO (Artificial Intelligence Research and Transfer Institute), the Quebec Ministry of Health and Sainte-Justine Hospital. M.Y. is financed by an award from the Fonds de Recherche en Santé du Québec (FRQS) Chair in Artificial Intelligence and Health. P.J. earns a research salary from FRQS.

Institutional Review Board Statement

The study was approved by the Institutional Review Board of Sainte-Justine Hospital (approval number: 2023-5124).

Informed Consent Statement

The study was carried out on a research database and the Institutional Review Board did not require informed consent.

Data Availability Statement

Access to data can be requested from Philippe Jouvet. Specific institutional review board rules will apply.

Acknowledgments

The authors gratefully acknowledge Philippe Debanné for his assistance in reviewing the manuscript. Furthermore, the authors thank all the investigators (pediatric intensivists/Radiologists) who participated in the PARDIE V3 study (Country, site and investigator list): Argentina. Hospital De Ninos Ricardo Gutierrez: Rossana Poterala; Hospital de Ninos sor Maria Ludovica: Pablo Castellani/Martin Giampieri/Claudia Pedraza; Hospital Nacional Alejandro Posadas: Nilda Agueda Vidal/Deheza Rosemary/Gonzalo Turon/Cecilia Monjes; Hospital Pediatrico Juan Pablo II: Segundo F. Espanol; Hospital Universitario Austral: Alejandro Siaba Serrate/Thomas Iolster/Silvio Torres; Sanatorio de Ninos de Rosario: Fernando Paziencia. Australia. Princess Margaret Hospital for Children: Simon Erickson/Samantha Barr/Sara Shea. Bolivia. Hospital del Nino Manuel Ascencio Villaroel: Alejandro F. Martinez Leon/Gustavo A. Guzman Rivera. Canada. CHU Sainte-Justine: Philippe Jouvet/Guillaume Emeriaud/Mariana Dumitrascu/Mary Ellen French. Chile. Hospital Base de Valdivia: Daniel Caro I/Andrés A Retamal Caro; Hospital El Carmen de Maipu: Pablo Cruces Romero/Tania Medina; Hospital Luis Calvo Mackenna: Carlos Acuna; Hospital Padre Hurtado: Franco Diaz/Maria Jose Nunez. China. Children’s Hospital of Fudan Univ: Yang Chen. Colombia. Clinica Infantil de Colsubsidio: Rosalba Pardo Carrero; Hospital General de Medellin: Yurika P. Lopez Alarcon; Hospital Militar Central: Ledys María Izquierdo; Hospital Pablo Tobon Uribe (HPTU): Byron E. Piñeres Olave. France. CHU de Nantes: Pierre Bourgoin; Hopital d’enfants de Brabois–CHU de Nancy: Matthieu Maria. Greece. University of Crete, University Hospital PICU: George Briassoulis/Stavroula Ilia. Italy. Children’s Hospital Bambino Gesu: Matteo Di Nardo/Fabrizio Chiusolo/Ilaria Erba/Orsola Gawronski; Children’s Hospital Vittore Buzzi: Anna Camporesi. Japan. Hiroshima University: Nobuaki Shime/Shinichiro Ohshimo/Yoshiko Kida/Michihito Kyo. Malaysia. Universiti Kebangsaan Malaysia: Swee Fong Tang/Chian Wern Tai; University Malaya Medical Center: Lucy Chai See Lum/Ismail Elghuwael. Mexico. Hospital Espanol De Mexico: Nestor J. Jimenez Rivera. Peru. Hospital de Emergencias Pediatricas: Daniel Vasquez Miranda/Grimaldo Ramirez Cortez; Instituto Nacional de Salud del Nino: Jose Tantalean. Portugal. Hospital Santa Maria–Centro Hospitalar Lisboa Norte: Cristina Camilo. Saudi Arabia. King Abdullah Specialist Children’s Hospital, King Abdulaziz Medical City: Tarek Hazwani/Nedaa Aldairi/Ahmed Al Amoudi/Ahmad Alahmadti. Spain. Cruces University Hospital: Yolanda Lopez Fernandez/Juan Ramon Valle/Lidia Martinez/Javier Pilar Orive; Hospital Regional Universitario de Malaga: Jose Manuel Gonzalez Gomez/Antonio Morales Martinez; Hospital Universitari I Politecnic La Fe: Vicent Modesto I Alapont; Sant Joan de Deu University Hospital: Marti Pons Odena; Hospital Universitario Central De Asturias: Alberto Medina; Virgen de la Arrixaca University Hospital: Susana Reyes Dominguez. Turkey. Akdeniz University School of Medicine: Oguz Dursun/Ebru Atike Ongun; Izmir Katip Celebi University Medical School and Tepecik Research and Training Hospital: Fulya Kamit Can/Ayse Berna Anil. UK. Evelina London Children’s Hospital: Jon Lillie/Shane Tibby/Paul Wellman/Holly Belfield/Claire Lloyd; Great Ormond St. Children’s Hospital: Joe Brierley/Troy E. Dominguez/Eugenia Abaleke/Yael Feinstein; Noah’s Ark Children’s Hospital for Wales: Siva Oruganti/Sara Harrison; Nottingham University Hospitals: Catarina Silvestre; Oxford Radcliffe Hospitals NHS Foundation Trust: James Weitz; Royal Manchester Children’s Hospital: Peter-Marc Fortune/Gayathri Subramanian/Claire Jennings; St. Mary’s Hospital: David Inwald/Calandra Feather/May-Ai Seah/Joanna Danin. USA. Arkansas Children’s Hospital: Ron Sanders/ Glenda Hefley/Katherine Irby/Lauren Edwards/Robert F Buchmann; Children’s Hospital and Medical Center: Sidharth Mahapatra/Edward Truemper/Lucinda Kustka; Children’s Hospital at Dartmouth: Sholeen T. Nett/Marcy Singleton/J. Dean Jarvis; Children’s Hospital Colorado: Aline B. Maddux/Peter M. Mourani/Kimberly Ralston/Yamila Sierra/Jason Weinman/Zach VanRheen/Christopher Newman; Children’s Hospital Los Angeles: Robinder Khemani/Christopher Newth/Jeni Kwok/Rica Morzov/Natalie Mahieu; Children’s Hospital of Philadelphia: Nadir Yehya/Natalie Napolitano/Marie Murphy/Laurie Ronan/Ryan Morgan/Sherri Kubis/Elizabeth Broden; Children’s Hospital of Wisconsin: Rainer Gedeit/Kathy Murkowski/Katherine Woods/Mary Kasch; Children’s Mercy Hospital and Clinics: Yong Y. Han/Jeremy T. Affolter/Kelly S. Tieves/Amber Hughes-Schalk; Cincinnati Children’s Hospital Medical Center: Ranjit S. Chima/Kelli Krallman/Erin Stoneman/Laura Benken/Toni Yunger; Connecticut Children’s Medical Center: Christopher L Carroll/James Santanelli; Inova Children’s Hospital: W. Keith Dockery/Shirin Jafari-Namin/Dana Barry/Keary Jane’t; Joseph M Sanzari Children’s Hospital at Hackensack University Medical Center: Shira Gertz; Nicklaus Children’s Hospital: Fernando Beltramo/Balagangadhar Totapally/Beatriz Govantes; Northwestern University, Ann & Robert H Lurie Children’s Hospital of Chicago: Bria Coates/Lawren Wellisch/Kiona Allen/Avani Shukla; Penn State Hershey Children’s Hospital: Neal J. Thomas/Debbie Spear; Rainbow Babies and Children’s Hospital, Steven L. Shein/Pauravi Vasavada; Saint Barnabas Medical Center: Shira Gertz; Stony Brook Children’s Hospital: Margaret M. Parker/Daniel Sloniewsky; The Children’s Hospital of Oklahoma; Christine Allen/Amy Harrell; UCSF Benioff Children’s Hospital Oakland: Natalie Cvijanovich; University of Miami/Holtz Children’s Hospital: Asumthia S. Jeyapalan/Alvaro Coronado-Munoz; University of Michigan–C.S. Mott Children’s Hospital: Heidi Flori/Mary K. Dahmer/Chaandini Jayachandran/Joseph Kohne; University of Minnesota Masonic Children’s Hospital: Janet Hume/Dan Nerheim/Kelly Dietz; University of WA/Seattle Children’s Hospital: Lincoln Smith/Silvia Hartmann/Erin Sullivan/Courtney Merritt; Weill Cornell Medical College: Deyin D. Hsing/Steve Pon/Jim Brian Estil/Richa Gautam; Yale School of Medicine: John S. Giuliano Jr./Joana Ta.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AIArtificial Intelligence
ARDSAcute Respiratory distress Syndrome
AUROCArea Under the ROC Curve
CDSSClinical Decision Support System
CHUSJCentre Hospitalier Universitaire Sainte-Justine (Sainte-Justine Hospital)
CXRChest X-ray
CPUCentral Processing Unit
DLDeep Learning
GPUGraphical Processing Unit
LLLeft Lower
LULeft Upper
MLMachine Learning
NLPNatural Language Processing
PALICCPediatric Acute Lung Injury Consensus Conference
PARDSPediatric Acute Respiratory Distress Syndrome
PICUPediatric Intensive Care Unit
ReLURectified Linear Unit
RLRight Lower
ROCReceiver Operating Characteristic
RURight Upper
UIUser Interface

References

  1. Force, A.D.T.; Ranieri, V.; Rubenfeld, G.; Thompson, B.; Ferguson, N.; Caldwell, E.; Fan, E.; Camporota, L.; Slutsky, A. Acute respiratory distress syndrome. JAMA 2012, 307, 2526–2533. [Google Scholar]
  2. Sedhai, Y.R.; Yuan, M.; Ketcham, S.W.; Co, I.; Claar, D.D.; McSparron, J.I.; Prescott, H.C.; Sjoding, M.W. Validating measures of disease severity in acute respiratory distress syndrome. Ann. Am. Thorac. Soc. 2021, 18, 1211–1218. [Google Scholar] [CrossRef]
  3. Pediatric Acute Lung Injury Consensus Conference Group. Pediatric acute respiratory distress syndrome: Consensus recommendations from the Pediatric Acute Lung Injury Consensus Conference. Pediatr. Crit. Care Med. J. Soc. Crit. Care Med. World Fed. Pediatr. Intensive Crit. Care Soc. 2015, 16, 428. [Google Scholar]
  4. Khemani, R.G.; Smith, L.; Lopez-Fernandez, Y.M.; Kwok, J.; Morzov, R.; Klein, M.J.; Yehya, N.; Willson, D.; Kneyber, M.C.; Lillie, J.; et al. Paediatric acute respiratory distress syndrome incidence and epidemiology (PARDIE): An international, observational study. Lancet Respir. Med. 2019, 7, 115–128. [Google Scholar] [CrossRef]
  5. Emeriaud, G.; López-Fernández, Y.M.; Iyer, N.P.; Bembea, M.M.; Agulnik, A.; Barbaro, R.P.; Baudin, F.; Bhalla, A.; de Carvalho, W.B.; Carroll, C.L.; et al. Executive Summary of the Second International Guidelines for the Diagnosis and Management of Pediatric Acute Respiratory Distress Syndrome (PALICC-2). Pediatr. Crit. Care Med. 2023, 24, 143–168. [Google Scholar] [CrossRef] [PubMed]
  6. López-Fernández, Y.M.; Smith, L.S.; Kohne, J.G.; Weinman, J.P.; Modesto-Alapont, V.; Reyes-Dominguez, S.B.; Medina, A.; Piñeres-Olave, B.E.; Mahieu, N.; Klein, M.J.; et al. Prognostic relevance and inter-observer reliability of chest-imaging in pediatric ARDS: A pediatric acute respiratory distress incidence and epidemiology (PARDIE) study. Intensive Care Med. 2020, 46, 1382–1393. [Google Scholar] [CrossRef]
  7. Sjoding, M.W.; Hofer, T.P.; Co, I.; Courey, A.; Cooke, C.R.; Iwashyna, T.J. Interobserver reliability of the Berlin ARDS definition and strategies to improve the reliability of ARDS diagnosis. Chest 2018, 153, 361–367. [Google Scholar] [CrossRef] [PubMed]
  8. Zaglam, N.; Jouvet, P.; Flechelles, O.; Emeriaud, G.; Cheriet, F. Computer-aided diagnosis system for the Acute Respiratory Distress Syndrome from chest radiographs. Comput. Biol. Med. 2014, 52, 41–48. [Google Scholar] [CrossRef]
  9. Plourde, F.; Cheriet, F.; Dansereau, J. Semiautomatic Detection of Scoliotic Rib Borders From Posteroanterior Chest Radiographs. IEEE Trans. Biomed. Eng. 2012, 59, 909–919. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Sjoding, M.W.; Taylor, D.; Motyka, J.; Lee, E.; Co, I.; Claar, D.; McSparron, J.I.; Ansari, S.; Kerlin, M.P.; Reilly, J.P.; et al. Deep learning to detect acute respiratory distress syndrome on chest radiographs: A retrospective study with external validation. Lancet Digit. Health 2021, 3, e340–e348. [Google Scholar] [CrossRef] [PubMed]
  11. Kexugit. Using Microsoft AI to Build a Lung-Disease Prediction Model Using Chest X-ray Images. 2018. Available online: https://learn.microsoft.com/en-ca/archive/blogs/machinelearning/using-microsoft-ai-to-build-a-lung-disease-prediction-model-using-chest-x-ray-images (accessed on 15 January 2023).
  12. Cohen, J.P.; Bertin, P.; Frappier, V. Chester: A web delivered locally computed chest X-ray disease prediction system. arXiv 2019, arXiv:1901.11210. [Google Scholar]
  13. Wang, X.; Peng, Y.; Lu, L.; Lu, Z.; Bagheri, M.; Summers, R.M. ChestX-ray8: Hospital-scale Chest X-ray Database and Benchmarks on Weakly-Supervised Classification and Localization of Common Thorax Diseases. arXiv 2017, arXiv:1705.02315. [Google Scholar]
  14. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE international Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
  15. Arun, N.; Gaw, N.; Singh, P.; Chang, K.; Aggarwal, M.; Chen, B.; Hoebel, K.; Gupta, S.; Patel, J.; Gidwani, M.; et al. Assessing the trustworthiness of saliency maps for localizing abnormalities in medical imaging. Radiol. Artif. Intell. 2021, 3, e200267. [Google Scholar] [CrossRef] [PubMed]
  16. Ahmed, K.B.; Goldgof, G.M.; Paul, R.; Goldgof, D.B.; Hall, L.O. Discovery of a generalization gap of convolutional neural networks on COVID-19 X-rays classification. IEEE Access 2021, 9, 72970–72979. [Google Scholar] [CrossRef] [PubMed]
  17. Çallı, E.; Sogancioglu, E.; van Ginneken, B.; van Leeuwen, K.G.; Murphy, K. Deep learning for chest X-ray analysis: A survey. Med. Image Anal. 2021, 72, 102125. [Google Scholar] [CrossRef] [PubMed]
  18. Jaeger, S.; Candemir, S.; Antani, S.; Wáng, Y.X.J.; Lu, P.X.; Thoma, G. Two public chest X-ray datasets for computer-aided screening of pulmonary diseases. Quant. Imaging Med. Surg. 2014, 4, 475. [Google Scholar]
  19. Teixeira, L.O.; Pereira, R.M.; Bertolini, D.; Oliveira, L.S.; Nanni, L.; Cavalcanti, G.D.; Costa, Y.M. Impact of lung segmentation on the diagnosis and explanation of COVID-19 in chest X-ray images. Sensors 2021, 21, 7116. [Google Scholar] [CrossRef]
  20. de Sousa Freire, N.; de Souza Leão, P.P.; Tiago, L.A.; Gonçalves, A.d.A.C.; Pinto, R.A.; dos Santos, E.M.; Souto, E. Generalizability of CNN on Predicting COVID-19 from Chest X-ray Images. In 2022: Proceedings of the 22nd Brazilian Symposium on Computing Applied to Health; SBCAS 2022 Companion Proceedings Series; Brazilian Computing Society (SBC): Porto Alegre, Brazil, 2022; pp. 36–47. [Google Scholar]
  21. Yahyatabar, M.; Jouvet, P.; Cheriet, F. Dense-Unet: A light model for lung fields segmentation in Chest X-ray images. In Proceedings of the 2020 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Montreal, QC, Canada, 20–24 July 2020; pp. 1242–1245. [Google Scholar]
  22. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar] [CrossRef] [Green Version]
  23. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. Pytorch: An imperative style, high-performance deep learning library. Adv. Neural Inf. Process. Syst. 2019, 32, 1–12. [Google Scholar]
  24. Rajpurkar, P.; Irvin, J.; Zhu, K.; Yang, B.; Mehta, H.; Duan, T.; Ding, D.; Bagul, A.; Langlotz, C.; Shpanskaya, K.; et al. Chexnet: Radiologist-level pneumonia detection on chest X-rays with deep learning. arXiv 2017, arXiv:1711.05225. [Google Scholar]
  25. Johnson, A.E.; Pollard, T.J.; Berkowitz, S.J.; Greenbaum, N.R.; Lungren, M.P.; Deng, C.y.; Mark, R.G.; Horng, S. MIMIC-CXR, a de-identified publicly available database of chest radiographs with free-text reports. Sci. Data 2019, 6, 317. [Google Scholar] [CrossRef] [Green Version]
  26. Towfighi, S.; Agarwal, A.; Mak, D.; Verma, A. Labelling chest X-ray reports using an open-source NLP and ML tool for text data binary classificatio. medRxiv 2019. [Google Scholar] [CrossRef]
  27. Oakden-Rayner, L. Exploring the ChestXray14 Dataset: Problems. 2017. Available online: https://laurenoakdenrayner.com/2017/12/18/the-chestxray14-dataset-problems/ (accessed on 15 January 2023).
  28. Le, T.D.; Noumeir, R.; Rambaud, J.; Sans, G.; Jouvet, P. Detecting of a patient’s condition from clinical narratives using natural language representation. IEEE Open J. Eng. Med. Biol. 2022, 3, 142–149. [Google Scholar] [CrossRef] [PubMed]
  29. Sauthier, M.; Tuli, G.; Jouvet, P.A.; Brownstein, J.S.; Randolph, A.G. Estimated Pao2: A continuous and noninvasive method to estimate Pao2 and oxygenation index. Crit. Care Explor. 2021, 3, e0546. [Google Scholar] [CrossRef] [PubMed]
  30. Brossier, D.; El Taani, R.; Sauthier, M.; Roumeliotis, N.; Emeriaud, G.; Jouvet, P. Creating a high-frequency electronic database in the PICU: The perpetual patient. Pediatr. Crit. Care Med. 2018, 19, e189–e198. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Organization of our study into four main phases. The data are collected from several data sources and annotated at Saint-Justine Hospital, Montreal, Canada. The DL model is trained using quadrant-level labels and lung segmentation maps. It is then evaluated on a set of previously unseen images; both the classification and segmentation performances are assessed. Finally, a web-based platform is designed and made available through the internet.
Figure 1. Organization of our study into four main phases. The data are collected from several data sources and annotated at Saint-Justine Hospital, Montreal, Canada. The DL model is trained using quadrant-level labels and lung segmentation maps. It is then evaluated on a set of previously unseen images; both the classification and segmentation performances are assessed. Finally, a web-based platform is designed and made available through the internet.
Diagnostics 13 00933 g001
Figure 2. The Dense-Ynet model takes advantage of the interaction between the segmentation and classification tasks by performing them simultaneously. The features from the original and lung-segmented images are concatenated and utilized to classify ARDS cases.
Figure 2. The Dense-Ynet model takes advantage of the interaction between the segmentation and classification tasks by performing them simultaneously. The features from the original and lung-segmented images are concatenated and utilized to classify ARDS cases.
Diagnostics 13 00933 g002
Figure 3. Confusion matrix for a binary classification problem. The matrix contains four elements that, together, evaluate the system’s predictions versus the real labels.
Figure 3. Confusion matrix for a binary classification problem. The matrix contains four elements that, together, evaluate the system’s predictions versus the real labels.
Diagnostics 13 00933 g003
Figure 4. Final confusion matrix obtained from the combination of network instances using hard voting. The numbers (percentages) are obtained by taking the average of several tests (five-fold cross validation).
Figure 4. Final confusion matrix obtained from the combination of network instances using hard voting. The numbers (percentages) are obtained by taking the average of several tests (five-fold cross validation).
Diagnostics 13 00933 g004
Figure 5. ROC curves for classification of lung quadrants regardless of their position in the lungs.
Figure 5. ROC curves for classification of lung quadrants regardless of their position in the lungs.
Diagnostics 13 00933 g005
Figure 6. Confusion matrix for classification of ARDS severity with three levels (none, mild, severe).
Figure 6. Confusion matrix for classification of ARDS severity with three levels (none, mild, severe).
Diagnostics 13 00933 g006
Figure 7. Main interface of the PARDS-CxR web application. In the standard mode, a single CXR image is analyzed according to the Berlin definition.
Figure 7. Main interface of the PARDS-CxR web application. In the standard mode, a single CXR image is analyzed according to the Berlin definition.
Diagnostics 13 00933 g007
Figure 8. PARDS-CxR interface in image comparison mode. The platform can analyze two CXR images to detect ARDS progression based on the PALICC-2 definition.
Figure 8. PARDS-CxR interface in image comparison mode. The platform can analyze two CXR images to detect ARDS progression based on the PALICC-2 definition.
Diagnostics 13 00933 g008
Table 1. ARDS severities in the Berlin definition and associated oxygenation levels and mortality rates [1].
Table 1. ARDS severities in the Berlin definition and associated oxygenation levels and mortality rates [1].
SeverityPaO2/FiO2Mortality
Mild200–30027%
Moderate100–20032%
Severe≤10045%
Table 2. Severity scoring scheme based on affected lung quadrants.
Table 2. Severity scoring scheme based on affected lung quadrants.
Affected QuadrantsScoreSeverity
4 quadrants5Severe
3 quadrants4
2 quadrants (Different sides)3Mild
2 quadrants (Same side)2Non-ARDS
1 quadrant1
No affected quadrant0
Table 3. Evaluation of the six models and the result of their combination (ensemble model) for classification.
Table 3. Evaluation of the six models and the result of their combination (ensemble model) for classification.
ModelAccuracyRecallPrecision F 1
Network 1 92.95 % 88.45 % 91.99 % 90.19 %
Network 2 93.54 % 96.41 % 84.37 % 89.99 %
Network 3 92.04 % 94.42 % 87.89 % 91.03 %
Network 4 92.96 % 100.0 % 83.33 % 90.91 %
Network 5 87.32 % 100.0 % 74.29 % 85.25 %
Network 6 88.74 % 80.01 % 80.01 % 80.02 %
Ensemble model 94.35 % 95.25 % 88.02 % 91.49 %
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yahyatabar, M.; Jouvet, P.; Fily, D.; Rambaud, J.; Levy, M.; Khemani, R.G.; Cheriet, F., on behalf of the Pediatric Acute Respiratory Distress Syndrome Incidence and Epidemiology (PARDIE) V3 Investigators and PALISI Network. A Web-Based Platform for the Automatic Stratification of ARDS Severity. Diagnostics 2023, 13, 933. https://doi.org/10.3390/diagnostics13050933

AMA Style

Yahyatabar M, Jouvet P, Fily D, Rambaud J, Levy M, Khemani RG, Cheriet F on behalf of the Pediatric Acute Respiratory Distress Syndrome Incidence and Epidemiology (PARDIE) V3 Investigators and PALISI Network. A Web-Based Platform for the Automatic Stratification of ARDS Severity. Diagnostics. 2023; 13(5):933. https://doi.org/10.3390/diagnostics13050933

Chicago/Turabian Style

Yahyatabar, Mohammad, Philippe Jouvet, Donatien Fily, Jérome Rambaud, Michaël Levy, Robinder G. Khemani, and Farida Cheriet on behalf of the Pediatric Acute Respiratory Distress Syndrome Incidence and Epidemiology (PARDIE) V3 Investigators and PALISI Network. 2023. "A Web-Based Platform for the Automatic Stratification of ARDS Severity" Diagnostics 13, no. 5: 933. https://doi.org/10.3390/diagnostics13050933

APA Style

Yahyatabar, M., Jouvet, P., Fily, D., Rambaud, J., Levy, M., Khemani, R. G., & Cheriet, F., on behalf of the Pediatric Acute Respiratory Distress Syndrome Incidence and Epidemiology (PARDIE) V3 Investigators and PALISI Network. (2023). A Web-Based Platform for the Automatic Stratification of ARDS Severity. Diagnostics, 13(5), 933. https://doi.org/10.3390/diagnostics13050933

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop