Next Article in Journal
Exploring Named Entity Recognition via MacBERT-BiGRU and Global Pointer with Self-Attention
Previous Article in Journal
Application of Task Allocation Algorithms in Multi-UAV Intelligent Transportation Systems: A Critical Review
Previous Article in Special Issue
An Improved Deep Learning Framework for Multimodal Medical Data Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Multimodal Machine Learning Model in Pneumonia Patients Hospital Length of Stay Prediction

1
Sub-Intensive Care Unit, Respiratory Physiopathology Department, Cotugno-Monaldi Hospital, AORN Ospedali dei Colli, 80131 Naples, Italy
2
Department of Precision Medicine, University of Campania Luigi Vanvitelli, 80138 Naples, Italy
3
Department of Electrical Engineering and Information Technologies (DIETI), University of Naples Federico II, 80125 Naples, Italy
4
Department of Mental and Physical Health and Preventive Medicine, University of Campania Luigi Vanvitelli, 80138 Naples, Italy
5
Department of Advanced Medical and Surgical Sciences, University of Campania Luigi Vanvitelli, 80138 Naples, Italy
*
Authors to whom correspondence should be addressed.
Big Data Cogn. Comput. 2024, 8(12), 178; https://doi.org/10.3390/bdcc8120178
Submission received: 30 September 2024 / Revised: 26 November 2024 / Accepted: 28 November 2024 / Published: 3 December 2024

Abstract

:
Hospital overcrowding, driven by both structural management challenges and widespread medical emergencies, has prompted extensive research into machine learning (ML) solutions for predicting patient length of stay (LOS) to optimize bed allocation. While many existing models simplify the LOS prediction problem to a classification task, predicting broad ranges of hospital days, an exact day-based regression model is often crucial for precise planning. Additionally, available data are typically limited and heterogeneous, often collected from a small patient cohort. To address these challenges, we present a novel multimodal ML framework that combines imaging and clinical data to enhance LOS prediction accuracy. Specifically, our approach uses the following: (i) feature extraction from chest CT scans via a convolutional neural network (CNN), (ii) their integration with clinically relevant tabular data from patient exams, refined through a feature selection system to retain only significant predictors. As a case study, we applied this framework to pneumonia patient data collected during the COVID-19 pandemic at two hospitals in Naples, Italy—one specializing in infectious diseases and the other general-purpose. Under our experimental setup, the proposed system achieved an average prediction error of only three days, demonstrating its potential to improve patient flow management in critical care environments.

1. Introduction

Since the healthcare environment faced the emergency brought by the COVID-19 pandemic, researchers and doctors understood how crucial it is to properly organize hospital beds availability in order not to have overcrowded hospital wards and fatal delays. For this reason, in order to prevent these unpleasant and dramatic situations, providing a solution which enables medical staff to acknowledge beforehand how long patients affected by diseases like pneumonia are required to stay in hospital is crucial. Even though there have already been studies on the estimation of hospital length of stay (LOS) with machine learning (ML) approaches, only with the advent of the COVID-19 global pandemic the urge to find a system for this task became prominent, with most of the works released since 2020. The implementation of a prediction system for hospital admissions, as well as other technical solutions aimed at speeding up the hospital sector functioning, could represent an important step towards greater resilience of the national health system. Regarding Italy, the COVID-19 health emergency demonstrated how unprepared the country was to manage such a massive influx of patients, highlighting the potential support such a system could have provided. Planning solutions in advance generally allows for the prevention of many problems and a faster application of timely solutions in emergency contexts.
This specific task, given the limited amount of tabular data and the inherent challenges of regression in this context—where daily precision is crucial—it is particularly difficult to address with existing machine learning models. Most current solutions for length of stay (LOS) prediction rely on classification rather than regression models, with targets that discretize the number of hospital days into ranges. For instance, some models from the literature [1] categorize LOS into short (1–3 days), medium (4–7 days), and long (8+ days) ranges. However, such categorical predictions may be insufficient in settings wherein daily management of hospital capacity is essential to ensure bed availability. Beyond the demands heightened by COVID-19, an automated system capable of optimizing hospital resources could enhance patient care during hospitalization and recovery while reducing stress for healthcare staff.
For this reason, the aim of the proposed work is to provide a solution which, starting from heterogeneous data containing images and tabular data, can understand how long the patient should stay in the hospital. The implemented architecture is designed to deliver accurate LOS predictions, even in the face of limited and diverse data. Importantly, it exploits data heterogeneity as an advantage, integrating information from both medical imaging and clinical records. This unique feature distinguishes our approach from existing methods, which typically rely on a single data type [2,3,4,5]. Moreover, the architecture has the potential for broader application beyond LOS prediction. Given that patient features in the dataset capture their current health status, the system, with targeted modifications in its final layers, could also function as an outcome prediction tool, particularly in predicting patient survival following therapy. This extension may also provide insight into the severity of a patient’s condition. Thus, the system could serve a dual purpose: acting as a classifier for outcome prediction and as a regression model for LOS estimation.
The data used in this study were collected during the COVID-19 pandemic in two Neapolitan hospitals when computed tomography (CT) scans were commonly performed on patients’ chests. As such, the patients may have been diagnosed with either generic pneumonia or COVID-19. However, this variability does not affect the primary goal of this research, which is to analyze generic chest CT scans and clinical data to predict LOS, irrespective of the specific underlying lung pathology. The only limitation is that the illness must affect the lungs, with pneumonia being the most common condition in this dataset.
The proposed ML system accounts for data heterogeneity, both across different data types and within the same data type from varying sources. It also addresses challenges such as dataset imbalance and the varying importance of features. In this work, we describe the design and implementation details of the system.
To sum up, the contribution of this paper can be summarized as follows:
  • We illustrate an approach to extract useful features from lungs CT scans and volumes for LOS prediction;
  • We provide a reliable solution for LOS prediction powered by CT features, whose results are more efficient than SOTA solutions;
  • We expand the solution to be compatible with clinical tabular data, available along with patients’ CT scans, and compared performances given the different types of input;
  • We adapt the proposed methodology for the task of post-therapy survival prediction, showing the generalization ability of the approach;
  • We illustrate and implement a multi-modal architecture capable of compensating the typical lack of data in the biomedical domain.
The rest of this paper is organized as follows: Section 2 briefly analyses the current literature, with an emphasis on the limits of current proposals; Section 3 introduces the considered datasets; Section 4 describes the implemented methodology; Section 5 reports the obtained results; Section 6 highlights some ethical consideration; finally, Section 7 provides some conclusions.

2. Related Works

Several works have been released and promoted in order to provide solutions to the complex task of hospitalization days, or LOS, prediction, for different kinds of diseases.
Since 2014, several studies have addressed the problem of predicting length of stay (LOS) using ML techniques. The such study by Morton et al. [6] presents a comparative analysis of ML models, including Linear Regressors, Support Vector Machines, Multi-Task Learning Models, and Random Forests, applied to clinical data for LOS prediction. Although this study focuses on diabetic patients, the authors suggest that the approach can be generalized to other diseases. This generalization is explored by Pendharkar et al. [7], who compare three different models, namely Classification and Regression Trees (CARTs), Chi-Square Automatic Interaction Detection (CHAID), and Support Vector Regressors (SVRs), all trained on clinical (tabular) data. Similarly, Turgeman et al. [8] employed a cubist regression tree to predict LOS using static clinical data from a broader patient population. In addition, Lorenzoni et al. [9] applied eight different ML models to predict LOS, focusing on patients with heart disease. Similarly, Daghistani et al. [1] developed a system using various ML classification models to predict LOS in terms of day ranges, specifically for cardiovascular disease patients. Saadatmand et al. [10] propose a system with five different ML algorithms for ICU admission, mortality, and LOS discretized in two classes, while Khajehali et al. [11] applied several ensembling models to detect the LOS discretized in five classes.
While the aforementioned works represent some of the earliest contributions to the field, numerous additional studies have emerged since 2020, including several notable works [12,13,14,15,16,17,18,19,20,21] and systematic reviews [2,3,4,5], all of which leverage ML techniques for LOS prediction. However, regardless of the publication year, few studies attempt to address LOS prediction using diverse data types beyond standard clinical data, or employ deep learning approaches. One example is the work by Zebin et al. [22], which uses an autoencoder combined with a deep neural network trained on the MIMIC-III dataset. It is worth noting that the task in this study was simplified to a binary classification problem, with LOS predictions limited to two possible day ranges.
From 2020 onward, predicting LOS in hospitals became an urgent problem, particularly in response to a unique, unprecedented challenge: COVID-19. The urge to understand how long patients would stay in hospital, before leaving beds for new patients, became particularly compelling during the emergency state the world lived in 2020 and 2021. Therefore, not only was there an immediate spread of works on LOS prediction for COVID-19 patients, but in general a huge increase in these kinds of works. One of the earliest studies addressing LOS prediction for COVID-19 patients was conducted by Roimi et al. [23], who proposed a system that used daily clinical data to predict the required hospital utilization for each patient. Similarly, Ebinger et al. [24] developed a system consisting of three ML models specifically designed to predict LOS for COVID-19 patients. In the following year, Alabbad et al. [25] introduced four ML models to predict the LOS of COVID-19 patients in Saudi Arabia, while Etu et al. [26] presented a similar approach with four ML models for patients at a Detroit hospital.
A common feature in these studies is their reliance almost exclusively on clinical tabular data, despite the frequent use of imaging techniques such as chest CT to monitor disease progression. This underutilization of imaging data represents a potential missed opportunity, as scans could provide valuable insights that might enhance the models’ ability to predict how long patients will require hospital care.

3. Dataset

Through collaborations with the University of Campania Luigi Vanvitelli Hospital (Azienda Ospedaliera Universitaria Vanvitelli) and Cotugno Hospital (Azienda Ospedaliera dei Colli Cotugno), we compiled a heterogeneous dataset of 84 patients. Each patient is represented by tabular data from clinical exams and a chest CT scan. Specifically, the dataset from the Hospital of University Vanvitelli (denoted as Vanvitelli dataset), comprising 34 patients, includes a CT scan session per patient, 71 tabular features, and two outcome variables: patient outcome and hospitalization duration. The dataset from the Hospital Cotugno (denoted as Cotugno dataset) consists of 50 patients, each with CT scan sessions, 75 tabular features, and three outcome variables: patient outcome, mortality status (exitus), and length of stay (LOS).
Besides analysing each dataset separately, we also merged them, analyzing the overall performance on a more variegated collection. Due to differences in data sources, the tabular features were not fully aligned, resulting in a merged dataset with only a subset of the available features. Prior to merging, both datasets underwent pre-processing to ensure they contained significant, independent, normalized features without missing values. As a result, the Vanvitelli dataset was reduced to 36 features, while the Cotugno dataset retained at 49 features. After merging, the final dataset included 14 common features along with two shared outputs (LOS and outcome). Additionally, an extra feature, derived from the LOS output, was created for model training. This feature, termed “range”, represents a discretized version of the LOS, with 7-day intervals (e.g., 01–07 days, 08–14 days). The final category included patients with 50 or more days of hospitalization.
Table 1 reports the details of the datasets acquired from Vanvitelli and Cotugno hospitals, respectively, after pre-processing and filtering procedures. The features that are present in both datasets are marked in bold, resulting in the Complete dataset.
Before adopting the listed datasets, the tabular features from clinical exams must be pre-processed in order to be properly balanced and populated. Specifically, features that were empty or nearly empty were removed. Only complete features, or those with minimal missing values that could be reasonably inferred (e.g., missing values in the “COVID therapy” or “Anticoagulant therapy” columns were assumed to indicate no therapy, hence assigned a value of 0), were retained after pre-processing. Furthermore, all non-binary features were normalized based on the range limits derived from the training subsets. Figure 1a,b display the dataset distribution for the two common outcomes, namely patient outcome and length of hospitalization, while Figure 2 illustrates the distribution across the defined hospitalization day ranges. An analysis of the original datasets revealed that, while both are balanced in terms of sex distribution, they are imbalanced with respect to other features, such as age, as well as the output variables. Feature distributions from both datasets are presented in Figure 3, Figure 4 and Figure 5.
Following pre-processing and merging of the datasets, the correlation matrices for the Vanvitelli, Cotugno, and Complete datasets are shown in Figure 6, Figure 7 and Figure 8, respectively. In these matrices, cells colored green represent strong positive correlations between feature pairs, while blue cells indicate negative correlations, meaning the features are still correlated but in an inverse manner. By identifying feature pairs at both extremes—highly correlated and highly anti-correlated—the dataset can be further filtered for model optimization.

4. Proposed Approach

In this paper, we propose an automatic approach to predict the LOS together with the outcome of the patient’s recovery. The proposed methodology is outlined in the following steps, which are described in detail in the subsequent sections:
  • Features extraction and collection;
  • Features selection;
  • LOS and outcome prediction.
The overall system architecture comprises two primary workflows: the dataset generation workflow (Figure 9) and the model training workflow (Figure 10). The dataset generation workflow focuses on feature extraction, as shown in Figure 9. The model training workflow, illustrated in Figure 10, includes steps for feature selection, dataset splitting, normalization, application of balancing algorithms, and the training of machine learning models for specific tasks. These trained models are then used for inference.
Specifically, in the dataset generation workflow, the original datasets are processed to produce an elaborated and filtered version, which contains tables of clinical data, features extracted from CT scans using CNNs, and their concatenation. At this stage, a task can be selected, and the corresponding dataset is prepared for input into the model training workflow.
The training workflow applies pre-processing steps such as normalization and balancing and, based on the selected configuration, determines the approaches to adopt on data, like ensembling or partitioning. In other words, once identified which prediction task to perform, it is possible to choose the dataset to take as input, which features to adopt, whether to apply an ensembling method and a feature selection approach, and the data partitioning approach. The whole pipeline, where each step can be changed according to the preferred configuration, is shown and detailed in Figure 10. In this way, it is possible to have a set of independent combinations, where each combination will provide its performance metrics for a specific task. Once the optimal model and configuration are identified, the system is ready for inference on new data.

4.1. Features Extraction and Collection

The initial intent of the activities for this work was to rely only on deep learning models for the required tasks, in particular by training them on the images’ side of the merged dataset, i.e., the 3D chest CT scans. However, given the low amount of available patients, even after the merging process, performances of any 2D/3D CNN model were unsatisfying for the proposed system, both in the case of fine-tuning and training from scratches. In particular, we tested a 2D model, developed and trained with the PyTorch framework, analyzing the slide wherein the amount of non-zero pixels is the largest, often corresponding to the slice wherein the lungs are at their maximum expansion. On the other hand, 3D models use all the available slices; therefore, no extra processing is needed, apart from the one used to uniform the dataset patients.
Since there were no chances of expanding the dataset with new patients to improve model performances, the focus moved towards the implementation of a machine learning model exploiting pre-trained CNNs as feature extractors for the chest CT scan. In order to accomplish this new task, depending on the original input adopted, which could be either the single slide with the greatest number of non-zero pixels or the whole 3D CT scan, a 2D and a 3D version of ResNet50 were chosen. In particular, the 2D version had pre-trained weights from ImageNet, while the 3D version had pre-trained weights from MedicalNet [27]. Both models were modified with the addition of an extra fully connected layer placed before the output layer, which is useful to fix the number of outputs representing the desired features obtained from a given input scan. A representation of this procedure on the original ResNet50 architecture is shown in Figure 11. These models were fine-tuned for 80 epochs (which resulted in the best amount before resulting in overfitting) and then adopted as feature extractors, thanks to the added fully connected layer. In both cases, the extra fully connected layer has an output vector of length 16, meaning that 16 “features” can be extracted from both single slices and CT volumes. Given these two sets, for each patient of every dataset, there are
  • A variable number of features (depending on the source) from clinical exams;
  • A total of 16 features from 2D slices;
  • A total of 16 features from 3D volumes.

4.2. Features Selection

Focusing on the hospitalization days prediction task, it can be faced as a regression problem where the purpose is, given the available features, to predict the number, in a continuous domain, of days a patient should probably stay in the hospital before being discharged. In order to maximize the performances of adopted regression models, a feature selection workflow was implemented by using the KNIME software (version 5.2.5) [28], which enabled the opportunity to develop a system wherein all the possible attempts and features combinations can be executed effortlessly in a single run. The adopted approaches for feature selection are the following:
  • Backward Feature Elimination (BFE): the dimensionality reduction is performed by an algorithm that, for each iteration with n features, removes one feature at a time in order to understand which set of n 1 input features results in the minimum increase in the error rate, and then removes the missing feature and continues with the next iteration;
  • Forward Feature Selection (FFS): the dimensionality reduction is performed by a dual algorithm that, starting from one feature, at each iteration adds a new feature adopting the same strategy of the BFE.
These approaches can be applied by making inferences with ML models, as they are wrapper methods. For this reason, two ML models were chosen for this purpose: the Linear Regression model and the M5P model. Therefore, for each dataset there are four possible attempts to perform the following:
  • BFE with linear regressor (LR);
  • FFS with linear regressor (LR);
  • BFE with M5P;
  • FFS with M5P.
For the outcome prediction task, where the model aims to predict whether a patient will either be discharged or succumb to the disease, a binary classification model is employed. Due to the simpler nature of this task, with only two possible outcomes, classification models generally demonstrate higher performance. As a result, a feature selection process is not necessary for this task.

4.3. LOS and Outcome Prediction

As mentioned before, the LOS prediction consists in determining the length of patients staying in the hospitals. This specific task, given the desired output, has to be treated as a regression problem. However, in order to simplify this task, it is possible to discretize the LOS in a set of days ranges (for instance, the LOS value “18” is equivalent to the days range value “15–21”). In this way, the regression problem becomes a classification problem.
On the other hand, the Outcome prediction consists in determining whether a patient, after the clinical therapy, will survive or not. This specific task is a classification task, and the model in particular shall be a binary predictor.
Depending on the chosen task, there are sets of models which are trained and tested in parallel with the same selected features. In particular, for the LOS prediction, the adopted models are Regression Tree, K-Nearest Neighbor (KNN), Multi-Layer Perceptron (MLP), Linear Regression model, M5P and SMO. For the classification tasks (Days range and Outcome), instead, the adopted models are Decision Tree, Naïve Bayes, Gradient Boosted Tree, Random Forest, MLP, KNN, LibLINEAR and ADABoostM1. All these models are provided by the KNIME environment.
All these models can take as input patients from one of the available datasets, selecting one of the features sets described before and following a specific approach, covering both the ensembling method and the feature selection type. All these possibilities are described and formalized in Table 2:
In particular, regarding the chosen approaches, the standard one does not include any particular condition or configuration for the selected models, while the ensemble one refers to a 4-Bagging Ensemble, where each model is replicated four times and trained on different subsets.
A joint look on Figure 10 and Table 2 makes clear all the possible independent experiments that can be performed with this system. The final user can choose the Prediction Task (Outcome, days range or regression), the Extracted Dataset to adopt (Complete, Vanvitelli or Cotugno) with a specific set of features (Tabular, 2D CT or 3D CT), and a specific Ensembling and Features Selection approach (Standard, 4-Bagging Ensemble, BFE with LR, FFS with LR, BFE with M5P, or FFS with M5P).
For the training process, all datasets are split with both a 80/20 hold-out approach and a 10-fold cross-validation approach [29], where in any case the training set is balanced with SMOTE (Synthetic Minority Oversampling Technique) [30] on the range feature so that the training set could be balanced by LOS, even if not directly, despite its continuous nature.
Figure 12 shows how the SMOTE algorithm, thanks to the oversampling process over the days range feature, can improve the dataset distribution over all its partitions.

5. Results

All the results obtained with the combinations resulting from Figure 10 and Table 2 are reported in this section, along with sets of best features from the selection process. Since the two tasks require different kinds of models, there will be two dedicated subsections covering the obtained results. As mentioned before, for the Outcome prediction task, the feature selection process described before is not required, as results obtained without feature selection are already satisfying. On the other hand, the hospitalization days regression task required this further step in order to retrieve some promising results. In order to understand the difference with the introduction of this extra step, results without feature selection will be provided as well.
Thanks to the design choice of implementing and training several SOTA models in the architecture, it is possible to have a direct and immediate comparison among them for the evaluation process. Therefore, in the next subsections, evaluation metrics from all the available models are reported, and the best ones are chosen for their respective tasks. In all the tables of this section, the hold-out approach is referred as HO (80-20), where 80-20 is the split ratio, while the cross-validation approach is referred to as CV (10), where 10 is the number of folds. Tabular clinical features are referred as Tabular, CT scan features as 2D CT, and CT volume features as 3D CT.

5.1. Outcome Classification Task

As aforementioned, outcome prediction is a binary classification task. The results are most promising regardless of the approach used (standard or bagging ensemble) and the dataset selected (Complete, Vanvitelli, or Cotugno). The following tables display the accuracy results for the chosen models, specifically using the original clinical data (Tabular Data), the 2D features extracted from CT scans (2D CT), and the 3D features extracted from CT volumes (3D CT).
In particular, Table 3 and Table 4 show the results of adopted classification models trained on the complete dataset for the outcome prediction, respectively, with the standard approach and with the 4-bagging ensemble approach, while Table 5 and Table 6 show the results of models trained on the Cotugno dataset for outcome prediction with the same approaches. There is no result table for the Vanvitelli dataset since, in this dataset, all the patients have an outcome value of 0, meaning that every patient has been dismissed, and consequently a model cannot be trained to detect different outcomes.
As noticeable in these tables, for the Cotugno dataset there are some configurations where the models appear to be almost perfect or perfect, while once the datasets (Vanvitelli and Cotugno) are merged, performances slightly drop. This is expected since new data have only one type of class for the outcome output.

5.2. Days Range Classification Task

On the other hand, the days range prediction task appears to be harder to solve, since it is ascribable to a regression task. Indeed, the results reflect this complexity. Table 7 and Table 8 show the results of the same models analyzed in the previous section, trained on the complete dataset. Table 9 and Table 10 show the results with models trained with the Vanvitelli dataset. Table 11 and Table 12 show the results with models trained with the Cotugno dataset.
In this case, accuracy scores of the whole workflow have a very low average value, but with few cases where specific models perform way better. Moreover, in any case, the 4-bagging ensemble approach improves the performances of the selected models.

5.3. Length of Stay Regression Task

Once completed the experiments for the classification tasks, some updates on the workflow were implemented for the hospitalization days prediction task. In particular:
  • Models were replaced with regression ones;
  • SMOTE algorithm with balancing on range feature was added;
  • Feature selection algorithms with utility models were added.
Given these differences, the test to be performed expanded significantly. For length reasons, we do not report all the combinations of features selection strategies and used models, but only the best models for each feature selection model and approach combination. Since there are two adopted models (Linear Regressor and M5P) and two approaches (Backward Feature Elimination and Forward Feature Selection), for each data type and for each dataset source, there are four best models.
In order to make proper evaluations of system performances on LOS prediction, the mean absolute error (MAE) will be adopted [31]. In particular, this evaluation metric can be obtained by making the overall sum of differences between predicted values (in this case LOS) and real values divided by the number of occurrences. The formula to obtain this metric is
M A E = | y i y ¯ i | n
where y i is the prediction value and y ¯ i is the real value.
The following tables contain performances, in terms of MAE, of models trained with the feature selection approach, with a comparison with the MAE obtained with the standard approach. In particular, these tables contain the attempts made with the complete dataset (Table 13), the Vanvitelli dataset (Table 14), and the Cotugno dataset (Table 15).
As expected, the feature selection approach significantly improves the performances obtained with the standard approach, even with less data from small datasets and simple models. Considering all the feature selection approaches, the most recurrent tabular features that appear to be significant are the COPD, COVID therapy, the ventilation, the LUS score, the TC score and the LINF. Thus, in general, these should be collected and taken into account with priority when new LOS predictions will be performed.

6. Ethical Considerations

While the introduction of AI technologies in the national healthcare system can be a valuable operational ally, it is also necessary to consider the possible risks associated with the use of these tools in a high-risk infrastructure context. Indeed, the use of sensitive data, such as those used in this study, enables the implementation of useful technologies for sensitive infrastructures like healthcare, but their use requires important ethical considerations regarding the nature of the data themselves.
First and foremost, to ensure an ethical and responsible use of these technologies, it is essential to respect patients’ privacy and rights. This implies that, whenever data are collected or reused, the process should be conducted transparently and in compliance with current regulations. Even in the case of retrospective data, such as those used in this study, it is crucial to implement measures that protect data security and confidentiality, in order to maintain public trust in healthcare systems and the prudent use of AI. Secondly, such a model should preferably be designed for an emergency context rather than as a preventive measure. While its use could certainly be beneficial even in non-risk situations (primarily in terms of healthcare service speed), adopting such a system on a daily basis presents potential issues. For instance, the collection and retention of sensitive data pose a greater risk in terms of cybersecurity, for the reasons mentioned above [32]. Even if the protection of these data were guaranteed through advanced security measures to prevent unauthorized access and potential risks, the accumulation of sensitive data still creates a “honeypot” for potential cyber violations, thereby increasing the overall vulnerability of the healthcare system. Although this represents an increased risk rather than a new issue, this consideration is important when evaluating the large-scale implementation of such technologies.
Furthermore, the constant use of such systems might be hindered by habitual reliance on these models, potentially leading to a lack of responsibility among medical personnel towards individual patients [33]. As a retrospective study, this cannot happen in our analysis. Nonetheless, this risk must be taken into account for future studies. Indeed, over-reliance on machine predictions could result in medical personnel treating patients considered by the machine to be at lower risk with less precision, relying on algorithms rather than comprehensive, ongoing medical evaluations. The risk of stigmatization and approximate medical decisions, inherent in any application of the model, must be mitigated by maintaining human supervision and integrating the system’s predictions with clinical judgment. Therefore, such a tool should be used exclusively for bed management and not as a measure of patient severity.
For these reasons, regardless of the use case, it is strongly recommended that healthcare facilities adopting these technologies also establish an ethical committee. This committee, preferably multidisciplinary in nature regarding the involved parties, should be able to manage any ethical and practical issues arising from the use of such tools, basing its decisions also on the establishment of predefined protocols and knowledge of relevant regulations (e.g., ISO/IEC, IEEE, etc.). The ethical committee should monitor the use of the system, evaluate the ethical implications of the decisions made, and ensure that the adoption of these technologies occurs responsibly and respectfully of patients’ rights, considering individual cases.
In addition to the previous considerations, it is also crucial to point out the risk of potential bias in an artificial intelligence model. In particular, in the present case, it is important to emphasize that possible biases could arise if the model is trained on datasets that are limited in size or unrepresentative. Indeed, if the data used for training do not adequately reflect the diversity of the patient population, there is a risk that the model will produce biased or inaccurate predictions, with potentially negative consequences for certain groups of patients. For example, if the model is trained only on data from patients of the same age, gender, or ethnicity, it may not be able to generalize correctly to patients with different characteristics. In addition to an imperfect model, this could lead to a reduction in the quality of service for certain categories of patients or, in extreme cases, perpetuate existing inequalities in access to care and health outcomes, with consequences that would make the system both ethically problematic and unhelpful. While the risk of bias in a model cannot be completely eliminated, it can certainly be reduced. To minimize this risk, it seems essential to ensure that the datasets used are as large and representative as possible and that validation studies are carried out in different populations. In order to have greater control over performance, it may also be useful to provide for the possibility of periodically reviewing the model and improving it by incorporating new data to correct any biases that may emerge over time.

7. Discussion and Conclusions

The aim of this work was to identify the best approach, starting from a multimodal dataset collected thanks to the effort of the Hospital of University of Campania Luigi Vanvitelli and of the Cotugno Hospital, for the prediction of pneumonia patients’ hospital length of stay, so that healthcare infrastructures can schedule and manage in advance beds and doctors’ availability and prevent emergencies. In particular, the task was splitted into three subtasks, which are the outcome prediction, the LOS range prediction, and the LOS prediction (in order of increasing complexity).
The results obtained from the three proposed subtasks highlight their varying levels of difficulty and underscore the importance of feature filtering and selection. While the outcomes for the outcome prediction and days range prediction tasks are consistent with those reported in the literature, a significant improvement was achieved in the Length of Stay (LOS) prediction task. This improvement can be attributed to the specific considerations made in this study, despite the challenges posed by the limited number of observations—especially when the datasets were analyzed separately—and the reduced number of features in the merged dataset. To summarize, the best results for each task across the three datasets (Complete, Vanvitelli, Cotugno) are as follows:
  • The best outcome prediction accuracy is of 0.875 on the Complete dataset by adopting the standard approach and 3D features, not available on the Vanvitelli dataset since all the patients have the same outcome, and 1.0 on the Cotugno dataset by adopting the standard approach and clinical data.
  • The best days range prediction accuracy is of 0.870 on the Complete dataset by adopting the bagging ensemble approach and 3D features, of 0.667 on the Vanvitelli dataset by adopting the bagging ensemble approach and 2D features, and of 0.571 on the Cotugno dataset by adopting the bagging ensemble approach and 2D features.
  • The best length of stay prediction MAE is 3.102 on the Complete dataset by adopting the feature selection approach and 2D features, 1.43 on the Vanvitelli dataset by adopting the feature qelection approach and 2D features, and 2.25 on the Cotugno dataset by adopting the feature selection approach and clinical data.
It is remarkable to notice several different aspects. First of all, the outcome prediction task is easy enough to adopt the standard approach and, if other methods are adopted, the system will overfit. Unlike the outcome task, the days range prediction task performs better with an ensemble approach, meaning that it is difficult enough not to make the system overfit during the training process. On the other hand, the regression task relies on feature selection, providing results which are way better than the ones obtained with the simplified version of the task; that is, the days range prediction, considering that the average error is of only 3 days. Moreover, even though, for the regression task, the best results are obtained with a selection of tabular features, the most common features adopted by the most performing systems are the 2D scan features retrieved from the ResNet, meaning that the convolutional neural network (CNN) was capable of extracting some extremely useful information from the images. An interesting aspect is that, when 2D features are adopted, models require only few features (sometimes just one) to perform at their best.
Further experiments were performed in order to understand whether the system behaves better with different fold values for cross-validation (2, 5, 20). However, experiments performed with k = 10 revealed to be the best among the other attempts. Moreover, despite the fact that there is a single case wherein cross-validation performed better than any other with hold-out (Vanvitelli dataset, CT scan features), where the MAE reached a value of 1.43, when the Complete dataset was adopted, it is still preferable to adopt a hold-out approach, since the best cross-validation model is about two days worse than the best hold-out model.
Finally, it is worth mentioning that the feature selection process, despite selecting different features based on the task and on the considered dataset, highlighted that some features (the COPD, COVID therapy, ventilation, the LUS score, the TC score, and the LINF) tend to be prominent in the evaluation of LOS. This is interesting, as it suggests that these metrics should always be considered and acquired during the early stages.
After evaluating the system’s performance on the available datasets, a comparison with existing methods can be conducted to assess whether the proposed approach provides a significant contribution to the research on LOS prediction for hospital patients. To this aim, Table 16 shows similarities and differences between the proposed solution and some other significant work in the literature [4,10,11].
We follow the details provided by the authors in [4,10,11] to implement the state-of-the-art solutions, and we report in Table 17 the results of the comparison in 10-fold CV with the workflow described in our work.
All the studies from the literature exhibit an average performance decrease of 0.3 in accuracy for classification tasks and 21 days for the regression task. This can primarily be attributed to two factors, namely the absence of key pre-processing steps, such as handling missing values or balancing the dataset, and the use of clinical data alone, rather than data extracted from CT scans and volumes.
Indeed, these values are compatible with the ones obtained with the proposed work itself, but with the adoption of standard clinical data.
Starting from these observations, the consequent future works would focus on the expansion of the dataset, trying to collect from the new ones’ features which are already present in the available datasets, on the adoption of other CNNs for features extraction from images, with consequent XAI analysis to understand where the different models focus more, and a study of the best features from clinical tests which doctors and researchers should provide to retrieve the best results possible.
This first result, obtained with the precious collaboration with the University of Campania Luigi Vanvitelli and the Cotugno Hospital, provides an extremely useful tool in hospital structures management, both for emergencies and for the daily routine, and a clear vision of how powerful machine learning models, if careful considerations and precise analysis on data are made, can be and how impactful they are in the biomedical context, even with a limited amount of data.

Author Contributions

All authors contributed to the study conception and design; S.C. (Salvatore Capuozzo), M.G. and S.M. contributed in setting up the technical environment, collecting data, and testing models; S.C. (Salvatore Capuozzo) prepared figures; A.A., S.C. (Salvatore Cappabianca), N.C., C.D.S., L.D., G.F., D.P., G.E.P., A.R. and C.S. (Caterina Sagnelli) contributed with the clinical aspects, supported the experimental setup, and provided data from patients; All authors contributed in analyzing results. L.M. contributed with the ethical aspects; S.C. (Salvatore Capuozzo) and S.M. contributed in writing the manuscript text; S.M., M.G. and C.S. (Carlo Sansone) contributed in reviewing the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki. Waiver for written patient consent was not sought from the institutional review board because it is not required by the national legislature for retrospective studies of existing data (e.g., registry data). Institutional review board review (approval or waiver) was not sought, because it is not required by the national legislature for retrospective studies of existing data.

Informed Consent Statement

Informed consent for participation is not required as per local legislation, in accordance with Italian law on retrospective studies using anonymized data (G.U. n. 76, 31 March 2008, implementing EU Directive 95/46/EC).

Data Availability Statement

The data presented in this study are available upon request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Daghistani, T.A.; Elshawi, R.; Sakr, S.; Ahmed, A.M.; Al-Thwayee, A.; Al-Mallah, M.H. Predictors of in-hospital length of stay among cardiac patients: A machine learning approach. Int. J. Cardiol. 2019, 288, 140–147. [Google Scholar] [CrossRef] [PubMed]
  2. Mpanya, D.; Celik, T.; Klug, E.; Ntsinjana, H. Predicting mortality and hospitalization in heart failure using machine learning: A systematic literature review. IJC Heart Vasc. 2021, 34, 100773. [Google Scholar] [CrossRef] [PubMed]
  3. Lequertier, V.; Wang, T.; Fondrevelle, J.; Augusto, V.; Duclos, A. Hospital length of stay prediction methods: A systematic review. Med. Care 2021, 59, 929–938. [Google Scholar] [CrossRef]
  4. Mekhaldi, R.N.; Caulier, P.; Chaabane, S.; Chraibi, A.; Piechowiak, S. A Comparative Study of Machine Learning Models for Predicting Length of Stay in Hospitals. J. Inf. Sci. Eng. 2021, 37, 1025–1038. [Google Scholar]
  5. Stone, K.; Zwiggelaar, R.; Jones, P.; Mac Parthaláin, N. A systematic review of the prediction of hospital length of stay: Towards a unified framework. PLoS Digit. Health 2022, 1, e0000017. [Google Scholar] [CrossRef] [PubMed]
  6. Morton, A.; Marzban, E.; Giannoulis, G.; Patel, A.; Aparasu, R.; Kakadiaris, I.A. A comparison of supervised machine learning techniques for predicting short-term in-hospital length of stay among diabetic patients. In Proceedings of the 2014 13th International Conference on Machine Learning and Applications, Detroit, MI, USA, 3–6 December 2014; pp. 428–431. [Google Scholar]
  7. Pendharkar, P.C.; Khurana, H. Machine learning techniques for predicting hospital length of stay in pennsylvania federal and specialty hospitals. Int. J. Comput. Sci. Appl. 2014, 11, 45–56. [Google Scholar]
  8. Turgeman, L.; May, J.H.; Sciulli, R. Insights from a machine learning model for predicting the hospital Length of Stay (LOS) at the time of admission. Expert Syst. Appl. 2017, 78, 376–385. [Google Scholar] [CrossRef]
  9. Lorenzoni, G.; Sabato, S.S.; Lanera, C.; Bottigliengo, D.; Minto, C.; Ocagli, H.; De Paolis, P.; Gregori, D.; Iliceto, S.; Pisanò, F. Comparison of machine learning techniques for prediction of hospitalization in heart failure patients. J. Clin. Med. 2019, 8, 1298. [Google Scholar] [CrossRef]
  10. Saadatmand, S.; Salimifard, K.; Mohammadi, R.; Kuiper, A.; Marzban, M.; Farhadi, A. Using machine learning in prediction of ICU admission, mortality, and length of stay in the early stage of admission of COVID-19 patients. Ann. Oper. Res. 2023, 328, 1043–1071. [Google Scholar] [CrossRef]
  11. Khajehali, N.; Alizadeh, S. Extract critical factors affecting the length of hospital stay of pneumonia patient by data mining (case study: An Iranian hospital). Artif. Intell. Med. 2017, 83, 2–13. [Google Scholar] [CrossRef]
  12. Mekhaldi, R.N.; Caulier, P.; Chaabane, S.; Chraibi, A.; Piechowiak, S. Using machine learning models to predict the length of stay in a hospital setting. In Proceedings of the World Conference on Information Systems and Technologies, Budva, Montenegro, 7–10 April 2020; pp. 202–211. [Google Scholar]
  13. Alsinglawi, B.; Alnajjar, F.; Mubin, O.; Novoa, M.; Alorjani, M.; Karajeh, O.; Darwish, O. Predicting length of stay for cardiovascular hospitalizations in the intensive care unit: Machine learning approach. In Proceedings of the 2020 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Montreal, QC, Canada, 20–24 July 2020; pp. 5442–5445. [Google Scholar]
  14. Alsinglawi, B.; Alshari, O.; Alorjani, M.; Mubin, O.; Alnajjar, F.; Novoa, M.; Darwish, O. An explainable machine learning framework for lung cancer hospital length of stay prediction. Sci. Rep. 2022, 12, 607. [Google Scholar] [CrossRef]
  15. Bacchi, S.; Gluck, S.; Tan, Y.; Chim, I.; Cheng, J.; Gilbert, T.; Menon, D.K.; Jannes, J.; Kleinig, T.; Koblar, S. Prediction of general medical admission length of stay with natural language processing and deep learning: A pilot study. Intern. Emerg. Med. 2020, 15, 989–995. [Google Scholar] [CrossRef] [PubMed]
  16. Bacchi, S.; Tan, Y.; Oakden-Rayner, L.; Jannes, J.; Kleinig, T.; Koblar, S. Machine learning in the prediction of medical inpatient length of stay. Intern. Med. J. 2022, 52, 176–185. [Google Scholar] [CrossRef] [PubMed]
  17. Zheng, L.; Wang, J.; Sheriff, A.; Chen, X. Hospital length of stay prediction with ensemble methods in machine learning. In Proceedings of the 2021 International Conference on Cyber-Physical Social Intelligence (ICCSI), Beijing, China, 18–20 December 2021; pp. 1–5. [Google Scholar]
  18. Jaotombo, F.; Pauly, V.; Fond, G.; Orleans, V.; Auquier, P.; Ghattas, B.; Boyer, L. Machine-learning prediction for hospital length of stay using a French medico-administrative database. J. Mark. Access Health Policy 2023, 11, 2149318. [Google Scholar] [CrossRef]
  19. Zhong, H.; Wang, B.; Wang, D.; Liu, Z.; Xing, C.; Wu, Y.; Gao, Q.; Zhu, S.; Qu, H.; Jia, Z.; et al. The application of machine learning algorithms in predicting the length of stay following femoral neck fracture. Int. J. Med. Inform. 2021, 155, 104572. [Google Scholar] [CrossRef] [PubMed]
  20. Alghatani, K.; Ammar, N.; Rezgui, A.; Shaban-Nejad, A. Predicting intensive care unit length of stay and mortality using patient vital signs: Machine learning model development and validation. JMIR Med. Inform. 2021, 9, e21347. [Google Scholar] [CrossRef]
  21. Chrusciel, J.; Girardon, F.; Roquette, L.; Laplanche, D.; Duclos, A.; Sanchez, S. The prediction of hospital length of stay using unstructured data. BMC Med. Inform. Decis. Mak. 2021, 21, 351. [Google Scholar] [CrossRef]
  22. Zebin, T.; Rezvy, S.; Chaussalet, T.J. A deep learning approach for length of stay prediction in clinical settings from medical records. In Proceedings of the 2019 IEEE Conference on Computational Intelligence in Bioinformatics and Computational Biology (CIBCB), Siena, Italy, 9–11 July 2019; pp. 1–5. [Google Scholar]
  23. Roimi, M.; Gutman, R.; Somer, J.; Ben Arie, A.; Calman, I.; Bar-Lavie, Y.; Gelbshtein, U.; Liverant-Taub, S.; Ziv, A.; Eytan, D.; et al. Development and validation of a machine learning model predicting illness trajectory and hospital utilization of COVID-19 patients: A nationwide study. J. Am. Med. Inform. Assoc. 2021, 28, 1188–1196. [Google Scholar] [CrossRef]
  24. Ebinger, J.; Wells, M.; Ouyang, D.; Davis, T.; Kaufman, N.; Cheng, S.; Chugh, S. A machine learning algorithm predicts duration of hospitalization in COVID-19 patients. Intell.-Based Med. 2021, 5, 100035. [Google Scholar] [CrossRef]
  25. Alabbad, D.A.; Almuhaideb, A.M.; Alsunaidi, S.J.; Alqudaihi, K.S.; Alamoudi, F.A.; Alhobaishi, M.K.; Alaqeel, N.A.; Alshahrani, M.S. Machine learning model for predicting the length of stay in the intensive care unit for COVID-19 patients in the eastern province of Saudi Arabia. Inform. Med. Unlocked 2022, 30, 100937. [Google Scholar] [CrossRef]
  26. Etu, E.E.; Monplaisir, L.; Arslanturk, S.; Masoud, S.; Aguwa, C.; Markevych, I.; Miller, J. Prediction of length of stay in the emergency department for COVID-19 patients: A machine learning approach. IEEE Access 2022, 10, 42243–42251. [Google Scholar] [CrossRef]
  27. Chen, S.; Ma, K.; Zheng, Y. Med3D: Transfer Learning for 3D Medical Image Analysis. arXiv 2019, arXiv:1904.00625. [Google Scholar]
  28. Berthold, M.R.; Cebron, N.; Dill, F.; Gabriel, T.R.; Kötter, T.; Meinl, T.; Ohl, P.; Sieb, C.; Thiel, K.; Wiswedel, B. KNIME: The Konstanz Information Miner. In Data Analysis, Machine Learning and Applications, Proceedings of the 31st Annual Conference of the Gesellschaft für Klassifikation e.V., Freiburg, Germany, 7–9 March 2007; Studies in Classification, Data Analysis, and Knowledge Organization (GfKL 2007); Springer: Berlin/Heidelberg, Germany, 2007. [Google Scholar]
  29. Gorriz, J.M.; Segovia, F.; Ramirez, J.; Ortiz, A.; Suckling, J. Is K-fold cross validation the best model selection method for Machine Learning? arXiv 2024, arXiv:2401.16407. [Google Scholar]
  30. Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic minority over-sampling technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  31. Willmott, C.; Matsuura, K. Advantages of the Mean Absolute Error (MAE) over the Root Mean Square Error (RMSE) in Assessing Average Model Performance. Clim. Res. 2005, 30, 79. [Google Scholar] [CrossRef]
  32. Razaque, A.; Amsaad, F.; Khan, M.J.; Hariri, S.; Chen, S.; Siting, C.; Ji, X. Survey: Cybersecurity vulnerabilities, attacks and solutions in the medical domain. IEEE Access 2019, 7, 168774–168797. [Google Scholar] [CrossRef]
  33. Goddard, K.; Roudsari, A.; Wyatt, J.C. Automation bias: Empirical results assessing influencing factors. Int. J. Med. Inform. 2014, 83, 368–375. [Google Scholar] [CrossRef]
Figure 1. (a) Complete dataset distribution on outcome output, where patients with value 0 have been dismissed, with value 1 are deceased without being in intensive care unit, with value 2 are deceased after being in intensive care unit. (b) Complete dataset distribution on LOS output.
Figure 1. (a) Complete dataset distribution on outcome output, where patients with value 0 have been dismissed, with value 1 are deceased without being in intensive care unit, with value 2 are deceased after being in intensive care unit. (b) Complete dataset distribution on LOS output.
Bdcc 08 00178 g001
Figure 2. Complete dataset distribution on range output.
Figure 2. Complete dataset distribution on range output.
Bdcc 08 00178 g002
Figure 3. (a) Vanvitelli dataset distribution on gender feature. (b) Cotugno dataset distribution on gender feature.
Figure 3. (a) Vanvitelli dataset distribution on gender feature. (b) Cotugno dataset distribution on gender feature.
Bdcc 08 00178 g003
Figure 4. (a) Vanvitelli dataset distribution on age feature. (b) Cotugno dataset distribution on age feature.
Figure 4. (a) Vanvitelli dataset distribution on age feature. (b) Cotugno dataset distribution on age feature.
Bdcc 08 00178 g004
Figure 5. (a) Vanvitelli dataset distribution on CT machine model feature. (b) Cotugno dataset distribution on CT machine model feature.
Figure 5. (a) Vanvitelli dataset distribution on CT machine model feature. (b) Cotugno dataset distribution on CT machine model feature.
Bdcc 08 00178 g005
Figure 6. Vanvitelli dataset correlation matrix.
Figure 6. Vanvitelli dataset correlation matrix.
Bdcc 08 00178 g006
Figure 7. Cotugno dataset correlation matrix.
Figure 7. Cotugno dataset correlation matrix.
Bdcc 08 00178 g007
Figure 8. Complete dataset correlation matrix.
Figure 8. Complete dataset correlation matrix.
Bdcc 08 00178 g008
Figure 9. The original Vanvitelli and Cotugno datasets consist of patients’ lung CT scans and raw clinical data. To extract relevant features, we process 3D CT volumes using a 3D-CNN to obtain three-dimensional tabular features. Additionally, the slice with the highest count of non-zero pixels is identified from each CT volume and processed via a 2D-CNN to extract two-dimensional tabular features. Concurrently, the raw clinical tabular data are pre-processed to generate clinically useful features. The final, consolidated dataset—referred to as the Complete dataset—is formed by concatenating patient data from both Vanvitelli and Cotugno sources across all feature sets. Features are color-coded by dataset origin: green for Vanvitelli, orange for Cotugno, and sky blue for the Complete dataset.
Figure 9. The original Vanvitelli and Cotugno datasets consist of patients’ lung CT scans and raw clinical data. To extract relevant features, we process 3D CT volumes using a 3D-CNN to obtain three-dimensional tabular features. Additionally, the slice with the highest count of non-zero pixels is identified from each CT volume and processed via a 2D-CNN to extract two-dimensional tabular features. Concurrently, the raw clinical tabular data are pre-processed to generate clinically useful features. The final, consolidated dataset—referred to as the Complete dataset—is formed by concatenating patient data from both Vanvitelli and Cotugno sources across all feature sets. Features are color-coded by dataset origin: green for Vanvitelli, orange for Cotugno, and sky blue for the Complete dataset.
Bdcc 08 00178 g009
Figure 10. Models training workflow. The last block contains the list of all the state-of-the-art ML models, in particular classification ones for Outcome and Days Range prediction and regression ones for Length of Stay estimation. The prediction task can be Length of stay estimation, Outcome or Days range prediction, the extracted features can be 3D, 2D or Tabular, according to the extraction process shown in Figure 9, and the dataset source can be Vanvitelli, Cotugno, or Complete.
Figure 10. Models training workflow. The last block contains the list of all the state-of-the-art ML models, in particular classification ones for Outcome and Days Range prediction and regression ones for Length of Stay estimation. The prediction task can be Length of stay estimation, Outcome or Days range prediction, the extracted features can be 3D, 2D or Tabular, according to the extraction process shown in Figure 9, and the dataset source can be Vanvitelli, Cotugno, or Complete.
Bdcc 08 00178 g010
Figure 11. Architecture of the ResNet50 model with the extra fully connected layer at the end, marked in orange.
Figure 11. Architecture of the ResNet50 model with the extra fully connected layer at the end, marked in orange.
Bdcc 08 00178 g011
Figure 12. Distributions of a subset sample over days range and LOS before and after applying SMOTE on days range.
Figure 12. Distributions of a subset sample over days range and LOS before and after applying SMOTE on days range.
Bdcc 08 00178 g012
Table 1. Table containing the features for each of the original datasets (Vanvitelli and Cotugno) after pre-processing and filtering. The features that are present in both datasets (9) are marked in bold, resulting in the Complete dataset.
Table 1. Table containing the features for each of the original datasets (Vanvitelli and Cotugno) after pre-processing and filtering. The features that are present in both datasets (9) are marked in bold, resulting in the Complete dataset.
DatasetSizeBinary FeaturesNumeric FeaturesFeatures Num.
Vanvitelli34Gender, Remdesivie, Neutrophils,
Ischemic heart disease, COPD,
COVID therapy, Ventilation,
Anticoagulant therapy, chronic liver disease,
Fever, Cough, Diarrhea/vomiting,
Arterial hypertension, Atrial fibrillation,
Active cancer within the past 5 years,
Dyspnea, Ageusia/anosmia,
Type II diabetes, Dementia, Lymphocytes
Age, WBC, HB, Platelets,
Azotemia, Creatinine,
Sodium, Potassium,
GOT or GPT,
Blood glucose, LDH,
INR, Fibrinogen,
D dimer, p/F,
LUS score
36
Cotugno50Gender, Remdesivie, Neutrophils,
Ischemic heart disease, COPD,
COVID therapy, Ventilation,
CDI, Complication, TEP,
PNX, PMD, MB, ICU, Pentaglobin,
CS, I dose, II dose, III dose,
Complete vaccine cycle, Comorbidity,
Cardio, DM, Nephro, Pneumo,
Emato, K, Tiroid, Stroke, IOP2
Age, WBC, TC score, PF,
Complication type,
Max ventilation,
Charlson, IL6, IL2R,
NEU, %, LINF, % (# 1),
n/l, ESR, DD,
PCR, PCT, Ferritin
49
Table 2. Table of the combinations performed during the training and inference process. For instance, one possible pipeline is days range classification task, 2D features from CT scans, the Cotugno dataset, and using an ensemble approach.
Table 2. Table of the combinations performed during the training and inference process. For instance, one possible pipeline is days range classification task, 2D features from CT scans, the Cotugno dataset, and using an ensemble approach.
TaskFeaturesDatasetApproach
Outcome (Classification)Tabular Data,
CT Scan Features (2D),
CT Volume Features (3D)
Complete,
Vanvitelli,
Cotugno
Standard,
Ensemble
Days range (Classification)Tabular Data,
CT Scan Features (2D),
CT Volume Features (3D)
Complete,
Vanvitelli,
Cotugno
Standard,
Ensemble
Length of stay (Regression)Tabular Data,
CT Scan Features (2D),
CT Volume Features (3D)
Complete,
Vanvitelli,
Cotugno
Standard,
Ensemble,
Feature Selection
Table 3. Accuracy scores of trained models for outcome prediction on complete dataset with the standard approach. Best results for each data type and split approach are in bold.
Table 3. Accuracy scores of trained models for outcome prediction on complete dataset with the standard approach. Best results for each data type and split approach are in bold.
Model
(Outcome, Complete, Standard)
Tabular
HO (80-20)
Tabular
CV (10)
2D CT
HO (80-20)
2D CT
CV (10)
3D CT
HO (80-20)
3D CT
CV (10)
Decision Tree0.6820.7260.5450.6710.5420.701
Naïve Bayes0.7730.6850.5450.4380.3750.416
Gradient Boosted Tree0.6820.8670.5450.6990.8750.701
Random Forest0.6820.8080.5910.7260.7920.753
Multi-Layer Perceptron0.6820.7400.4550.7260.7920.779
LibLINEAR0.6820.7810.6360.7400.7920.753
ADABoostM10.6820.7950.6360.7260.7920.766
K-Nearest Neighbor 30.7270.7260.6360.6850.7080.727
K-Nearest Neighbor 50.6820.7260.4550.6710.7920.753
K-Nearest Neighbor 70.6820.7670.4550.6990.7920.766
Table 4. Accuracy scores of trained models for outcome prediction on complete dataset with the 4-bagging ensemble approach. Best results for each data type and split approach are in bold.
Table 4. Accuracy scores of trained models for outcome prediction on complete dataset with the 4-bagging ensemble approach. Best results for each data type and split approach are in bold.
Model
(Outcome, Complete, Bagging)
Tabular
HO (80-20)
Tabular
CV (10)
2D CT
HO (80-20)
2D CT
CV (10)
3D CT
HO (80-20)
3D CT
CV (10)
Decision Tree0.6670.7120.6670.6990.7080.714
Naïve Bayes0.7080.8080.7080.5620.6250.662
Gradient Boosted Tree0.7920.7670.7500.6710.7080.649
Random Forest0.7500.8220.6670.6850.6670.623
Multi-Layer Perceptron0.6250.8360.7080.7260.7500.727
LibLINEAR0.7500.7810.6670.6300.5830.662
ADABoostM10.7920.7810.7080.6300.6670.675
K-Nearest Neighbor 30.7080.7400.7080.7670.6250.649
K-Nearest Neighbor 50.6670.7400.6670.7530.7080.688
K-Nearest Neighbor 70.6670.7810.7080.7530.7500.714
Table 5. Accuracy scores of trained models for outcome prediction on Cotugno dataset with the standard approach. Best results for each data type and split approach are in bold.
Table 5. Accuracy scores of trained models for outcome prediction on Cotugno dataset with the standard approach. Best results for each data type and split approach are in bold.
Model
(Outcome, Cotugno, Standard)
Tabular
HO (80-20)
Tabular
CV (10)
2D CT
HO (80-20)
2D CT
CV (10)
3D CT
HO (80-20)
3D CT
CV (10)
Decision Tree0.9230.9760.5380.6190.6150.535
Naïve Bayes0.9230.7860.5380.6430.5380.419
Gradient Boosted Tree1.0000.9760.5380.6670.5380.558
Random Forest0.9230.9290.6920.6430.6920.581
Multi-Layer Perceptron0.9230.6900.5380.6190.6920.674
LibLINEAR0.8460.7860.4620.5480.6920.605
ADABoostM11.0000.9760.5380.5240.6920.581
K-Nearest Neighbor 30.8460.6670.6150.6670.6150.605
K-Nearest Neighbor 50.8460.6900.5380.5710.7690.628
K-Nearest Neighbor 70.8460.6900.5380.5950.6920.651
Table 6. Accuracy scores of trained models for outcome prediction on Cotugno dataset with the 4-bagging ensemble approach. Best results for each data type and split approach are in bold.
Table 6. Accuracy scores of trained models for outcome prediction on Cotugno dataset with the 4-bagging ensemble approach. Best results for each data type and split approach are in bold.
Model
(Outcome, Cotugno, Bagging)
Tabular
HO (80-20)
Tabular
CV (10)
2D CT
HO (80-20)
2D CT
CV (10)
3D CT
HO (80-20)
3D CT
CV (10)
Decision Tree0.9290.9050.7140.6190.5330.581
Naïve Bayes0.7860.6670.5000.4760.5330.581
Gradient Boosted Tree0.8570.8570.5000.5480.5330.605
Random Forest0.6430.6670.5000.4760.4670.581
Multi-Layer Perceptron0.7140.7860.5710.6190.3330.488
LibLINEAR0.7860.6670.5710.5480.6670.535
ADABoostM10.8570.8570.4290.5000.6000.581
K-Nearest Neighbor 30.5710.6430.5000.4050.6000.558
K-Nearest Neighbor 50.7140.6900.6430.5480.6670.535
K-Nearest Neighbor 70.6430.6430.6430.5000.4670.581
Table 7. Accuracy scores of trained models for days range prediction on complete dataset with the standard approach. Best results for each data type and split approach are in bold.
Table 7. Accuracy scores of trained models for days range prediction on complete dataset with the standard approach. Best results for each data type and split approach are in bold.
Model
(Days Range, Complete, Standard)
Tabular
HO (80-20)
Tabular
CV (10)
2D CT
HO (80-20)
2D CT
CV (10)
3D CT
HO (80-20)
3D CT
CV (10)
Decision Tree0.1820.3010.3640.2330.1670.143
Naïve Bayes0.2270.2190.0910.2880.2500.169
Gradient Boosted Tree0.1360.2050.4090.2880.2500.182
Random Forest0.1820.2050.4090.2330.1250.104
Multi-Layer Perceptron0.1360.1510.3640.2880.2500.117
LibLINEAR0.3180.2330.3180.2470.2500.130
ADABoostM10.2730.1510.0450.1230.0830.039
K-Nearest Neighbor 30.1820.1640.2730.2880.0830.182
K-Nearest Neighbor 50.1820.1230.2730.2330.1250.130
K-Nearest Neighbor 70.2270.1640.1820.1920.0830.117
Table 8. Accuracy scores of trained models for days range prediction on complete dataset with the 4-bagging ensemble approach. Best models for each data type and split approach are in bold.
Table 8. Accuracy scores of trained models for days range prediction on complete dataset with the 4-bagging ensemble approach. Best models for each data type and split approach are in bold.
Model
(Days Range, Complete, Bagging)
Tabular
HO (80-20)
Tabular
CV (10)
2D CT
HO (80-20)
2D CT
CV (10)
3D CT
HO (80-20)
3D CT
CV (10)
Decision Tree0.2920.2050.7920.7810.7080.870
Naïve Bayes0.1250.1510.6250.5070.3750.364
Gradient Boosted Tree0.2080.2880.5830.7530.5830.675
Random Forest0.1250.1370.3750.3840.3750.364
Multi-Layer Perceptron0.2080.1230.7500.6580.3750.455
LibLINEAR0.2080.2050.5000.5070.2500.286
ADABoostM10.0830.0820.5000.3150.5000.403
K-Nearest Neighbor 30.2920.1370.2500.3420.2500.260
K-Nearest Neighbor 50.2920.1780.3750.3010.1250.234
K-Nearest Neighbor 70.1670.1640.2920.3290.2080.247
Table 9. Accuracy scores of trained models for days range prediction on Vanvitelli dataset with the standard approach. Best results for each data type and split approach are in bold.
Table 9. Accuracy scores of trained models for days range prediction on Vanvitelli dataset with the standard approach. Best results for each data type and split approach are in bold.
Model
(Days Range, Vanvitelli, Standard)
Tabular
HO (80-20)
Tabular
CV (10)
2D CT
HO (80-20)
2D CT
CV (10)
3D CT
HO (80-20)
3D CT
CV (10)
Decision Tree0.4440.3330.4440.3000.4000.212
Naïve Bayes0.0000.2670.4440.2670.3000.273
Gradient Boosted Tree0.3330.2670.4440.2670.4000.273
Random Forest0.2220.3330.3330.2000.5000.242
Multi-Layer Perceptron0.2220.1330.2220.2330.3000.273
LibLINEAR0.2220.2330.1110.3670.2000.273
ADABoostM10.2220.1000.1110.1000.0000.030
K-Nearest Neighbor 30.1110.2330.3330.2330.2000.273
K-Nearest Neighbor 50.2220.1330.2220.1330.0000.152
K-Nearest Neighbor 70.2220.1330.1110.2670.0000.091
Table 10. Accuracy scores of trained models for days range prediction on Vanvitelli dataset with the 4-bagging ensemble approach. Best results for each data type and split approach are in bold.
Table 10. Accuracy scores of trained models for days range prediction on Vanvitelli dataset with the 4-bagging ensemble approach. Best results for each data type and split approach are in bold.
Model
(Days Range, Vanvitelli, Bagging)
Tabular
HO (80-20)
Tabular
CV (10)
2D CT
HO (80-20)
2D CT
CV (10)
3D CT
HO (80-20)
3D CT
CV (10)
Decision Tree0.4170.3000.5830.6670.5000.576
Naïve Bayes0.1670.1330.5830.3670.2500.303
Gradient Boosted Tree0.2500.1000.5000.4670.2500.394
Random Forest0.2500.1000.5000.3670.2500.303
Multi-Layer Perceptron0.2500.1000.5000.4000.3330.394
LibLINEAR0.3330.2670.2500.2000.5000.455
ADABoostM10.3330.2330.4170.5000.3330.394
K-Nearest Neighbor 30.5000.2000.4170.2000.1670.333
K-Nearest Neighbor 50.3330.1330.3330.3670.2500.273
K-Nearest Neighbor 70.3330.1330.3330.1330.2500.364
Table 11. Accuracy scores of trained models for days range prediction on Cotugno dataset with the standard approach. Best results for each data type and split approach are in bold.
Table 11. Accuracy scores of trained models for days range prediction on Cotugno dataset with the standard approach. Best results for each data type and split approach are in bold.
Model
(Days Range, Cotugno, Standard)
Tabular
HO (80-20)
Tabular
CV (10)
2D CT
HO (80-20)
2D CT
CV (10)
3D CT
HO (80-20)
3D CT
CV (10)
Decision Tree0.3850.1900.3850.1670.2310.186
Naïve Bayes0.0770.1190.3080.2140.2310.209
Gradient Boosted Tree0.3850.1430.1540.3330.2310.209
Random Forest0.3080.2380.3850.2140.1540.116
Multi-Layer Perceptron0.0770.0950.3080.2860.2310.209
LibLINEAR0.1540.1430.3080.2620.1540.116
ADABoostM10.0770.0240.1540.1190.0770.093
K-Nearest Neighbor 30.0000.0480.2310.2140.1540.256
K-Nearest Neighbor 50.2310.1430.2310.2140.1540.233
K-Nearest Neighbor 70.2310.0950.3080.1430.0770.093
Table 12. Accuracy scores of trained models for days range prediction on Cotugno dataset with the 4-bagging ensemble approach. Best results for each data type and split approach are in bold.
Table 12. Accuracy scores of trained models for days range prediction on Cotugno dataset with the 4-bagging ensemble approach. Best results for each data type and split approach are in bold.
Model
(Days Range, Cotugno, Bagging)
Tabular
HO (80-20)
Tabular
CV (10)
2D CT
HO (80-20)
2D CT
CV (10)
3D CT
HO (80-20)
3D CT
CV (10)
Decision Tree0.2860.2860.5710.5710.4670.512
Naïve Bayes0.4290.2620.5000.2860.3330.465
Gradient Boosted Tree0.5710.3810.2860.4760.3330.372
Random Forest0.2860.3100.2860.3330.3330.349
Multi-Layer Perceptron0.2860.2140.2860.5000.4000.465
LibLINEAR0.2140.1900.2140.2620.3330.326
ADABoostM10.2860.2140.3570.2860.3330.442
K-Nearest Neighbor 30.3570.3100.2860.3570.4670.419
K-Nearest Neighbor 50.3570.3570.3570.4290.2670.349
K-Nearest Neighbor 70.3570.4050.3570.3100.2670.233
Table 13. MAE scores of trained models for hospitalization days prediction on complete dataset with feature selection approaches. The best combination is the row in bold.
Table 13. MAE scores of trained models for hospitalization days prediction on complete dataset with feature selection approaches. The best combination is the row in bold.
Split
Approach
Features
Type
FS ModelFS
Approach
Selected FeaturesBest Model
with FS (MAE)
Best Model
w/o FS (MAE)
HO (80-20)TabularLinear Reg.BFEIschemic heart dis., COPD,
COVID therapy
Reg. Tree (7.528)SMO (20.357)
CV (10)COPD, AgeSMO (18.648)SMO (20.373)
HO (80-20)TabularLinear Reg.FFSGender, Ischemic heart dis.,
COPD, COVID therapy
Linear Reg. (8.073)SMO (20.357)
CV (10)COPD, AgeSMO (18.229)SMO (20.373)
HO (80-20)TabularM5PBFEIschemic heart dis.,
COVID therapy
M5P (7.576)SMO (20.357)
CV (10)COPD, White blood cells,
Gender
SMO (18.567)SMO (20.373)
HO (80-20)TabularM5PFFSCOPD, COVID therapy,
Ventilation
KNN 7 (7.674)SMO (20.357)
CV (10)Neutrophils, GenderSMO (17.717)SMO (20.373)
HO (80-20)2D CTLinear Reg.BFE6 featuresLinear Reg. (3.102)SMO (4.078)
CV (10)7 featuresSMO (5.067)SMO (5.11)
HO (80-20)2D CTLinear Reg.FFS9 featuresLinear Reg. (3.336)SMO (4.078)
CV (10)14 featuresSMO (5.106)SMO (5.11)
HO (80-20)2D CTM5PBFE6 featuresSMO (4.025)SMO (4.078)
CV (10)3 featuresSMO (5.111)SMO (5.11)
HO (80-20)2D CTM5PFFS1 featureSMO (4.0)SMO (4.078)
CV (10)1 featureSMO (5.001)SMO (5.11)
HO (80-20)3D CTLinear Reg.BFE4 features SMO (12.602)SMO (12.686)
CV (10)4 featuresSMO (16.796)SMO (16.898)
HO (80-20)3D CTLinear Reg.FFS8 featuresSMO (12.712)SMO (12.686)
CV (10)8 featuresSMO (16.6)SMO (16.898)
HO (80-20)3D CTM5PBFE3 featuresSMO (12.516)SMO (12.686)
CV (10)3 featuresSMO (17.599)SMO (16.898)
HO (80-20)3D CTM5PFFS3 featuresSMO (12.472)SMO (12.686)
CV (10)3 featuresMLP (9.526)SMO (16.898)
Table 14. MAE scores of trained models for hospitalization days prediction on Vanvitelli dataset with feature selection approaches. The best combination is the row in bold.
Table 14. MAE scores of trained models for hospitalization days prediction on Vanvitelli dataset with feature selection approaches. The best combination is the row in bold.
Split
Approach
Features
Type
FS ModelFS
Approach
Selected FeaturesBest Model
with FS (MAE)
Best Model
w/o FS (MAE)
HO (80-20)TabularLinear Reg.BFESex, Vent., Dementia, Lymph.,
GOT or GPT, Neutrophils
Linear Reg. (5.319)GBTree (36.856)
CV (10)Fever, Cough, Dementia, COPD,
Active cancer last 5 years,
Chronic liver disease, Neutrophils,
Lymphocytes, Ageusia/Anosmia,
Type 2 diabetes, p/F
MLP (27.485)GBTree (46.564)
HO (80-20)TabularLinear Reg.FFSVentilation, Age, Fever, Dyspnea,
Fibrinogen, LUS score, Neutrophils
Linear Reg. (4.334)GBTree (36.856)
CV (10)21 featuresM5P (40.781)GBTree (46.564)
HO (80-20)TabularM5PBFEHB, Platelets, Fibrinogen, LUS score,
Cough, Art. hypert.
Reg. Tree (6.833)GBTree (36.856)
CV (10)Cough, Dyspnea, Dementia,
GOT or GPT, Blood glucose, INR,
Ventilation
M5P (31.259)GBTree (46.564)
HO (80-20)TabularM5PFFSCOPD, Atrial fibril., HB, Sodium,
LUS score, Neutrophils
Reg. Tree (6.667)GBTree (36.856)
CV (10)Fever, Cough, Lymphocytes,
Blood glucose, Fibrinogen,
Ventilation
M5P (26.589)GBTree (46.564)
HO (80-20)2D CTLinear Reg.BFE6 featuresLinear Reg. (4.866)MLP (6.639)
CV (10)6 featuresMLP (4.511)MLP (3.731)
HO (80-20)2D CTLinear Reg.FFS2 featuresM5P (7.281)MLP (6.639)
CV (10)2 featuresMLP (4.998)MLP (3.731)
HO (80-20)2D CTM5PBFE1 featureGBTree (6.09)MLP (6.639)
CV (10)1 featureMLP (1.43)MLP (3.731)
HO (80-20)2D CTM5PFFS1 featureGBTree (6.09)MLP (6.639)
CV (10)1 featureMLP (4.148)MLP (3.731)
HO (80-20)3D CTLinear Reg.BFE5 features Linear Reg. (28.287)KNN3 (36.09)
CV (10)5 featuresSMO (23.993)SMO (28.749)
HO (80-20)3D CTLinear Reg.FFS6 featuresLinear Reg. (34.16)KNN3 (36.09)
CV (10)6 featuresMLP (14.611)SMO (28.749)
HO (80-20)3D CTM5PBFE4 featuresGBTree (34.113)KNN3 (36.09)
CV (10)4 featuresMLP (18.1)SMO (28.749)
HO (80-20)3D CTM5PFFS1 featureGBTree (33.734)KNN3 (36.09)
CV (10)1 featureSMO (26.879)SMO (28.749)
Table 15. MAE scores of trained models for hospitalization days prediction on Cotugno dataset with feature selection approaches. The best combination is the row in bold.
Table 15. MAE scores of trained models for hospitalization days prediction on Cotugno dataset with feature selection approaches. The best combination is the row in bold.
Split
Approach
Features
Type
FS ModelFS
Approach
Selected FeaturesBest Model
with FS (MAE)
Best Model
w/o FS (MAE)
HO (80-20)TabularLinear Reg.BFEIschemic heart dis., Age, TC score,
PF, Compl., MB, ICU, I dose,
NEFRO, NEU, LINF, % (#1), n/l
Linear Reg. (3.781)Simple Reg. (6.951)
CV (10)Age, Gender, Score TC, Compl. type,
TEP, ICU, CS, III dose, CARDIO,
THYROID, ictus
Linear Reg. (8.236)GBTree (8.421)
HO (80-20)TabularLinear Reg.FFSTC score, CS, IL6,
%, LINF, % (#1), Neutr.
Linear Reg. (2.25)Simple Reg. (6.951)
CV (10)Age, Compl. type, TEP,
CARDIO, PNEUMO, LINF,
% (#1), Ischemic heart dis.
MLP (7.779)GBTree (8.421)
HO (80-20)TabularM5PBFEAge, TC score, Compl., Max vent.,
CS, I dose, II dose, DM,
IL6, NEU, % (#1), VES, DD
KNN 5 (5.225)Simple Reg. (6.951)
CV (10)Gender, PF, Compl. type,
Max ventilation, PENTAGLOBIN,
CARDIO, IL6, %, VES, PCR, COPD,
White blood cells, Neutrophils
MLP (7.307)GBTree (8.421)
HO (80-20)TabularM5PFFSCOVID therapy, CS, IL6,
WBC, %, % (#1), n/l, ictus
Linear Reg. (3.772)Simple Reg. (6.951)
CV (10)Age, TEP, PNX, PENTAGLOBIN,
PMD, CS, Cardio, PNEUMO,
THYROID, NEU, VES, COPD,
White blood cells, Neutrophils
Simple Reg. (10.274)GBTree (8.421)
HO (80-20)2D CTLinear Reg.BFE1 featureLinear Reg. (2.659)Simple Reg. (2.636)
CV (10)1 featureMLP (4.701)MLP (4.633)
HO (80-20)2D CTLinear Reg.FFS3 featuresLinear Reg. (2.455)Simple Reg. (2.636)
CV (10)3 featuresLinear Reg. (4.845)MLP (4.633)
HO (80-20)2D CTM5PBFE1 featureM5P (2.556)Simple Reg. (2.636)
CV (10)1 featureSMO (4.896)MLP (4.633)
HO (80-20)2D CTM5PFFS1 featureM5P (2.556)Simple Reg. (2.636)
CV (10)1 featureMLP (4.26)MLP (4.633)
HO (80-20)3D CTLinear Reg.BFE4 features Linear Reg. (5.997)SMO (7.666)
CV (10)4 featuresGBTree (10.479)GBTree (11.555)
HO (80-20)3D CTLinear Reg.FFS6 featuresGBTree (6.078)SMO (7.666)
CV (10)6 featuresSMO (11.463)GBTree (11.555)
HO (80-20)3D CTM5PBFE5 featuresM5P (7.391)SMO (7.666)
CV (10)3 featuresMLP (6.069)GBTree (11.555)
HO (80-20)3D CTM5PFFS4 featuresGBTree (5.621)SMO (7.666)
CV (10)3 featuresMLP (2.947)GBTree (11.555)
Table 16. Comparison table containing some aspects of the proposed work and of three relevant works in the literature for hospitalization prediction tasks.
Table 16. Comparison table containing some aspects of the proposed work and of three relevant works in the literature for hospitalization prediction tasks.
WorkPrediction TasksDiseaseData TypeMultiple
Datasets
Missing
Values
Handling
Dataset
Balancing
Proposed
work
Outcome (Classification)
Days Range (Classification)
LOS (Regression)
Pneumonia
COVID-19
Clinical data
CT scans
CT volumes
YesYesSMOTE
Saadatmand
et al. [10]
Outcome (Classification)
Days Range (Classification)
COVID-19Clinical dataYesYesROSE
Khajehali
 et al. [11]
Days Range (Classification)PneumoniaClinical dataNoYesNo
Mekhaldi
 et al. [4]
LOS (Regression)GenericClinical dataNoNoSMOTE
Table 17. Performances of best models of compared works in all their available tasks with the Complete dataset. The best results for each task are in bold.
Table 17. Performances of best models of compared works in all their available tasks with the Complete dataset. The best results for each task are in bold.
TaskModelData TypeMetricValue
Outcome
(Classification)
Gradient Boosted Tree (Our)CT volumesAccuracy0.875
XGBoost Tree Ensemble [10]Clinical dataAccuracy0.765
Days Range
(Classification)
4-Bagging Ensemble of Decision Trees (Our)CT volumesAccuracy0.870
Bagged CART [10]Clinical dataAccuracy0.294
Voting Classifier [11]Clinical dataAccuracy0.235
LOS
(Regression)
BFE + Linear Regression (Our)CT scansMAE3.102
Multiple Linear Regression [4]Clinical dataMAE24.45
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Annunziata, A.; Cappabianca, S.; Capuozzo, S.; Coppola, N.; Di Somma, C.; Docimo, L.; Fiorentino, G.; Gravina, M.; Marassi, L.; Marrone, S.; et al. A Multimodal Machine Learning Model in Pneumonia Patients Hospital Length of Stay Prediction. Big Data Cogn. Comput. 2024, 8, 178. https://doi.org/10.3390/bdcc8120178

AMA Style

Annunziata A, Cappabianca S, Capuozzo S, Coppola N, Di Somma C, Docimo L, Fiorentino G, Gravina M, Marassi L, Marrone S, et al. A Multimodal Machine Learning Model in Pneumonia Patients Hospital Length of Stay Prediction. Big Data and Cognitive Computing. 2024; 8(12):178. https://doi.org/10.3390/bdcc8120178

Chicago/Turabian Style

Annunziata, Anna, Salvatore Cappabianca, Salvatore Capuozzo, Nicola Coppola, Camilla Di Somma, Ludovico Docimo, Giuseppe Fiorentino, Michela Gravina, Lidia Marassi, Stefano Marrone, and et al. 2024. "A Multimodal Machine Learning Model in Pneumonia Patients Hospital Length of Stay Prediction" Big Data and Cognitive Computing 8, no. 12: 178. https://doi.org/10.3390/bdcc8120178

APA Style

Annunziata, A., Cappabianca, S., Capuozzo, S., Coppola, N., Di Somma, C., Docimo, L., Fiorentino, G., Gravina, M., Marassi, L., Marrone, S., Parmeggiani, D., Polistina, G. E., Reginelli, A., Sagnelli, C., & Sansone, C. (2024). A Multimodal Machine Learning Model in Pneumonia Patients Hospital Length of Stay Prediction. Big Data and Cognitive Computing, 8(12), 178. https://doi.org/10.3390/bdcc8120178

Article Metrics

Back to TopTop