Next Article in Journal
Pelvic Sentinel Lymph Node Biopsy for Endometrial Cancer with Multi-Modal Infrared Signal Technology: A Video Article
Previous Article in Journal
Becoming Self Employed: Israeli Family Physicians’ Push and Pull Factors
Previous Article in Special Issue
Vertebral Column Pathology Diagnosis Using Ensemble Strategies Based on Supervised Machine Learning Techniques
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing Emergency Department Management: A Data-Driven Approach to Detect and Predict Surge Persistence

by
Kang Heng Lim
1,2,†,
Francis Ngoc Hoang Long Nguyen
1,†,
Ronald Wen Li Cheong
1,
Xaver Ghim Yong Tan
1,3,
Yogeswary Pasupathy
4,
Ser Chye Toh
3,
Marcus Eng Hock Ong
1,4,5 and
Sean Shao Wei Lam
1,5,6,*
1
Health Services Research Centre, Singapore Health Services Pte Ltd., Singapore 169856, Singapore
2
NUS Business Analytics Centre, NUS Business School, National University of Singapore, Singapore 119245, Singapore
3
Ngee Ann Polytechnic, Singapore 599489, Singapore
4
Department of Emergency Medicine, Singapore General Hospital, Singapore 169608, Singapore
5
Health Services and Systems Research, Duke-NUS Medical School, National University of Singapore, Singapore 169857, Singapore
6
Lee Kong Chian School of Business, Singapore Management University, Singapore 178899, Singapore
*
Author to whom correspondence should be addressed.
Joint first authors. These authors contributed equally to this work.
Healthcare 2024, 12(17), 1751; https://doi.org/10.3390/healthcare12171751
Submission received: 1 July 2024 / Revised: 24 August 2024 / Accepted: 26 August 2024 / Published: 2 September 2024

Abstract

:
The prediction of patient attendance in emergency departments (ED) is crucial for effective healthcare planning and resource allocation. This paper proposes an early warning system that can detect emerging trends in ED attendance, offering timely alerts for proactive operational planning. Over 13 years of historical ED attendance data (from January 2010 till December 2022) with 1,700,887 data points were used to develop and validate: (1) a Seasonal Autoregressive Integrated Moving Average with eXogenous factors (SARIMAX) forecasting model; (2) an Exponentially Weighted Moving Average (EWMA) surge prediction model, and (3) a trend persistence prediction model. Drift detection was achieved with the EWMA control chart, and the slopes of a kernel-regressed ED attendance curve were used to train various machine learning (ML) models to predict trend persistence. The EWMA control chart effectively detected significant COVID-19 events in Singapore. The surge prediction model generated preemptive signals on changes in the trends of ED attendance over the COVID-19 pandemic period from January 2020 until December 2022. The persistence of novel trends was further estimated using the trend persistence model, with a mean absolute error of 7.54 (95% CI: 6.77–8.79) days. This study advanced emergency healthcare management by introducing a proactive surge detection framework, which is vital for bolstering the preparedness and agility of emergency departments amid unforeseen health crises.

1. Introduction

The efficient management of emergency departments (EDs) is critical to providing timely and effective healthcare services to needy patients. EDs often face numerous challenges due to increasing patient volumes, inpatient bed shortages, and unpredictable patient arrivals at varying severity levels [1,2,3]. When poorly managed, lengthy patient waiting times can lead to patients leaving the ED without being treated, foster violence against healthcare staff, and result in increased morbidity and mortality [4,5]. ED overcrowding similarly compromises the quality of treatment and prognosis by medical staff, decreasing physician job satisfaction and reducing patient safety [6,7].
To address the challenges of ED overcrowding, healthcare professionals have turned to making reliable ED attendance predictions in order to optimize resource allocation. Traditional statistical methods such as moving averages, regression analysis, and time series analysis are standard implementations for ED predictions [8,9,10,11,12], with errors ranging between 4.2% and 14.4% for daily attendance predictions [13]. There has been comprehensive coverage of static model building, validation, and testing in the context of ED attendance forecasting [8,9,10,11,12,13,14,15,16,17,18,19]. However, research on model deployment for continuous training and testing is relatively scarce [8,9,10,11,12,13,14,15,16,17,18,19]. Previous research has proposed change point detection methods and control charts to detect outliers and changes in trends of stochastic processes [20]. However, there have been limited methods to address situations with high uncertainty [21,22,23,24,25].
In times of uncertainty, the ability to pre-emptively detect shifts in ED trends is useful for guiding effective policy response strategies. These shifts are notably exemplified by the challenges posed by the COVID-19 pandemic [26,27] and related changes in nationwide healthcare policies [28,29,30]. These interconnected occurrences underscore the need to continuously monitor trends in ED attendance so that stakeholders can appropriately respond to emergent situations and ensure the continued provision of high levels of care.
We aimed to forecast ED attendances using a Seasonal Autoregressive Integrated Moving Average with eXogenous factors (SARIMAX) model that captures calendar fixed effects [13,14,15]. In a comparative analysis of statistical models (e.g., Holt–Winters, ARIMA, SARIMAX) and machine learning models (e.g., LSTM, Gradient Boosting, Random Forest) within the context of ED attendance prediction, the SARIMAX model was chosen due to its offering comparable performances along with advantages in model interpretability and ease of implementation [8,9,10,11,12,13,16]. We propose an Early Warning System for ED attendance predictions (EWS-ED), which allows for continuous monitoring of forecasts for changes to the underlying stochastic processes that drive ED attendance. EWS-ED includes a predictive model for predicting the persistence of these changes. An Exponentially Weighted Moving Average (EWMA) control chart [30,31] was used to first detect anomalies in the quality of forecasts based on the SARIMAX model. A machine learning (ML) model was then trained and validated to predict the persistence of the anomalous conditions. This can enable pre-emptive detection of anomalous trends and allow analysts to decide whether retraining of the model is required.

2. Materials and Methods

The study hospital (SH) is one of the largest comprehensive public hospitals in Singapore, comprising more than 30 clinical disciplines and approximately 1900 inpatient beds in 2022. In the same year, the SH saw more than 100,000 ED attendances annually. Daily ED attendance data between 1 January 2010 and 31 December 2022 were extracted from the ED administrative database of the SH. Other data collected for the study included exogenous factors that may affect ED attendance (Table 1) [9,10,11]. Meteorological factors such as ambient temperature, air quality, and relative humidity were omitted, as Singapore is a tropical country with low variability in weather conditions [10].
The structure of the EWS-ED, containing three separate models, is shown in Figure 1. Data from January 2010 to December 2013 were first used to build a SARIMAX model to predict and compare against 2014 data (i.e., four years of training and one year of validation). The prediction errors from 2014 were used to define and initialize an EWMA control chart (the process mean and control limits). This control chart was subsequently used for drift detection for data from January 2015 onwards.

2.1. ED Attendance Forecasting Model (EFM)

Univariate analysis was carried out. Time series analysis for ED attendance forecasting was conducted using a SARIMAX model, which is represented as SARIMA(p, d, q) × (P, D, Q)s and is written for a time series X t as [30]: ϕ p B Φ p B s Y t = δ + θ q B Θ Q B s Z t + i = 1 n β i w i , where B denotes the backward shift operation; ϕ p , Φ p , θ q , and Θ Q are polynomials of order p, P, q, and Q, respectively; s is the seasonal period; δ is the drift constant, Z t ~ W N ( 0 , σ 2 ) ; β i corresponds to the weights of n exogenous factors w i ; and Y t = d D s X t = 1 B d 1 B s D X t , where ∇ is the differencing operator, d is the trend difference order, and D is the seasonal differencing order [32]. The exogenous factors describe the types of calendar days. The proposed SARIMAX model was identified by grid search across values of the p, d, q, P, D, and Q parameters that minimized the Akaike Information Criterion (AIC).
Data from 2010 to 2019 were used for training and data from 2020 to 2022 were used for testing. A moving window was used to allocate four years of data for training and one year of data for validation. The accuracy of the retraining framework was evaluated through an incremental learning validation process. During validation, the model parameters were updated with each week of data realization (i.e., incremental learning). Candidate models were diagnostically tested for adequacy using the Ljung–Box Test [33] and Heteroskedasticity Test [34] as well as graphically through a Quantile–Quantile (QQ) Plot and an Autocorrelation (ACF) Plot of the residuals [35]. The underlying assumption of an adequate SARIMAX model is that the residuals will follow a white noise process (i.e., zero mean, constant variance, and uncorrelated) [13,14,15]. The Ljung–Box Test checks for the presence of autocorrelation in the model residuals. It helps to determine whether there are correlation patterns left in the residuals that the model has not captured. The Heteroskedasticity Test assesses whether the variance of the residuals is constant over time. The detection of heteroskedasticity may indicate model misspecification and unreliability of predictive intervals [34]. The QQ Plot compares the distribution of residuals to a normal distribution, helping to visually identify deviations from normality. The ACF Plot shows the autocorrelation of residuals at different lags, providing insight into any remaining patterns that the model has not sufficiently captured.
Model performance was evaluated based on the Mean Absolute Error (MAE), Mean Squared Error (MSE), Mean Absolute Percentage Error (MAPE), and Root Mean Squared Error (RMSE) of annual forecasts. The MAE quantifies the average magnitude of prediction errors, providing a straightforward and interpretable measure of accuracy. The MSE emphasizes larger discrepancies by penalizing larger errors more heavily, making it sensitive to outliers, which is crucial for ED management applications where large errors are undesirable. The MAPE offers a scale-independent relative measure of error, facilitating easy comparison across different datasets. The RMSE is expressed in the same units as the data, and combines the interpretability of MAE with the sensitivity to large errors of MSE. Together, these metrics ensure a balanced assessment of model performance capturing both the average error and the impact of larger deviations [36].

2.2. Surge Prediction Model (SPM)

EWMA control charts serve as an extension to traditional Shewhart control charts [37,38] by providing emphasis on recent data points, and on average have better ability to quickly detect small shifts in the process means compared to the Shewhart Chart [21,39]. EWMA control charts were also shown to be useful for monitoring COVID-19 phases in a recent case study [40]. The upper and lower control limits were set at two standard deviations from the process mean. The EWMA statistics were monitored using resizable windows, for which the size was determined using the Window Resize Algorithm for Batch Data (WRABD) [41]. Initially, past prediction errors were used to establish the upper and lower control limits of the EWMA control chart. As new prediction errors are generated from the weekly streaming data, they can be compared against these control limits. When the prediction errors remain within the control limits, no drift is detected; the errors are stored in a list and accumulate over time. When the majority of prediction errors fall outside the control limits, this signals an out-of-control condition, indicating potential data drift. Therefore, the accumulated errors are added to the existing list of prediction errors and the oldest errors (the tail of the list) are removed in order to maintain a consistent data length. This updated set of errors is used to establish the new control limits. Subsequent prediction errors are monitored, accumulated, and compared against the updated control limits. This iterative process ensures that the control limits evolve in response to data drift. The out-of-control (OOC) rules dictate the sensitivity of the chart to detect drift. The Western Electric rules are often used to detect OOC signals in process control for industrial processes [42]. For a particular week of predictions, we assumed an OOC signal to be detected if at least four out of seven prediction errors (i.e., the majority) were more than two standard deviations from the mean (Figure 2a). The median timestamp of this OOC signal was noted, and is indicated by the dotted green vertical line in Figure 2b.

2.3. Trend Persistence Prediction Model (TPPM)

The general trends and turning points in ED attendance were derived through kernel regression [21,39]. A Gaussian kernel regression model was estimated from the ED attendance data and local extrema were noted (Figure 2b). Extrema were defined as the maximum or minimum value within a 30-day neighborhood before and after the point of interest. The bandwidth, which controls the width of the kernel function, was set as 20 to balance the trade-off between new data sensitivity and generalizability.
The OOC condition identified by the EWMA control chart based on the Western Electric rules signals a potential failure of the model to forecast accurately. When drift is detected, the gradients of a kernel regression are computed using local linear regression on various days after the detected drift. These gradients serve as a feature input to a prediction model, with the prediction target being the duration until the next turning point. The gradients from an OOC signal (or drift point) and their corresponding duration to the next turning point serve as predictor–target pairs for the trend persistence prediction model (TPPM). Detected drift points fall into one of two categories: either a drift point near an extremum (i.e., within a 7-day neighborhood vicinity) or a drift point on a trending line. As the interest of the TPPM is to predict the persistence of new trends, drift points in the former context were omitted from the training set.
The number of drift signals (i.e., data points) in the original dataset was insufficient to build a robust TPPM model. Hence, block bootstrapping was performed to resample and increase the training size while preserving the temporal correlations among the attendance data [43,44,45]. Using a block size of 30 (i.e., monthly resampling), seven block bootstrap samples were drawn within each year and concatenated across the years to generate new datasets synthetically.
The drift detection algorithm was applied to each dataset between 2015 and 2019 to generate predictor–target pairs, which were subsequently trained using various prediction models (e.g., Random Forest, XGBoost, Support Vector Machine). The hyperparameters for each potential model were tuned based on a randomized grid search through a 10-fold cross-validation for each hyperparameter candidate set. The performance of each TPPM was tested against the original dataset between 2020 and 2022, then evaluated using the MAE, MSE, RMSE, and MAPE metrics. The bias and variance of the MAE metric were estimated using the jack-knife resampling technique, where the jack-knife estimator was built by aggregating parameter estimators from each subsample obtained by omitting one observation [46]. This approach aims to provide intuition on the stability of model predictions when there are changes in the training set.

3. Results

The data comprised 1,700,887 ED attendances spanning a period of 13 years (i.e., 2010–2022). The patients came from diverse community settings, including differing genders, races, residency statuses, and triage classes. The aggregated daily ED attendance was used as the primary source of data. Table 2 summarizes a sample of the study cohort for 2020–2022. Summary statistics for the exogenous variables are listed in Table A1 in Appendix A.
The best-fit models and model performance for the EFM are summarized in Table 3. The validation process converged to two SARIMAX models with parameters (1,1,2)(0,0,0)[7] and (0,1,1)(0,0,0)[7], and the highest error occurred during the test period of 2020. There was no significant evidence of autocorrelation in the errors; all errors were normally distributed, and the variances in the errors were constant across all levels (Table A2 and Figure A1). An example of the time plot of the model forecasts for 2022 is shown in Figure 3. Time plots of the model forecasts for 2014–2021 are shown in Figure A2, Figure A3, Figure A4, Figure A5, Figure A6, Figure A7, Figure A8 and Figure A9. The SARIMAX model allows for the identification of exogenous calendar day-related variables that are significantly associated with ED attendance volumes (Table A3).
Table 4 summarizes the performance metrics of different machine learning models for the TPPM on the test set (i.e., 2020–2022). The bias and variability of the MAE performance were estimated using the jack-knife resampling method; Figure 4 shows the 95% confidence interval (CI) constructed for each MAE estimate, while Figure 5 summarizes the feature importance of the various slopes (or gradients) from kernel regression computed using local linear regression over a seven-day period from the point of drift detection.

4. Discussion

This study presents an EWS-ED early warning framework, including a prediction model for the forecasting of ED attendance, a drift detection framework based on process control charts [33], and the ability to predict the persistence of trends after an OOC signal is detected. The time series model enables the identification of independent variables that are significantly associated with ED attendance volumes (Table A3). Compared to the benchmark (i.e., a Sunday that does not fall on a public holiday, post-holiday, or pre-holiday), the ED attendance on post-public holidays was larger on average, with positive and statistically significant magnitudes in the regression coefficients. This concurs with findings reported in the extant literature [13,14]. Similar conclusions can be drawn for Monday, with statistically significant positive associations throughout the test data [47]. The higher ED attendance on these occasions can be attributed to the Monday Effect (i.e., days following a day off) reported in the literature [10]. Possible explanations for the Monday effect include patients returning to the ED from a weekend absence or the return of primary care practitioners to their office and sending their patients to the ED [48]. These analyses support the observation that attendance data falling on post-public holidays and Mondays are essential for accurate ED forecasting.
The time series model in the EFM yielded MAPE values between 5.3% and 6.6%. This result corresponds with the results of similar studies [13]. The time series plot revealed the ability of the SARIMAX model to predict the direction, peak, and troughs consistently and accurately, albeit often at conservative levels (i.e., underestimation) (Figure 3). The model’s worst performance was seen in 2020 (Table 3). The relatively higher error could be due to the COVID-19 outbreak, when EDs worldwide saw a reduction in general attendance [49]; the SARIMAX model may have faced a challenge in addressing the sharp decline in attendance. Similar to findings described in Duarte et al. [50], given the inherent uncertainties resulting from sudden surges or declines in ED attendance volumes due to pandemics or other health emergencies, these results point to the need for a drift detection model to accompany any ED attendance forecasting model [26,27]. Changes in healthcare systems, policies, and other external factors can affect the model’s generalization. The effects of these temporal changes on the ED forecasting framework were accounted for using a time-series-based SARIMAX model for the EFM. From the performance metrics reported in Table A4, the MAE ranged from the smallest in 2017 at 14.6 to the largest in 2015 at 17.2, with an average attendance difference of 2.6, which is relatively small compared to the hundreds of attendances forecasted daily.
The drift detection functionality introduced by the SPM leverages the EWMA control chart. EWMA control charts are sensitive to shifts or changes in the distribution of the prediction errors. An OOC signal suggests that the model’s prediction errors exceed the expected variations. This could be due to unusual spikes/dips in ED attendance (e.g., public health crisis, natural disaster), seasonal or temporal shifts in ED attendance (e.g., seasons, holidays), and changes in healthcare policies or resources (e.g., changes in patient flow and capacity management). EWMA and other process control charts have been introduced to detect drifts in the predictive capability of prediction algorithms [51]. By monitoring drifts in the prediction errors, the EWMA chart can provide a simple technique for continuously monitoring predictive accuracy, as the OOC signals provide preemptive information that can identify the potential of sudden and significant changes in data distributions. Figure 6 shows the ability of the EWMA to accurately pick out drift signals where changes in ED attendance are visually apparent.
Although the use of statistical process control methods in the ED setting is not novel [52], the SPM developed on accurate data across the COVID-19 pandemic reveals merits in detecting apparent changes in data trends attributable to events that happened throughout the pandemic in Singapore. A summary of nationwide events that may have led to the corresponding signals picked up by the EWMA chart is shown in Table 5. The period of January 2020 to April 2020 is corroborated with the onset of the COVID-19 pandemic. Health advisories to restrict public movements across workplaces and schools were introduced to curb the rising number of cases [53,54,55,56]. The restrictive measures (i.e., lockdowns, stay-at-home orders, social distancing) reduced the number of patients seeking medical attention at the ED [57]. This reduction may be attributed to decreased non-urgent medical issues, concerns about virus exposure, and increased public awareness about avoiding unnecessary visits to the ED. The drastic declines in ED attendance during the onset of these events were picked up by the EWMA control chart in multiple instances. Nurses and staff at the bed management unit can utilize the signals from the SPM to make preemptive adjustments to the bed resource allocations and staffing requirements.
In 2021 and 2022, Singapore began adapting to the new situation, and drift detections picked up by the EWMA chart occurred at relatively fewer persisting intervals. These signals may represent a temporal shift in ED attendance resulting from holiday effects or external shocks, as compared to changes in ED attendance resulting from the COVID-19 pandemic in early 2020. In these cases, the SARIMAX model alone failed to accurately capture the external effects, leading to higher prediction errors. For example, the drift signal picked up on 23 October 2022 corresponds to the day before the Deepavali public holiday, when the ED often observes reduced attendance. In the same period, local news agencies reported a bed crunch at Singapore hospitals, urging the public to avoid non-emergency visits to the ED [58,59]. The combination of these events led to a notable decrease in ED attendance, complemented by overestimation of ED attendance predictions, causing more significant errors and a corresponding drift signal detected in the EWMA chart. These anomalies indicate the need for a complementary model to inform hospital management of the persistence of trend changes whenever OOC signals are detected.
Table 5. Nationwide events.
Table 5. Nationwide events.
Detection Date RangeEvent DateEvent
8–14 January 20202 January 2020The Ministry of Health (MOH) issued health advisories and temperature checks for passengers at Changi Airport [53]
1–4 February 202030 January 2020Masks are issued to every household to encourage the wearing of masks for those who are unwell [55]
12–18 February 20207 February 2020Disease Outbreak Response System Condition (DOSCORN) level raised from Yellow to Orange [54]
19–24 March 202021 March 2020The first two confirmed deaths due to COVID-19 [60]
15–21 April 20203 April 2020The Circuit Breaker was announced. All non-essential workplaces were closed, and schools were moved to home-based learning [56]
25–28 April 2020
12–16 May 202116 May 2021Following an uptick in COVID-19 cases, Singapore reverts to stricter restrictions under the name of “Phase 2 (Heightened Alert)” [61]
21–25 May 2021
22–28 September 202127 September 2021Singapore entered a “Stabilization Phase” [62]
2–6 February 20221/2 February 2022Chinese New Year public holiday, and increasing number of Omicron cases [63]
25 February–1 March 202222 February 2022Singapore hit record-high infection numbers that topped 26,000, signaling the peak and end of the Omicron wave [63]
21–25 October 202224 October 2022Deepavali public holiday. Longer waiting times at hospitals [35,36]
In predicting the extent of trend shifts due to system shocks, the Extra Trees Regressor (ET) model in the TPPM yielded the lowest MAE of 7.54 days, MAPE of 27.4%, MSE of 142.46 days squared, and MAPE of 11.94 days for the TPPM (Table 4). The linear regression (LR) model performed the worst among the tested models. The predictions made by the ET model provided the smallest average MAE of 7.77 days, while the Extreme Gradient Boosting (XGB) model provided the smallest range of 95% CI (Figure 4). The predictive accuracy results of the TPPM points to the fact that pre-emptive planning can be activated if trends are expected to persist for an extended period, helping to deal with ED overcrowding arising from higher ED attendance and bed crunch [64]. Several operational measures can be implemented in the Bed Management Unit (BMU) to ensure optimal resource allocation and patient care. These include increasing staffing levels (i.e., number of doctors and nurses on calls) to manage anticipated rise in patient volume, expanding bed capacity and preparation of non-traditional spaces as additional patient care areas, accelerating discharge planning to facilitate timely bed turnover, and enhancing communication and coordination with other hospitals and care centers to manage patient transfers and support capacity needs effectively [64]. These measures are proposed actions for managing anticipated increases in patient volume. A stricter review of existing policies and thorough evaluation of the efficacy of these measures will be necessary in order to ensure that they effectively address the operational challenges and meet patient care standards.
As part of data preprocessing for the TPPM, drift points detected within a 7-day neighborhood to an extremum (estimated with kernel regression) were omitted in the TPPM training set. This preprocessing aligns with the key aim of the TPPM, which is to predict the persistence of new trends, for which turning points or changes in the direction of trends are not relevant. Nonetheless, further testing was performed on the case where these extreme drift points were included in the training set. Table 6 provides a summary of the test statistics. Compared to the metrics shown in Table 4, deterioration in the performance metrics, for example MAE, is observed for the ET, RF, XGB, SVM, and DT models. This comparison validates the improvement of the proposed data preprocessing procedure to achieve the key objective of the TPPM. The model predictions reveal higher errors for drift points within the vicinity of an extremum (Table A4), which reinforces the fact that trend persistence predictions at extreme drifts can be viewed as false positives. In the operational context (i.e., week-to-week realization of ED attendance) it is not possible to identify whether a set of gradients constitutes a turning point due to the need for future data. As such, a limitation of the algorithm is that predictions are still made for every drift point. Nevertheless, in due time the TPPM could warn users about plausible erroneous predictions when drifts are ascertained as being near an extremum.
Integrating the SPM and the TPPM within the EWS-ED framework takes inspiration from the concept of the drift detection problem [51], extending from pure drift detection to predicting the extent of drift over time. Predicting the extent of drift provides additional information to decision-makers, aiding in the sustainable deployment of ML models and helping to monitor their performance over time [65]. Application of the TPPM for concept drift detection may extend beyond the context of ED attendance. In separate contexts, developers and users can develop models specific to their contexts and utilize the information as a decision support tool to calibrate the smoothing parameter of the EWMA statistic, redefine the OOC conditions, and evaluate appropriate model updating or retraining strategies that suit the use case, among other considerations. This study contributes to the body of knowledge around monitoring concept drift in time series models. It emphasizes the importance of evaluating concept drift from the conflicting problem arising due to the stability versus the plasticity of predictive models over time [66]. Future research will need to quantify the impact of these conflicting objectives in a more theoretical manner.
A limitation of the present study is that the framework has been developed based on a single study site, which may limit the generalizability of results beyond the SH. However, the SH is one of the largest comprehensive public hospitals in Singapore, while the cohort comprised 1,700,887 ED attendances spanning 13 years (i.e., 2010–2022) and appears to be representative of the national population (Table 2). As of 2023, the three largest ethnic groups in Singapore are Chinese, Malay, and Indian, comprising 75.6%, 15.1%, and 7.6% of the total population, respectively [67]. These ratios are consistent with the distribution in the study cohort. The gender distribution of Singaporeans is 97.6 males per 100 females, compared to the study cohort of 104.6 males per 100 females in 2022. The average percentage of Singaporean citizens who attended the hospital emergency department was 86.5% (Table 2), compared to the national proportion of 61.0% [67]. A relatively higher percentage of Singaporean citizens visit public hospitals due to the extensive support and subsidies provided by the government. With the contextual validity of the cohort and the model developed with the SH data, the SH has piloted the EFM module for operational usage since January 2023. Nonetheless, given the consistent public health policies governing emergency care services across Singapore, the modeling framework can potentially be generalized nationally. Future research will look at the external validation of the modeling framework across other public hospitals in Singapore. In order to further validate the model and to ensure its applicability across diverse settings, future research could look into including data from multiple hospitals across different regions to capture a wider range of patient demographics and hospital practices. Briefly, the multi-site study design includes: first, the identification of hospitals across different geographic locations and healthcare systems; second, curation of standardized protocols for data collection, data processing, model implementation and performance evaluation for consistency across participating sites; third, fine-tuning and customization of the EWS-ED framework to account for site-specific variations and unique operational conditions; fourth, analysis comparing model accuracy and effectiveness; and lastly, conducting pilot programs at selected sites to test model feasibility and effectiveness prior to full-scale development. Through a robust multi-site validation approach, this study remains committed to advancing research with enhanced generalizability and ensuring its effectiveness in diverse operational environments [68].

5. Conclusions

This research advances emergency healthcare management by introducing a proactive surge detection framework, which is vital for bolstering the preparedness and agility of emergency departments amid unforeseen health crises. The EWS-ED framework is designed to detect drifts and predict the extent of these drifts, thereby allowing pre-emptive signals to be detected in order to consider model updating or retraining. This capability underpins agile operational planning by offering critical signals indicating the need for model recalibration.

Author Contributions

Conceptualization, S.S.W.L., K.H.L., M.E.H.O. and F.N.H.L.N.; methodology, K.H.L., F.N.H.L.N., R.W.L.C., X.G.Y.T., S.C.T. and S.S.W.L.; software, K.H.L., X.G.Y.T. and R.W.L.C.; validation, K.H.L., F.N.H.L.N., X.G.Y.T. and S.S.W.L.; formal analysis, K.H.L., F.N.H.L.N., X.G.Y.T. and S.S.W.L.; investigation, K.H.L., F.N.H.L.N., X.G.Y.T., Y.P., S.C.T. and S.S.W.L.; resources, K.H.L., F.N.H.L.N., M.E.H.O. and S.S.W.L.; data curation, K.H.L., F.N.H.L.N., R.W.L.C., X.G.Y.T., Y.P., M.E.H.O. and S.S.W.L.; writing—original draft preparation, K.H.L., F.N.H.L.N., X.G.Y.T., M.E.H.O. and S.S.W.L.; visualization, K.H.L.; supervision, F.N.H.L.N., S.C.T., M.E.H.O. and S.S.W.L.; project administration, F.N.H.L.N. and S.S.W.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data from this research is available via: https://datadryad.org/stash (accessed on 14 June 2023).

Acknowledgments

We would like to thank the Bed Management Unit of the SH for supporting this project. We would also like to thank WQ See for support in project management of this project.

Conflicts of Interest

Authors Kang Heng Lim, Francis Ngoc Hoang Long Nguyen, Ronald Wen Li Cheong, Xaver Ghim Yong Tan, Marcus Eng Hock Ong and Sean Shao Wei Lam were employed by the company Singapore Health Services Pte Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Appendix A

Table A1. Summary statistics of exogenous variables.
Table A1. Summary statistics of exogenous variables.
FactorDaily Mean (Standard Deviation)Daily Median (Interquartile Range)
202020212022202020212022
Chinese New Year334.7 (38.0)264.0 (43.8)326.5 (44.5)336.0 (38.0)264.0 (31.0)326.5 (31.5)
Christmas Day268.0 (-)265.0 (-)290.0 (1.4)268.0 (-)265.0 (-)290.0 (1.0)
Deepavali283.0 (-)246.0 (-)281.0 (-)283.0 (-)246.0 (-)281.0 (-)
Good Friday283.0 (-)266.0 (-)245.0 (-)283.0 (-)266.0 (-)245.0 (-)
Hari Raya Haji260.0 (-)270.0 (-)350.5 (38.9)260.0 (-)270.0 (-)350.5 (27.5)
Hari Raya Puasa284.5 (54.4)330.0 (-)296.0 (-)284.5 (38.5)330.0 (-)296.0 (-)
Labour Day238.0 (-)303.0 (-)284.0 (14.1)238.0 (-)303.0 (-)284.0 (10.0)
National Day318.0 (45.3)278.0 (-)280.0 (-)318.0 (32.0)278.0 (-)280.0 (-)
New Year’s Day339.0 (-)251.0 (-)281.0 (-)339.0 (-)251.0 (-)281.0 (-)
Vesak Day282.0 (-)281.0 (-)305.5 (13.4)282.0 (-)281.0 (-)305.5 (9.5)
Pre-Chinese New Year289.0 (-)260.0 (-)301.0 (-)289.0 (-)260.0 (-)301.0 (-)
Pre-Christmas Day285.0 (-)232.0 (-)288.0 (-)285.0 (-)232.0 (-)288.0 (-)
Pre-Deepavali317.0 (-)254.0 (-)264.0 (-)317.0 (-)254.0 (-)264.0 (-)
Pre-Good Friday318.0 (-)343.0 (-)309.0 (-)318.0 (-)343.0 (-)309.0 (-)
Pre-Hari Raya Haji338.0 (-)377.0 (-)320.0 (-)338.0 (-)377.0 (-)320.0 (-)
Pre-Hari Raya Puasa254.0 (-)344.0 (-)274.0 (-)254.0 (-)344.0 (-)274.0 (-)
Pre-Labour Day243.0 (-)320.0 (-)286.0 (-)243.0 (-)320.0 (-)286.0 (-)
Pre-National Day310.0 (-)231.0 (-)354.0 (-)310.0 (-)231.0 (-)354.0 (-)
Pre-New Year’s Day271.0 (-)253.0 (-)291.0 (-)271.0 (-)253.0 (-)291.0 (-)
Pre-Vesak Day252.0 (-)309.0 (-)314.0 (-)252.0 (-)309.0 (-)314.0 (-)
Post-Chinese New Year398.0 (-)402.0 (-)365.0 (-)398.0 (-)402.0 (-)365.0 (-)
Post-Christmas Day372.0 (-)355.0 (-)329.0 (-)372.0 (-)355.0 (-)329.0 (-)
Post-Deepavali380.0 (-)340.0 (-)348.0 (-)380.0 (-)340.0 (-)348.0 (-)
Post-Good Friday327.0 (-)369.0 (-)315.0 (-)327.0 (-)369.0 (-)315.0 (-)
Post-Hari Raya Haji362.0 (-)355.0 (-)407.0 (-)362.0 (-)355.0 (-)407.0 (-)
Post-Hari Raya Puasa331.0 (-)422.0 (-)332.0 (-)331.0 (-)422.0 (-)332.0 (-)
Post-Labour Day290.0 (-)389.0 (-)- (-)290.0 (-)389.0 (-)- (-)
Post-National Day344.0 (-)348.0 (-)363.0 (-)344.0 (-)348.0 (-)363.0 (-)
Post-New Year’s Day445.0 (-)336.0 (-)346.0 (-)445.0 (-)336.0 (-)346.0 (-)
Post-Vesak Day230.0 (-)312.0 (-)389.0 (-)230.0 (-)312.0 (-)389.0 (-)
Working Day310.8 (36.3)311.0 (34.8)321.0 (32.1)307.0 (46.25)307.0 (46.0)318.0 (39.0)
Friday296.6 (31.8)295.2 (32.0)305.2 (24.1)293.0 (37.0)298.0 (35.0)305.5 (29.5)
Monday349.3 (28.6)352.8 (30.5)354.6 (31.2)350.0 (34.0)353.5 (31.25)360.0 (38.5)
Saturday282.7 (32.0)281.3 (22.9)296.6 (21.1)284.0 (31.0)282.5 (29.25)293.0 (28.0)
Sunday268.8 (27.8)268.5 (25.8)276.9 (24.4)266.0 (38.25)270.0 (39.25)276.0 (34.25)
Thursday300.0 (32.6)295.6 (27.4)309.4 (30.0)298.0 (36.0)293.5 (40.5)309.0 (36.75)
Tuesday308.3 (33.2)304.9 (24.4)321.4 (27.4)308.0 (39.0)305.5 (34.0)320.0 (31.25)
Table A2. Diagnostic tests.
Table A2. Diagnostic tests.
Model (Test Set)Heteroskedasticity TestLjung-Box Test
Test Statisticp-ValueTest Statisticp-Value
20201.020.841.710.19
20211.440.000.210.65
20221.020.790.380.54
Table A3. Univariate analysis of SARIMAX models.
Table A3. Univariate analysis of SARIMAX models.
Testing Period202020212022
PredictorsCoefficientp-ValueCoefficientp-ValueCoefficientp-Value
Intercept367.40.0 *397.80.0 *429.30.0 *
Chinese New Year5.90.415.10.0 *8.20.2
Christmas Day−0.11.0−4.10.8−0.51.0
Deepavali−0.81.0−0.71.0−3.60.9
Good Friday−2.40.9−4.70.8−2.70.9
Hari Raya Haji−2.30.8−10.70.2−11.00.2
Hari Raya Puasa−13.70.1−12.30.1−0.90.9
Labour Day−9.80.6−5.40.7−3.70.8
National Day3.30.90.61.0−7.30.6
New Year’s Day13.30.114.40.42.70.9
Vesak Day1.90.811.30.210.70.3
Pre-Chinese New Year−49.30.0 *−48.00.0 *−51.50.0 *
Pre-Christmas Day−24.20.1−21.80.2−25.40.0 *
Pre-Deepavali−18.00.1−8.30.3−6.70.6
Pre-Good Friday−19.30.0 *−14.10.17.10.6
Pre-Hari Raya Haji0.71.07.40.610.80.5
Pre-Hari Raya Puasa−31.50.0 *−14.80.0 *−9.70.2
Pre-Labour Day1.80.94.50.72.50.8
Pre-National Day−0.31.04.10.7−7.60.4
Pre-New Year’s Day−24.90.0 *−29.90.0 *−26.50.1
Pre-Vesak Day6.10.67.30.6−0.31.0
Post-Chinese New Year58.70.0 *59.30.0 *55.80.0 *
Post-Christmas Day46.40.0 *40.10.0 *35.50.0 *
Post-Deepavali18.90.0 *27.60.0 *41.40.0 *
Post-Good Friday−3.60.8−8.90.3−2.20.8
Post-Hari Raya Haji21.40.0 *13.50.128.60.0 *
Post-Hari Raya Puasa25.10.232.50.150.80.0 *
Post-Labour Day39.70.0 *25.90.0 *14.80.1
Post-National Day60.20.148.30.0 *44.60.0 *
Post-New Year’s Day28.50.0 *47.60.0 *34.70.0 *
Post-Vesak Day27.00.0 *22.90.0 *21.90.0 *
Friday−20.40.0 *−5.80.3−7.20.3
Monday40.10.0 *52.50.0 *48.90.0 *
Saturday3.20.17.30.0 *9.90.0 *
Thursday−14.50.0 *−1.20.8−3.50.6
Tuesday−1.60.88.90.14.50.5
Wednesday−14.80.0 *−3.10.6−4.40.5
Working Day34.30.0 *25.80.0 *31.60.0 *
AR1--1.00.0 *--
MA1−0.90.0 *−1.80.0 *−0.80.0 *
MA2--0.80.0 *--
* Statistically significant (p-value < 0.05).
Table A4. Summary of performance metrics on the validation set.
Table A4. Summary of performance metrics on the validation set.
Train PeriodTest PeriodMAEMAPEMSERMSE
2010–2013201416.3300.044441.30321.007
2011–2014201517.2070.046489.65922.128
2012–2015201616.2150.046430.95520.759
2013–2016201714.6140.042340.41418.450
2014–2017201815.6260.046406.28520.157
2015–2018201917.1080.050456.60821.368
Figure A1. Diagnostic tests: (a) Q-Q Plot (2020); (b) ACF Plot (2020); (c) Q-Q Plot (2021); (d) ACF Plot (2021); (e) Q-Q Plot (2022); (f) ACF Plot (2022). The red line of the Q-Q Plot is a reference line to represent the case where the data is normally distributed. The shaded blue region of the ACF Plot denotes the 95% confidence band to determine if autocorrelation values are statistically significant.
Figure A1. Diagnostic tests: (a) Q-Q Plot (2020); (b) ACF Plot (2020); (c) Q-Q Plot (2021); (d) ACF Plot (2021); (e) Q-Q Plot (2022); (f) ACF Plot (2022). The red line of the Q-Q Plot is a reference line to represent the case where the data is normally distributed. The shaded blue region of the ACF Plot denotes the 95% confidence band to determine if autocorrelation values are statistically significant.
Healthcare 12 01751 g0a1
Figure A2. Model forecasts for 2014.
Figure A2. Model forecasts for 2014.
Healthcare 12 01751 g0a2
Figure A3. Model forecasts for 2015.
Figure A3. Model forecasts for 2015.
Healthcare 12 01751 g0a3
Figure A4. Model forecasts for 2016.
Figure A4. Model forecasts for 2016.
Healthcare 12 01751 g0a4
Figure A5. Model forecasts for 2017.
Figure A5. Model forecasts for 2017.
Healthcare 12 01751 g0a5
Figure A6. Model forecasts for 2018.
Figure A6. Model forecasts for 2018.
Healthcare 12 01751 g0a6
Figure A7. Model forecasts for 2019.
Figure A7. Model forecasts for 2019.
Healthcare 12 01751 g0a7
Figure A8. Model forecasts for 2020.
Figure A8. Model forecasts for 2020.
Healthcare 12 01751 g0a8
Figure A9. Model forecasts for 2021.
Figure A9. Model forecasts for 2021.
Healthcare 12 01751 g0a9

References

  1. Forster, A.J. The Effect of Hospital Occupancy on Emergency Department Length of Stay and Patient Disposition. Acad. Emerg. Med. 2003, 10, 127–133. [Google Scholar] [CrossRef] [PubMed]
  2. Salway, R.; Valenzuela, R.; Shoenberger, J.; Mallon, W.; Viccellio, A. Emergency department (ED) overcrowding: Evidence-based answers to frequently asked questions. Rev. Médica Clínica Las Condes 2017, 28, 213–219. [Google Scholar] [CrossRef]
  3. Forero, R.; McCarthy, S.; Hillman, K. Access Block and Emergency Department Overcrowding. In Annual Update in Intensive Care and Emergency Medicine 2011; Vincent, J.L., Ed.; Springer: Berlin/Heidelberg, Germany, 2011; Volume 1, pp. 720–728. Available online: http://link.springer.com/10.1007/978-3-642-18081-1_63 (accessed on 29 May 2023).
  4. Sprivulis, P.C.; Da Silva, J.; Jacobs, I.G.; Jelinek, G.A.; Frazer, A.R.L. The association between hospital overcrowding and mortality among patients admitted via Western Australian emergency departments. Med. J. Aust. 2006, 184, 208–212. [Google Scholar] [CrossRef]
  5. Richardson, D.B. Increase in patient mortality at 10 days associated with emergency department overcrowding. Med. J. Aust. 2006, 184, 213–216. [Google Scholar] [CrossRef]
  6. Lin, B.Y.J.; Hsu, C.P.C.; Chao, M.C.; Luh, S.P.; Hung, S.W.; Breen, G.M. Physician and Nurse Job Climates in Hospital-Based Emergency Departments in Taiwan: Management and Implications. J. Med. Syst. 2008, 32, 269–281. [Google Scholar] [CrossRef] [PubMed]
  7. Hall, L.H.; Johnson, J.; Watt, I.; Tsipa, A.; O’Connor, D.B. Healthcare Staff Wellbeing, Burnout, and Patient Safety: A Systematic Review. PLoS ONE 2016, 11, e0159015. [Google Scholar] [CrossRef]
  8. Calegari, R.; Fogliatto, F.S.; Lucini, F.R.; Neyeloff, J.; Kuchenbecker, R.S.; Schaan, B.D. Forecasting Daily Volume and Acuity of Patients in the Emergency Department. Comput. Math. Methods Med. 2016, 2016, 1–8. [Google Scholar] [CrossRef]
  9. Afilal, M.; Yalaoui, F.; Dugardin, F.; Amodeo, L.; Laplanche, D.; Blua, P. Forecasting the Emergency Department Patients Flow. J. Med. Syst. 2016, 40, 175. [Google Scholar] [CrossRef]
  10. Sun, Y.; Heng, B.H.; Seow, Y.T.; Seow, E. Forecasting daily attendances at an emergency department to aid resource planning. BMC Emerg. Med. 2009, 9, 1. [Google Scholar] [CrossRef]
  11. Ekström, A.; Kurland, L.; Farrokhnia, N.; Castrén, M.; Nordberg, M. Forecasting Emergency Department Visits Using Internet Data. Ann. Emerg. Med. 2015, 65, 436–442.e1. [Google Scholar] [CrossRef]
  12. Etu, E.E.; Monplaisir, L.; Masoud, S.; Arslanturk, S.; Emakhu, J.; Tenebe, I.; Miller, J.B.; Hagerman, T.; Jourdan, D.; Krupp, S. A Comparison of Univariate and Multivariate Forecasting Models Predicting Emergency Department Patient Arrivals during the COVID-19 Pandemic. Healthcare 2022, 10, 1120. [Google Scholar] [CrossRef]
  13. Wargon, M.; Guidet, B.; Hoang, T.D.; Hejblum, G. A systematic review of models for forecasting the number of emergency department visits. Emerg. Med. J. 2009, 26, 395–399. [Google Scholar] [CrossRef] [PubMed]
  14. McCarthy, M.L.; Zeger, S.L.; Ding, R.; Aronsky, D.; Hoot, N.R.; Kelen, G.D. The Challenge of Predicting Demand for Emergency Department Services. Acad. Emerg. Med. 2008, 15, 337–346. [Google Scholar] [CrossRef] [PubMed]
  15. Jones, S.S.; Thomas, A.; Evans, R.S.; Welch, S.J.; Haug, P.J.; Snow, G.L. Forecasting Daily Patient Volumes in the Emergency Department. Acad. Emerg. Med. 2008, 15, 159–170. [Google Scholar] [CrossRef] [PubMed]
  16. Vollmer, M.A.C.; Glampson, B.; Mellan, T.; Mishra, S.; Mercuri, L.; Costello, C.; Klaber, R.; Cooke, G.; Flaxman, S.; Bhatt, S. A unified machine learning approach to time series forecasting applied to demand at emergency departments. BMC Emerg. Med. 2021, 21, 9. [Google Scholar] [CrossRef] [PubMed]
  17. Jilani, T.; Housley, G.; Figueredo, G.; Tang, P.S.; Hatton, J.; Shaw, D. Short and Long term predictions of Hospital emergency department attendances. Int. J. Med. Inf. 2019, 129, 167–174. [Google Scholar] [CrossRef]
  18. Graham, B.; Bond, R.; Quinn, M.; Mulvenna, M. Using Data Mining to Predict Hospital Admissions From the Emergency Department. IEEE Access. 2018, 6, 10458–10469. [Google Scholar] [CrossRef]
  19. Champion, R.; Kinsman, L.D.; Lee, G.A.; Masman, K.A.; May, E.A.; Mills, T.M.; Taylor, M.D.; Thomas, P.R.; Williams, R.J. Forecasting emergency department presentations. Aust. Health Rev. 2007, 31, 83. [Google Scholar] [CrossRef]
  20. Croux, C.; Gelper, S.; Mahieu, K. Robust control charts for time series data. Expert. Syst. Appl. 2011, 38, 13810–13815. [Google Scholar] [CrossRef]
  21. Souza, G.; Samohyl, R. Monitoring Forecast Errors with Combined CUSUM and Shewhart Control Charts. In Proceedings of the International Symposium of Forecasting, Nice, France, 22–25 June 2008. [Google Scholar]
  22. Jiang, W.; Tsui, K.-L. A New SPC Monitoring Method: The ARMA Chart; Taylor & Francis, Ltd.: Abingdon, UK, 2000. [Google Scholar]
  23. Alwan, L.C.; Roberts, H.V. Time-Series Modeling for Statistical Process Control. J. Bus. Econ. Stat. 1988, 6, 87–95. [Google Scholar] [CrossRef]
  24. Koehler, A.B.; Marks, N.B.; O’Connell, R.T. EWMA control charts for autoregressive processes. J. Oper. Res. Soc. 2001, 52, 699–707. [Google Scholar] [CrossRef]
  25. Dann, L.; Fitzsimons, J.; Gorman, K.M.; Hourihane, J.; Okafor, I. Disappearing act: COVID-19 and paediatric emergency department attendances. Arch. Dis. Child. 2020, 105, 810–811. [Google Scholar] [CrossRef] [PubMed]
  26. Honeyford, K.; Coughlan, C.; Nijman, R.; Expert, P.; Burcea, G.; Maconochie, I.; Kinderlerer, A.; Cooke, G.S.; Costelloe, C.E. Changes in Emergency Department Activity and the First COVID-19 Lockdown: A Cross-sectional Study. West. J. Emerg. Med. 2021, 22, 603–607. Available online: https://escholarship.org/uc/item/2s7818zq (accessed on 29 August 2023). [CrossRef] [PubMed]
  27. Kalucy, R.; Thomas, L.; King, D. Changing Demand for Mental Health Services in the Emergency Department of a Public Hospital. Aust. N. Z. J. Psychiatry 2005, 39, 74–80. [Google Scholar] [CrossRef] [PubMed]
  28. Brownlea, S.; Miller, J.; Taylor, N.; Miller, P.; Coomber, K.; Baldwin, R.; Palmer, D. Impact of alcohol policy changes on substance-affected patients attending an emergency department in the Northern Territory with police. Emerg. Med. Australas. 2023, 35, 390–397. [Google Scholar] [CrossRef]
  29. Van Den Heede, K.; Van De Voorde, C. Interventions to reduce emergency department utilisation: A review of reviews. Health Policy 2016, 120, 1337–1349. [Google Scholar] [CrossRef]
  30. Gan, F.F. Designs of One- and Two-Sided Exponential EWMA Charts. J. Qual. Technol. 1998, 30, 55–69. [Google Scholar] [CrossRef]
  31. Rasheed, Z.; Zhang, H.; Anwar, S.M. Reassessment of performance evaluation of EWMA control chart for exponential process. Qual. Reliab. Eng. Int. 2023, 40, 1685–1697. [Google Scholar] [CrossRef]
  32. Box, G.E.P.; Jenkins, G.M.; Reinsel, G.C. Time Series Analysis: Forecasting and Control, 3rd ed.; Prentice Hall: Englewood Cliffs, NJ, USA, 1994; 598p. [Google Scholar]
  33. Ljung, G.M.; Box, G.E.P. On a measure of lack of fit in time series models. Biometrika 1978, 65, 297–303. [Google Scholar] [CrossRef]
  34. Breusch, T.S.; Pagan, A.R. A Simple Test for Heteroscedasticity and Random Coefficient Variation. Econometrica 1979, 47, 1287. [Google Scholar] [CrossRef]
  35. Hyndman, R.J.; Athanasopoulos, G. Forecasting: Principles and Practice, 2nd ed.; Otexts: Lexington, KY, USA, 2018; 382p. [Google Scholar]
  36. Willmott, C.J. Some Comments on the Evaluation of Model Performance. Bull. Am. Meteorol. Soc. 1982, 63, 1309–1313. [Google Scholar] [CrossRef]
  37. Nelson, L.S. The Shewhart Control Chart—Tests for Special Causes. J. Qual. Technol. 1984, 16, 237–239. [Google Scholar] [CrossRef]
  38. Rasheed, Z.; Zhang, H.; Anwar, S.M.; Noor-ul-Amin, M.; Adegoke, N.A.; Abbasi, S.A. Designing efficient dispersion control charts under various ranked-set sampling approaches. J. Comput. Appl. Math. 2024, 441, 115680. [Google Scholar] [CrossRef]
  39. Lone, S.A.; Rasheed, Z.; Anwar, S.; Khan, M.; Anwar, S.M.; Shahab, S. Enhanced fault detection models with real-life applications. AIMS Math. 2023, 8, 19595–19636. [Google Scholar] [CrossRef]
  40. Waqas, M.; Xu, S.H.; Anwar, S.M.; Rasheed, Z.; Shabbir, J. The optimal control chart selection for monitoring COVID-19 phases: A case study of daily deaths in the USA. Int. J. Qual. Health Care 2023, 35, mzad058. [Google Scholar] [CrossRef] [PubMed]
  41. Klinkenberg, R.; Renz, I. Adaptive Information Filtering: Learning in the Presence of Concept Drifts. In Learning for Text Categorization; 1998; pp. 33–40. Available online: https://www.researchgate.net/publication/2786410_Adaptive_Information_Filtering_Learning_in_the_Presence_of_Concept_Drifts (accessed on 19 February 2024).
  42. Montgomery, D.C. Introduction to Statistical Quality Control, 8th ed.; John Wiley & Sons, Inc.: Hoboken, NJ, USA, 2020. [Google Scholar]
  43. Nadaraya, E.A. On Estimating Regression. Theory Probab. Its Appl. 1964, 9, 141–142. [Google Scholar] [CrossRef]
  44. Bierens, H.J. Topics in Advanced Econometrics: Estimation, Testing, and Specification of Cross-Section and Time Series Models; Cambridge University Press: Cambridge, UK; New York, NY, USA, 1994; 258p. [Google Scholar]
  45. Radovanov, B.; Marcikić, A. A comparison of four different block bootstrap methods. Croat. Oper. Res. Rev. 2014, 5, 189–202. [Google Scholar] [CrossRef]
  46. Efron, B. The Jackknife, the Bootstrap, and Other Resampling Plans; Society for Industrial and Applied Mathematics: Philadelphia, PA, USA, 1982. [Google Scholar]
  47. Downing, A. Temporal and demographic variations in attendance at accident and emergency departments. Emerg. Med. J. 2002, 19, 531–535. [Google Scholar] [CrossRef]
  48. Batal, H.; Tench, J.; McMillan, S.; Adams, J.; Mehler, P.S. Predicting Patient Visits to an Urgent Care Clinic Using Calendar Variables. Acad. Emerg. Med. 2001, 8, 48–53. [Google Scholar] [CrossRef]
  49. Lateef, F. The impact of the COVID 19 pandemic on emergency department attendance: What seems to be keeping the patients away? J. Emerg. Trauma Shock 2020, 13, 246. [Google Scholar] [CrossRef]
  50. Duarte, D.; Walshaw, C.; Ramesh, N. A Comparison of Time-Series Predictions for Healthcare Emergency Department Indicators and the Impact of COVID-19. Appl. Sci. 2021, 11, 3561. [Google Scholar] [CrossRef]
  51. Bayram, F.; Ahmed, B.S.; Kassler, A. From concept drift to model degradation: An overview on performance-aware drift detectors. Knowl.-Based Syst. 2022, 245, 108632. [Google Scholar] [CrossRef]
  52. Harrou, F.; Kadri, F.; Sun, Y.; Khadraoui, S. Monitoring patient flow in a hospital emergency department: ARMA-based nonparametric GLRT scheme. Health Inform. J. 2021, 27, 146045822110216. [Google Scholar] [CrossRef]
  53. Goh, T. Travellers Arriving at Changi Airport from Wuhan to Undergo Temperature Screening after Pneumonia Outbreak. The Straits Times, 3 February 2020. [Google Scholar]
  54. MOH. Risk Assessment Raised to Dorscon Orange; MOH: Singapore, 2020.
  55. Salma, K.; Goh, T.; Kurohi, R. Wuhan Virus: Every Household in Singapore to Get 4 Masks; Collection Starts on Feb 1. The Straits Times, 30 January 2020. [Google Scholar]
  56. MOH. Circuit Breaker to Minimise Further Spread of COVID-19; MOH: Singapore, 2020.
  57. Chong, S.L.; Soo, J.S.L.; Allen, J.C.; Ganapathy, S.; Lee, K.P.; Tyebally, A.; Yung, C.F.; Thoon, K.C.; Ng, Y.H.; Oh, J.Y. Impact of COVID-19 on pediatric emergencies and hospitalizations in Singapore. BMC Pediatr. 2020, 20, 562. [Google Scholar] [CrossRef]
  58. Lim, V. Longer Waiting Times at Hospitals with Some Patients Told to Wait Up to 50 Hours for a Bed. CAN, 20 October 2022. [Google Scholar]
  59. Salma, K. Bed Crunch at Singapore Hospitals: Some Patients Are Stuck in Emergency Departments. The Straits Times, 20 October 2022. [Google Scholar]
  60. Liu, V. Singapore Reports First Two Coronavirus Deaths: A 75-Year-Old Singaporean Woman and 64-Year Old Indonesian Man. The Straits Times, 22 March 2020. [Google Scholar]
  61. MOH. Updates on Local Situation and Heightened Alert to Minimise Transmission; MOH: Singapore, 2021.
  62. MOH. Stabilising Our COVID-19 Situation and Protecting Our Overall Healthcare Capacity; MOH: Singapore, 2021.
  63. Salma, K. High Numbers Indicate COVID-19 Omicron Wave Will Likely Peak Soon, Say Experts. The Straits Times, 23 February 2022. p. 31.
  64. Olshaker, J.S. Managing Emergency Department Overcrowding. Emerg. Med. Clin. N. Am. 2009, 27, 593–603. [Google Scholar] [CrossRef] [PubMed]
  65. Hoens, T.R.; Polikar, R.; Chawla, N.V. Learning from streaming data with concept drift and imbalance: An overview. Prog. Artif. Intell. 2012, 1, 89–101. [Google Scholar] [CrossRef]
  66. Grossberg, S. Nonlinear neural networks: Principles, mechanisms, and architectures. Neural Netw. 1988, 1, 17–61. [Google Scholar] [CrossRef]
  67. National Population and Talent PMO; Singapore Department of Statistics; Ministry of Home Affairs; Immigration and Checkpoints Authority; Ministry of Manpower. Population in Brief. Singapore Government. 2023. Available online: https://www.population.gov.sg/files/media-centre/publications/population-in-brief-2023.pdf (accessed on 19 February 2024).
  68. Goodlett, D.; Hung, A.; Feriozzi, A.; Lu, H.; Bekelman, J.E.; Mullins, C.D. Site engagement for multi-site clinical trials. Contemp. Clin. Trials Commun. 2020, 19, 100608. [Google Scholar] [CrossRef]
Figure 1. EWS-ED Framework (EFM, SPM, and TPPM) [31].
Figure 1. EWS-ED Framework (EFM, SPM, and TPPM) [31].
Healthcare 12 01751 g001
Figure 2. Drift Detection: (a) EWMA control chart with OOC condition (SPM) and (b) drift signal along the kernel-regressed curve (TPPM filters). The red ‘x’ denotes the local extrema of the kernel-regressed curve.
Figure 2. Drift Detection: (a) EWMA control chart with OOC condition (SPM) and (b) drift signal along the kernel-regressed curve (TPPM filters). The red ‘x’ denotes the local extrema of the kernel-regressed curve.
Healthcare 12 01751 g002
Figure 3. Model forecasts for 2022.
Figure 3. Model forecasts for 2022.
Healthcare 12 01751 g003
Figure 4. Jack-knife confidence interval for MAE estimates.
Figure 4. Jack-knife confidence interval for MAE estimates.
Healthcare 12 01751 g004
Figure 5. Feature importance for gradients estimated from drift detection within seven days.
Figure 5. Feature importance for gradients estimated from drift detection within seven days.
Healthcare 12 01751 g005
Figure 6. Drift signals between 2020 and 2022 detected by the SPM. The black ‘x’ denotes the local extrema of the kernel-regressed curve.
Figure 6. Drift signals between 2020 and 2022 detected by the SPM. The black ‘x’ denotes the local extrema of the kernel-regressed curve.
Healthcare 12 01751 g006
Table 1. Exogenous factors.
Table 1. Exogenous factors.
FactorDescription
HolidayPublic holidays declared by the Ministry of Manpower in Singapore (e.g., Good Friday)
Post-HolidayThe working day following a public holiday
Pre-HolidayThe day preceding a public holiday
Working DayIndicator for weekdays and non-public holidays
Day of the WeekIndicator for Monday, Tuesday, Wednesday, Thursday, Friday, Saturday, or Sunday
Table 2. Cohort summary data of daily ED attendance.
Table 2. Cohort summary data of daily ED attendance.
Characteristic *Daily Mean (Standard Deviation)Daily Median (Interquartile Range)
202020212022202020212022
Male160.8 (23.0)156.5 (20.8)160.5 (19.6)158 (144–231)154 (143–168)158.5 (147–172)
Female141.3 (22.4)143.3 (19.6)152.0 (18.9)140 (125–155)142 (130–156)151.5 (138–164)
Chinese192.2 (28.4)194.7 (25.9)204.2 (24.2)191 (174–209)192 (178–211)202 (188–220)
Malay31.5 (7.74)31.2 (6.58)33.5 (7.59)36 (31–61)31 (27–35)33 (28–38)
Indian42.8 (8.31)42.6 (8.23)41.0 (7.39)48 (42–48)43 (37–48)41 (36–46)
Others15.9 (4.92)15.8 (3.80)17.8 (4.63)19 (15–42)16 (13–18)17 (14–21)
Singapore Residents256.3 (37.0)261.7 (34.4)272.9 (33.0)255 (230–277)258 (239–281)271 (250–292)
Non-Residents45.8 (13.7)38.2 (7.30)39.5 (7.47)43 (36–53)38 (33–43)39 (34–45)
P130.9 (7.91)33.5 (7.97)40.4 (9.44)31 (25–35.8)33 (28–39)39 (34–46)
P1F0.83 (1.12)1.09 (1.17)1.55 (1.67)0.0 (0.0–1.0)1.0 (0.0–2.0)1.0 (0.0–2.0)
P283.2 (19.3)92.6 (18.8)103.6 (19.9)82 (70–94)92 (79–105)102.5 (90–117)
P2+85.3 (14.2)87.8 (13.9)86.9 (12.4)85 (76.2–94.0)87 (79–97)86 (79–95)
P2F6.02 (3.33)7.74 (4.55)9.02 (6.12)6.0 (4.0–8.0)7.0 (4.0–10)7.0 (5.0–12.3)
P356.9 (19.0)54.8 (11.8)58.0 (13.1)54.5 (42–66.8)54 (47–62)58 (48–67)
P3F36.3 (24.5)21.7 (7.27)10.4 (9.81)31 (23–43)21 (17–27)7.0 (4.0–14)
P40.60 (0.94)0.40 (0.65)0.51 (0.75)0.0 (0.0–1.0)0.0 (0.0–1.0)0.0 (0.0–1.0)
* Characteristics are listed in the following order: Gender, Race, Residency Status, Triage Class. Missing data are present under the Race and Triage Class categories: 0.1% and 2.6%, respectively.
Table 3. Summary of forecasting performance.
Table 3. Summary of forecasting performance.
Train PeriodTest PeriodBest Fit ModelMAEMAPEMSERMSE
2016–20192020SARIMAX(0,1,1)(0,0,0)[7]19.50.066263425.2
2017–20202021SARIMAX(1,1,2)(0,0,0)[7]17.40.059250222.4
2018–20212022SARIMAX(0,1,1)(0,0,0)[7]16.30.052942120.5
Table 4. Summary of TPPM performance.
Table 4. Summary of TPPM performance.
ModelMAEMAPEMSERMSE
Support Vector Regression (SVM)9.080.35144.0012.00
K Neighbors Regressor (KNN)8.620.30152.6212.35
Extra Trees Regressor (ET)7.540.27142.4611.94
Random Forest Regressor (RF)8.380.29154.3812.43
Extreme Gradient Boosting (XGB)8.690.37162.5412.75
Decision Tree Regressor (DT)9.460.34149.0012.21
Gradient Boosting Regressor (GBR)8.850.33180.8513.45
Linear Regression (LR)2.72 × 1071.232.32 × 10154.82 × 107
Table 6. Summary of TPPM performance without dropping extreme drift points).
Table 6. Summary of TPPM performance without dropping extreme drift points).
ModelMAEMAPEMSERMSE
Support Vector Regression (SVM)7.850.31126.9211.27
K Neighbors Regressor (KNN)8.380.28148.2312.18
Extra Trees Regressor (ET)9.920.46203.3114.26
Random Forest Regressor (RF)12.081.96315.3117.76
Extreme Gradient Boosting (XGB)12.081.20319.4617.87
Decision Tree Regressor (DT)13.462.71351.3118.74
Gradient Boosting Regressor (GBR)8.620.45178.3113.35
Linear Regression (LR)3.38 × 1071.833.59 × 10155.99 × 107
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lim, K.H.; Nguyen, F.N.H.L.; Cheong, R.W.L.; Tan, X.G.Y.; Pasupathy, Y.; Toh, S.C.; Ong, M.E.H.; Lam, S.S.W. Enhancing Emergency Department Management: A Data-Driven Approach to Detect and Predict Surge Persistence. Healthcare 2024, 12, 1751. https://doi.org/10.3390/healthcare12171751

AMA Style

Lim KH, Nguyen FNHL, Cheong RWL, Tan XGY, Pasupathy Y, Toh SC, Ong MEH, Lam SSW. Enhancing Emergency Department Management: A Data-Driven Approach to Detect and Predict Surge Persistence. Healthcare. 2024; 12(17):1751. https://doi.org/10.3390/healthcare12171751

Chicago/Turabian Style

Lim, Kang Heng, Francis Ngoc Hoang Long Nguyen, Ronald Wen Li Cheong, Xaver Ghim Yong Tan, Yogeswary Pasupathy, Ser Chye Toh, Marcus Eng Hock Ong, and Sean Shao Wei Lam. 2024. "Enhancing Emergency Department Management: A Data-Driven Approach to Detect and Predict Surge Persistence" Healthcare 12, no. 17: 1751. https://doi.org/10.3390/healthcare12171751

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop