Next Article in Journal
The Impact of Cultural Immersion Experience on Identity Transformation Process
Next Article in Special Issue
Facebook Reviews as a Supplemental Tool for Hospital Patient Satisfaction and Its Relationship with Hospital Accreditation in Malaysia
Previous Article in Journal
Study on the Relationship between Drivers’ Personal Characters and Non-Standard Traffic Signs Comprehensibility
Previous Article in Special Issue
A Deep Learning Model for Classification of Endoscopic Gastroesophageal Reflux Disease
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Predicting Survival in Veterans with Follicular Lymphoma Using Structured Electronic Health Record Information and Machine Learning

1
Veritas, Division of Epidemiology, School of Medicine, University of Utah, Salt Lake City, UT 84112, USA
2
George E. Wahlen Veterans Health Administration, Salt Lake City, UT 84148, USA
3
Hematology & Hematologic Malignancies, Huntsman Cancer Institute, Salt Lake City, UT 84112, USA
*
Author to whom correspondence should be addressed.
Int. J. Environ. Res. Public Health 2021, 18(5), 2679; https://doi.org/10.3390/ijerph18052679
Submission received: 7 January 2021 / Revised: 1 March 2021 / Accepted: 2 March 2021 / Published: 7 March 2021
(This article belongs to the Special Issue Disease Prediction, Machine Learning, and Healthcare)

Abstract

:
The most accurate prognostic approach for follicular lymphoma (FL), progression of disease at 24 months (POD24), requires two years’ observation after initiating first-line therapy (L1) to predict outcomes. We applied machine learning to structured electronic health record (EHR) data to predict individual survival at L1 initiation. We grouped 523 observations and 1933 variables from a nationwide cohort of FL patients diagnosed 2006–2014 in the Veterans Health Administration into traditionally used prognostic variables (“curated”), commonly measured labs (“labs”), and International Classification of Diseases diagnostic codes (“ICD”) sets. We compared performance of random survival forests (RSF) vs. traditional Cox model using four datasets: curated, curated + labs, curated + ICD, and curated + ICD + labs, also using Cox on curated + POD24. We evaluated variable importance and partial dependence plots with area under the receiver operating characteristic curve (AUC). RSF with curated + labs performed best, with mean AUC 0.73 (95% CI: 0.71–0.75). It approximated, but did not surpass, Cox with POD24 (mean AUC 0.74 [95% CI: 0.71–0.77]). RSF using EHR data achieved better performance than traditional prognostic variables, setting the foundation for the incorporation of our algorithm into the EHR. It also provides for possible future scenarios in which clinicians could be provided an EHR-based tool which approximates the predictive ability of the most accurate known indicator, using information available 24 months earlier.

1. Introduction

Follicular lymphoma (FL), the most common indolent non-Hodgkin lymphoma [1], accounts for about 20% of non-Hodgkin lymphoma [2,3]. Patients with FL have highly heterogeneous prognoses; some patients experience an indolent course of disease, while others endure a more aggressive disease with a trajectory that can include frequent progression, relapse, and early demise [4,5,6]. FL treatments can be associated with morbidity and, rarely, mortality [7,8]. Thus, patients and clinicians must calibrate therapy choice to the risk posed by FL. Otherwise, treatment risks could outweigh the benefits [4,9]. In order to apply risk-adapted treatment strategies effectively, clinicians must be able to accurately identify high-risk patients early in the course of disease, but the methods currently available for this task have significant drawbacks.
Specific patient-, disease-, and environment-related variables that can serve as prognostic factors in oncology play a pivotal role in understanding disease trajectory, designing clinical trials, making treatment decisions, and providing individual patients with accurate information about their health risks [10]. The most commonly used FL prognostic index is the Follicular Lymphoma International Prognostic Index (FLIPI) [11], which resulted from an international collaboration that collected clinical characteristics and associated outcomes from a large number of lymphoma patients treated with chemotherapy on various clinical trial protocols. The variables included in the final model were age (>60 years), disease stage (III–IV), hemoglobin (<12 g/dL), number of nodal areas (>4), and lactate dehydrogenase (LDH) level (>upper limit of normal). A study validating FLIPI [12] obtained a Harrell’s concordance index (C-index) of 0.66. As with FLIPI, most traditional prognostic indices in oncology have typically been based on human-abstracted variables collected during clinical trials [13,14]. These indices were optimized not only for predictive performance, but also for ease of use by clinicians at the point of care, constraining the number of variables that could be included and the complexity of score calculations. FLIPI’s optimization for ease of recollection and calculation in a pre-electronic health record (EHR) era makes it poorly suited for current clinical practice in the era of near-universal EHR use.
In recent years, disease progression within two years following initiation of first-line (L1) treatment with chemoimmunotherapy (aka progression of disease within 24 months [POD24]) [6,15] has risen to prominence as a simple and relatively accurate way of defining a high-risk group of patients with decreased overall survival. Patients whose response to L1 treatment lasts beyond two years demonstrate survival rates similar to those of the age-matched general population [15]. However, using POD24 to assess risk requires waiting two years after the start of L1 treatment in order to identify these patients; during that time, neither patient nor physician can be sure about the disease prognosis. Furthermore, using POD24 does not support a risk-adapted strategy in L1 treatment, whereas, ideally, patients and clinicians could take into account a patient’s risk status while choosing an L1 therapy.
EHR systems, now widespread, could theoretically automatically extract and calculate prognostic indices and present the resulting risk prediction to practicing physicians. Absent the need to provide for easy recollection and calculation by clinicians, EHR-based indices could instead be optimized for predictive performance and ease of extraction. Such an optimization would favor the use of structured or semi-structured information over information presented in unstructured narratives.
For survival analysis in oncology, the traditional Cox proportional hazard model investigates the relationship between survival time and one or more variables. Although well established and easily interpretable, the Cox model carries key limitations. It must satisfy linearity and proportional hazards assumptions and can be applied only to a restricted number of predictors, thus limiting its ability to account for interactions between variables. Compared with the Cox model, machine learning can better manage large numbers of predictors—even handling cases with more predictors than observations—and can account for both non-linearity and variable interactions [16]. The random survival forest (RSF) method, a flexible, non-parametric ensemble tree machine learning approach extended from Breiman’s [17] random forest method in order to analyze right-censored survival data [18], shows promise as a way to use high-dimensional, structured EHR data to identify factors for predicting patient risk [19]. RSF’s reliance on only the data in seeking a predictive model offers real advantages: it does not require model assumptions, and it can be applied in exploratory investigations, allowing the discovery of unexpected risk factors even with a paucity of prior survival data [18,20]. RSF avoids overfitting [21], while effectively handling outliers [22]. Its out-of-bag prediction also provides reliable inferences for training data, and it can measure relative contributions of different variables (i.e., variable importance) to the survival prediction [20].
Machine learning techniques have been widely used in cancer research [23,24,25]. However, to our knowledge, no prior study has used RSF to predict patients’ prognoses in FL and investigated how structured and semi-structured EHR data could benefit prediction.
In this study, we applied RSF to predict FL patients’ survival and risks using EHR data from a cohort of patients treated within the Veterans Health Administration (VHA). Our findings supported our hypothesis that this method would outperform traditional approaches and achieve predictive performance similar to that of POD24. Our results provide clinicians with insights into the relative importance of traditional and new variables in predicting FL patients’ risk.

2. Materials and Methods

This research was approved by the University of Utah Institutional Review Board (#00083982) and the VA Salt Lake City Human Research Protection Program.

2.1. Cohort Selection

Using data from the Veterans Affairs (VA) Cancer Registry System and pharmacy dispensation records from the VA Corporate Data Warehouse, we identified a nationwide cohort of patients with grade 1–3a, stage II–IV FL diagnosed from 1 January 2006 to 31 December 2014 in the VHA, who received any of three widely used L1 therapies—rituximab combined with cyclophosphamide, doxorubicin, vincristine, and prednisone; rituximab combined with cyclophosphamide, vincristine, and prednisone; or bendamustine combined with rituximab [26]. Table 1 shows patients’ characteristics by treatment. We excluded patients who received maintenance therapy (see Appendix A for rationale), as well as patients without a hematology/oncology visit within 6 months of diagnosis, as these latter patients likely received treatment outside the VHA which we would have limited ability to identify. Patients with a VA Cancer Registry System record of another malignancy prior to FL diagnosis were also excluded. The final cohort included 523 patients. Figure 1 shows the study flow diagram.
The dataset included 523 patients, with 1933 variables. Median follow-up time was 5.03 years, and median survival time 9.32 years. At five years, 150 (28.68%) patients had died, 111 (21.22%) patients had been administratively censored at the end of the study observation period, and 262 (50.10%) patients were alive. Within two years, 99 patients (22.81%) experienced progression of disease (POD24). As in Casulo et al. [15], we excluded from our POD24 analysis 89 patients who were administratively censored (n = 28) or died (n = 61) from causes unrelated to their disease relapse, to the best of our knowledge. Figure 2 illustrates cohort distribution for POD24 and reference groups.

2.2. Feature Extractions

We grouped covariates into three groups:
  • A curated clinical set (“curated”) comprised patient demographics and disease-specific characteristics commonly recognized to be associated with survival, which were available in structured form in the VA Cancer Registry System or Corporate Data Warehouse (Table A1). Patient characteristics included age, sex, and modified Charlson comorbidity index; disease-specific characteristics included stage, grade, and lactate dehydrogenase at L1 initiation. We also included treatment used for L1.
  • The second group (“labs”) included results of 33 lab values typically obtained prior to initiation of L1, extracted from the EHR lab domain. These data included most of the labs (available in 70% or more of patients) in the complete blood count and comprehensive metabolic panels (Table A1). We included medians and ranges of lab results for the period starting three months prior to start of L1 and ending just prior to start of L1. RSF handles missing data itself; for the Cox model, missing data were imputed by random forest imputation algorithm [27] using randomForestSRC R package [28].
  • Finally, a larger group of variables (“ICD”) included any International Classification of Diseases (ICD) diagnostic codes present from one year prior to L1 initiation to three months prior to L1, with information indicating presence or absence of ICD codes as well as how many times each individual code was present during this nine-month period. There were 1841 ICD codes overall.

2.3. Outcomes

Vital status and date of death for patients who died before the end of observation period on 31 December 2018 were obtained from the Corporate Data Warehouse, which aggregates vital status from multiple sources. Survival was measured from first dispensation of L1 to date of death as recorded in Corporate Data Warehouse records. Patients were censored if they were recorded to have another cancer in the VA Cancer Registry System, or at the end of observation period. Five-year overall survival is an established clinical milestone in FL studies [29,30,31] and POD24 predicts a high-risk group with a median survival of five years; we compared Cox and RSF methods’ relative ability to predict five-year survival. We defined as belonging to the high-risk group any patient who died within five years after L1 initiation, with the low-risk group defined as patients who lived beyond the five-year time point.

2.4. Models

The Cox model estimates the hazard function h(t), which gives instantaneous risk of event, by a linear function (i.e., the log risk of failure is a linear combination of covariates [see (2) below]). Given the hazard function, survival function in terms of time, the probability that an observation survives beyond a specific time, can be obtained. The model parameters are estimated through optimization of the cox partial likelihood.
The hazard at time t given covariates x =   ( x 0 ,   x 1 ,   ,   x p ) is defined as
h ( t | x ) =   h 0 ( t ) exp ( j p x j β j )
From (1), we can also obtain
ln ( h ( t | x ) ) =   ln ( h 0 ( t ) ) +   ( j p x j β j ) .
where h 0 ( t ) is the baseline hazard function and does not need to be specified while estimating the parameters β   =   ( β 0 , β 1 ,   ,   β 2 ) T through maximizing the partial likelihood defined as the following formula:
L ( β ) =   f F exp ( β T x j f ) j R exp ( β T x j )  
where F is the set of indices of failures, R is the set of indices of individuals at risk at time t, and j f is the index of failure at time t.
For individual risk prediction in Cox model, the estimated hazard for individual i, with covariate vector x i follows:
h i ^ ( t ) =   h 0 ^ ( t ) exp ( x i β ^   ) ,
where β ^ is the estimated coefficients by maximizing partial likelihood, h i ^ ( t ) is the estimated hazard for individual i and h 0 ^ ( t ) is the estimated baseline hazard.
Similarly, the survival function of individual i at time t is then is calculated with the following formula:
S i ^ ( t ) =   S 0 ^ ( t ) exp ( x i β ^ ) ,
with S 0 ^ ( t ) = exp ( Λ 0 ^ ( t ) ) and Λ 0 ^ ( t ) is the cumulated hazard until time t [32].
However, the Cox model assumes that the data have linear proportional hazard, which assumption, in many cases—especially for high dimensional data with non-linear interactions—is not satisfied. Therefore, a more complicated non-linear model, such as a regularized Cox model or a machine learning model, is needed.
We used a regularized Cox model with lasso approach when the dimensionality of the data increases. The lasso regularization method is a variable selection and shrinkage method in Cox model, where the log partial likelihood is minimized subject to the sum of the absolute values of the parameter bounded by a given constant [33]. The constraint shrinks the coefficients; some are shrunk to zero, thus reducing the estimation variance [33]. We used a five-fold cross-validation approach to obtain the optimal penalty parameter [34].
Among machine learning approaches, RSF is a flexible non-parametric ensemble tree method, extended from Breiman’s [17] random forest method for analyzing right-censored survival data [18]. Random forest is a popular tree-based ensemble method and each tree is built on randomly selected variables [35]. Bootstrap samples of the data are selected to build the trees. Each node of the tree is split using a randomly selected set of variables. The randomization decorrelates the trees and keeps the variance relatively small, since combining the predictions of all these uncorrelated trees built with different samples and subsets of features leads to lower variance [17]. The trees are grown to the full extent, which reduces bias. The forest prediction is based on the majority of votes, with each tree receiving one vote for the observation that is not in the bootstrap sample (out-of-bag observation).
RSF is extended from the random forest method. For the RSF approach,
  • Bootstrap samples of the training data are selected to build the trees. For each bootstrap sample, about 2/3 of the observations are selected and 1/3 are left out.
  • In each bootstrap sample, a survival tree is constructed. Each node, p candidate variables are randomly selected to build the tree. The split of the nodes maximizes the survival difference between daughter nodes; in this study, a log-rank splitting rule is used to determine the split of the nodes [18].
  • The tree is grown to full size under the constraint that there should be at least one event with unique survival times at each terminal node.
  • Survival curves are estimated for the out-of-bag observations and the average survival curves are calculated as the survival curve for each subject. The cumulative hazard functions in terminal nodes are time-dependent. Performance is assessed based on the testing set using the RSF model obtained from the training and parameter tuning process.
The nodes of the trees are split using a survival criterion considering survival time and censoring status information [18]. The cumulative hazard function is an ensemble estimate of all the trees.
To train each model, patients were randomly split into a training set (80%) for model training with five-fold cross-validation approach and a testing set (20%) for performance evaluation. Optimal parameters for RSF were selected using grid search [36] with the cross-validation process. The details of parameter tuning process are provided in Appendix B. To calculate average performance and 95% confidence intervals (CI), we repeated this process 30 times. The mean AUC is the mean of the 30 C-indices and AUCs. The baseline characteristic summarization, as well as variable importance and partial dependence plots, are shown based on the split with performance closest to the average of 30 splits [37].
For each approach, we fit models with four different sets of input variables with increasing granularity: curated set, curated+ labs set, curated + ICD set, and curated + labs + ICD set. Only the curated set satisfied proportional hazard and linearity assumptions, i.e., had survival curves for two strata with hazard functions are roughly proportional over time for each categorical covariate, and the relationship between the ln(hazard of death) and each of the continuous variables in the model were close to linear after adjusting other covariates. No interactions were fit or tested. A traditional Cox model is unable to model the curated + ICD and curated + labs + ICD sets due to the high dimensionality, so a “regularized” Cox model with lasso approach [34] was used for these two sets. We also used the regularized Cox model for the curated + labs set where the Cox model assumptions were violated.

2.5. Use of RSF to Predict High or Low Risk

We set the survival function S(t) as the probability that a patient would survive beyond five years after the start of L1. We used predictSurvProb function in the pec R package [38] for the survival probability calculation [39]. When S(t) < an optimal cut-off point, where the optimal cut-off point was identified using Youden’s index (J), defined as J = max{sensitivity + specificity − 1} [40], we classified a patient as belonging to the high-risk group; otherwise, we categorized the patient as being at low risk.
We used training data to “grow” the RSF, in which each patient in the test set was passed down the trees. Individual survival functions for each patient were extracted by evaluating the obtained cumulative hazard functions over a five-year time interval. High- and low-risk groups were then determined by the individual survival functions with the optimal cut-off point.

2.6. Model Performance Measures

We evaluated the predictive performance of the models using area under the receiver operating characteristic curve (AUC), which is also referred to as a ‘C-statistic’ or C-index [41], a rank-order statistic for assessing predictions against true outcomes (defined as the ratio of the concordant pairs to the total comparable pairs [42]), served as a standard performance measure for model assessment in survival analysis. AUC provides information about how well a model can distinguish between classifications, where a higher AUC reflects a better model.

2.7. Model Interpretation

We used variable importance and partial dependence plots to explain the best-performing models, measuring a given variable’s importance by increase or decrease in prediction error after dropping the predictor variable at RSF nodes. Friedman 2001 [43] proposed the partial dependence plot to interpret machine learning algorithms; the plot illustrates the average change of the predicted value when the variable changes over its marginal distribution. The partial dependence plot serves as a useful tool in understanding the relationship between predictors and predictions in various predictive settings [44].

3. Results

We used R [45] to perform the analyses. Model performances on four datasets are shown in Table 2 and Table 3. Using the curated clinical variable set, the baseline Cox model yielded a mean AUC of 0.64 (95% CI: 0.61–0.67), while the RSF model achieved a mean AUC of 0.67 (95% CI: 0.65–0.70). The addition of lab results had little effect on the performance of the Cox model, with a mean AUC of 0.62 (95% CI: 0.60–0.64). Using a regularized Cox model for curated + labs set boosted AUC to 0.71 (95% CI: 0.69–0.73). With more variables available (ICD codes), a regularized Cox model achieved comparable performance to RSF. The variables retained in the regularized Cox model are reported in Appendix C. The RSF model showed marked improvement in prediction performance on the curated + labs set, with a mean AUC of 0.73 (95% CI: 0.71–0.75). The curated + labs set with RSF model turned out to be the best-performing combination of dataset and approach. We found no significant difference between RSF performance on the curated + labs set (95% CI for AUC: 0.71–0.75) and curated + ICD + labs set (95% CI for AUC: 0.65–0.73). RSF performance on curated + ICD + labs showed much higher variance than performance on the curated + labs dataset. The best-performing combination of approach and dataset—the RSF model with curated + labs set—achieved a performance that was close to but unable to beat that of a Cox model based on POD24 (mean AUC 0.74 [95% CI: 0.71–0.77]).
Receiver operating characteristic (ROC) curves are presented in Figure 3 for the Cox model applied to curated set, RSF model applied to the curated + labs set, and Cox model applied to the curated + POD24 dataset. The ROC curves show the tradeoff between sensitivities and specificities for each model.
Figure 4 provides variable importance plots for the RSF model applied to the curated + labs set. The nine most influential predictors were examined with partial dependence plots to explore their relationship with predicted survival probabilities. These nine most important variables, in order of decreasing influence, were albumin, age at L1 initiation, erythrocytes, urea nitrogen, bilirubin, protein, aspartate aminotransferase, alanine aminotransferase, and calcium. Figure 5 provides partial dependence plots for these nine variables. The vertical axis shows the predicted survival probability, and the horizontal axis indicates change in predictor values. The partial dependence plots illustrate that the predicted survival probabilities decrease with age above 60 years, dropping faster with age over 70 years. Similarly, survival decreased as albumin and erythrocytes decreased and urea nitrogen, bilirubin, protein, aspartate aminotransferase, alanine aminotransferase, and calcium increased beyond their respective normal ranges, indicating worsening renal (urea nitrogen), liver (albumin, bilirubin, alanine aminotransferase), and marrow (erythrocyte, hemoglobin) functions.
Our best-performing model and data set, RSF with curated + labs set, identified patients at elevated risk of death within five years (hazard ratio: 4.39 [95% CI: 2.11–9.14]) (Table 4) and with greater likelihood to show factors traditionally considered indicative of high risk (hemoglobin, lactate dehydrogenase), even though these variables were not among the top variables used by the RSF algorithm to predict risk. Lactate dehydrogenase and hemoglobin were the only traditional indicators found to be statistically significant risk factors (p-value < 0.05). Table 5 provides baseline characteristics of predicted high- and low-risk groups. Table 4 shows how the high-/low-risk classifications for RSF at the beginning of L1 compared to the classifications for POD24 two years later. The estimated hazard ratio for high risk is slightly (but not significantly) higher with POD24 compared to RSF.

4. Discussion

In this study, we sought to compare the predictive performance of an RSF model versus a Cox model on datasets including traditional curated variables (i.e., patient demographics and disease-specific characteristics) along with higher-dimensional data (i.e., lab values and/or ICD diagnostic codes) from the EHRs of a population of patients with FL. We also compared performance of the RSF model against the most accurate known indicator of FL risk, POD24. Note that this comparison is inherently biased towards POD24 since our model considered information available at the time of L1 treatment initiation, while POD24, based on progression-free survival in the 24 months following L1 treatment initiation, incorporates information about the disease course a full two years after our model’s prediction is made.
We demonstrate that a limited set of EHR-extractable variables can provide similar performance to those generated from clinical trials [12]. While these limited EHR-derived variables would not necessarily provide a predictive gain over established clinical indices, they offer the advantage of being automatically extractable and computable, such that a modern EHR could simply provide them to the clinician without the need for the clinician to recall them, calculate them manually, and then include them in the chart. These prognostic factors would also be identifiable in the EHR from discrete structured data without needing to deploy information extraction or natural language processing approaches on unstructured data.
Our findings show that a broader set of data extracted from the EHR can improve prediction performance. However, not all data achieved such an effect. ICD codes added relatively little to prediction performance. On the other hand, lab values, available in EHRs but not other claims-based healthcare databases, substantially improved prediction. These results suggest that studies of the utility of EHR data in predicting clinical endpoints should emphasize EHR information that reflects the biology of patient and disease and deemphasize administrative data.
In addition, while our model included disease-specific variables commonly considered a reflection of disease risk (e.g., stage and grade), the variable importance analysis identified none of these variables as being among the most useful variables for prediction. Instead, the top variables reflected age, along with kidney, liver, and marrow function. This finding suggests that change in these variables may serve as the best indicator of disease risk, host fitness, and the host’s ability to tolerate the stresses introduced by the disease and its associated treatments. The correlation between age and survival probability in our model corroborates findings that a model including age in three categories (≤60 years old/61–70 years old/>70 years old) improves the predictive performance of FLIPI [12].
The best-performing combination of model and data we found, RSF applied to the curated + labs dataset, could not outperform POD24, although we found no significant difference between the AUCs of the two models. We report a relative risk for the predicted high- vs. low-risk groups using our best model of 4.39, lower than the previously reported risk ratio of POD24 vs. reference group of 6.44 [15]. However, our model offers the distinct advantage of using information available at the time of L1 treatment initiation, without the need to wait for two years while patients are observed for progression. An earlier prognosis would not only provide more information to patients about the potential impact of their disease on survival, it would also lead to significant differences in their management. Patients with high risk could be monitored more closely and would likely receive imaging surveillance more promptly if they developed symptoms that could indicate a recurrence. In addition, identification of high risk at the time of L1 initiation allows for adapting L1 to the disease risk, either by choosing a more aggressive treatment regimen and/or by adopting maintenance treatment (additional immunotherapy) following the completion of L1 treatment.

Limitations

There are several limitations in this study. First, the majority of VHA patients are male, limiting generalizability of our results to a subset of the population of patients with FL. However, we implemented our methods in a nationwide cohort using structured EHR data commonly available in any EHR system, making it feasible to widely adopt our methods to automatically predict individual patient risk at the beginning of L1. Secondly, we limited our analysis to variables available in structured format; we did not include unstructured data in radiology reports, which might improve prognostic ability. With advances in precision medicine, genetic profiling and deep sequencing of tumors could likely become a widely adopted standard of care. Such information would probably improve prognosis as well. Lastly, we did not include deep learning approaches in our preliminary comparison of machine learning methods due to their computational cost and the need for specialized hardware.

5. Conclusions

RSF using EHR data achieved better performance than traditional prognostic variables. Our method sets the foundation for the incorporation of our algorithm into the EHR. It also allows for possible future scenarios to provide clinicians with an EHR-based tool which approximates the predictive ability of the most accurate known indicator, using information available 24 months earlier in the disease trajectory.

Author Contributions

Conceptualization, C.Y., A.S.H. and C.L.; methodology, A.S.H., B.C.S. and C.L.; software, V.P.; validation, V.P., K.M.R., D.M. and J.H.; formal analysis, C.L.; investigation, B.C.S., J.H. and H.-C.C.; resources, B.C.S., K.M.R., C.Y. and H.-C.C.; data curation, V.P., Z.B. and D.M., writing—original draft preparation, A.S.H., C.L. and C.Y.; writing—review and editing, C.Y., K.M.R. and H.-C.C.; visualization, V.P., K.M.R. and C.L.; supervision, A.S.H., B.C.S. and Z.B.; project administration, K.M.R.; funding acquisition, not applicable. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki, and approved by the University of Utah Institutional Review Board #00083982 and the VA Salt Lake City Human Research Protection Program (protocol #00083982 approved on 31 July 2019).

Informed Consent Statement

Patient consent was waived due to the fact that the study used existing data not linked to identifiable individuals, involved no more than minimal risk to subjects, did not adversely affect subjects’ rights and welfare, and could not practicably be performed without a waiver of consent.

Data Availability Statement

The data that support the findings of this study are available through the Veterans Health Administration. Restrictions applying to the availability of these data were approved under the University of Utah Institutional Review Board #00083982 and the VA Salt Lake City Human Research Protection Program. Please contact the corresponding author for data inquiries.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Clinical Rationale for Inclusion and Exclusion Criteria

Patients with Grade 3B and Stage I disease are treated differently: Stage I patients mostly receive radiation, while Grade 3B is thought to be a different clinical and biological entity whose treatments are more aligned with diffuse large B-cell lymphoma than FL. We excluded patients who receive single-agent rituximab since this therapy is less efficacious than the other three treatments and is typically given either in patients who do not meet treatment indication criteria or are too fragile to receive a more aggressive standard treatment. We also excluded maintenance rituximab since this is shown to prolong progression-free survival and would not allow for POD24 calculation, the latter reported mainly in patients who did not receive maintenance rituximab. Prior malignancy is a common exclusion criteria as a prior malignancy recurrence could lead to increased mortality independent from FL.

Appendix B. RSF Parameter Tuning

In the training process, we used a grid search in order to find the optimal parameters;, i.e., an extensive grid search is performed with 5-fold cross-validation of each split. The parameters we tuned are: average number of observations in terminal nodes (nodesize), number of trees in the forest (ntree), and the maximum number of splits for continuous variables (nsplit). The tuning process is based on the C-index of cross-validation test results.
The parameters we specified using R code with the grid search are:
  • Nodesize: seq(100, 1000, by = 50)
  • Ntree: seq(1, ncol(train_data), length.out = 100)
  • Nsplit: c(1:9, seq(10, 100, by = 5))

Appendix C. Tables and Figure

Table A1. Variables included in “labs” and “curated” datasets.
Table A1. Variables included in “labs” and “curated” datasets.
Labs Set
Alanine aminotransferase
Albumin
Alkaline phosphatase
Basophils
Basophils per 100 leukocytes
Bilirubin
Calcium
Carbon dioxide
Chloride
Creatinine
Eosinophils
Eosinophils per 100 leukocytes
Erythrocyte distribution width
Erythrocyte mean corpuscular hemoglobin
Erythrocyte mean corpuscular hemoglobin concentration
Erythrocyte mean corpuscular volume
Erythrocytes
Glomerular filtration rate per 1.73 square meters predicted
Glucose
Hematocrit
Lymphocytes
Lymphocytes per 100 leukocytes
Monocytes
Monocytes per 100 leukocytes
Neutrophils
Platelet mean volume
Platelets
Potassium
Protein
Sodium
Urea nitrogen
Curated Set
Sex
Race
Disease stage
Disease grade
Modified Charlson comorbidity index prior to first-line treatment
First-line treatment regimen
Age at first-line treatment initiation
Hemoglobin at first-line treatment initiation
Lactate dehydrogenase
Region of residence
Days from diagnosis to starting first-line treatment
Table A2. Cox model performance (AUC) using “curated” dataset with two-way interactions and restricted cubic spline.
Table A2. Cox model performance (AUC) using “curated” dataset with two-way interactions and restricted cubic spline.
Model Curated (95% CI)
Cox0.59 (0.56–0.61)
AUC: area under the receiver operating characteristic curve; CI: confidence interval.
Table A3. Cox model performance (AUC) on different datasets using multiple imputation instead of a random forest imputation algorithm.
Table A3. Cox model performance (AUC) on different datasets using multiple imputation instead of a random forest imputation algorithm.
Model Curated (95% CI)Curated + Labs (95% CI)POD24 (95% CI)
Cox0.64 (0.61–0.66)0.64 (0.62–0.67)0.75 (0.73–0.78)
AUC: area under the receiver operating characteristic curve; CI: confidence interval; POD24: progression of disease within 24 months of starting first-line treatment.
Table A4. Cox model performance (AUC) with two-way interactions using a stepwise variable selection approach with different datasets.
Table A4. Cox model performance (AUC) with two-way interactions using a stepwise variable selection approach with different datasets.
ModelCurated + Labs (95% CI)Curated + ICD (95% CI)Curated + ICD + Labs (95% CI)
Cox0.64 (0.62–0.67)0.57 (0.55–0.59)0.59 (0.55–0.63)
AUC: area under the receiver operating characteristic curve; CI: confidence interval; ICD: International Classification of Diseases diagnostic codes.
Table A5. Number of times (out of 30 splits) that different variables were retained in the regularized Cox model after lasso regularization, with different datasets.
Table A5. Number of times (out of 30 splits) that different variables were retained in the regularized Cox model after lasso regularization, with different datasets.
LabsICDLabs + ICD
VariableTimes RetainedVariableTimes RetainedVariableTimes Retained
Urea nitrogen30Age at L130Age at L130
Age at L1 initiation30Hemoglobin *30Albumin *30
Albumin30424.119Urea nitrogen *29
Erythrocytes30v45.8119Erythrocytes *27
Chloride16362.0218v45.8124
Protein11414.0114424.123
Calcium10250.8013v81.114
Lymphocytes6782.313305.0313
Carbon dioxide5v58.6113v58.6113
Aspartate aminotransferase4v68.8911216.512
Potassium4305.0310414.0112
Northeast US residence4443.97v68.8912
LDH3366.154Potassium *11
Sodium3523.424285.810
Alanine aminotransferase2v58.833362.0210
Alkaline phosphatase1v76.433250.809
Basophils1173.92782.39
Bilirubin1244.92362.056
Eosinophils1427.312443.96
Hematocrit1721.32Protein *6
Neutrophils1v58.662366.155
Platelets1250.001427.315
RCVP as L11295.321v15.824
Male sex1300.021v49.894
427.891340.3
428.01371.53
715.001785.23
721.21v43.13
722.01295.322
724.31362.012
785.21726.732
998.831v16.12
LDH*1v58.832
v15.821v76.432
v43.11244.91
v65.191266.21
v72.311362.041
v72.61366.161
369.91
427.91
523.421
586.1
702.191
713.51
721.21
721.31
726.321
786.91
793.991
998.831
Aspartate aminotransferase *1
Chloride *1
E878.21
Glucose *1
v43.31
v65.191
72.311
* denotes imputed result. ICD: International Classification of Diseases diagnostic codes. L1: first-line treatment; LDH: lactate dehydrogenase; RCVP: rituximab, cyclophosphamide, vincristine, and prednisone.
Figure A1. Calibration plots for Cox applied to curated set (AUC [95% CI]: 0.64 [0.61–0.67]), RSF models applied to the curated + labs set (AUC [95% CI]: 0.73 [0.71–0.75]), and Cox applied to curated + POD24 set (AUC [95% CI]: 0.74 [0.71–0.77]).
Figure A1. Calibration plots for Cox applied to curated set (AUC [95% CI]: 0.64 [0.61–0.67]), RSF models applied to the curated + labs set (AUC [95% CI]: 0.73 [0.71–0.75]), and Cox applied to curated + POD24 set (AUC [95% CI]: 0.74 [0.71–0.77]).
Ijerph 18 02679 g0a1

References

  1. Jemal, A.; Bray, F.; Center, M.M.; Ferlay, J.; Ward, E.; Forman, D. Global cancer statistics. CA Cancer J. Clin. 2011, 61, 69–90. [Google Scholar] [CrossRef] [Green Version]
  2. Ekström-Smedby, K. Epidemiology and etiology of non-Hodgkin lymphoma—A review. Acta Oncol. 2006, 45, 258–271. [Google Scholar] [CrossRef] [Green Version]
  3. Monga, N.; Nastoupil, L.; Garside, J.; Quigley, J.; Hudson, M.; O’Donovan, P.; Parisi, L.; Tapprich, C.; Thieblemont, C. Burden of illness of follicular lymphoma and marginal zone lymphoma. Ann. Hematol. 2019, 98, 175–183. [Google Scholar] [CrossRef]
  4. Kahl, B.S. Follicular lymphoma: Are we ready for a risk-adapted approach? Hematol. Am. Soc. Hematol. Educ. Program 2017, 2017, 358–364. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Teras, L.R.; DeSantis, C.E.; Cerhan, J.R.; Morton, L.M.; Jemal, A.; Flowers, C.R. 2016 US lymphoid malignancy statistics by World Health Organization subtypes. CA Cancer J. Clin. 2016, 66, 443–459. [Google Scholar] [CrossRef] [PubMed]
  6. Maurer, M.J.; Bachy, E.; Ghesquiéres, H.; Ansell, S.M.; Nowakowski, G.S.; Thompson, C.A.; Inwards, D.J.; Allmer, C.; Chassagne-Clément, C.; Nicolas-Virelizier, E.; et al. Early event status informs subsequent outcome in newly diagnosed follicular lymphoma. Am. J. Hematol. 2016, 91, 1096–1101. [Google Scholar] [CrossRef] [Green Version]
  7. Rummel, M.J.; Niederly, N.; Maschmeyer, G.; Banat, G.A.; von Grünhagen, U.; Losem, C.; Kofahl-Krause, D.; Heil, G.; Welslau, M.; Balser, C.; et al. Bendamustine plus rituximab versus CHOP plus rituximab as first-line treament for patients with indolent and mantle-cell lymphomas: An open-label, multicentre, randomised, phase 3 non-inferiority trial. Lancet 2013, 381, 1203–1210. [Google Scholar] [CrossRef]
  8. Flinn, I.W.; van der Jagt, R.; Kahl, B.; Wood, P.; Hawkins, T.; MacDonald, D.; Simpson, D.; Kolibaba, K.; Issa, S.; Chang, J.; et al. First-line treatment of patients with indolent non-Hodgkin lymphoma or mantle-cell lymphoma with bendamustine plus rituximab versus R-CHOP or R-CVP: Results of the BRIGHT 5-year follow-up study. J. Clin. Oncol. 2019, 37, 984–991. [Google Scholar] [CrossRef] [PubMed]
  9. Casulo, C.; Nastoupil, L.; Fowler, N.H.; Friedberg, J.W.; Flowers, C.R. Unmet needs in the first-line treatment of follicular lymphoma. Ann. Oncol. 2017, 28, 2094–2106. [Google Scholar] [CrossRef] [PubMed]
  10. Halabi, S.; Owzar, K. The importance of identifying and validating prognostic factors in oncology. Semin. Oncol. 2010, 37, e9–e18. [Google Scholar] [CrossRef] [Green Version]
  11. Solal-Céligny, P.; Roy, P.; Colombat, P.; White, J.; Armitage, J.O.; Arranz-Saez, R.; Au, W.Y.; Bellei, M.; Brice, P.; Caballero, D.; et al. Follicular Lymphoma International Prognostic Index. Blood 2004, 104, 1258–1265. [Google Scholar] [CrossRef] [Green Version]
  12. Van de Schans, S.A.M.; Steyerberg, E.W.; Nijziel, M.R.; Creemers, G.-J.; Janssen-Heijnen, M.L.; van Spronsen, D.J. Validation, revision and extension of the Follicular Lymphoma International Prognostic Index (FLIPI) in a population-based setting. Ann. Oncol. 2009, 20, 1697–1702. [Google Scholar] [CrossRef] [PubMed]
  13. Haybittle, J.L.; Blamey, R.W.; Elston, C.W.; Johnson, J.; Doyle, P.J.; Campbell, F.C.; Nicholson, R.I.; Griffiths, K. A prognostic index in primary breast cancer. Br. J. Cancer 1982, 45, 361–366. [Google Scholar] [CrossRef] [Green Version]
  14. Petersen, V.C.; Baxter, K.J.; Love, S.B.; Shepherd, N.A. Identification of objective pathological prognostic determinants and models of prognosis in Dukes‘ B colon cancer. Gut 2002, 51, 65–69. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Casulo, C.; Byrtek, M.; Dawson, K.L.; Zhou, X.; Farber, C.M.; Flowers, C.R.; Hainsworth, J.D.; Maurer, M.J.; Cerhan, J.R.; Link, B.K.; et al. Early relapse of follicular lymphoma after rituximab plus cyclophosphamide, doxorubicin, vincristine, and prednisone defines patients at high risk for death: An analysis from the National LymphoCare Study. J. Clin. Oncol. 2015, 33, 2516–2522. [Google Scholar] [CrossRef] [PubMed]
  16. Obermeyer, Z.; Emanuel, E.J. Predicting the future-big data, machine learning, and clinical medicine. NEJM 2016, 375, 1216–1219. [Google Scholar] [CrossRef] [Green Version]
  17. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  18. Ishwaran, H.; Kogalur, U.B.; Blackstone, E.H.; Lauer, M.S. Random survival forests. Ann. Appl. Stat. 2008, 2, 841–860. [Google Scholar] [CrossRef]
  19. Wang, H.; Li, G. A selective review on random survival forests for high dimensional data. Quant. Biosci. 2017, 36, 85–96. [Google Scholar] [CrossRef]
  20. Ishwaran, H.; Lu, M. Standard errors and confidence intervals for variable importance in random forest regression, classification, and survival. Stat. Med. 2018, 38, 558–582. [Google Scholar] [CrossRef]
  21. Dietrich, S.; Floegel, A.; Troll, M.; Kühn, T.; Rathmann, W.; Peters, A.; Sookthai, D.; von Bergen, M.; Kaaks, R.; Adamski, J.; et al. Random Survival Forest in practice: A method for modelling complex metabolomics data in time to event analysis. Int. J. Epidemiol. 2016, 45, 1406–1420. [Google Scholar] [CrossRef] [Green Version]
  22. LeBlanc, M. Regression trees. In Encyclopedia of Environmentrics; El-Shaarawi, A.H., Piegorsch, W.W., Zhang, H.H., Eds.; Wiley Online Library: Hoboken, NJ, USA, 2006. [Google Scholar] [CrossRef]
  23. Gupta, S.; Tran, T.; Luo, W.; Phung, D.; Kennedy, R.L.; Broad, A.; Campbell, D.; Kipp, D.; Singh, M.; Khasraw, M.; et al. Machine-learning prediction of cancer survival: A retrospective study using electronic administrative records and a cancer registry. BMJ Open 2014, 4, e004007. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Elfiky, A.A.; Pany, M.J.; Parikh, R.B.; Obermeyer, Z. Development and application of a machine learning approach to assess short-term mortality risk among patients with cancer starting chemotherapy. JAMA Netw. Open 2018, 1, e180926. [Google Scholar] [CrossRef] [Green Version]
  25. Parikh, R.B.; Manz, C.; Chivers, C.; Regli, S.H.; Braun, J.; Draugelis, M.E.; Schuchter, L.M.; Shulman, L.N.; Navathe, A.S.; Patel, M.S.; et al. Machine learning approaches to predict 6-month mortality among patients with cancer. JAMA Netw. Open 2019, 2, e1915997. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Halwani, A.S.; Rasmussen, K.M.; Patil, V.; Morreall, D.; Li, C.; Yong, C.; Burningham, Z.; Dawson, K.; Masaquel, A.; Henderson, K.; et al. Maintenance rituximab in Veterans with follicular lymphoma. Cancer Med. 2020, 9, 7537–7547. [Google Scholar] [CrossRef] [PubMed]
  27. Tang, F.; Ishwaran, H. Random forest missing data algorithms. Stat. Anal. Data Min. 2017, 10, 363–377. [Google Scholar] [CrossRef]
  28. Ishwaran, H.; Kogalur, U.B.; Fast Unified Random Forests for Survival, Regression, and Classification (RF-SRC). CRAN R-Project. Available online: https://cran.r-project.org/web/packages/randomForestSRC/randomForestSRC.pdf (accessed on 10 November 2020).
  29. Monfardini, S.; Banfi, A.; Bonadonna, G.; Rilke, F.; Milani, F.; Valagussa, P.; Lattuada, A. Improved five year survival after combined radiotherapy-chemotherapy for state I-II non-Hodgkin‘s lymphoma. Int. J. Radiat. Oncol. Biol. Phys. 1980, 6, 125–134. [Google Scholar] [CrossRef]
  30. Guadagnolo, B.A.; Li, S.; Neuberg, D.; Ng, A.; Hua, L.; Silver, B.; Stevenson, M.A.; Mauch, P. Long-term outcome and mortality trends in early-stage, Grade 1-2 follicular lymphoma treated with radiation therapy. Int. J. Radiat. Oncol. Biol. Phys. 2006, 64, 928–934. [Google Scholar] [CrossRef]
  31. Carreras, J.; Lopez-Guillermo, A.; Fox, B.C.; Colomo, L.; Martinez, A.; Roncador, G.; Montserrat, E.; Campo, E.; Banham, A.H. High numbers of tumor-infiltrating FOXP3-positive regulatory T cells are associated with improved overall survival in follicular lymphoma. Blood 2006, 108, 2957–2964. [Google Scholar] [CrossRef]
  32. Tibshirani, R. The lasso method for variable selection in the Cox model. Stat. Med. 1997, 16, 385–395. [Google Scholar] [CrossRef] [Green Version]
  33. Stackexchange. Available online: https://stats.stackexchange.com/questions/36015/prediction-in-cox-regression (accessed on 20 August 2020).
  34. Tay, K.; Simon, N.; Friedman, J.; Hastie, T.; Tibshirani, R.; Naraimhan, B. Regularized Cox Regression. Available online: https://cran.r-project.org/web/packages/glmnet/vignettes/Coxnet.pdf (accessed on 11 February 2021).
  35. Cutler, A.; Cutler, D.R.; Stevens, J.R. Random forests. In Ensemble Machine Learning; Zhang, C., Ma, Y., Eds.; Springer: Boston, MA, USA, 2012; pp. 157–175. [Google Scholar]
  36. Shi, Y.; Graves, J.A.; Garbett, S.P.; Zhou, Z.; Marathi, R.; Wang, X.; Harrell, F.E.; Lasko, T.A.; Denny, J.C.; Roden, D.M.; et al. A decision-theoretic approach to panel-based, preemptive genotyping. MDM Policy Prac. 2019, 4. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Konerman, M.A.; Beste, L.A.; Van, T.; Liu, B.; Zhang, X.; Zhu, J.; Saini, S.D.; Su, G.L.; Nallamothu, B.K.; Ioannou, G.N.; et al. Machine learning models to predict disease progression among Veterans with hepatitis C virus. PLoS ONE 2019, 14, e0208141. [Google Scholar] [CrossRef] [PubMed]
  38. Gerds, T.A. Prediction Error Curves for Survival Models, R package pec version 1.1.5; CRAN: Vienna, Austria, 2009. [Google Scholar]
  39. Mogensen, U.B.; Ishwaran, H.; Gerds, T.A. Evaluating random forests for survival analysis using prediction error curves. J. Stat. Softw. 2012, 50, 1–23. [Google Scholar] [CrossRef]
  40. Uno, H.; Cai, T.; Pencina, M.J.; D’Agostino, R.B.; Wei, L.J. On the C-statistics for evaluating overall adequacy of risk prediction procedures with censored survival data. Stat. Med. 2011, 30, 1105–1117. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  41. Ruopp, M.D.; Perkins, N.J.; Whitcomb, B.W.; Schisterman, E.F. Youden Index and optimal cut-point estimated from observations affected by a lower limit of detection. Biomed. J. 2008, 50, 419–430. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  42. Harrell, F.E. “with Contributions from Many Others”. Harrell Miscellaneous. R Package Version 4.1-1. Available online: https://CRAN.R-project.org/package=Hmisc (accessed on 11 February 2021).
  43. Friedman, J.H. Greedy function approximation: A gradient boosting machine. Ann. Stat. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
  44. Goldstein, A.; Kapelner, A.; Bleich, J.; Pitkin, E. Peeking inside the black box: Visualizing statistical learning with plots of individual conditional expectation. J. Comp. Graph. Stat. 2015, 45, 44–65. [Google Scholar] [CrossRef]
  45. R Core Team. R: A language and environment for statistical computing. R Foundation for Statistical Computing. Vienna, Austria. Available online: https://www.R-project.org (accessed on 11 February 2021).
Figure 1. Study cohort attrition. 1L: first-line treatment; BR: bendamustine and rituximab; MR: maintenance rituximab; RCHOP: rituximab, cyclophosphamide, doxorubicin, vincristine, and prednisone; RCVP: rituximab, cyclophosphamide, vincristine, and prednisone; VHA: Veterans Health Administration.
Figure 1. Study cohort attrition. 1L: first-line treatment; BR: bendamustine and rituximab; MR: maintenance rituximab; RCHOP: rituximab, cyclophosphamide, doxorubicin, vincristine, and prednisone; RCVP: rituximab, cyclophosphamide, vincristine, and prednisone; VHA: Veterans Health Administration.
Ijerph 18 02679 g001
Figure 2. Participant selection diagram for Cox model with curated + POD24 dataset. L1: first-line treatment; POD24: progression of disease within 24 months of initiation of first-line treatment.
Figure 2. Participant selection diagram for Cox model with curated + POD24 dataset. L1: first-line treatment; POD24: progression of disease within 24 months of initiation of first-line treatment.
Ijerph 18 02679 g002
Figure 3. ROC plots for Cox applied to curated dataset, RSF applied to curated + labs set, and Cox applied to curated + POD24 set. ROC: receiver operating characteristic; RSF: random survival forest; POD24: progression of disease within 24 months of initiation of first-line treatment.
Figure 3. ROC plots for Cox applied to curated dataset, RSF applied to curated + labs set, and Cox applied to curated + POD24 set. ROC: receiver operating characteristic; RSF: random survival forest; POD24: progression of disease within 24 months of initiation of first-line treatment.
Ijerph 18 02679 g003
Figure 4. Variable importance plot for RSF analysis of the curated + labs dataset. RSF: random survival forest; L1: time of initiation of first-line therapy; RBC: erythrocytes: BUN: urea nitrogen; AST: aspartate aminotransferase; ALT: alanine aminotransferase; vimp: variable importance.
Figure 4. Variable importance plot for RSF analysis of the curated + labs dataset. RSF: random survival forest; L1: time of initiation of first-line therapy; RBC: erythrocytes: BUN: urea nitrogen; AST: aspartate aminotransferase; ALT: alanine aminotransferase; vimp: variable importance.
Ijerph 18 02679 g004
Figure 5. Partial-dependence plot for the top nine variables in the RSF model applied to curated + labs dataset. Vertical axis shows predicted five-year survival probability. RSF: random survival forest; L1: time of initiation of first-line therapy; RBC: erythrocytes: BUN: urea nitrogen; AST: aspartate aminotransferase; ALT: alanine aminotransferase.
Figure 5. Partial-dependence plot for the top nine variables in the RSF model applied to curated + labs dataset. Vertical axis shows predicted five-year survival probability. RSF: random survival forest; L1: time of initiation of first-line therapy; RBC: erythrocytes: BUN: urea nitrogen; AST: aspartate aminotransferase; ALT: alanine aminotransferase.
Ijerph 18 02679 g005
Table 1. Baseline characteristics of patients by first-line treatment regimen.
Table 1. Baseline characteristics of patients by first-line treatment regimen.
BRRCHOPRCVPp-Value
N120235168
Sex = male * (%)113 (94.2)220 (93.6)165 (98.2)0.085
Race (%)0.177
Hispanic3 (2.5)10 (4.3)3 (1.8)
Non-Hispanic Black6 (5.0)26 (11.1)9 (5.4)
Non-Hispanic White109 (90.8)194 (82.6)153 (91.1)
Other2 (1.7)5 (2.1)3 (1.8)
Disease stage (%)0.254
II12 (10.0)42 (17.9)20 (11.9)
III54 (45.0)97 (41.3)78 (46.4)
IV54 (45.0)96 (40.9)70 (41.7)
Disease grade (%)<0.001
138 (31.7)55 (23.4)70 (41.7)
1–211 (9.2) 7 (3.0)11 (6.5)
258 (48.3)76 (32.3)69 (41.1)
38 (6.7)63 (26.8)12 (7.1)
3a5 (4.2)34 (14.5)6 (3.6)
Region of residence (%)0.110
Midwest33 (27.5)60 (25.5)38 (22.6)
Northwest17 (14.2)26 (11.1)36 (21.4)
South46 (38.3)86 (36.6)53 (31.5)
West24 (20.0)63 (26.8)41 (24.4)
Pre-L1 CCI (mean [SD])2.36 (2.56)2.51 (2.58)2.03 (2.33)0.161
Age > 60 years at L1 (%)85 (70.8)149 (63.4)117 (69.6)0.259
Hemoglobin at L1 < 12 g/dL (%)32 (26.7)76 (32.3)52 (31.0)0.544
LDH at L1 > upper limit of normal39 (32.5)91 (38.7)48 (28.6)0.097
Days from diagnosis to L1 (mean [SD])227.47 (321.72)116.53 (314.30)168.35 (328.04)0.008
* The high percentage of male patients reflects the demographics of the Veterans Health Administration’s patient population. BR: bendamustine and rituximab; CCI: modified Charlson comorbidity index; L1: first-line; LDH: lactate dehydrogenase; RCHOP: rituximab, cyclophosphamide, doxorubicin, vincristine, and prednisone; RCVP: rituximab, cyclophosphamide, vincristine, and prednisone; SD: standard deviation.
Table 2. Cox and RSF model performance with different predictor datasets.
Table 2. Cox and RSF model performance with different predictor datasets.
Model (AUC)Curated (95% CI)Curated + Labs (95% CI)Curated + ICD (95% CI)Curated + ICD + Labs (95% CI)
Cox (regularized Cox denoted by *)0.64 (0.61–0.67)0.61 (0.59–0.64)
* 0.71 (0.69–0.73)
* 0.69 (0.67–0.71)* 0.73 (0.70–0.75)
RSF0.68 (0.65–0.70)0.73 (0.71–0.75)0.63 (0.61–0.65)0.71 (0.63–0.79)
* Normal Cox model cannot handle high-dimensional data such as ICD codes; we provide performance measures of both normal and regularized Cox models for the curated + labs dataset for comparison. AUC: area under the receiver operating characteristic curve; CI: confidence interval; ICD: International Classification of Diseases diagnostic codes; RSF: random survival forest.
Table 3. Cox model performance with curated set and POD24.
Table 3. Cox model performance with curated set and POD24.
Model (AUC)Curated + POD24 (95% CI)
Cox0.74 (0.71–0.77)
AUC: area under the receiver operating characteristic curve; CI: confidence interval; POD24: progression of disease within 24 months of starting first-line treatment.
Table 4. Relative risk for the predicted high- and low-risk groups.
Table 4. Relative risk for the predicted high- and low-risk groups.
Risk GroupNHazard Ratio (95% CI)5-Year Overall Survival (95% CI)
RSF Model
Low6210.83 (0.73–0.94)
High434.39 (2.11–9.14)0.44 (0.30–0.63)
POD24 Model
Low6110.87 (0.78–0.98)
High255.55 (3.27–9.35)0.41 (0.26–0.68)
CI: confidence interval; POD24: progression of disease within 24 months of starting first-line treatment; RSF: random survival forest.
Table 5. Baseline characteristics of predicted high- and low-risk groups.
Table 5. Baseline characteristics of predicted high- and low-risk groups.
Low-RiskHigh-Riskp-Value
N6243
Sex = male * (%)59 (95.2)40 (93.0)0.971
Race (%)0.520
Hispanic1 (1.6)1 (2.3)
Non-Hispanic Black5 (8.1)1 (2.3)
Non-Hispanic White52 (83.9)40 (93.0)
Other2 (3.2)0
Unknown2 (3.2)1 (2.3)
Disease stage (%)0.503
II8 (12.9)5 (11.6)
III26 (41.9)13 (30.2)
IV24 (38.7)23 (53.5)
Unknown4 (6.5)2 (4.7)
Disease grade (%)0.208
113 (21.0)12 (27.9)
1–23 (4.8) 0
218 (29.0)7 (16.3)
38 (12.9)7 (16.3)
3a8 (12.9)3 (7.0)
Unknown12 (19.4)14 (32.6)
L1 treatment regimen (%)0.483
BR13 (21.0)6 (14.0)
RCHOP25 (40.3)22 (51.2)
RCVP24 (38.7)15 (34.9)
Region of residence (%)0.789
Midwest11 (17.7)10 (23.3)
Northwest6 (9.7)6 (14.0)
South17 (27.4)8 (18.6)
West13 (21.0)8 (18.6)
Unknown15 (24.2)11 (25.6)
Pre-L1 CCI (mean [SD])2.37 (2.42)2.93 (2.81)0.284
Days from diagnosis to L1 (mean [SD])135.31 (310.77)102.49 (200.15)0.543
Age > 60 years at L1 (%)37 (59.7)32 (79.1)0.061
Hemoglobin at L1 (%)<0.001
<12 g/dL3 (4.8)26 (60.5)
≥12 g/dL58 (93.5)17 (39.5)
Unknown1 (1.6)0
LDH at L1 (%)0.009
≤Upper limit of normal47 (75.8)21 (48.8)
>Upper limit of normal9 (14.5)17 (39.5)
Unknown6 (9.7)5 (11.6)
* The high percentage of male patients reflects the demographics of the Veterans Health Administration’s patient population. BR: bendamustine and rituximab; CCI: modified Charlson comorbidity index; L1: first-line; LDH: lactate dehydrogenase; RCHOP: rituximab, cyclophosphamide, doxorubicin, vincristine, and prednisone; RCVP: rituximab, cyclophosphamide, vincristine, and prednisone; SD: standard deviation.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, C.; Patil, V.; Rasmussen, K.M.; Yong, C.; Chien, H.-C.; Morreall, D.; Humpherys, J.; Sauer, B.C.; Burningham, Z.; Halwani, A.S. Predicting Survival in Veterans with Follicular Lymphoma Using Structured Electronic Health Record Information and Machine Learning. Int. J. Environ. Res. Public Health 2021, 18, 2679. https://doi.org/10.3390/ijerph18052679

AMA Style

Li C, Patil V, Rasmussen KM, Yong C, Chien H-C, Morreall D, Humpherys J, Sauer BC, Burningham Z, Halwani AS. Predicting Survival in Veterans with Follicular Lymphoma Using Structured Electronic Health Record Information and Machine Learning. International Journal of Environmental Research and Public Health. 2021; 18(5):2679. https://doi.org/10.3390/ijerph18052679

Chicago/Turabian Style

Li, Chunyang, Vikas Patil, Kelli M. Rasmussen, Christina Yong, Hsu-Chih Chien, Debbie Morreall, Jeffrey Humpherys, Brian C. Sauer, Zachary Burningham, and Ahmad S. Halwani. 2021. "Predicting Survival in Veterans with Follicular Lymphoma Using Structured Electronic Health Record Information and Machine Learning" International Journal of Environmental Research and Public Health 18, no. 5: 2679. https://doi.org/10.3390/ijerph18052679

APA Style

Li, C., Patil, V., Rasmussen, K. M., Yong, C., Chien, H. -C., Morreall, D., Humpherys, J., Sauer, B. C., Burningham, Z., & Halwani, A. S. (2021). Predicting Survival in Veterans with Follicular Lymphoma Using Structured Electronic Health Record Information and Machine Learning. International Journal of Environmental Research and Public Health, 18(5), 2679. https://doi.org/10.3390/ijerph18052679

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop