Next Article in Journal
Effects of Weight Status and Related Metabolic Disorders on Fertility-Sparing Treatment Outcomes in Endometrial Atypical Hyperplasia and Endometrial Cancer: A Retrospective Study
Next Article in Special Issue
The Impact of the COVID Pandemic on the Incidence of Presentations with Cancer-Related Symptoms in Primary Care
Previous Article in Journal
Infectious Complications of Targeted Therapies in Children with Leukemias and Lymphomas
Previous Article in Special Issue
Urine CA125 and HE4 for the Triage of Symptomatic Women with Suspected Endometrial Cancer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Machine Learning for Risk Prediction of Oesophago-Gastric Cancer in Primary Care: Comparison with Existing Risk-Assessment Tools

1
School of Computing, University of Leeds, Leeds LS2 9JT, UK
2
Leeds Institute for Data Analytics, University of Leeds, Leeds LS2 9NL, UK
3
The Alan Turing Institute, London NW1 2DB, UK
4
Department of Health and Community Sciences, University of Exeter, Exeter EX1 2LU, UK
5
Academic Unit of Primary Medical Care, University of Sheffield, Sheffield S10 2TN, UK
*
Author to whom correspondence should be addressed.
Cancers 2022, 14(20), 5023; https://doi.org/10.3390/cancers14205023
Submission received: 14 September 2022 / Revised: 12 October 2022 / Accepted: 12 October 2022 / Published: 14 October 2022
(This article belongs to the Special Issue Cancer Detection in Primary Care)

Abstract

:

Simple Summary

Oesophago-gastric cancer is one of the commonest cancers worldwide, yet it can be particularly difficult to diagnose given that initial symptoms are often non-specific and routine screening is not available. Cancer risk-assessment tools, which calculate cancer risk based on symptoms and other risk factors present in the primary care record, can aid decisions on referrals for cancer investigations, facilitating earlier diagnosis. Diagnosing common cancers earlier could help improve survival rates. Using UK primary care electronic health record data, we compared five different machine learning techniques for probabilistic classification of cancer patients against a current widely used UK primary care cancer risk-assessment tool. The machine learning algorithms outperformed the current risk-assessment tool, with a higher overall accuracy and an ability to reasonably identify 11–25% more cancer patients. We conclude that machine-learning-based risk-assessment tools could help better identify suitable patients for further investigation and support earlier diagnosis.

Abstract

Oesophago-gastric cancer is difficult to diagnose in the early stages given its typical non-specific initial manifestation. We hypothesise that machine learning can improve upon the diagnostic performance of current primary care risk-assessment tools by using advanced analytical techniques to exploit the wealth of evidence available in the electronic health record. We used a primary care electronic health record dataset derived from the UK General Practice Research Database (7471 cases; 32,877 controls) and developed five probabilistic machine learning classifiers: Support Vector Machine, Random Forest, Logistic Regression, Naïve Bayes, and Extreme Gradient Boosted Decision Trees. Features included basic demographics, symptoms, and lab test results. The Logistic Regression, Support Vector Machine, and Extreme Gradient Boosted Decision Tree models achieved the highest performance in terms of accuracy and AUROC (0.89 accuracy, 0.87 AUROC), outperforming a current UK oesophago-gastric cancer risk-assessment tool (ogRAT). Machine learning also identified more cancer patients than the ogRAT: 11.0% more with little to no effect on false positives, or up to 25.0% more with a slight increase in false positives (for Logistic Regression, results threshold-dependent). Feature contribution estimates and individual prediction explanations indicated clinical relevance. We conclude that machine learning could improve primary care cancer risk-assessment tools, potentially helping clinicians to identify additional cancer cases earlier. This could, in turn, improve survival outcomes.

1. Introduction

Oesophago-gastric cancer is one of the commonest cancers worldwide [1], with approximately 15,800 cases diagnosed annually in the UK (statistics from 2019) [2,3]. It is also associated with some of the poorest survival rates, in part attributable to late-stage diagnosis [4]. International comparisons highlight particularly inferior cancer survival rates in the UK amongst economically developed countries [5]. When comparing cancer survival rates across stage at diagnosis, five-year net survival for oesophago-gastric cancer in the UK is up to 65.0% when diagnosed at the earliest stage, compared with between 16.0–25.0% at stage 3, and as low as 2.0% at stage 4 [6].
Diagnosing oesophago-gastric cancer can be difficult and mainly depends upon presentation with symptoms to primary care, given that there is no routine screening besides the monitoring of patients with Barrett’s oesophagus [1]. Often, patients present with non-specific symptoms indiscernible from symptoms of more minor issues, with the two main alarm symptoms—dysphagia and iron-deficiency anaemia—only accounting for a small amount of initial presentations [7]. Diagnostic intervals are generally much greater for patients with no alarm symptoms, usually those with early-stage and localised disease [8].
The latest long-term plan for the National Health Service (NHS) in England, launched in 2019, aspires to diagnose 75.0% of cancers in England at stage 2 or earlier by 2028. Taking steps towards improving diagnostic performance for oesophago-gastric cancer would help meet this objective [9]. Furthermore, expediting diagnosis of symptomatic patients may have benefits for both mortality and morbidity [10,11].
Clinical decision-support tools can assist primary care clinicians in diagnosing cancer earlier [12]. Being able to more accurately detect warning signs and symptoms of oesophago-gastric cancer to strategically refer high-risk patients for endoscopic tests could see great benefits [7]. Additionally, adopting a more ‘personalised’ approach, as opposed to a population-based approach, to cancer risk assessment and staging could be a valuable approach in helping to raise survival rates [13].
Current UK primary care interventions to accelerate cancer diagnosis include cancer Risk-Assessment Tools (RATs): a series of 2 × 2 matrices which display risk scores associated with individual and pairwise combinations of diagnostic factors based on positive predictive values [14,15]. These were derived from the CAncer Prevention in ExetER studies (CAPER) [16] published in 2009: a series of case-control studies designed to quantify the risk of certain cancers using diagnostic risk factors available in the primary care record, such as symptoms and lab test results. Another predominant diagnostic tool is QCancer® [17], first created in 2012. The QCancer® tool consists of two risk-prediction algorithms (one for men and one for women) which generate individual risk scores based on symptomatic, baseline, and lifestyle-related diagnostic factors across 13 different types of common cancer [18,19]. These studies have been instrumental in influencing current UK guidelines on recognising and referring suspected cancer cases in primary care [20]. Efforts have been made to digitise these tools; however, the tools themselves remain in relative infancy and have certain limitations: the risk scores generated by the oesophago-gastric RAT (ogRAT), for example, take into account a maximum of two symptoms simultaneously, and only give actionable risk scores for patients over 55 [7]. Additionally, sets of diagnostic factors used in QCancer® are inexhaustive [18,19] and the oesophago-gastric QCancer® model may have a tendency to over-predict high risk scores [21]. At the time of writing, the Electronic RIsk assessment for CAncer (ERICA) trial is underway to assess the clinical usefulness of digitised RATs integrated into major primary care clinical systems [22].
Given the heterogeneous nature of cancer, using simple scoring techniques, relying on known diagnostic factors, or placing more emphasis on alarm symptoms are likely to be insufficient strategies to assess risk comprehensively [8,23]. Machine Learning (ML) may be able to elevate diagnostic performance, exploiting large-scale electronic health record (EHR) data with diverse feature sets to automatically determine more individually tailored risk scores [24]. However, attempts so far lack clinical validation, and major challenges—such as improving robustness, interpretability, and clinical relevance—abound [24]. The vast majority of efforts employing ML for cancer diagnosis have concentrated on improving diagnostic accuracy for medical imaging [25,26], with few studies performed on primary care data [26]. This is despite evidence suggesting that some of the greatest benefits could be seen by reducing diagnostic delays at the point of primary care [27].
To the extent of the authors’ knowledge, based on our investigation of the literature, to date there has been no direct comparison of ML-based risk-assessment tools for cancer diagnosis in primary care against currently implemented tools in the UK [26,28]. Furthermore, there has been a distinct lack of focus on upper gastrointestinal cancers—including oesophago-gastric cancer—in ML-based clinical risk-prediction tools [28]. The aim of this study, therefore, was to assess suitability of an ML-based approach, by benchmarking the performance of probabilistic, supervised, ML classifiers against the UK Cancer ogRAT for risk prediction of oesophago-gastric cancer, using data derived from primary care EHRs.
The study seeks to address the question as to whether ML models could help with the strategic referral of patients for further investigation and the earlier detection of a greater number of cancer cases. In terms of model performance, this translates to the extent to which recall values might be increased whilst roughly preserving precision values, thereby detecting a larger number of cases yet minimising the number of unnecessary procedures.
We focus on traditional ML approaches, which can perform comparatively well in the context of machine learning and deep learning [24,29] and are therefore favoured for the purpose of this study in the interest of minimising computational complexity and supporting interpretability [29]. Explainability and interpretability should be emphasised from the outset to allow greater insight into clinical relevance of models and garner clinician trust and understanding [30,31]. The UK Department of Health and Social Care recommends that such qualities are built into data-driven health technologies by design in order to maximise transparency [31].

2. Materials and Methods

2.1. Data Overview

The dataset employed in this study was a primary care EHR dataset, derived from a 2013 case-control study investigating risk factors of oesophago-gastric cancer, the results of which were used to develop the ogRAT model [7]. The data originated from the General Practice Research Database (GPRD) in the UK, a high-quality database containing anonymised EHR data on approximately 4.8 million patients from around 600 general practices across the UK [32,33].
The resulting dataset comprises information on 40,348 patients (7471 cases and 32,877 controls), selected from the database using a list of 42 tumour diagnostic codes, with up to five (mean of four) matched controls (matched according to age, sex, and practice) for each case [7]. Inclusion criteria consisted of a diagnosis during or after the year 2000, age of ≥40 years, and availability of consultations in the year prior to diagnosis of the case [7]. All medical evidence was collected in the year prior to diagnosis. Variables include basic demographic information, ICPC-coded symptoms, and various lab test results (See Supplementary Material Table S1 for further information). Repeat presentations of symptoms (up to one repeat, where available) were also recorded. The mean age of patients was 73 years (±standard deviation of 10), with a 64.2/35.8% split of male and female patients respectively. Oesophageal cancer diagnoses constituted 65.3% of the cases, with 34.7% being gastric cancer cases (which roughly corresponds to the national incidence rate [34]). See Table 1 for the full list of patient characteristics.
Regarding pre-processing, much of this had already been performed prior to acquisition of the dataset, including the generalisation of potentially identifiable information such as age and date of birth [7]. Symptoms had been one-hot encoded, and all lab test results with continuous values were converted to binary (normal/abnormal), except for cholesterol (which was included as both a categorical and numerical variable) [7]. Additional pre-processing performed specifically for this study involved min-max normalisation of continuous variables and conversion of binary categorical (string) variables into dummy variables (0 or 1).

2.2. Ethical Approval

This secondary data analysis study was performed in line with the original ethical approval remit. This was approved by the Independent Scientific Advisory Committee (ISAC). The dataset does not contain any personally identifiable information.

2.3. Feature Selection

Initial feature selection was based on known diagnostic factors for oesophago-gastric cancer [35,36,37] and basic demographic variables (age, gender), to be included as baseline. Additional features were selected following a series of experiments with different subsets: initially constraining the set to just those present in the ogRAT, and then testing combinations including other symptoms. The final selection of features corresponded to the best-performing model on the merit of model performance and clinical suitability of predictions. Model-based feature selection fine-tuned this set further.

2.4. Predictive Models

Five probabilistic ML classifiers were chosen: Random Forest, Support Vector Machine (with probability calibration using Platt scaling), Logistic Regression, Naïve Bayes, and Extreme Gradient Boosted Decision Trees. These are commonly used ML techniques which have been shown to be successful for other clinical risk-prediction tasks [24,29].
Model hyperparameters were fine-tuned using fivefold grid search cross-validation (see Supplementary Material Table S2). The problem was formulated as a binary classification task between ‘cancer’ (case) and ‘no cancer’ (control), with a corresponding probability of cancer given.

2.5. Evaluation

The dataset was split into a train/test split of 75.0%/25.0%, respectively, allocated randomly. Fivefold cross-validation on the training set was used for model selection with the final performance evaluated on the unseen test set.
Classification performance was evaluated on the basis of commonly used ML performance metrics including overall accuracy, precision, recall, F1 score, and Area Under the Receiver Operating Characteristic curve (AUROC). Particular attention was paid to the trade-off between precision and recall as the aim was to build models that optimise the proportion of cancer cases detected (maximising recall) whilst simultaneously minimising the proportion of unnecessary referrals (maximising precision). Using these performance measures, the ML models’ predictions were compared against the ogRAT’s predictions on the same test set. Clinical justifiability was assessed using prediction explanations and feature importance estimation.
Data was processed using the popular software development toolkit Jupyter Notebooks in the programming language Python (v3.7.10) using the pandas (v1.2.4), scikit-learn (v0.24.2), xgboost (v1.5.0), and numpy (v1.20.2) libraries for model building, training, testing, and performance evaluation; matplotlib (v3.3.4) to generate plots; and the Local Interpretable Model-Agnostic Explanations (LIME) package for model explanations [38].

3. Results

3.1. Model Performance

All five ML approaches classified oesophago-gastric cancer cases with a high level of accuracy, surpassing that of the ogRAT. All models achieved an accuracy of 0.89: the accuracy of the ogRAT when employed at the typical 2.0% threshold came in slightly lower at 0.87. The highest-performing models in terms of AUROC were Linear Support Vector Machine, Logistic Regression, and XGBoost (0.87 AUROC). Table 2 shows the performance of all models in comparison with the ogRAT when employed at the 1.0% or 2.0% risk threshold for cancer (the typical thresholds at which the ogRAT was designed to be used). See Supplementary Material Table S3 for results of the full performance evaluation for all models across a range of classification thresholds. The classification threshold is flexible; a couple of examples are given here for each model (Table 2) to demonstrate thresholds at which performances are comparable to the ogRAT—in terms of precision and recall—and overall accuracy is maximised. The aim was to improve upon overall accuracy and surpass the current recall rate offered by the ogRAT model—which equates to detecting a larger proportion of cases—whilst forfeiting as little as possible in terms of precision, thereby limiting false positives. The recall rate offered by the current risk-assessment strategy, the ogRAT, stands at 33.0% when adopting the typical 2.0% risk threshold (Table 2).
For example, employing Logistic Regression at the 0.425 threshold improves upon the accuracy and AUROC score of the ogRAT, with a dramatic increase in recall of cancer patients (potentially recalling, in absolute terms, from 17.0% up to 25.0% more cancer patients depending on the chosen risk threshold for the ogRAT) and little cost to precision. Even comparing the performance with the most lenient (1.0%) risk boundary for the ogRAT results in the successful detection of 17.0% more cancer patients with little cost to precision (a 5.0% reduction) and a slight improvement in overall accuracy (Table 2). Adjusting the Logistic Regression risk threshold to 0.75 to roughly preserve the higher false positive rate of the ogRAT would still see the identification of up to 11.0% more cases (Table 2).
Table 3 displays the recall offered by two of the best-performing ML models (Linear Support Vector Machine and Logistic Regression) when exactly preserving the precision rate of the ogRAT model. Using these precision values as fixed reference points, it is clear that the ML models could offer a greater recall (between 8.0–11.0% higher) at the same precision as the ogRAT. This translates to the potential to detect a greater number of cancer cases without the need for any extra investigations in the cancer-free population.

3.2. Feature Contribution Estimates

The final feature set included all 16 previously acknowledged risk factors from the Stapley et al. (2013) study [7], with the addition of age (categorised as ≤55 and >55), and serum cholesterol level. Figure 1 shows the estimated importance rankings corresponding to the Support Vector Machine (linear) model, for this feature set. From the feature importance estimates (Supplementary Material Figures S1–S6), it appears that all models assigned a similar ranking to the features, with dysphagia (considered to be the classical ‘alarm’ symptom for OG cancer) accruing the highest risk contribution across all models, closely followed by epigastric pain, weight loss, dyspepsia, and reflux, in varying order. The remaining features were assigned similar importance across all models, with one notable exception being cholesterol level: for the Random Forest model, high cholesterol corresponded to increased cancer risk; however, Support Vector Machine and Logistic Regression associated low cholesterol levels with elevated OG cancer risk. Ostensibly, this seems inaccurate, but appears to be consistent with the medical literature which suggests that abnormally low levels of low-density lipoprotein cholesterol—and in turn, serum cholesterol—are in fact associated with increased OG cancer risk [39,40].

4. Discussion

We developed and compared five ML-based probabilistic classifiers for risk prediction of oesophago-gastric cancer using data derived from the primary care EHR consisting of basic demographic information, symptomatic evidence, and various lab test results. The performance of these models was benchmarked against an oesophago-gastric cancer risk-assessment tool—the ‘ogRAT’—currently employed in primary care clinics across the UK.
Our models selected a combination of previously established clinical features in oesophago-gastric cancer risk-prediction models [7,20] and other less commonly used diagnostic factors [39,40] to deliver predictions with a high overall accuracy (0.89 for Linear Support Vector Machine). The best-performing techniques were Linear Support Vector Machine, Logistic Regression, and XGBoost, but all techniques displayed similar performances. Models performed relatively consistently across demographic groups (according to age and sex) and were slightly better performing for oesophageal cases than gastric cases, as was also observed in the ogRAT model (see Supplementary Material Table S4 for results stratified according to demographic groups and cancer sites).
Feature importance also appeared similar across techniques, with dysphagia consistently established as the biggest contributor to risk (see Figure 1 for the feature importance ranking estimate for Linear Support Vector Machine, and Supplementary Material Figures S2–S6 for all other models). Overall, models showed good calibration to the test dataset (Supplementary Material Figure S7), with the Naïve Bayes model being notably the least well calibrated. It will be more enlightening to assess aspects such as calibration and clinical utility during further validation phases on external datasets [42], in order to verify that these characteristics are maintained in general.
Such models could be highly influential in meeting current challenges faced in risk prediction for common cancers in primary care. One major challenge is being able to strategically select a greater number of patients for referral for cancer investigation, achieving an optimal balance between detection and conversion rates [7]. ML-based models could help achieve this: the current recall rate at the point of primary care is low, with only 31.0% of oesophago-gastric cancer cases identified by stage 2 and 34.0% of all cases missed completely by primary care clinicians [43]. The current recommended risk threshold for urgent investigation of suspected oesophago-gastric cancer is 3.0% [20]; however, there have been suggestions to potentially liberalise this threshold to 2.0%, or even 1.0%, in order to detect more cases earlier [7,44]. For oesophago-gastric cancer, increasing the referral rate to allow for the detection of a greater number of cases needs to be done in such a way which minimises any resulting increase in false positive rates [44]. From the results of this study, using the ogRAT model at the recommended risk threshold of 3.0% would detect around 31.0% of cases from symptomatic presentation to primary care (Supplementary Material Table S3). Employing the ogRAT at the more liberal threshold of 2.0% would encourage the referral of around 33.0% of cases. ML could increase this further: detecting, in this study, around 58.0% of these cases (with a slight cost to precision), 44.0% if roughly preserving the false positive rate (a decrease of 1.0% in precision), or 41.0% if exactly preserving the false positive rate. Achieving an overall increase in recall of 10.0%, for example, would theoretically equate to the earlier detection of over 1500 additional oesophago-gastric cancer cases in the UK annually [2,3].
Investigation into individual prediction explanations using LIME [38] highlighted cases which ML models would flag as high risk of cancer whereas the ogRAT would not, revealing that ML picks up on certain cases with vaguer symptoms which would otherwise go undetected using RATs (see Supplementary Material Figure S8 for one such example). Results also showed that ML could improve upon areas in which RATs are prone to over-generalisation: for example, the ogRAT would not flag any cases in under-55s, who are deemed collectively to have a low risk [7].
Therefore, integrating these ML-based classifiers into primary care cancer risk-prediction tools could assist in prioritisation of suspected cancer patients for referral. Prompting early testing could expedite diagnosis and treatment, and could improve outcomes in terms of both mortality and morbidity [45]. Such a tool would, for example, take the form of a digitised risk-prediction tool embedded in the primary care clinician’s workflow. Based on symptomatic and lab test information input with clinical codes such as SNOMED or Read codes, the ML algorithm would calculate a cancer risk score for each patient and subsequently flag patients with a high risk score according to a predetermined risk threshold. Implementing the tools in this way does risk bias related to coding inconsistencies and unreported symptoms, for example, and future work should seek to address the mitigation of such biases. Determining an optimal risk threshold is not a trivial task and requires a clinical utility assessment to quantify the trade-off between detecting more cases and minimising false positives [42,46]. The extent to which an increase in recall is worth a slight cost to precision is currently unclear. Risk thresholds are easily adjustable depending on the clinical context and should be initiated with a view to maximising clinical benefits (for example, in terms of costs and test harms) [42,46].
Regarding current progress with implementation, cancer risk-assessment tools are not currently widely used in primary care and there are significant hurdles to increasing their adoption [47]. As of 2019, cancer risk tools were available to approximately a third of all primary care practices in the UK, with 18.5% of practices having access to these in electronic form [47]. However, they are currently underused: only 16.7% of practices were likely to use any form of cancer risk-assessment tool [47]. Likely explanations for this profound underuse include an aversion to using them due to time constraints during consultations and ‘prompt fatigue’, whereby clinicians become overburdened with on-screen alerts [48]. Another reason is an overall lack of awareness about these tools [48,49]. This is despite the fact that many clinicians find them very helpful, and the tools have the potential to assist them in cases with vaguer presentations, influence their decisions on referral, and serve as a reminder of the possibility of cancer [49]. Furthermore, they have been shown to have an educational effect [12,48], with one study finding that the use of cancer risk-assessment tools gradually improved primary care clinicians’ decision-making over time [12]. This suggests that incorporating new electronic cancer decision-support tools into primary care clinical systems could prove very valuable provided that they are designed and implemented in such a way which supports clinicians’ work processes, and that awareness and education surrounding such tools is prioritised. Finding optimal ways of effectively integrating such tools remains challenging [48].

Limitations

Availability of data posed certain limitations. Firstly, although the dataset includes all the key known diagnostic factors for oesophago-gastric cancer, it does not provide a comprehensive set of all potential diagnostic factors, omitting symptomatic factors such as collapse or aetiological factors such as gastritis or obesity, as highlighted in other pieces of medical literature [23,50]. Furthermore, other types of data are not included, such as lifestyle data, family history, or previous surgical procedures, all of which could contribute to oesophago-gastric cancer risk [23,51]. Additionally, the coarse granularity of the majority of the feature-variable values restricts the precision of predictions. Limited availability of demographic variables is also an issue. In order to avoid discriminatory biases moving forward, variables such as socioeconomic deprivation and racial characteristics should also be taken into consideration [52,53]. Including other characteristics of symptomatic presentations, such as duration or severity of symptoms, could also improve predictive performance [54].
Limitations regarding performance evaluation were also experienced. Firstly, comorbidities and other cancers were not accounted for and the models’ only task was to differentiate between the presence and absence of oesophago-gastric cancer. Quantifying how much earlier these tools might be able to detect cases using temporal and/or staging information will be an essential next step. Assessing performance in comparison with other frontrunners for oesophago-gastric risk assessment, such as QCancer® [55], will also be important. Naturally, any clinical benefits of such models would be highly dependent on model–clinician interaction, which forthcoming implementation trials such as the ERICA trial [22] may shed some light on. It is important to note that implementation phases should emphasise patient safety, adopting a systems approach to risk, safety, and accident avoidance, particularly since current regulations of health informatics solutions are likely insufficient [56].

5. Conclusions

In conclusion, these results favour the use of an ML-based approach for primary care risk prediction for oesophago-gastric cancer. These risk-prediction models could assist GPs in diagnostic decision-making and highlight potential cancer cases which might otherwise be missed. Such techniques could potentially be extended to other common—and possibly uncommon—cancers. These would be particularly useful for those cancers typically expressed by non-specific symptoms and with a lack of available routine screening. External validation, comparison against other existing cancer risk-prediction tools, and incorporation of a more diverse feature set involving other indicators of health are necessary moving forward.
Assessing the performance of models in combination with clinicians will be imperative to gauge practicality and usefulness in a real-world scenario. Future iterations of AI-enhanced cancer risk-prediction models could involve dynamic versions which update in real time as new information is accumulated, or sequential models exploiting temporal data and accounting for time intervals between symptoms. Implementation of usable decision-support tools could be extended to include algorithms which better reflect uncertainty, dynamically adapt to new information, and incorporate additional functionality such as user prompts for additional information to improve the quality of predictions.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/cancers14205023/s1, Table S1. Feature variables. N.B.: Where a binary value has been used for symptoms and lab test results, 0 indicates normal and 1 indicates abnormal; Table S2. Hyperparameter tuning strategy. Includes list of hyperparameters searched during grid-search cross-validation for fine-tuning models and the corresponding results of each search. 5-fold cross-validation was used for all models. The scoring metric used to select optimal values was mean accuracy; Table S3. Performance for all machine learning based probabilistic classifiers, across a range of thresholds, in comparison with oesophago-gastric cancer risk assessment tool (ogRAT) for prediction of oesophago-gastric cancer incidence, on test dataset. For machine learning models, the classification threshold range is given between 0.3 and 0.8 (in increments of 0.1) to represent best performance and a trade-off between precision and recall which is comparable to that of the ogRAT. ogRAT performance is displayed at the 0.01, 0.02, and 0.03 risk thresholds which are the risk thresholds realistically considered in practice when using the ogRAT. Table S4. Model performances stratified across different patient groups according to demographics (sex, age group) and cancer site, demonstrated for a selection of some of the best-performing models (linear support vector machine and logistic regression) in comparison to the oesophago-gastric cancer Risk-Assessment Tool (ogRAT). N.B: Since the ogRAT only gives risk scores for over 55s, recall is 0 for the under 55 age group, and precision cannot be calculated. Figure S1. Feature contribution estimation for Support Vector Machine (Linear kernel). Feature contribution values approximated using model coefficients. Figure S2. Feature contribution estimation for Support Vector Machine (Radial Basis Function kernel). Feature contribution values approximated using permutation feature importance (i.e., the mean relative decrease in the model accuracy score when a single feature value is randomly shuffled). Figure S3. Feature contribution estimation for Logistic Regression. Feature contribution values approximated using model coefficients. Figure S4. Feature contribution estimation for Naïve Bayes (Bernoulli). Feature contribution values determined using permutation feature importance (i.e., the mean relative decrease in the model accuracy score when a single feature value is randomly shuffled). Figure S5. Feature contribution estimation for Random Forest. Feature contribution values correspond to mean decrease in impurity (Gini). Figure S6. Feature contribution estimation for Extreme Gradient Boosted Decision Trees. Feature contribution values correspond to mean decrease in impurity (Gini). Figure S7. Calibration plot demonstrating the fraction of observed positives in the test dataset across the range of predicted probabilities, and the Brier score, for all models. SVM—Support Vector Machine. RBF—Radial Basis Function. XGBoost—eXtreme Gradient Boosted decision trees. Figure S8. Example of an explanation for an individual prediction, demonstrating a cancer case with vaguer symptoms to which ML-based tools would assign a high risk score, whereas the current oesophago-gastric cancer Risk Assessment Tool (ogRAT) would not. Explanation generated using the Local Interpretable Model-Agnostic Explanations package [38]. Model: Linear Support Vector Machine. The code used in this project can be found at: https://github.com/emmalucybriggs/oesophago-gastric-cancer-risk-prediction-ml (accessed on 16 August 2022).

Author Contributions

E.B., C.D.M., M.d.K., O.J. and R.D.N. contributed to the conceptualisation of the project. E.B. prepared the initial draft of the manuscript. W.H. was responsible for the data curation. E.B. performed the data analysis and machine learning. C.D.M., M.d.K., O.J. and R.D.N. supervised the execution of the project. All authors provided critical comments and edited the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research was financially supported by the UK Research and Innovation Engineering and Physical Sciences Research Council (grant number EP/S024336/1/). The study was also partly funded by the National Institute for Health Research (NIHR) Yorkshire and Humber Patient Safety Translational Research Centre (NIHR Yorkshire and Humber PSTRC). The research was also supported by The Phoenix Partnership (TPP) Ltd., and Macmillan Cancer Support.

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki, and approved by the Independent Scientific Advisory Committee for MHRA database research (ISAC) (protocol number 09_110).

Informed Consent Statement

Not applicable (patient informed consent was not required as the dataset consists of unidentifiable patient information collected from primary care records, with freedom to opt out).

Data Availability Statement

The data are not publicly available due to restricted access.

Acknowledgments

This work was supported by the UKRI Centre for Doctoral Training Programme in Artificial Intelligence for Medical Diagnosis and Care at the University of Leeds (grant number EP/S024336/1/). This research is linked to the CanTest Collaborative which is funded by Cancer Research UK [C8640/A23385], of which R.D.N. and W.H. are members.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results. The views expressed in this article are those of the author(s) and not necessarily those of the NHS, the NIHR, UKRI EPSRC, TPP, Cancer Research UK, Macmillan Cancer Support, or the Department of Health and Social Care.

References

  1. Allum, W.; Lordick, F.; Alsina, M.; Andritsch, E.; Ba-Ssalamah, A.; Beishon, M.; Braga, M.; Caballero, C.; Carneiro, F.; Cassinello, F.; et al. ECCO essential requirements for quality cancer care: Oesophageal and gastric cancer. Crit. Rev. Oncol. Hematol. 2018, 122, 179–193. [Google Scholar] [CrossRef]
  2. Cancer Research UK. UK Oesophageal Cancer Statistics. London, UK. 2019. Available online: https://www.cancerresearchuk.org/health-professional/cancer-statistics/statistics-by-cancer-type/oesophageal-cancer (accessed on 1 December 2021).
  3. Cancer Research UK. UK Stomach Cancer Statistics. London, UK. 2019. Available online: https://www.cancerresearchuk.org/health-professional/cancer-statistics/statistics-by-cancer-type/stomach-cancer (accessed on 1 December 2021).
  4. Swann, R.; McPhail, S.; Witt, J.; Shand, B.; Abel, G.A.; Hiom, S.; Rashbass, J.; Lyratzopoulos, G.; Rubin, G.; The National Cancer Diagnosis Audit Steering Group. Diagnosing cancer in primary care: Results from the National Cancer Diagnosis Audit. Br. J. Gen. Pract. 2018, 68, e63–e72. [Google Scholar] [CrossRef]
  5. Arnold, M.; Rutherford, M.J.; Bardot, A.; Ferlay, J.; Andersson, T.M.L.; Myklebust, T.Å.; Tervonen, H.; Thursfield, V.; Ransom, D.; Shack, L.; et al. Progress in cancer survival, mortality, and incidence in seven high-income countries 1995–2014 (ICBP SURVMARK-2): A population-based study. Lancet Oncol. 2019, 20, 1493–1505. [Google Scholar] [CrossRef] [Green Version]
  6. Office for National Statistics (ONS). Cancer Survival by Stage at Diagnosis for England. 2019. Available online: https://www.ons.gov.uk/peoplepopulationandcommunity/healthandsocialcare/conditionsanddiseases/datasets/cancersurvivalratescancersurvivalinenglandadultsdiagnosed (accessed on 5 July 2021).
  7. Stapley, S.; Peters, T.J.; Neal, R.D.; Rose, P.W.; Walter, F.M.; Hamilton, W. The risk of oesophago-gastric cancer in symptomatic patients in primary care: A large case–control study using electronic records. Br. J. Cancer 2013, 108, 25–31. [Google Scholar] [CrossRef] [Green Version]
  8. Bowrey, D.J.; Griffin, S.M.; Wayman, J.; Karat, D.; Hayes, N.; Raimes, S.A. Use of alarm symptoms to select dyspeptics for endoscopy causes patients with curable esophagogastric cancer to be overlooked. Surg. Endosc. 2006, 20, 1725–1728. [Google Scholar] [CrossRef] [PubMed]
  9. NHS. NHS Long-Term Plan NHS. 2019. Available online: https://www.longtermplan.nhs.uk/publication/nhs-long-term-plan/ (accessed on 15 July 2021).
  10. Neal, R.D.; Tharmanathan, P.; France, B.; Din, N.U.; Cotton, S.; Fallon-Ferguson, J.; Hamilton, W.; Hendry, A.; Hendry, M.; Lewis, R.; et al. Is increased time to diagnosis and treatment in symptomatic cancer associated with poorer outcomes? Systematic review. Br. J. Cancer 2015, 112, S92–S107. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Hamilton, W.; Walter, F.M.; Rubin, G.; Neal, R.D. Improving early diagnosis of symptomatic cancer. Nat. Rev. Clin. Oncol. 2016, 13, 740–749. [Google Scholar] [CrossRef]
  12. Kostopoulou, O.; Arora, K.; Pálfi, B. Using cancer risk algorithms to improve risk estimates and referral decisions. Commun. Med. 2022, 2, 2. [Google Scholar] [CrossRef]
  13. Amin, M.B.; Greene, F.L.; Edge, S.B.; Compton, C.C.; Gershenwald, J.E.; Brookland, R.K.; Meyer, L.; Gress, D.M.; Byrd, B.R.; Winchester, D.P. The Eighth Edition AJCC Cancer Staging Manual: Continuing to build a bridge from a population-based to a more “personalized” approach to cancer staging: The Eighth Edition AJCC Cancer Staging Manual. CA Cancer J. Clin. 2017, 67, 93–99. [Google Scholar] [CrossRef]
  14. Cancer Research UK. Risk Assessment Tools (RATs). Available online: https://www.cancerresearchuk.org/sites/default/files/rats_pdf_1.pdf (accessed on 6 January 2022).
  15. Hamilton, W.; Green, T.; Martins, T.; Elliott, K.; Rubin, G.; Macleod, U. Evaluation of risk assessment tools for suspected cancer in general practice: A cohort study. Br. J. Gen. Pract. 2013, 63, e30–e36. [Google Scholar] [CrossRef] [PubMed]
  16. Hamilton, W. The CAPER studies: Five case-control studies aimed at identifying and quantifying the risk of cancer in symptomatic primary care patients. Br. J. Cancer 2009, 101, S80–S86. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Hippisley-Cox, J.; Coupland, C. QCancer. Available online: https://www.qcancer.org (accessed on 12 August 2021).
  18. Hippisley-Cox, J.; Coupland, C. Symptoms and risk factors to identify men with suspected cancer in primary care: Derivation and validation of an algorithm. Br. J. Gen. Pract. 2013, 63, e1–e10. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Hippisley-Cox, J.; Coupland, C. Symptoms and risk factors to identify women with suspected cancer in primary care: Derivation and validation of an algorithm. Br. J. Gen. Pract. 2013, 63, e11–e21. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. National Institute for Health and Care Excellence (NICE). Suspected Cancer: Recognition and Referral. NICE Guideline [NG12]. NICE. 2021. Available online: www.nice.org.uk/guidance/ng12 (accessed on 10 July 2021).
  21. Collins, G.S.; Altman, D.G. Identifying patients with undetected gastro-oesophageal cancer in primary care: External validation of QCancer® (Gastro-Oesophageal). Eur. J. Cancer 2013, 49, 1040–1048. [Google Scholar] [CrossRef] [PubMed]
  22. Hamilton, W.; Calitri, R.; Shepherd, L.; Fletcher, E.; Shakespeare, M. The Electronic RIsk Assessment for CAncer (ERICA) Trial. 2022. Available online: https://www.theericatrial.co.uk/the-trial-team/ (accessed on 15 July 2021).
  23. Allum, W.H.; Blazeby, J.M.; Griffin, S.M.; Cunningham, D.; Jankowski, J.A.; Wong, R. Guidelines for the management of oesophageal and gastric cancer. Gut 2011, 60, 1449–1472. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Richter, A.N.; Khoshgoftaar, T.M. A review of statistical and machine learning methods for modeling cancer risk using structured clinical data. Artif. Intell. Med. 2018, 90, 1–14. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Niu, P.H.; Zhao, L.L.; Wu, H.L.; Zhao, D.B.; Chen, Y.T. Artificial intelligence in gastric cancer: Application and future perspectives. World J. Gastroenterol. 2020, 26, 5408–5419. [Google Scholar] [CrossRef] [PubMed]
  26. Jones, O.T.; Calanzani, N.; Saji, S.; Duffy, S.W.; Emery, J.; Hamilton, W.; Singh, H.; de Wit, N.J.; Walter, F.M. Artificial Intelligence Techniques That May Be Applied to Primary Care Data to Facilitate Earlier Diagnosis of Cancer: Systematic Review. J. Med. Internet. Res. 2021, 23, e23483. [Google Scholar] [CrossRef]
  27. Lyratzopoulos, G.; Abel, G.A.; McPhail, S.; Neal, R.D.; Rubin, G.P. Measures of promptness of cancer diagnosis in primary care: Secondary analysis of national audit data on patients with 18 common and rarer cancers. Br. J. Cancer 2013, 108, 686–690. [Google Scholar] [CrossRef] [PubMed]
  28. Alfayez, A.A.; Kunz, H.; Grace Lai, A. Predicting the risk of cancer in adults using supervised machine learning: A scoping review. BMJ Open. 2021, 11, e047755. [Google Scholar] [CrossRef] [PubMed]
  29. Uddin, S.; Khan, A.; Hossain, M.E.; Moni, M.A. Comparing different supervised machine learning algorithms for disease prediction. BMC Med. Inform. Decis. Mak. 2019, 19, 281. [Google Scholar] [CrossRef] [PubMed]
  30. Zihni, E.; Madai, V.I.; Livne, M.; Galinovic, I.; Khalil, A.A.; Fiebach, J.B.; Frey, D. Opening the black box of artificial intelligence for clinical decision support: A study predicting stroke outcome. Stoean R, editor. PLoS ONE 2020, 15, e0231166. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. Department of Health and Social Care UK. A Guide to Good PRACTICE for Digital and Data-Driven Health Technologies. 2021. Available online: https://www.gov.uk/government/publications/code-of-conduct-for-data-driven-health-and-care-technology/initial-code-of-conduct-for-data-driven-health-and-care-technology (accessed on 28 May 2022).
  32. Walley, T.; Mantgani, A. The UK General Practice Research Database. Lancet 1997, 350, 1097–1099. [Google Scholar] [CrossRef]
  33. Lawson, D.; Sherman, V.; Hollowell, J. The General Practice Research Database. Scientific and Ethical Advisory Group. QJM 1998, 91, 445–452. [Google Scholar] [CrossRef] [Green Version]
  34. PHE. Public Health England (PHE) Data and Analysis Tools: Cancer Incidence. 2019. Available online: https://www.cancerdata.nhs.uk/incidence_and_mortality (accessed on 10 July 2021).
  35. Thrumurthy, S.G.; Chaudry, M.A.; Hochhauser, D.; Mughal, M. The diagnosis and management of gastric cancer. BMJ 2013, 347, f6367. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  36. Arnal, M.J.D.; Arenas, Á.; Arbeloa, Á. Esophageal cancer: Risk factors, screening and endoscopic treatment in Western and Eastern countries. World J. Gastroenterol. 2015, 21, 7933. [Google Scholar] [CrossRef]
  37. NICE. Oesophago-Gastric Cancer: Assessment and Management in Adults (NICE Guideline NG83). National Institute for Health and Care Excellence (NICE). Available online: https://www.nice.org.uk/guidance/ng83 (accessed on 5 July 2021).
  38. Ribeiro, M.T.; Singh, S.; Guestrin, C. ‘Why Should I Trust You?’: Explaining the Predictions of Any Classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; ACM: New York, NY, USA, 2016; pp. 1135–1144. Available online: https://dl.acm.org/doi/10.1145/2939672.2939778 (accessed on 12 August 2021).
  39. Benn, M.; Tybjærg-Hansen, A.; Stender, S.; Frikke-Schmidt, R.; Nordestgaard, B.G. Low-Density Lipoprotein Cholesterol and the Risk of Cancer: A Mendelian Randomization Study. JNCI J. Natl. Cancer Inst. 2011, 103, 508–519. [Google Scholar] [CrossRef] [Green Version]
  40. Tomiki, Y.; Suda, S.; Tanaka, M.; Okuzawa, A.; Matsuda, M.; Ishibiki, Y.; Sakamoto, K.; Kamano, T.; Tsurumaru, M.; Watanabe, Y.; et al. Reduced low-density-lipoprotein cholesterol causing low serum cholesterol levels in gastrointestinal cancer: A case control study. J. Exp. Clin. Cancer Res. CR 2004, 23, 233–240. [Google Scholar]
  41. Hunter, J.D. Matplotlib: A 2D Graphics Environment. Comput. Sci. Eng. 2007, 9, 90–95. [Google Scholar] [CrossRef]
  42. Wynants, L.; van Smeden, M.; McLernon, D.J.; Timmerman, D.; Steyerberg, E.W.; Van Calster, B. Three myths about risk thresholds for prediction models. BMC Med. 2019, 17, 192. [Google Scholar] [CrossRef] [Green Version]
  43. Cromwell, D.; Wahedally, H.; Park, M.H.; Maynard, N.; Crosby, T.; Trudgill, N.; Gaskell, J.; Napper, R. National Oesophago-Gastric Cancer Audit. Healthcare Quality Improvement Partnership (HQIP). 2019. Available online: https://www.nogca.org.uk/content/uploads/2019/12/REF150_NOGCA_2019-Annual-Report-FINAL_19Dec.pdf (accessed on 28 May 2022).
  44. Moore, S.F.; Price, S.J.; Chowienczyk, S.; Bostock, J.; Hamilton, W. The impact of changing risk thresholds on the number of people in England eligible for urgent investigation for possible cancer: An observational cross-sectional study. Br. J. Cancer 2021, 125, 1593–1597. [Google Scholar] [CrossRef] [PubMed]
  45. Hamilton, W.; Stapley, S.; Campbell, C.; Lyratzopoulos, G.; Rubin, G.; Neal, R.D. For which cancers might patients benefit most from expedited symptomatic diagnosis? Construction of a ranking order by a modified Delphi technique. BMC Cancer 2015, 15, 820. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  46. Van Calster, B.; McLernon, D.J.; van Smeden, M.; Wynants, L.; Steyerberg, E.W. Calibration: The Achilles heel of predictive analytics. BMC Med. 2019, 17, 230. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  47. Price, S.; Spencer, A.; Medina-Lara, A.; Hamilton, W. Availability and use of cancer decision-support tools: A cross-sectional survey of UK primary care. Br. J. Gen. Pract. 2019, 69, e437–e443. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  48. Dikomitis, L.; Green, T.; Macleod, U. Embedding electronic decision-support tools for suspected cancer in primary care: A qualitative study of GPs’ experiences. Prim. Health Care Res. Dev. 2015, 16, 548–555. [Google Scholar] [CrossRef] [Green Version]
  49. Green, T.; Martins, T.; Hamilton, W.; Rubin, G.; Elliott, K.; Macleod, U. Exploring GPs’ experiences of using diagnostic tools for cancer: A qualitative study in primary care. Fam. Pract. 2015, 32, 101–105. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  50. Panter, S.J.; Bramble, M.G.; O’Flanagan, H.; Hungin, A.P.S. Urgent cancer referral guidelines: A retrospective cohort study of referrals for upper gastrointestinal adenocarcinoma. Br. J. Gen. Pract. J R Coll Gen. Pract. 2004, 54, 611–613. [Google Scholar]
  51. Dong, J.; Thrift, A.P. Alcohol, smoking and risk of oesophago-gastric cancer. Best Pract. Res. Clin. Gastroenterol. 2017, 31, 509–517. [Google Scholar] [CrossRef]
  52. Martins, T.; Hamilton, W.; Ukoumunne, O. Ethnic inequalities in time to diagnosis of cancer: A systematic review. BMC Fam. Pract. 2013, 14, 197. Available online: http://www.biomedcentral.com/1471-2296/14/197 (accessed on 12 August 2021). [CrossRef] [Green Version]
  53. Gupta, S.; Tao, L.; Murphy, J.D.; Camargo, M.C.; Oren, E.; Valasek, M.A.; Gomez, S.L.; Martinez, M.E. Race/Ethnicity-, Socioeconomic Status-, and Anatomic Subsite-Specific Risks for Gastric Cancer. Gastroenterology 2019, 156, 59–62.e4. [Google Scholar] [CrossRef] [PubMed]
  54. Hamilton, W.T.; Round, A.P.; Sharp, D.; Peters, T.J. The quality of record keeping in primary care: A comparison of computerised, paper and hybrid systems. Br. J. Gen. Pract. J R Coll Gen. Pract. 2003, 53, 929–933, discussion 933. [Google Scholar]
  55. Hippisley-Cox, J.; Coupland, C. Identifying patients with suspected gastro-oesophageal cancer in primary care: Derivation and validation of an algorithm. Br. J. Gen. Pract. 2011, 61, e707–e714. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  56. McInerney, C.D.; Scott, B.C.; Johnson, O.A. Are Regulations Safe? Reflections From Developing a Digital Cancer Decision-Support Tool. JCO Clin. Cancer Inform. 2021, 5, 353–363. [Google Scholar] [CrossRef]
Figure 1. Feature importance graph displaying relative contribution of diagnostic factors for oesophago-gastric cancer to model risk score. Model: Support Vector Machine (with linear kernel). Feature weights represented by model coefficients. Generated with matplotlib [41].
Figure 1. Feature importance graph displaying relative contribution of diagnostic factors for oesophago-gastric cancer to model risk score. Model: Support Vector Machine (with linear kernel). Feature weights represented by model coefficients. Generated with matplotlib [41].
Cancers 14 05023 g001
Table 1. Patient characteristics of 40,348 individuals according to basic demographic information, symptoms, lab test results, and cancer site, organised by case/control. Figures given as number of occurrences (in total cohort/case group/control group) and percentages.
Table 1. Patient characteristics of 40,348 individuals according to basic demographic information, symptoms, lab test results, and cancer site, organised by case/control. Figures given as number of occurrences (in total cohort/case group/control group) and percentages.
Patient CharacteristicTotal Cohort
Count (%)
N = 40,348
Case
Count (%)
N = 7471 (18.5%)
Control
Count (%)
N = 32,877 (81.5%)
Age
Under 552550 (6.3)514 (6.9)2036 (6.2)
Over 5537,798 (93.7)6957 (93.1)30,841 (93.8)
Sex
Male14,860 (36.4)2672 (35.8)12,188 (37.1)
Female25,488 (63.6)4799 (64.2)20,689 (62.9)
Cancer Site
Oesophagus26,360 (65.3)4854 (65.0)21,506 (65.4)
Stomach13,988 (34.7)2617 (35.0)11,371 (34.6)
Symptoms
Abdominal Pain2215 (5.5)905 (12.1)1310 (4.0)
Chest Pain2316 (5.7)727 (9.7)1589 (4.8)
Constipation1681 (4.2)608 (8.1)1073 (3.3)
Cough4782 (11.9)1005 (13.4)3777 (11.4)
Dyspepsia2085 (5.2)1294 (17.3)764 (2.3)
Dyspepsia (repeat)699 (1.7)532 (7.1)167 (0.5)
Dysphagia2605 (6.5)2420 (32.3)185 (0.6)
Dysphagia (repeat)678 (1.7)635 (8.5)43 (0.1)
Epigastric Pain883 (2.2)617 (8.3)266 (0.8)
Fatigue1362 (3.4)388 (5.2)974 (3.0)
Nausea/Vomiting1616 (4.0)979 (13.1)637 (1.9)
Nausea/Vomiting (repeat)534 (1.3)386 (5.2)148 (0.5)
Reflux1355 (3.4)842 (11.3)513 (1.6)
Shortness of breath2621 (6.5)629 (8.4)1992 (6.1)
Weight loss 891 (2.2)615 (8.2)276 (0.8)
Lab Test Results
Cholesterol (high)6100 (15.1)920 (12.3)5180 (15.8)
Haemoglobin (low)5398 (13.3)2045 (27.3)3353 (10.2)
Inflammatory Markers (high)2431 (6.0)1010 (13.5)1421 (4.3)
Liver Function Test (high)4751 (11.8)1272 (17.0)3479 (10.6)
Mean Corpuscular Volume (low)1007 (2.5)640 (8.6)367 (1.1)
Platelet Count (high)1274 (3.2)706 (9.4)568 (1.7)
White Cell Count (high)1533 (3.8)671 (9.0)862 (2.6)
Table 2. Performances of machine-learning-based probabilistic classifiers on test dataset in comparison with oesophago-gastric cancer risk-assessment tool (ogRAT) matrix (performance calculated on same test set) for prediction of oesophago-gastric cancer incidence.
Table 2. Performances of machine-learning-based probabilistic classifiers on test dataset in comparison with oesophago-gastric cancer risk-assessment tool (ogRAT) matrix (performance calculated on same test set) for prediction of oesophago-gastric cancer incidence.
ClassifierAUROCClassification ThresholdAccuracyPrecisionRecallF1
Linear Support Vector Machine0.870.500
0.800
0.89
0.88
0.85
0.90
0.53
0.41
0.65
0.57
Logistic Regression0.870.425
0.750
0.89
0.88
0.81
0.90
0.58
0.44
0.68
0.59
Random Forest0.860.600
0.700
0.89
0.88
0.86
0.92
0.48
0.39
0.62
0.55
Bernoulli Naïve Bayes0.860.700
0.900
0.89
0.86
0.80
0.84
0.55
0.34
0.65
0.49
XGBoost0.870.500
0.800
0.89
0.88
0.85
0.91
0.54
0.39
0.66
0.55
ogRAT0.810.010
0.020
0.87
0.87
0.86
0.91
0.41
0.33
0.56
0.49
Table 3. Performance of machine-learning-based probabilistic classifiers on test dataset (results shown for some of the best-performing techniques: Support Vector Machine with linear kernel and Logistic Regression with stochastic average gradient solver) in comparison with oesophago-gastric cancer risk-assessment tool (ogRAT) matrix (performance calculated on same test set) for prediction of oesophago-gastric cancer incidence. Performances shown when making comparisons using the exact precision values achieved by the ogRAT.
Table 3. Performance of machine-learning-based probabilistic classifiers on test dataset (results shown for some of the best-performing techniques: Support Vector Machine with linear kernel and Logistic Regression with stochastic average gradient solver) in comparison with oesophago-gastric cancer risk-assessment tool (ogRAT) matrix (performance calculated on same test set) for prediction of oesophago-gastric cancer incidence. Performances shown when making comparisons using the exact precision values achieved by the ogRAT.
ClassifierAUROCClassification ThresholdAccuracyPrecisionRecallF1
Linear Support Vector Machine0.870.56
0.82
0.89
0.88
0.86
0.91
0.52
0.41
0.65
0.57
Logistic Regression0.870.57
0.82
0.89
0.88
0.86
0.91
0.51
0.41
0.64
0.57
ogRAT0.810.01
0.02
0.87
0.87
0.86
0.91
0.41
0.33
0.56
0.49
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Briggs, E.; de Kamps, M.; Hamilton, W.; Johnson, O.; McInerney, C.D.; Neal, R.D. Machine Learning for Risk Prediction of Oesophago-Gastric Cancer in Primary Care: Comparison with Existing Risk-Assessment Tools. Cancers 2022, 14, 5023. https://doi.org/10.3390/cancers14205023

AMA Style

Briggs E, de Kamps M, Hamilton W, Johnson O, McInerney CD, Neal RD. Machine Learning for Risk Prediction of Oesophago-Gastric Cancer in Primary Care: Comparison with Existing Risk-Assessment Tools. Cancers. 2022; 14(20):5023. https://doi.org/10.3390/cancers14205023

Chicago/Turabian Style

Briggs, Emma, Marc de Kamps, Willie Hamilton, Owen Johnson, Ciarán D. McInerney, and Richard D. Neal. 2022. "Machine Learning for Risk Prediction of Oesophago-Gastric Cancer in Primary Care: Comparison with Existing Risk-Assessment Tools" Cancers 14, no. 20: 5023. https://doi.org/10.3390/cancers14205023

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop