Next Article in Journal
Type 2 Diabetes Mellitus in Low- and Middle-Income Countries: The Significant Impact of Short-Chain Fatty Acids and Their Quantification
Previous Article in Journal
Three-Stage Framework for Accurate Pediatric Chest X-ray Diagnosis Using Self-Supervision and Transfer Learning on Small Datasets
Previous Article in Special Issue
Evolving and Novel Applications of Artificial Intelligence in Thoracic Imaging
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Early Detection of Heart Failure with Autonomous AI-Based Model Using Chest Radiographs: A Multicenter Study

1
Department of Radiology, Massachusetts General Hospital and Harvard Medical School, Boston, MA 02114, USA
2
Mass General Brigham Data Science Office, Boston, MA 02114, USA
3
Qure AI, Mumbai 400063, India
*
Author to whom correspondence should be addressed.
Diagnostics 2024, 14(15), 1635; https://doi.org/10.3390/diagnostics14151635 (registering DOI)
Submission received: 22 May 2024 / Accepted: 18 June 2024 / Published: 30 July 2024
(This article belongs to the Special Issue Applications of Artificial Intelligence in Thoracic Imaging)

Abstract

:
The opportunistic use of radiological examinations for disease detection can potentially enable timely management. We assessed if an index created by an AI software to quantify chest radiography (CXR) findings associated with heart failure (HF) could distinguish between patients who would develop HF or not within a year of the examination. Our multicenter retrospective study included patients who underwent CXR without an HF diagnosis. We included 1117 patients (age 67.6 ± 13 years; m:f 487:630) that underwent CXR. A total of 413 patients had the CXR image taken within one year of their HF diagnosis. The rest (n = 704) were patients without an HF diagnosis after the examination date. All CXR images were processed with the model (qXR-HF, Qure.AI) to obtain information on cardiac silhouette, pleural effusion, and the index. We calculated the accuracy, sensitivity, specificity, and area under the curve (AUC) of the index to distinguish patients who developed HF within a year of the CXR and those who did not. We report an AUC of 0.798 (95%CI 0.77–0.82), accuracy of 0.73, sensitivity of 0.81, and specificity of 0.68 for the overall AI performance. AI AUCs by lead time to diagnosis (<3 months: 0.85; 4–6 months: 0.82; 7–9 months: 0.75; 10–12 months: 0.71), accuracy (0.68–0.72), and specificity (0.68) remained stable. Our results support the ongoing investigation efforts for opportunistic screening in radiology.

1. Introduction

According to the Centers for Disease Control and Prevention (CDC), about 6.2 million adults suffer from heart failure (HF) in the United States [1]. In 2018, 379,800 death certificates mentioned HF. In 2012, HF had an estimated cost of USD 30.7 billion, encompassing healthcare services, medical treatments, and missed workdays. The HF cost is likely significantly greater now than in 2012 estimates [1]. Global statistics suggest that 64.3 million people worldwide suffered from HF in 2017, with an incidence rate of 1–20 HF cases per 1000 population, a 5-year mortality of 50–75%, and an annual healthcare cost of EUR 25,500 per year [2]. With increasing overall life expectancy and continued improvements in life-saving treatment options, the prevalence of HF is expected to rise [1,3]. The high prevalence of HF is likely related to its common risk factors, such as coronary artery disease, diabetes, hypertension, and obesity [3].
Timely diagnosis and initiation of treatment can slow the progression of HF and improve patient outcomes. Although HF can present with several symptoms, it is frequently first diagnosed in acute care settings of an emergency department or a hospital, where it carries a high mortality. A recent study suggested that among 959,438 patients with a new HF diagnosis, 38% of patients were diagnosed in acute care settings, and 46% of these patients had HF symptoms about 6 months prior to their diagnosis [4]. Underdiagnosis is more common in patients with lower socioeconomic status, black individuals, women, and those with comorbidities [4] Additionally, multiple studies have demonstrated a high prevalence of unknown or unrecognized heart failure and asymptomatic ventricular dysfunction in patients with known history of chronic degenerative diseases like diabetes, hypertension or COPD [5,6,7,8,9,10,11,12,13]. In such a context of delayed and overlooked HF diagnoses, we hypothesized that the most common imaging test of chest radiography (CXR) could be used as an objective and automatic tool for the opportunistic screening of HF. Therefore, we used a CXR-based AI tool to automate the quantification of cardiac silhouette size and pleural effusion to create an index for early HF detection. We performed a multicenter study to determine if the autonomous AI model can help in the early detection of HF using CXR signs.

2. Materials and Methods

2.1. Patients

The study included two groups of patients older than 45 years without a known HF diagnosis who underwent CXR. The inclusion criteria for the first group were as follows: baseline posterior–anterior projection CXR of patient without diagnosis of HF before or at the time of the examination; a minimum of one-year follow-up in the institutional medical records; and a clinical diagnosis of HF within a year of the CXR (by pro-B-type natriuretic peptide (NT-proBNP) levels and echocardiography at the time of diagnosis). The control group comprised >45-year-old patients without a known diagnosis of HF in their electronic medical records, either before or after the acquisition of their chest radiographs. Patients with the diagnosis of HF at the time of their CXR or before were excluded. Patients with only portable or anterior–posterior projection chest radiographs were excluded. There were 413 chest radiographs taken in the HF group and 704 chest radiographs taken in the non-HF group (Figure 1).
All patients were identified using an SQL-based query on the Epic data warehouse (EDW, Epic, Verona, WI, USA) for querying diagnosis time, lab values, and imaging examinations in electronic medical records. All retrieved and eligible patients were included in the study in a consecutive manner without any exclusion. Two physician study coinvestigators (PKK, with three years of post-doctoral research experience, and EGF, with one year of post-doctoral research experience) manually reviewed the medical records of each patient to further confirm the SQL query results against the study eligibility criteria.

2.2. Chest Radiographs

Once the two study groups, with and without the future diagnosis of HF, were confirmed, we exported and de-identified posterior–anterior projection chest radiographs from PACS (Visage 7, Visage Imaging Inc., San Diego, CA, USA). Chest radiographs with evaluation-limiting technical or quality issues (such as field of view not included, overlying anatomy, or oblique projection) were not included. Only one chest radiograph was included for each patient.
Each radiograph was processed with a CXR-based AI algorithm (Qure.AI, Mumbai, India). The algorithm provided output on the probability of the presence of an enlarged cardiac silhouette and pleural effusion in addition to an index for heart failure risk based on the presence of enlarged heart size and pleural effusion. The heart failure risk index was calculated as a weighted sum of the probabilities of enlarged cardiac silhouette and pleural effusion. The algorithm exported the outputs into a comma-separated value (CSV, Microsoft EXCEL, Microsoft Inc., Redmond, WA, USA).

2.3. AI Algorithm

The enlarged cardiac silhouette algorithm (Qure.AI, India) is a UNET++-based segmentation model with an Efficientnet-v2 backbone (Figure 2). The model was trained to output segmentation masks of the heart and lungs to estimate the cardiothoracic ratio. The pleural effusion algorithm is a hybrid classification and segmentation model with the same UNET++ architecture, and outputs a probability score between 0 and 1 based on the probability of the presence of pleural effusion on the radiograph. It also outputs a segmentation map to mark the location of pleural effusions.
The algorithms for both enlarged cardiac silhouette and pleural effusion are part of an AI software (qXR version 4.0) that can detect multiple other radiological findings in a frontal CXR image, in addition to enlarged cardiac silhouette and pleural effusion. The AI software device was trained with a large dataset of 4.2 million CXR images (about 70% from India, and the rest from the US, EU, Middle East, and South Asia) labeled by radiologists. The data were extensively filtered and cleaned to ensure quality, and then curated to minimize any racial or gender bias. Within the large training set, there were approximately 60,000 confirmed cases of pleural effusions and a similar number of cases with enlarged cardiac silhouettes. The device was internally tested using approximately 300,000 CXR images, with half the CXR images from similar sources and the other half from sites not included in the training process.

2.4. Statistical Analysis

Data were analyzed with the statistical software R Studio (V2023.06.1). We calculated the area under the curve (AUC) for receiver operating characteristic (ROC) analysis for the entire dataset (HF versus no-HF groups) and its optimal threshold. Next, we calculated stratified AUCs after dividing the chest radiographs into four subgroups according to the time between the CXR date and the date of HF diagnosis (subgroups—A: <3 months; B: 4 to 6 months; C: 7 to 9 months; D: 10 to 12 months) and three subgroups by left ventricle ejection fraction (<40% for low ejection fraction; 41–49 for mildly reduced ejection fraction; >50% for normal ejection fraction) reported in the cardiac ultrasound, when available. We used the optimal threshold obtained from the entire dataset analysis when calculating the AUC by time groups and ejection fraction groups. Mann–Whitney U and Kruskal–Wallis tests were performed to assess the statistical differences in the AI algorithm’s performance between the different subgroups. A p-value less than 0.05 was considered statistically significant.

3. Results

3.1. Overall Statistics

The study included a total of 1117 chest radiographs from 1117 patients (female:male, 630:487; mean age, 67.60 years) from 11 hospitals. For the entire HF and non-HF groups, the algorithm had an AUC of 0.798 (95% confidence interval 0.772–0.824) using an optimal decision threshold of 0.389, with a sensitivity of 0.813, specificity of 0.684, and accuracy of 0.732. Two examples from each group are shown in Figure 3.
The respective positive and the negative likelihood ratios were +1.888 and –0.471. There was no difference in model performance based on patient gender (female AUC = 0.807 vs. male AUC = 0.784) (p > 0.1).
There were significant differences in the three AI outputs (cardiac silhouette size, pleural effusion, and HF risk) between the HF and non-HF groups (p < 0.001). The average probability scores for enlarged cardiac silhouette and pleural effusions in the HF group were 0.537 ± 0.05 and 0.678 ± 0.29, and 0.487 ± 0.05 and 0.369 ± 0.32 for the non-HF group. The corresponding average HF risk scores between the two groups were 0.552 ± 0.19 and 0.292 ± 0.23, respectively.

3.2. Stratified Analysis

The AUCs for the subgroups based on the time intervals between CXR and HF decreased with increasing time intervals: <3 months—0.854; 3–6 months—0.821; 7–9 months—0.756; 10–12 months—0.717. The corresponding sensitivities exhibited a similar trend: 0.897, 0.851, 0.768, and 0.688 for the <3 months, 3–6 months, 7–9 months, and 10–12 months subgroups. The respective accuracies for these time subgroups were 0.717, 0.705, 0.695, and 0.685. The specificity for all groups was 0.684. We then used the heart failure probability scores in each group to perform a Kruskal–Wallis test. The differences in algorithm performance at four different time intervals were significantly different (p < 0.001). In the pairwise assessment, there was a strong significant difference (p-values < 0.003) between group A (<3 months) and groups C (7–9 months) and D (10–12 months), as well as between group B (3–6 months) and group D (10–12 months), but not between any of the two consecutive groups, such as groups A vs. B or groups C vs. D (p > 0.221). The contingency values for each group are depicted in Table 1.
The AUCs of the low ejection fraction, mildly reduced ejection fraction, and normal ejection fraction groups were as follows: 0.830, 0.804, and 0.795. The respective results for sensitivity were 0.857, 0.805, and 0.815, with a constant specificity for all groups of 0.684. The contingency table for this is depicted in Table 2. The respective accuracies across the three ejection fraction groups were 0.695, 0.690, and 0.720. There was no statistical difference in model performance across the three ejection fraction groups with HF (p = 0.061).
A pairwise analysis demonstrated a significant difference (p < 0.019) between the low ejection fraction groups and the mildly reduced ejection fraction group, as well as between the low ejection fraction group and the normal ejection fraction group (p < 0.027). The difference between the normal ejection fraction and the high ejection fraction groups was not significant (p < 0.854).

4. Discussion

Our study shows that the HF risk index based on the probability of enlarged cardiac silhouette and pleural effusions in chest radiography can help detect the development of HF. Although we noted substantial predictive AUC and accuracy values for up to 12 months before the clinical diagnosis of HF, the AI performance was higher within 3 months of HF diagnosis than at 12 months from HF diagnosis. Our findings add to the growing body of evidence for opportunistic screening applications of AI algorithms for cardiovascular diagnosis from chest radiography.
The importance of early diagnosing heart failure and the exponential interest on the clinical application of artificial intelligence has encouraged the scientific community to study their potential clinical aplication. The processing of bigdata from patient demographics, signs and symptoms, laboratory results, and other clinical information from electronic medical records and wearables has been evaluated not only for diagnosing heart failure but also for predicting hospital readmissions and mortality in patient with HF [14,15,16,17,18]. Celik et al. reported on the use of an AI algorithm on 212 chest radiographs from a single healthcare practice to detect patients with undiagnosed HF. Their model assessed enlarged cardiac silhouette and pleural effusions with an index for HF diagnosis. The authors achieved an accuracy of 0.84, sensitivity of 0.90, specificity of 0.80, and a positive likelihood ratio of detecting HF in the chest radiographs of 4.54 [19]. The differences between the results in our study and Celik et al.’s investigation might be related to patient demographics, race, ethnicity, comorbidities, or underlying risk factors, as well as radiographic quality. Since enlarged cardiac silhouette and pleural effusions can be related to non-HF diagnosis, it is not surprising that our larger, multi-center study found low specificity for the AI algorithm. A much smaller study (n = 50 patients) with a Keras-based model (with all radiographs with enlarged cardiac silhouette labeled as positive for HF) and data from the National Institutes of Health (NIH) reported a higher accuracy of 0.82 [20].
Other AI algorithms using chest radiographs have targeted cardiopulmonary circulation, such as for the detection of high pulmonary arterial capillary pressure. Saito et al. reported an AUC of 0.86 for their AI algorithm for differentiating normal and elevated capillary pressures, compared with an AUC of 0.83 for the participating cardiologists (p = 0.24) [21]. Hirata et al. developed an AI model to differentiate between 115 patients with normal and elevated pulmonary capillary pressures with an AUC of 0.77 [22]. The authors explored several serum and radiological values, such as BNP, pulmonary arterial wedge pressure, and left ventricle ejection fraction to improve their AI performance for the detection of elevated pulmonary artery wedge pressure. Unfortunately, their study did not assess the model’s performance based on the classification of preserved or low ejection fractions.
The use of chest radiographs has also been investigated for the classification of valvular diseases. Ueda et al. recently developed and validated an AI model that had an AUC from 0.83 to 0.92 to classify different valvular diseases, inferior vena cava dilatation (AUC of 0.85), and different groups of categories of left ventricular ejection fraction (AUC: 0.92) [23]. These results further support the research for disease detection in common radiological examinations.
We found that the low ejection fraction group had a significantly higher index than the mildly reduced and normal ejection fraction groups (AUC 0.830 vs. 0.804, and 0.795, respectively) (p < 0.01). This is biologically consistent with the pathophysiology of heart failure, as signs and symptoms are more frequent in the presence of noticeable structural changes in the low ejection fraction group. Since the heart can progresivelly adapt and pump excess volumes at the onset of cardiac insults due to the Frank–Starling mechanism, despite changes in cardiac size, some patients remain asymptomatic for a period of time. Over time, this mechanism becomes insufficient, and symptoms begin to appear [24,25]. Therefore, not every heart above normal size necessarily represents a case of heart failure, and other factors and results should be taken into account.
The major clinical implication of our study is the further reinforcement of the role of AI algorithms for detecting sub-clinical or undiagnosed HF with high AUC and accuracy. An early diagnosis of HF via chest radiography, the most frequently performed radiological test, can help initiate early treatment. The early treatment of HF can help improve the mortality and morbidity associated with HF, as reported in prior studies [26,27]. Our study also reinforces the need for radiologists to pay closer attention to the evaluation of cardiac silhouette on chest radiographs, particularly in posterior–anterior projections in the presence of concurrent pleural effusions. In the presence of such findings, referring physicians should consider triggering additonal investigations such as an echocardiogram or BNP to further study cardiac function in selected patients. [26,28,29,30]
Our study has several limitations. First, our study was retrospective and might not reflect the challenges of real-world clinical practice. Second, both enlarged cardiac silhouette and pleural effusions can exist without HF, a fact reflected by the low specificity noted in our study. Third, the AI algorithm did not specifically target the evaluation of pulmonary interstitial or airspace edema or the redistribution of pulmonary vasculature with pulmonary venous hypertension. Fourth, we did not specifically enrich the data with patients with non-HF causes of enlarged cardiac silhouette and/or pleural effusions. Fifth, our evaluation was limited to posterior–anterior projection radiographs only and did not include patients who underwent portable chest radiographs, which are extremely common in acutely ill patients and those visiting emergency departments or urgent care facilities. Sixth, we excluded patients with a diagnosis of HF at the time of their radiographs, which could have limited our ability to further stratify the results and gain a detailed understanding of the AI’s utility. Since our AI algorithm was built on a generic training dataset (without explicit annotation for radiographs with and without HF) for the detection of enlarged cardiac silhouette and pleural effusions, this could have resulted in lower-than-expected specificity.

5. Conclusions

We believe that future developments in the AI algorithm could target specific signs of pulmonary interstitial and airspace edema to improve performance, particularly from the point of view of specificity. Likewise, the AI-based evaluation of temporal changes in enlarged cardiac silhouette, pulmonary vasculature, and pleural effusions using serial radiographs can also help improve HF detection in underdiagnosed populations. Finally, a prospective study to assess the impact of the early radiographic prediction of HF will help determine if such an AI algorithm can lead to real financial gains and improvement in HF-related mortality and morbidity.
In conclusion, our autonomous AI model can help identify signs of HF up to 12 months prior to the clinical diagnosis of HF with a high and consistent AUC and accuracy, regardless of patient gender and imaging site. The detection of HF signs with an autonomous chest-radiography-based AI algorithm can help in the early diagnosis and treatment of unsuspected HF.

Author Contributions

Conceptualization: E.G.-F., B.B., K.J.D., M.K.K. and S.R.D.; Methodology, E.G.-F., P.K., L.K., R.F., S.H., B.B., K.J.D., M.K.K. and S.R.D.; Software, P.P., M.T., S.K., C.A. and D.R.; Validation, E.G.-F., P.K., L.K., R.F., S.H., B.B., K.J.D., M.K.K. and S.R.D.; Formal analysis, E.G.-F., P.K., L.K., R.F., S.H., B.B., K.J.D., M.K.K. and S.R.D.; Investigation, E.G.-F., P.K., L.K., R.F., S.H., B.B., K.J.D., M.K.K. and S.R.D.; Resources: P.P., M.T., S.K., C.A. and D.R.; Data curation, E.G.-F., P.K., L.K., R.F., S.H., B.B., K.J.D., M.K.K. and S.R.D.; Writing—original draft preparation, E.G.-F., B.B., K.J.D., M.K.K. and S.R.D.; Writing—review and editing, E.G.-F., P.K., L.K., R.F., S.H., B.B., K.J.D., M.K.K. and S.R.D.; Visualization, E.G.-F., P.K., L.K., R.F., S.H., B.B., K.J.D., M.K.K. and S.R.D.; Supervision, B.B., K.J.D., M.K.K. and S.R.D.; Project administration, B.B., K.J.D., M.K.K. and S.R.D.; Funding acquisition, P.P., M.T., S.K., C.A. and D.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Qure.AI, grant number [239717].

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Institutional Review Board of Massachusetts General Hospital (protocol number: 2020P003950; date: 12/09/2020) for studies involving humans. The retrospective study received approval from the institutional review board (protocol number: 2020P003950) with a waiver of informed consent. The study was HIPAA-compliant.

Informed Consent Statement

Patient consent was waived by our IRB committee due to the minimal risks of this retrospective study.

Data Availability Statement

The datasets presented in this article are not readily available due to restrictions from our Institutional Review Board.

Conflicts of Interest

SRD: provides independent image analysis for hospital-contracted clinical research trials programs for Merck, Pfizer, Bristol Myers Squibb, Novartis, Roche, Polaris, Cascadian, Abbvie, Gradalis, Bayer, Zai laboratories, Biengen, Resonance, Analise, Astra Zeneca, and Genentech, and has received research grants from Lunit Inc., Vuno, GE, and Qure AI, honorarium from Siemens, and royalties from Elsievier. MKK: received research grants from Siemens, Globus, Riverain, and Promedius.

References

  1. Virani, S.S.; Alonso, A.; Benjamin, E.J.; Bittencourt, M.S.; Callaway, C.W.; Carson, A.P.; Chamberlain, A.M.; Chang, A.R.; Cheng, S.; Delling, F.N.; et al. Heart Disease and Stroke Statistics-2020 Update: A Report from the American Heart Association. Circulation 2020, 141, e139–e596. [Google Scholar] [CrossRef]
  2. Savarese, G.; Becher, P.M.; Lund, L.H.; Seferovic, P.; Rosano, G.M.C.; Coats, A.J.S. Global burden of heart failure: A comprehensive and updated review of epidemiology. Cardiovasc. Res. 2023, 118, 3272–3287. [Google Scholar] [CrossRef] [PubMed]
  3. Malik, A.; Brito, D.; Vaqar, S.; Chhabra, L. Congestive Heart Failure. [Updated 2023 Nov 5]. In StatPearls [Internet]; StatPearls Publishing: Treasure Island, FL, USA, 2024. Available online: https://www.ncbi.nlm.nih.gov/books/NBK430873/ (accessed on 17 June 2024).
  4. Sandhu, A.T.; Tisdale, R.L.; Rodriguez, F.; Stafford, R.S.; Maron, D.J.; Hernandez-Boussard, T.; Lewis, E.; Heidenreich, P.A. Disparity in the Setting of Incident Heart Failure Diagnosis. Circ. Heart Fail. 2021, 14, e008538. [Google Scholar] [CrossRef] [PubMed]
  5. Boonman-de Winter, L.J.; Rutten, F.H.; Cramer, M.J.; Landman, M.J.; Liem, A.H.; Rutten, G.E.; Hoes, A.W. High prevalence of previously unknown heart failure and left ventricular dysfunction in patients with type 2 diabetes. Diabetologia 2012, 55, 2154–2162. [Google Scholar] [CrossRef]
  6. Patil, V.C.; Patil, H.V.; Shah, K.B.; Vasani, J.D.; Shetty, P. Diastolic dysfunction in asymptomatic type 2 diabetes mellitus with normal systolic function. J. Cardiovasc. Dis. Res. 2011, 2, 213–222. [Google Scholar] [CrossRef]
  7. Mishra, T.K.; Rath, P.K.; Mohanty, N.K.; Mishra, S.K. Left ventricular systolic and diastolic dysfunction and their relationship with microvascular complications in normotensive, asymptomatic patients with type 2 diabetes mellitus. Indian Heart J. 2008, 60, 548–553. [Google Scholar]
  8. Antakly-Hanon, Y.; Ben Hamou, A.; Garçon, P.; Moeuf, Y.; Banu, I.; Fumery, M.; Voican, A.; Abassade, P.; Oriez, C.; Chatellier, G.; et al. Asymptomatic left ventricular dysfunction in patients with type 2 diabetes free of cardiovascular disease and its relationship with clinical characteristics: The DIACAR cohort study. Diabetes Obes. Metab. 2021, 23, 434–443. [Google Scholar] [CrossRef]
  9. Hoek, A.G.; Dal Canto, E.; Wenker, E.; Bindraban, N.; Handoko, M.L.; Elders, P.J.M.; Beulens, J.W.J. Epidemiology of heart failure in diabetes: A disease in disguise. Diabetologia 2024, 67, 574–601. [Google Scholar] [CrossRef] [PubMed]
  10. Swiatkiewicz, I.; Patel, N.T.; Villarreal-Gonzalez, M.; Taub, P.R. Prevalence of diabetic cardiomyopathy in patients with type 2 diabetes in a large academic medical center. BMC Med. 2024, 22, 195. [Google Scholar] [CrossRef] [PubMed] [PubMed Central]
  11. Barrios, V.; Escobar, C.; De La Sierra, A.; Llisterri, J.L.; González-Segura, D. Detection of unrecognized clinical heart failure in elderly hypertensive women attended in primary care setting. Blood Press. 2010, 19, 301–307. [Google Scholar] [CrossRef]
  12. Rutten, F.H.; Cramer, M.-J.M.; Grobbee, D.E.; Sachs, A.P.; Kirkels, J.H.; Lammers, J.-W.J.; Hoes, A.W. Unrecognized heart failure in elderly patients with stable chronic obstructive pulmonary disease. Eur. Heart J. 2005, 26, 1887–1894. [Google Scholar] [CrossRef] [PubMed]
  13. Wong, C.W.; Tafuro, J.; Azam, Z.; Satchithananda, D.; Duckett, S.; Barker, D.; Patwala, A.; Ahmed, F.Z.; Mallen, C.; Kwok, C.S. Misdiagnosis of Heart Failure: A Systematic Review of the Literature. J. Card. Fail. 2021, 27, 925–933. [Google Scholar] [CrossRef] [PubMed]
  14. Stehlik, J.; Schmalfuss, C.; Bozkurt, B.; Nativi-Nicolau, J.; Wohlfahrt, P.; Wegerich, S.; Rose, K.; Ray, R.; Schofield, R.; Deswal, A.; et al. Continuous Wearable Monitoring Analytics Predict Heart Failure Hospitalization: The LINK-HF Multicenter Study. Circ. Heart Fail. 2020, 13, e006513. [Google Scholar] [CrossRef] [PubMed]
  15. Gardner, R.S.; Singh, J.P.; Stancak, B.; Nair, D.G.; Cao, M.; Schulze, C.; Thakur, P.H.; An, Q.; Wehrenberg, S.; Hammill, E.F.; et al. HeartLogic Multisensor Algorithm Identifies Patients During Periods of Significantly Increased Risk of Heart Failure Events: Results from the MultiSENSE Study. Circ. Heart Fail. 2018, 11, e004669. [Google Scholar] [CrossRef] [PubMed]
  16. Yoon, M.; Park, J.J.; Hur, T.; Hua, C.H.; Hussain, M.; Lee, S.; Choi, D.J. Application and Potential of Artificial Intelligence in Heart Failure: Past, Present, and Future. Int. J. Heart Fail. 2023, 6, 11–19. [Google Scholar] [CrossRef] [PubMed]
  17. Choi, D.-J.; Park, J.J.; Ali, T.; Lee, S. Artificial intelligence for the diagnosis of heart failure. NPJ Digit. Med. 2020, 3, 54. [Google Scholar] [CrossRef] [PubMed] [PubMed Central]
  18. Guo, A.; Pasque, M.; Loh, F.; Mann, D.L.; Payne, P.R.O. Heart Failure Diagnosis, Readmission, and Mortality Prediction Using Machine Learning and Artificial Intelligence Models. Curr. Epidemiol. Rep. 2020, 7, 212–219. [Google Scholar] [CrossRef]
  19. Celik, A.; Surmeli, A.O.; Demir, M.; Esen, K.; Camsari, A. The diagnostic value of chest X-ray scanning by the help of Artificial Intelligence in Heart Failure (ART-IN-HF). Clin. Cardiol. 2023, 46, 1562–1568. [Google Scholar] [CrossRef] [PubMed]
  20. Matsumoto, T.; Kodera, S.; Shinohara, H.; Ieki, H.; Yamaguchi, T.; Higashikuni, Y.; Kiyosue, A.; Ito, K.; Ando, J.; Takimoto, E.; et al. Diagnosing Heart Failure from Chest X-Ray Images Using Deep Learning. Int. Heart J. 2020, 61, 781–786. [Google Scholar] [CrossRef]
  21. Saito, Y.; Omae, Y.; Fukamachi, D.; Nagashima, K.; Mizobuchi, S.; Kakimoto, Y.; Toyotani, J.; Okumura, Y. Quantitative estimation of pulmonary artery wedge pressure from chest radiographs by a regression convolutional neural network. Heart Vessel. 2022, 37, 1387–1394. [Google Scholar] [CrossRef]
  22. Hirata, Y.; Kusunose, K.; Tsuji, T.; Fujimori, K.; Kotoku, J.; Sata, M. Deep Learning for Detection of Elevated Pulmonary Artery Wedge Pressure Using Standard Chest X-Ray. Can. J. Cardiol. 2021, 37, 1198–1206. [Google Scholar] [CrossRef] [PubMed]
  23. Ueda, D.; Matsumoto, T.; Ehara, S.; Yamamoto, A.; Walston, S.L.; Ito, A.; Shimono, T.; Shiba, M.; Takeshita, T.; Fukuda, D.; et al. Artificial intelligence-based model to classify cardiac functions from chest radiographs: A multi-institutional, retrospective model development and validation study. Lancet Digit. Health 2023, 5, e525–e533. [Google Scholar] [CrossRef] [PubMed]
  24. Kobirumaki-Shimozawa, F.; Inoue, T.; Shintani, S.A.; Oyama, K.; Terui, T.; Minamisawa, S.; Ishiwata, S.I.; Fukuda, N. Cardiac thin filament regulation and the Frank–Starling mechanism. J. Physiol. Sci. 2014, 64, 221–232. [Google Scholar] [CrossRef] [PubMed]
  25. Delicce, A.V.; Makaryus, A.N. Physiology, Frank Starling Law. [Updated 2023 Jan 30]. In StatPearls [Internet]; StatPearls Publishing: Treasure Island, FL, USA, 2024. Available online: https://www.ncbi.nlm.nih.gov/books/NBK470295/ (accessed on 17 June 2024).
  26. Abdin, A.; Anker, S.D.; Butler, J.; Coats, A.J.S.; Kindermann, I.; Lainscak, M.; Lund, L.H.; Metra, M.; Mullens, W.; Rosano, G.; et al. ‘Time is prognosis’ in heart failure: Time-to-treatment initiation as a modifiable risk factor. ESC Heart Fail. 2021, 8, 4444–4453. [Google Scholar] [CrossRef] [PubMed]
  27. Ng, K.; Steinhubl, S.R.; deFilippi, C.; Dey, S.; Stewart, W.F. Early Detection of Heart Failure Using Electronic Health Records: Practical Implications for Time Before Diagnosis, Data Diversity, Data Quantity, and Data Density. Circ. Cardiovasc. Qual. Outcomes 2016, 9, 649–658. [Google Scholar] [CrossRef]
  28. Galasko, G.I.; Barnes, S.C.; Collinson, P.; Lahiri, A.; Senior, R. What is the most cost-effective strategy to screen for left ventricular systolic dysfunction: Natriuretic peptides, the electrocardiogram, hand-held echocardiography, traditional echocardiography, or their combination? Eur. Heart J. 2006, 27, 193–200. [Google Scholar] [CrossRef]
  29. Dini, F.L.; Bajraktari, G.; Zara, C.; Mumoli, N.; Rosa, G.M. Optimizing Management of Heart Failure by Using Echo and Natriuretic Peptides in the Outpatient Unit. Adv. Exp. Med. Biol. 2018, 1067, 145–159. [Google Scholar] [CrossRef]
  30. Castiglione, V.; Aimo, A.; Vergaro, G.; Saccaro, L.; Passino, C.; Emdin, M. Biomarkers for the diagnosis and management of heart failure. Heart Fail Rev. 2022, 27, 625–643. [Google Scholar] [CrossRef] [PubMed] [PubMed Central]
Figure 1. The flow diagram for inclusion and exclusion criteria.
Figure 1. The flow diagram for inclusion and exclusion criteria.
Diagnostics 14 01635 g001
Figure 2. The UNET++ architecture for pleural effusion and enlarged cardiac silhouette algorithms (Qure.AI).
Figure 2. The UNET++ architecture for pleural effusion and enlarged cardiac silhouette algorithms (Qure.AI).
Diagnostics 14 01635 g002
Figure 3. On the (left), a patient with a positive heart failure score was diagnosed with heart failure 10 months after CXR. On the (right) is a low-score patient without a diagnosis of heart failure.
Figure 3. On the (left), a patient with a positive heart failure score was diagnosed with heart failure 10 months after CXR. On the (right) is a low-score patient without a diagnosis of heart failure.
Diagnostics 14 01635 g003
Table 1. Contingency table of the four groups depending on the lead time between the radiological examination and the heart failure diagnosis.
Table 1. Contingency table of the four groups depending on the lead time between the radiological examination and the heart failure diagnosis.
<3 months group
(n = 831; negatives:positives, 704:127)
Disease–Ground Truth
PositiveNegative
Model InferencePositive114222
Negative13482
4–6 months group
(n = 805; negatives:positives, 704:101)
Disease–Ground Truth
PositiveNegative
Model InferencePositive86222
Negative15482
7–9 months group
(n = 812; negatives:positives, 704:108)
Disease–Ground Truth
PositiveNegative
Model InferencePositive83222
Negative25482
10–12 months group
(n = 781; negatives:positives, 704:77)
Disease–Ground Truth
PositiveNegative
Model InferencePositive53222
Negative24482
Table 2. Contingency table of the three ejection fraction groups.
Table 2. Contingency table of the three ejection fraction groups.
Low ejection fraction group
(n = 753; negatives:positives, 704:49)
Disease–Ground Truth
PositiveNegative
Model InferencePositive42222
Negative7482
Mildly reduced ejection fraction group
(n = 740; negatives:positives, 704:36)
Disease–Ground Truth
PositiveNegative
Model InferencePositive29222
Negative7482
Normal ejection fraction group
(n = 1019; negatives:positives, 704:315)
Disease–Ground Truth
PositiveNegative
Model InferencePositive257222
Negative58482
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Garza-Frias, E.; Kaviani, P.; Karout, L.; Fahimi, R.; Hosseini, S.; Putha, P.; Tadepalli, M.; Kiran, S.; Arora, C.; Robert, D.; et al. Early Detection of Heart Failure with Autonomous AI-Based Model Using Chest Radiographs: A Multicenter Study. Diagnostics 2024, 14, 1635. https://doi.org/10.3390/diagnostics14151635

AMA Style

Garza-Frias E, Kaviani P, Karout L, Fahimi R, Hosseini S, Putha P, Tadepalli M, Kiran S, Arora C, Robert D, et al. Early Detection of Heart Failure with Autonomous AI-Based Model Using Chest Radiographs: A Multicenter Study. Diagnostics. 2024; 14(15):1635. https://doi.org/10.3390/diagnostics14151635

Chicago/Turabian Style

Garza-Frias, Emiliano, Parisa Kaviani, Lina Karout, Roshan Fahimi, Seyedehelaheh Hosseini, Preetham Putha, Manoj Tadepalli, Sai Kiran, Charu Arora, Dennis Robert, and et al. 2024. "Early Detection of Heart Failure with Autonomous AI-Based Model Using Chest Radiographs: A Multicenter Study" Diagnostics 14, no. 15: 1635. https://doi.org/10.3390/diagnostics14151635

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop