The Performance and Clinical Applicability of HER2 Digital Image Analysis in Breast Cancer: A Systematic Review
Abstract
:Simple Summary
Abstract
1. Introduction
2. Materials and Methods
2.1. Search Strategy
2.2. Study Selection
2.3. Data Extraction and Quality Assessment
2.4. Data Synthesis and Analysis
3. Results
3.1. Criteria for Performance Evaluation
3.2. Characteristics of the Applied Methods
3.3. Dataset Characteristics
3.4. Aspects of Clinical Use Perspectives
4. Discussion
5. Conclusions
Supplementary Materials
Author Contributions
Funding
Acknowledgments
Conflicts of Interest
References
- Sung, H.; Ferlay, J.; Siegel, R.L. Global Cancer Statistics 2020: GLOBOCAN Estimates of Incidence and Mortality Worldwide for 36 Cancers in 185 Countries. CA. Cancer J. Clin. 2021, 71, 209–249. [Google Scholar] [CrossRef] [PubMed]
- Wolff, A.C.; McShane, L.M.; Hammond, M.E.H.; Allison, K.H.; Fitzgibbons, P.; Press, M.F.; Harvey, B.E.; Mangu, P.B.; Bartlett, J.M.S.; Hanna, W.; et al. Human Epidermal Growth Factor Receptor 2 Testing in Breast Cancer: American Society of Clinical Oncology/College of American Pathologists Clinical Practice Guideline Focused Update. Arch. Pathol. Lab. Med. 2018, 142, 1364–1382. [Google Scholar] [CrossRef] [PubMed]
- Bradley, R.; Braybrooke, J.; Gray, R.; Hills, R.; Liu, Z.; Peto, R.; Davies, L.; Dodwell, D.; McGale, P.; Pan, H.; et al. Trastuzumab for Early-Stage, HER2-Positive Breast Cancer: A Meta-Analysis of 13 864 Women in Seven Randomised Trials. Lancet Oncol. 2021, 22, 1139–1150. [Google Scholar] [CrossRef]
- Mitri, Z.; Constantine, T.; O’Regan, R. The HER2 Receptor in Breast Cancer: Pathophysiology, Clinical Use, and New Advances in Therapy. Chemother. Res. Pract. 2012, 2012, 743193. [Google Scholar] [CrossRef] [PubMed]
- Goddard, K.A.B.; Weinmann, S.; Richert-Boe, K.; Chen, C.; Bulkley, J.; Wax, C. HER2 Evaluation and Its Impact on Breast Cancer Treatment Decisions. Public Health Genom. 2011, 15, 1. [Google Scholar] [CrossRef] [PubMed]
- Hammond, M.E.H.; Hayes, D.F.; Dowsett, M.; Allred, D.C.; Hagerty, K.L.; Badve, S.; Fitzgibbons, P.L.; Francis, G.; Goldstein, N.S.; Hayes, M.; et al. American Society of Clinical Oncology/College Of American Pathologists Guideline Recommendations for Immunohistochemical Testing of Estrogen and Progesterone Receptors in Breast Cancer. J. Clin. Oncol. 2010, 28, 2784–2795. [Google Scholar] [CrossRef]
- Gradishar, W.J.; Moran, M.S.; Abraham, J.; Aft, R.; Agnese, D.; Allison, K.H.; Anderson, B.; Burstein, H.J.; Chew, H.; Dang, C.; et al. Breast Cancer, Version 3.2022, NCCN Clinical Practice Guidelines in Oncology. J. Natl. Compr. Canc. Netw. 2022, 20, 691–722. [Google Scholar] [CrossRef] [PubMed]
- Slamon, D.; Leyland-Jones, B.; Shak, S.; Hank, F.; Paton, V.; Harm, P.D.; Bajamonde, A.; Fleming, T.; Eiermann, W.; Wolter, J.; et al. Use of Chemotherapy plus a Monoclonal Antibody against HER2 for Metastatic Breast Cancer That Overexpresses HER2. N. Engl. J. Med. 2001, 344, 783–792. [Google Scholar] [CrossRef]
- Swain, S.M.; Miles, D.; Kim, S.B.; Im, Y.H.; Im, S.A.; Semiglazov, V.; Ciruelos, E.; Schneeweiss, A.; Loi, S.; Monturus, E.; et al. Pertuzumab, Trastuzumab, and Docetaxel for HER2-Positive Metastatic Breast Cancer (CLEOPATRA): End-of-Study Results from a Double-Blind, Randomised, Placebo-Controlled, Phase 3 Study. Lancet. Oncol. 2020, 21, 519–530. [Google Scholar] [CrossRef]
- Kay, E.W.; Walsh, C.J.B.; Cassidy, M.; Curran, B.; Leader, M. C-ErbB-2 Immunostaining: Problems with Interpretation. J. Clin. Pathol. 1994, 47, 816–822. [Google Scholar] [CrossRef]
- Tubbs, R.R.; Pettay, J.D.; Roche, P.C.; Stoler, M.H.; Jenkins, R.B.; Grogan, T.M. Discrepancies in Clinical Laboratory Testing of Eligibility for Trastuzumab Therapy: Apparent Immunohistochemical False-Positives Do Not Get the Message. J. Clin. Oncol. 2001, 19, 2714–2721. [Google Scholar] [CrossRef] [PubMed]
- Rhodes, A.; Jasani, B.; Anderson, E.; Dodson, A.R.; Balaton, A.J. Evaluation of HER-2/Neu Immunohistochemical Assay Sensitivity and Scoring on Formalin-Fixed and Paraffin-Processed Cell Lines and Breast Tumors: A Comparative Study Involving Results from Laboratories in 21 Countries. Am. J. Clin. Pathol. 2002, 118, 408–417. [Google Scholar] [CrossRef] [PubMed]
- Grassini, D.; Cascardi, E.; Sarotto, I.; Annaratone, L.; Sapino, A.; Berrino, E.; Marchiò, C. Unusual Patterns of HER2 Expression in Breast Cancer: Insights and Perspectives. Pathobiology 2022, 89, 278–296. [Google Scholar] [CrossRef]
- Nitta, H.; Kelly, B.D.; Allred, C.; Jewell, S.; Banks, P.; Dennis, E.; Grogan, T.M. The Assessment of HER2 Status in Breast Cancer: The Past, the Present, and the Future. Pathol. Int. 2016, 66, 313–324. [Google Scholar] [CrossRef] [PubMed]
- Rosa, M.; Khazai, L. Comparison of HER2 Testing among Laboratories: Our Experience with Review Cases Retested at Moffitt Cancer Center in a Two-Year Period. Breast J. 2018, 24, 139–147. [Google Scholar] [CrossRef] [PubMed]
- Robbins, C.J.; Fernandez, A.I.; Han, G.; Wong, S.; Harigopal, M.; Podoll, M.; Singh, K.; Ly, A.; Kuba, M.G.; Wen, H.; et al. Multi-Institutional Assessment of Pathologist Scoring HER2 Immunohistochemistry. Mod. Pathol. 2023, 36, 100032. [Google Scholar] [CrossRef]
- Wolff, A.C.; Hammond, M.E.H.; Schwartz, J.N.; Hagerty, K.L.; Allred, D.C.; Cote, R.J.; Dowsett, M.; Fitzgibbons, P.L.; Hanna, W.M.; Langer, A.; et al. American Society of Clinical Oncology/College of American Pathologists Guideline Recommendations for Human Epidermal Growth Factor Receptor 2 Testing in Breast Cancer. J. Clin. Oncol. 2007, 25, 118–145. [Google Scholar] [CrossRef]
- Niazi, M.K.K.; Parwani, A.V.; Gurcan, M.N. Digital Pathology and Artificial Intelligence. Lancet Oncol. 2019, 20, e253–e261. [Google Scholar] [CrossRef]
- Hamilton, P.W.; Bankhead, P.; Wang, Y.; Hutchinson, R.; Kieran, D.; McArt, D.G.; James, J.; Salto-Tellez, M. Digital Pathology and Image Analysis in Tissue Biomarker Research. Methods 2014, 70, 59–73. [Google Scholar] [CrossRef]
- Rojo, M.G.; Bueno, G.; Slodkowska, J. Review of Imaging Solutions for Integrated Quantitative Immunohistochemistry in the Pathology Daily Practice. Folia Histochem. Cytobiol. 2009, 47, 349–354. [Google Scholar] [CrossRef]
- Pantanowitz, L.; Farahani, N.; Parwani, A. Whole Slide Imaging in Pathology: Advantages, Limitations, and Emerging Perspectives. Pathol. Lab. Med. Int. 2015, 2015, 23–33. [Google Scholar] [CrossRef]
- Bankhead, P.; Fernández, J.A.; McArt, D.G.; Boyle, D.P.; Li, G.; Loughrey, M.B.; Irwin, G.W.; Harkin, D.P.; James, J.A.; McQuaid, S.; et al. Integrated Tumor Identification and Automated Scoring Minimizes Pathologist Involvement and Provides New Insights to Key Biomarkers in Breast Cancer. Lab. Investig. 2018, 98, 15–26. [Google Scholar] [CrossRef] [PubMed]
- Qaiser, T.; Mukherjee, A.; Reddy, P.B.C.; Munugoti, S.D.; Tallam, V.; Pitkäaho, T.; Lehtimäki, T.; Naughton, T.; Berseth, M.; Pedraza, A.; et al. HER2 Challenge Contest: A Detailed Assessment of Automated HER2 Scoring Algorithms in Whole Slide Images of Breast Cancer Tissues. Histopathology 2018, 72, 227–238. [Google Scholar] [CrossRef] [PubMed]
- Bast, R.C.J.; Lilja, H.; Urban, N.; Rimm, D.L.; Fritsche, H.; Gray, J.; Veltri, R.; Klee, G.; Allen, A.; Kim, N.; et al. Translational Crossroads for Biomarkers. Clin. Cancer Res. 2005, 11, 6103–6108. [Google Scholar] [CrossRef] [PubMed]
- Stålhammar, G.; Fuentes Martinez, N.; Lippert, M.; Tobin, N.P.; Mølholm, I.; Kis, L.; Rosin, G.; Rantalainen, M.; Pedersen, L.; Bergh, J.; et al. Digital Image Analysis Outperforms Manual Biomarker Assessment in Breast Cancer. Mod. Pathol. 2016, 29, 318–329. [Google Scholar] [CrossRef] [PubMed]
- Roshan, N.M.; Salehabadi, S.; Forooshani, M.K.; Nikoozadeh, A. Accuracy of Digital Image Analysis (Dia) of Borderline Human Epidermal Growth Factor Receptor (Her2) Immunohistochemistry in Invasive Ductal Carcinoma. Int. J. Cancer Manag. 2020, 13, e101179. [Google Scholar] [CrossRef]
- Jakobsen, M.R.; Teerapakpinyo, C.; Shuangshoti, S.; Keelawat, S. Comparison between Digital Image Analysis and Visual Assessment of Immunohistochemical HER2 Expression in Breast Cancer. Pathol. Res. Pract. 2018, 214, 2087–2092. [Google Scholar] [CrossRef]
- Yue, M.; Zhang, J.; Wang, X.; Yan, K.; Cai, L.; Tian, K.; Niu, S.; Han, X.; Yu, Y.; Huang, J.; et al. Can AI-Assisted Microscope Facilitate Breast HER2 Interpretation? A Multi-Institutional Ring Study. Virchows Arch. 2021, 479, 443–449. [Google Scholar] [CrossRef] [PubMed]
- Bui, M.M.; Riben, M.W.; Allison, K.H.; Chlipala, E.; Colasacco, C.; Kahn, A.G.; Lacchetti, C.; Madabhushi, A.; Pantanowitz, L.; Salama, M.E.; et al. Quantitative Image Analysis of Human Epidermal Growth Factor Receptor 2 Immunohistochemistry for Breast Cancer: Guideline From the College of American Pathologists. Arch. Pathol. Lab. Med. 2019, 143, 1180–1195. [Google Scholar] [CrossRef]
- Page, M.J.; McKenzie, J.E.; Bossuyt, P.M.; Boutron, I.; Hoffmann, T.C.; Mulrow, C.D.; Shamseer, L.; Tetzlaff, J.M.; Akl, E.A.; Brennan, S.E.; et al. The PRISMA 2020 Statement: An Updated Guideline for Reporting Systematic Reviews. BMJ 2021, 372, n71. [Google Scholar] [CrossRef]
- Ouzzani, M.; Hammady, H.; Fedorowicz, Z.; Elmagarmid, A. Rayyan-a Web and Mobile App for Systematic Reviews. Syst. Rev. 2016, 5, 210. [Google Scholar] [CrossRef] [PubMed]
- Kabir, S.; Vranic, S.; Mahmood Al Saady, R.; Salman Khan, M.; Sarmun, R.; Alqahtani, A.; Abbas, T.O.; Chowdhury, M.E. The Utility of a Deep Learning-Based Approach in Her-2/Neu Assessment in Breast Cancer. Expert Syst. Appl. 2024, 238, 957–4174. [Google Scholar] [CrossRef]
- Bórquez, S.; Pezoa, R.; Salinas, L.; Torres, C.E. Uncertainty Estimation in the Classification of Histopathological Images with HER2 Overexpression Using Monte Carlo Dropout. Biomed. Signal Process. Control 2023, 85, 104864. [Google Scholar] [CrossRef]
- Mukundan, R. Analysis of Image Feature Characteristics for Automated Scoring of HER2 in Histology Slides. J. Imaging 2019, 5, 35. [Google Scholar] [CrossRef] [PubMed]
- Tewary, S.; Mukhopadhyay, S. AutoIHCNet: CNN Architecture and Decision Fusion for Automated HER2 Scoring. Appl. Soft Comput. 2022, 119, 108572. [Google Scholar] [CrossRef]
- Saha, M.; Chakraborty, C. Her2Net: A Deep Framework for Semantic Segmentation and Classification of Cell Membranes and Nuclei in Breast Cancer Evaluation. IEEE Trans. Image Process. 2018, 27, 2189–2200. [Google Scholar] [CrossRef]
- Mirimoghaddam, M.M.; Majidpour, J.; Pashaei, F.; Arabalibeik, H.; Samizadeh, E.; Roshan, N.M.; Rashid, T.A. HER2GAN: Overcome the Scarcity of HER2 Breast Cancer Dataset Based on Transfer Learning and GAN Model. Clin. Breast Cancer 2024, 24, 53–64. [Google Scholar] [CrossRef]
- Pham, M.D.; Balezo, G.; Tilmant, C.; Petit, S.; Salmon, I.; Hadj, S.B.; Fick, R.H.J. Interpretable HER2 Scoring by Evaluating Clinical Guidelines through a Weakly Supervised, Constrained Deep Learning Approach. Comput. Med. Imaging Graph. 2023, 108, 102261. [Google Scholar] [CrossRef]
- Wu, S.; Yue, M.; Zhang, J.; Li, X.; Li, Z.; Zhang, H.; Wang, X.; Han, X.; Cai, L.; Shang, J.; et al. The Role of Artificial Intelligence in Accurate Interpretation of HER2 Immunohistochemical Scores 0 and 1+ in Breast Cancer. Mod. Pathol. 2023, 36, 100054. [Google Scholar] [CrossRef]
- Che, Y.; Ren, F.; Zhang, X.; Cui, L.; Wu, H.; Zhao, Z. Immunohistochemical HER2 Recognition and Analysis of Breast Cancer Based on Deep Learning. Diagnostics 2023, 13, 263. [Google Scholar] [CrossRef]
- Cordova, C.; Muñoz, R.; Olivares, R.; Minonzio, J.G.; Lozano, C.; Gonzalez, P.; Marchant, I.; González-Arriagada, W.; Olivero, P. HER2 Classification in Breast Cancer Cells: A New Explainable Machine Learning Application for Immunohistochemistry. Oncol. Lett. 2023, 25, 44. [Google Scholar] [CrossRef] [PubMed]
- Yao, Q.; Hou, W.; Wu, K.; Bai, Y.; Long, M.; Diao, X.; Jia, L.; Niu, D.; Li, X. Using Whole Slide Gray Value Map to Predict HER2 Expression and FISH Status in Breast Cancer. Cancers 2022, 14, 6233. [Google Scholar] [CrossRef] [PubMed]
- Tewary, S.; Mukhopadhyay, S. HER2 Molecular Marker Scoring Using Transfer Learning and Decision Level Fusion. J. Digit. Imaging 2021, 34, 667–677. [Google Scholar] [CrossRef] [PubMed]
- Tewary, S.; Arun, I.; Ahmed, R.; Chatterjee, S.; Mukhopadhyay, S. AutoIHC-Analyzer: Computer-Assisted Microscopy for Automated Membrane Extraction/Scoring in HER2 Molecular Markers. J. Microsc. 2021, 281, 87–96. [Google Scholar] [CrossRef] [PubMed]
- Khameneh, F.D.; Razavi, S.; Kamasak, M. Automated Segmentation of Cell Membranes to Evaluate HER2 Status in Whole Slide Images Using a Modified Deep Learning Network. Comput. Biol. Med. 2019, 110, 164–174. [Google Scholar] [CrossRef] [PubMed]
- Yim, K.; Park, H.S.; Kim, D.M.; Lee, Y.S.; Lee, A. Image Analysis of HER2 Immunohistochemical Staining of Surgical Breast Cancer Specimens. Yonsei Med. J. 2019, 60, 158–162. [Google Scholar] [CrossRef]
- Vandenberghe, M.E.; Scott, M.L.J.; Scorer, P.W.; Söderberg, M.; Balcerzak, D.; Barker, C. Relevance of Deep Learning to Facilitate the Diagnosis of HER2 Status in Breast Cancer. Sci. Rep. 2017, 7, 45938. [Google Scholar] [CrossRef] [PubMed]
- Palm, C.; Connolly, C.E.; Masser, R.; Padberg Sgier, B.; Karamitopoulou, E.; Simon, Q.; Bode, B.; Tinguely, M. Determining HER2 Status by Artificial Intelligence: An Investigation of Primary, Metastatic, and HER2 Low Breast Tumors. Diagnostics 2023, 13, 168. [Google Scholar] [CrossRef]
- Koopman, T.; Buikema, H.J.; Hollema, H.; de Bock, G.H.; van der Vegt, B. What Is the Added Value of Digital Image Analysis of HER2 Immunohistochemistry in Breast Cancer in Clinical Practice? A Study with Multiple Platforms. Histopathology 2019, 74, 917–924. [Google Scholar] [CrossRef]
- Pedraza, A.; Gonzalez, L.; Deniz, O.; Bueno, G. Deep Neural Networks for HER2 Grading of Whole Slide Images with Subclasses Levels. Algorithms 2024, 17, 97. [Google Scholar] [CrossRef]
- Kabakçı, K.A.; Çakır, A.; Türkmen, İ.; Töreyin, B.U.; Çapar, A. Automated Scoring of CerbB2/HER2 Receptors Using Histogram Based Analysis of Immunohistochemistry Breast Cancer Tissue Images. Biomed. Signal Process. Control. 2021, 69, 102924. [Google Scholar] [CrossRef]
- Rashid, T.A.; Majidpour, J.; Thinakaran, R.; Batumalay, M.; Dewi, D.A.; Hassan, B.A.; Dadgar, H.; Arabi, H. NSGA-II-DL: Metaheuristic Optimal Feature Selection With Deep Learning Framework for HER2 Classification in Breast Cancer. IEEE Access 2024, 12, 38885–38898. [Google Scholar] [CrossRef]
- Marcuzzo, T.; Giudici, F.; Ober, E.; Rizzardi, C.; Bottin, C.; Zanconati, F. Her2 Immunohistochemical Evaluation by Traditional Microscopy and by Digital Analysis, and the Consequences for FISH Testing. Pathol. Res. Pract. 2016, 212, 911–918. [Google Scholar] [CrossRef] [PubMed]
- Shovon, M.S.H.; Mridha, M.F.; Hasib, K.M.; Alfarhood, S.; Safran, M.; Che, D. Addressing Uncertainty in Imbalanced Histopathology Image Classification of HER2 Breast Cancer: An Interpretable Ensemble Approach With Threshold Filtered Single Instance Evaluation (SIE). IEEE Access 2023, 11, 122238–122251. [Google Scholar] [CrossRef]
- Tran, N.T.; Tran, V.H.; Nguyen, N.B.; Nguyen, T.K.; Cheung, N.M. On Data Augmentation for GAN Training. IEEE Trans. Image Process. 2021, 30, 1882–1897. [Google Scholar] [CrossRef]
- Ménard, S.; Pupa, S.M.; Campiglio, M.; Tagliabue, E. Biologic and Therapeutic Role of HER2 in Cancer. Oncogene 2003, 22, 6570–6578. [Google Scholar] [CrossRef] [PubMed]
- Iqbal, N.; Iqbal, N. Human Epidermal Growth Factor Receptor 2 (HER2) in Cancers: Overexpression and Therapeutic Implications. Mol. Biol. Int. 2014, 2014, 852748. [Google Scholar] [CrossRef] [PubMed]
- Burstein, H.J. The Distinctive Nature of HER2-Positive Breast Cancers. N. Engl. J. Med. 2005, 353, 1652–1654. [Google Scholar] [CrossRef]
- Voigtlaender, M.; Schneider-Merck, T.; Trepel, M. Lapatinib. Recent Results Cancer Res. 2018, 211, 19–44. [Google Scholar] [CrossRef]
- von Minckwitz, G.; Procter, M.; de Azambuja, E.; Zardavas, D.; Benyunes, M.; Viale, G.; Suter, T.; Arahmani, A.; Rouchet, N.; Clark, E.; et al. Adjuvant Pertuzumab and Trastuzumab in Early HER2-Positive Breast Cancer. N. Engl. J. Med. 2017, 377, 122–131. [Google Scholar] [CrossRef]
- Aggarwal, R.; Sounderajah, V.; Martin, G.; Ting, D.S.W.; Karthikesalingam, A.; King, D.; Ashrafian, H.; Darzi, A. Diagnostic Accuracy of Deep Learning in Medical Imaging: A Systematic Review and Meta-Analysis. NPJ Digit. Med. 2021, 4, 65. [Google Scholar] [CrossRef]
- Tafavvoghi, M.; Bongo, L.A.; Shvetsov, N.; Busund, L.-T.R.; Møllersen, K. Publicly Available Datasets of Breast Histopathology H&E Whole-Slide Images: A Scoping Review. J. Pathol. Inform. 2024, 15, 100363. [Google Scholar] [CrossRef] [PubMed]
- Liu, X.; Faes, L.; Kale, A.U.; Wagner, S.K.; Fu, D.J.; Bruynseels, A.; Mahendiran, T.; Moraes, G.; Shamdas, M.; Kern, C.; et al. A Comparison of Deep Learning Performance against Health-Care Professionals in Detecting Diseases from Medical Imaging: A Systematic Review and Meta-Analysis. Lancet Digit. Health 2019, 1, e271–e297. [Google Scholar] [CrossRef] [PubMed]
- Langlotz, C.P.; Allen, B.; Erickson, B.J.; Kalpathy-Cramer, J.; Bigelow, K.; Cook, T.S.; Flanders, A.E.; Lungren, M.P.; Mendelson, D.S.; Rudie, J.D.; et al. A Roadmap for Foundational Research on Artificial Intelligence in Medical Imaging: From the 2018 NIH/RSNA/ACR/The Academy Workshop. Radiology 2019, 291, 781. [Google Scholar] [CrossRef] [PubMed]
- Kohli, M.D.; Summers, R.M.; Geis, J.R. Medical Image Data and Datasets in the Era of Machine Learning—Whitepaper from the 2016 C-MIMI Meeting Dataset Session. J. Digit. Imaging 2017, 30, 392. [Google Scholar] [CrossRef] [PubMed]
- Datasets. Available online: https://spacing.itu.edu.tr/itu-med-datasets (accessed on 14 March 2024).
- Madani, M.; Behzadi, M.M.; Nabavi, S. The Role of Deep Learning in Advancing Breast Cancer Detection Using Different Imaging Modalities: A Systematic Review. Cancers 2022, 14, 5334. [Google Scholar] [CrossRef] [PubMed]
- Willemink, M.J.; Koszek, W.A.; Hardell, C.; Wu, J.; Fleischmann, D.; Harvey, H.; Folio, L.R.; Summers, R.M.; Rubin, D.L.; Lungren, M.P. Preparing Medical Imaging Data for Machine Learning. Radiology 2020, 295, 4–15. [Google Scholar] [CrossRef]
- Rädsch, T.; Reinke, A.; Weru, V.; Tizabi, M.D.; Schreck, N.; Kavur, A.E.; Pekdemir, B.; Roß, T.; Kopp-Schneider, A.; Maier-Hein, L. Labelling Instructions Matter in Biomedical Image Analysis. Nat. Mach. Intell. 2023, 5, 273–283. [Google Scholar] [CrossRef]
- Gibson, E.; Hu, Y.; Huisman, H.J.; Barratt, D.C. Designing Image Segmentation Studies: Statistical Power, Sample Size and Reference Standard Quality. Med. Image Anal. 2017, 42, 44–59. [Google Scholar] [CrossRef]
- Moreno-Barea, F.J.; Jerez, J.M.; Franco, L. Improving Classification Accuracy Using Data Augmentation on Small Data Sets. Expert Syst. Appl. 2020, 161, 113696. [Google Scholar] [CrossRef]
- Althnian, A.; AlSaeed, D.; Al-Baity, H.; Samha, A.; Dris, A.B.; Alzakari, N.; Abou Elwafa, A.; Kurdi, H. Impact of Dataset Size on Classification Performance: An Empirical Evaluation in the Medical Domain. Appl. Sci. 2021, 11, 796. [Google Scholar] [CrossRef]
- Ivanova, M.; Porta, F.M.; D’Ercole, M.; Pescia, C.; Sajjadi, E.; Cursano, G.; De Camilli, E.; Pala, O.; Mazzarol, G.; Venetis, K.; et al. Standardized Pathology Report for HER2 Testing in Compliance with 2023 ASCO/CAP Updates and 2023 ESMO Consensus Statements on HER2-Low Breast Cancer. Virchows Arch. 2024, 484, 3–14. [Google Scholar] [CrossRef]
- Fotouhi, S.; Asadi, S.; Kattan, M.W. A Comprehensive Data Level Analysis for Cancer Diagnosis on Imbalanced Data. J. Biomed. Inform. 2019, 90, 103089. [Google Scholar] [CrossRef] [PubMed]
- Taha, A.A.; Hanbury, A. Metrics for Evaluating 3D Medical Image Segmentation: Analysis, Selection, and Tool. BMC Med. Imaging 2015, 15, 29. [Google Scholar] [CrossRef] [PubMed]
- Müller, D.; Soto-Rey, I.; Kramer, F. Towards a Guideline for Evaluation Metrics in Medical Image Segmentation. BMC Res. Notes 2022, 15, 210. [Google Scholar] [CrossRef]
- Cantaloni, C.; Tonini, R.E.; Eccher, C.; Morelli, L.; Leonardi, E.; Bragantini, E.; Aldovini, D.; Fasanella, S.; Ferro, A.; Cazzolli, D.; et al. Diagnostic Value of Automated Her2 Evaluation in Breast Cancer: A Study on 272 Equivocal (Score 2+) Her2 Immunoreactive Cases Using an FDA Approved System. Appl. Immunohistochem. Mol. Morphol. AIMM 2011, 19, 306–312. [Google Scholar] [CrossRef] [PubMed]
- Tuominen, V.J.; Ylinen, O.; Helin, H.J.; Isola, J. Free Digital Image Analysis Software Helps to Resolve Equivocal Scores in HER2 Immunohistochemistry. Virchows Arch. 2016, 468, 191–198. [Google Scholar] [CrossRef]
- Janowczyk, A.; Madabhushi, A. Deep Learning for Digital Pathology Image Analysis: A Comprehensive Tutorial with Selected Use Cases. J. Pathol. Inform. 2016, 7, 29. [Google Scholar] [CrossRef] [PubMed]
- Suzuki, K. Overview of Deep Learning in Medical Imaging. Radiol. Phys. Technol. 2017, 10, 257–273. [Google Scholar] [CrossRef]
- Duggento, A.; Conti, A.; Mauriello, A.; Guerrisi, M.; Toschi, N. Deep Computational Pathology in Breast Cancer. Semin. Cancer Biol. 2021, 72, 226–237. [Google Scholar] [CrossRef] [PubMed]
- Chen, X.; Wang, X.; Zhang, K.; Fung, K.M.; Thai, T.C.; Moore, K.; Mannel, R.S.; Liu, H.; Zheng, B.; Qiu, Y. Recent Advances and Clinical Applications of Deep Learning in Medical Image Analysis. Med. Image Anal. 2022, 79, 102444. [Google Scholar] [CrossRef] [PubMed]
- Jiang, X.; Hu, Z.; Wang, S.; Zhang, Y. Deep Learning for Medical Image-Based Cancer Diagnosis. Cancers 2023, 15, 3608. [Google Scholar] [CrossRef] [PubMed]
- Castiglioni, I.; Rundo, L.; Codari, M.; Di Leo, G.; Salvatore, C.; Interlenghi, M.; Gallivanone, F.; Cozzi, A.; D’Amico, N.C.; Sardanelli, F. AI Applications to Medical Images: From Machine Learning to Deep Learning. Phys. Medica 2021, 83, 9–24. [Google Scholar] [CrossRef] [PubMed]
- LeCun, Y.; Bengio, Y.; Hinton, G. Deep Learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
- van der Laak, J.; Litjens, G.; Ciompi, F. Deep Learning in Histopathology: The Path to the Clinic. Nat. Med. 2021, 27, 775–784. [Google Scholar] [CrossRef] [PubMed]
- Ker, J.; Wang, L.; Rao, J.; Lim, T. Deep Learning Applications in Medical Image Analysis. IEEE Access 2017, 6, 9375–9379. [Google Scholar] [CrossRef]
- Hou, Y.; Nitta, H.; Li, Z. HER2 Intratumoral Heterogeneity in Breast Cancer, an Evolving Concept. Cancers 2023, 15, 2664. [Google Scholar] [CrossRef]
- Nitta, H.; Li, Z. Breast HER2 Intratumoral Heterogeneity as a Biomarker for Improving HER2-Targeted Therapy. Crit. Rev. Oncog. 2020, 25, 233–240. [Google Scholar] [CrossRef]
- Tanei, T.; Seno, S.; Sota, Y.; Hatano, T.; Kitahara, Y.; Abe, K.; Masunaga, N.; Tsukabe, M.; Yoshinami, T.; Miyake, T.; et al. High HER2 Intratumoral Heterogeneity Is a Predictive Factor for Poor Prognosis in Early-Stage and Locally Advanced HER2-Positive Breast Cancer. Cancers 2024, 16, 1062. [Google Scholar] [CrossRef]
- Tarantino, P.; Hamilton, E.; Tolaney, S.M.; Cortes, J.; Morganti, S.; Ferraro, E.; Marra, A.; Viale, G.; Trapani, D.; Cardoso, F.; et al. HER2-Low Breast Cancer: Pathological and Clinical Landscape. J. Clin. Oncol. 2020, 38, 1951–1962. [Google Scholar] [CrossRef]
- Modi, S.; Park, H.; Murthy, R.K.; Iwata, H.; Tamura, K.; Tsurutani, J.; Moreno-Aspitia, A.; Doi, T.; Sagara, Y.; Redfern, C.; et al. Antitumor Activity and Safety of Trastuzumab Deruxtecan in Patients With HER2-Low-Expressing Advanced Breast Cancer: Results From a Phase Ib Study. J. Clin. Oncol. 2020, 38, 1887–1896. [Google Scholar] [CrossRef] [PubMed]
- Modi, S.; Jacot, W.; Yamashita, T.; Sohn, J.; Vidal, M.; Tokunaga, E.; Tsurutani, J.; Ueno, N.T.; Prat, A.; Chae, Y.S.; et al. Trastuzumab Deruxtecan in Previously Treated HER2-Low Advanced Breast Cancer. N. Engl. J. Med. 2022, 387, 9–20. [Google Scholar] [CrossRef] [PubMed]
- Zhang, H.; Peng, Y. Current Biological, Pathological and Clinical Landscape of HER2-Low Breast Cancer. Cancers 2022, 15, 126. [Google Scholar] [CrossRef] [PubMed]
- FDA. Approves Fam-Trastuzumab Deruxtecan-Nxki for HER2-Low Breast Cancer|FDA. Available online: https://www.fda.gov/drugs/resources-information-approved-drugs/fda-approves-fam-trastuzumab-deruxtecan-nxki-her2-low-breast-cancer (accessed on 12 March 2024).
- Xia, L.Y.; Cao, X.C.; Yu, Y. Survival Outcomes in HER2-Low versus HER2-Zero Breast Cancer after Neoadjuvant Chemotherapy: A Meta-Analysis. World J. Surg. Oncol. 2024, 22, 106. [Google Scholar] [CrossRef]
- Wei, T.; Wang, D.; Gao, S.; Wang, X.; Yue, J.; Kang, Y.; Ju, J.; Yang, Z.; Shuai, Y.; Yuan, P. Clinicopathologic Characteristics and Prognostic Significance of HER2-Low Expression in Patients with Early Breast Cancer: A Systematic Review and Meta-Analysis. Front. Oncol. 2023, 13, 1100332. [Google Scholar] [CrossRef]
Inclusion Criteria | Exclusion Criteria |
---|---|
Experimental or clinical studies on the use of AI in DIA of HER2 in patients with BC Availability of full-text studies Availability of criteria of performance evaluation (image, slide, patch levels) Availability of the components of DIA | Studies outside of the scope of the search Reviews (systematic, scoping) Studies with no criteria for performance evaluation |
Author | Brief Presentation of HER2 Classifying Method | Dataset (Public/Clinical) | Total Number of Cases/Number of Cases in Each Class | Features | Criteria of Evaluation | Key Findings | Limitations |
---|---|---|---|---|---|---|---|
Kabir, 2024 [32] | Three key stages: tumor patch classifier, patch score classifier, and WSI-level score classifier. | Public (Warwick) | 86 WSI (77 in final) /6641 patches
| DL models (DenseNet201, GoogleNet, MobileNet, Vision Transformer—ViTs). | Accuracy: ViTs: 92.6% (tumor patch classifier); patch score classifier: The RF: 91.15% (4 classes, patch), 88% (4 classes, WSI); 96% (3 classes, WSI); DenseNet201: 96.17%. | Annotated dataset (50 WSIs); internal validation; ground truth: 2 pathologists’ assessment. | No clinical dataset;no external validation. |
Bórquez, 2023 [33] | Patch-level classification with different dropout rates and aggregation methods to classify tissue objects. Patch-level predictions were combined for classifying HER2 images at the tissue object level. | Public (Warwick) | 52 WSI
| DL (Bayesian neural networks with Monte Carlo dropout). | Accuracy (4 classes, WS-tissue level): 0.89 on average. | Dataset, labeled previously; balanced dataset; Internal validation (5-fold CV); ground truth: pathologist’s assessment. | No clinical dataset; no external validation. |
Mukundan, 2019 [34] | Characteristic curves for representing the % of staining, rotation-invariant uniform local binary pattern curves as texture descriptors, and a connectedness measure as a morphological feature of the staining patterns. | Public (Warwick) | 52 WSI /4019 image patches
| DL for cell region detection and classification. ML (Logistic regression, SVM) for scoring. | Accuracy (4 classes, patch level): Average = 91% logistic regression algorithm: 93.86% SVM: 89%. | Dataset, labeled previously, balanced; ground truth: pathologist’s IHC assessment; internal validation (CV, 70%:30% images). | No clinical dataset; no external validation. |
Tewary, 2022 [35] | 2 CNN networks were compared with ImmunoMembrane. | Public (Warwick) | 40 WSI of 3 classes (from 52 WSI with 13 cases for 4 classes (0, 1+, 2+, and 3+)). | Transfer learning (Xception); DL CNN (AutoIHCNet); ImmunoMembrane. | Accuracy (3 classes): Patch-based score: Xception—95% AutoIHCNet—96% ROI image-based score: Xception—97% AutoIHCNet—98% ImmunoMembrane—87%. | Dataset, labeled previously; ground truth: pathologist’s IHC assessment; internal validation (train-30 labeled images, test—10 WSIs). | No clinical dataset; no external validation. |
Saha, 2018 [36] | Semantic segmentation of cell membrane and nucleus detection and scoring. | Public (Warwick) 188 for each score, i.e., 0, 1+, 2+, 3+ | 79 WSIs/752 core images
| DL (Her2net—LSTM recurrent network). | Accuracy (4 classes, patch level): 98.33%. | Dataset, labeled previously, balanced dataset; ground truth: pathologists assessment; internal validation (train: 51 WSIs; test: 28 WSIs). | No clinical dataset; no external validation. |
Mirimoghaddam, 2024 [37] | GAN-based model was used for generating high-quality HER2 images to overcome the scarcity of HER2 images; 5 different types of classifiers were used for HER2 classification (MobilenetV2, InceptionV3, InceptionResNetV2, ViT, and Swin-T). | Mixed dataset (Warwick, clinical). | Clinical—126 patients:
| Transfer learning (HER2GAN). | Accuracy (4 classes, patch level, with InceptionResNetV2): 98.8% (Warwick, synthetic train + test sets); 90.5% (Warwick, original train + test sets); 85.71% (clinical dataset, original train + test sets) 92.13% (clinical dataset, synthetic train + test sets). | Labeled datasets; ground truth in a clinical dataset: pathologist’s assessment; internal validation (5-fold CV, 80%:20%). | No external validation; Best accuracy rates were achieved on a fully synthetized dataset based on Warwick. |
Pham, 2023 [38] | An interpretable, weakly supervised constrained deep learning model for HER2 scoring. | Mixed dataset: Warwick, clinical (Erasme), and AIDPATH. |
Clinical (270 WSIs)/Warwick (50 WSIs)
AIDPATH (50 WSIs) 37 negative 6 equivocal 7 positive. | DL | (4 classes, patch level) F1 score: 0.78 Precision in the testing set (0) 0.822 (1+) 0.841 (2+) 0.909 (3+) 0.845 Recall in the testing set (0) 0.839 (1+) 0.905 (2+) 0.937 (3+) 0.938. | Clinical dataset (Erasme): labeled previously, annotated; AIDPATH datasets: labeled previously, annotated. Ground truth based on the clinical outcomes (negative, equivocal, positive). Erasme and AIDPATH were used to train the model to segment all tumor pixels. The patches are randomly split into 80%:10%: 10% for training, validation, and test set. Warwick: labeled previously as a part of the HER2 scoring contest training set. | No accuracy. |
Si Wu, 2023 [39] | The authors conducted 2 rounds of HER2 0 and 1+ assessment. The first ring study (RS1) involved 15 pathologists interpreting 246 HER2 IHC sections via conventional microscopic examination. The second ring study (RS2—pathologist review): pathologists reassessed images with AI assistance using an AI microscope (by embedding an augmented reality module under the microscope eyepiece). The study aimed to improve the accuracy of HER2 0 and 1+ assessment and evaluate the role of AI in assessing low HER2 heterogeneity. | Clinical | 246 cases
(0) 120, (1+) 126. | DL (Microscope with AI). | Accuracy (2 classes (0, 1+), WSI level). RS1 (pathologists review) 0.80 RS2 (pathologists + microscope with AI) 0.93. | Balanced dataset; Annotation was likely conducted for detecting tumor areas; ground truth: pathologist’s IHC assessment; internal validation (validation conducted through a multi-institutional two-round ring study involving 15 pathologists with varying levels of experience). | No external validation. |
Yuxuan Che, 2023 [40] | Binary classification of labeled patches (tumor patch/normal patch), WSI segmentation, scoring by integrated calculation of staining intensity, circumferential membrane staining pattern, and proportion of positive cells. | Clinical | 95 WSI
| DL (ResNet) | Accuracy (4 classes) 73.49% (segmentation, patch level) 97.9% (scoring, WSIs). | Annotation of concentrated tumor areas (23 WSIs); ground truth: pathologist’s IHC assessment; 16 WSI for training, 79 for test. | No external validation. |
Cordova, 2022 [41] | Classification of photomicrographs of HER2 using pathologists’ diagnoses (IHC only) vs. the final diagnosis (IHC + FISH) as training outputs, with applying of an explainability algorithm based on Shapley Additive exPlanations (SHAP) values to determine feature contributions (IHC only vs. IHC + FISH). | Clinical | 131 patient samples and 10 controls (423 photographs) + 30 control samples (with and without tumor). | Supervised ML (logistic regression-based). | Accuracy (2 classes (IHC model and IHC + FISH model), WSI level): 0.88 (IHC model) 0.93 (IHC + FISH). | Dataset, labeled previously; ground truth: IHC+ ISH (previous reporting of pathologist’s diagnosis by IHC and IHC + FISH) 0.65:0.35 a ratio of training/testing sets. | No external validation. |
Qian Yao, 2022 [42] | Predicting HER2 expression level in IHC and HER2 gene status in FISH analysis and comparing two models (GrayMax and its updates model, GrayMap + CNN). | Clinical | 228 biopsy cases of IBC-NST with both IHC and FISH information
| DL (GrayMap+ CNN, GrayMax). | Accuracy (3 classes, WSI level): 95.20% (GrayMap + CNN) 84.19% (GrayMax). | Labeled dataset; ground truth: IHC and FISH. For IHC, manual assessment of 3 “blinded” pathologists (2 times after a 4-week washout); for ISH, 2 pathologists evaluated HER2/CEP17 the HER2 ratios of 20 tumor cells independently and blinded to IHC results; internal validation (5-fold CV). | No external validation. |
Meng Yue, 2021 [28] | 1st ring study: 33 pathologists from 6 hospitals read 50 HER2 WSIs through an online system. 2nd ring study: pathologists read HER2 slides using a conventional microscope. 3rd ring study: the pathologists used our AI microscope (Sunnyoptic ARM50) for assisted interpretation. | Clinical | 50 WSIs consisted of 50% HER2-negative cases and 50% HER2-positive cases, with a total of 25 cases in each category.
| (AI)–assisted microscope Sunnyoptic ARM50: equipped with a conventional microscope and an augmented reality module. | For 3 classes, WSI level “AI”: accuracy κ = 0.86 [95% CI 0.84–0.89] “Pathologist Review” accuracy κ = 0.84 [95% CI 0.82–0.86]. | Labeled and annotated dataset; ground truth: IHC consensus scores of 2 pathologists and a 3rd pathologist for discordant cases. Annotated (identifying and delineating tumor areas as points on the image patches), approx. 500 WSI from the training dataset. | No external validation. |
Tewary, 2021 [43] | Transfer learning is applied using five pre-trained deep learning architectures (VGG16, VGG19, ResNet50, MobileNetV2, and NASNetMobile) with modified output layers for three-class classification. A statistical voting scheme using the mode operator is employed to combine the patch-based scores and generate the final image-based HER2 score. | Public (Warwick) | 40 cases
| Transfer learning; DL | Accuracy (3 classes, patch and image level, VGG19) 0.91 in training; 0.93 in testing (100 epochs) patch-based scoring; 0.98 in image-based scoring. | Balanced, previously labeled dataset; ground truth: IHC pathologist’s assessment; internal validation (30 cases in training = 2130 image patches; 10 test cases =100 images). | Combined classes of HER2 (0/1); no clinical dataset; no external validation. |
Tewary, 2021 * [44] | The approach AutoIHC-Analyzer and a publicly available open-source ImmunoMembrane software were compared with the scores of expert pathologists. | Clinical (from confusion matrix on page 5). | 180 images
| DL (AutoIHC-Analyzer); Classifiers: SVM with Gaussian kernel); ML; ImmunoMembrane. | Accuracy (3 classes) 93%—AutoIHC-Analyzer: 78%—Immuno Membrane: Accuracy. | Labeled dataset; ground truth: IHC score provided by the clinical experts; internal validation (90 images for validation). | Combined classes of HER2 (0/1); no external validation. |
Khameneh, 2019 ** [45] | The authors proposed an approach based on (1) Superpixel-based SVM classifies epithelial/stromal regions, (2) CNN segments membrane areas on epithelial regions, and (3) merged tiles evaluate slide scores. Experimental results compared with state-of-the-art handcraft and deep learning-based approaches. | Mixed: Warwick and clinical. Modified U-Net for classification. | total 127 WSIs Warwick dataset—79 WSIs.
Clinical dataset (from Acibadem)—48 WSIs. | Modified U-Net for classification ML (SVM) for segmenting, classifying and quantifying. DL (CNN) for segmentation. | Accuracy (3 classes, WSI level) 0.87%—classification accuracy 0.9482%—segmentation accuracy. | Warwick dataset: labeled previously, ground truth: FISH and HER2 IHC (pathologist’s assessment); used for testing. Clinical dataset (from Acibadem): annotated (on tumor areas, cell membrane staining patterns, epithelial and stromal regions); ground truth: pathologist’s assessment; used for training. | Combined classes of HER2 (0/1) |
Kwangil Yim, 2019 [46] | The results of the HER2 image analysis software (Companion Algorithm HER2 (4B5) image analysis software (Roche) compared with the manual scoring method and with HER2 SISH results (as the gold standard)). Previously, the authors found that at least 1000 tumor cells need to be examined in the most strongly stained areas (foci of view). | Clinical | 555 patients in main research: SISH: (negative) 451, (positive) 104.
32 HER2 2+ for preliminary test. | Companion Algorithm HER2 (4B5) image analysis software (Roche). | Accuracy (4 classes, foci of view level) 91.7%—manual scoring; 90.8%—image analysis. | Preliminary research resulted in using the approach of analyzing a certain area (40,457.64 μm2) until FOVs with at least 1000 tumor cells were assessed. Pathologists selected areas (“foci of view” (FOVs)) for further analysis. FOVs were chosen based on the following criteria: intense, thick, and complete membrane staining in the HER2 IHC-stained breast cancer specimens; ground truth: based on HER2 SISH. | No external validation. |
Vandenberghe, 2017 [47] | Results of ConvNets for HER2 cell assessment were evaluated and compared to classical machine learning techniques (hand-crafted features + LSVM; hand-crafted features + RF). | Private dataset: from the AstraZeneca BioBank or acquired from a commercial provider (Dako Denmark A/S). | 71 WSI Negative—43 Equivocal—17 Positive—11. | DL (ConvNets—Custom CNN). | Accuracy (four classes, WSI level) ConvNets 78% overall accuracy. | Annotated dataset (A total of 12,200 cells from a subset of 18 WSIs) was manually annotated by extracting 18 biologically relevant features (cell morphology, nuclear color, texture, and HER2 membrane staining), training classical machine learning models. The dataset was annotated using Definiens Developer XD for cell detection and feature extraction). Ground truth: manual annotation of cell features and HER2 scoring; internal validation (10-fold CV). | No external validation. |
Palm, 2023 [48] | Results of groups (“pathologists” and “pathologists + AI”) were compared with AI results and a ground truth. | Clinical | a preliminary cohort: 495 newly diagnosed primary IBCs and their 30 metastatic BC (475 in total):
a study cohort, 97 (all 30 metastatic tumors and their matched primaries and a further random selection of primary tumors from the preliminary cohort (67 primary tumors)). ISH on 55/97 samples of all cases with an IHC HER2 score of ≥1+ 26/67 from primary tumors with IHC 1+ or above were assessed by ISH. | HER2 4B5 algorithm in the uPath enterprise software (Roche Diagnostic International, Rotkreuz, Switzerland). | Sensitivity/specificity (slide level): 93.8%/96.1% for the IHC algorithm 100%/94.7% for the ISH algorithm. | Ground truth: consensus in the pathologists’ opinions. For IHC, a manual consensus score of three pathologists. For equivocal results of ISH (in HER2 2+), recounting the ISH signals of 20 cells by a second pathologist. As a result of testing the AI-IHC algorithm on the preliminary cohort, changes in incubation and counterstaining time of the automated slide stainer were applied. The adjusted protocol was used for a newly prepared study cohort. | No accuracy; no external validation. |
Koopman, 2019 [49] | HER2 image analysis was compared between two independent platforms (Visiopharm Integrator System (Denmark) and HALO (USA) for inter-platform agreement, as well as with the manual score. | Clinical | 152 Resection specimens of consecutive primary invasive breast carcinomas. 136 ISH Negative 16 ISH Positive
| Visiopharm; HALO | Sensitivity/specificity (3 classes, slide level): Visiopharm: 81.3%/100% HALO: 100%/100%. | Ground truth: manual scoring by two independent pathologists and ISH in 2+ cases. | Combined classes of HER2 (0/1); no external validation. |
Pedraza, 2024 [50] | Color transfer for data augmentation was employed on the initial dataset (DS1) to create a new dataset (DS2) with five classes: background, 0, 1, 2+, and 3+. Additionally, a separate dataset (DS3) was created with seven classes, including 1.5+ and 2.5+. The results from DS3 were then merged back into five classes for comparison. Multiple CNNs were applied for patch-wise grading of HER2. | AIDPATH | 306 WSIs from 153 BC from 3 centers: 172 WSI from NHS (Warwick); 104 WSI from SESCAM; 30 WSI from SAS
| DL: five different CNNs (AlexNet, GoogleNet, VGG, ResNet-101, DenseNet-201). | Average accuracy = 97% for DenseNet-201 on DS2 (dataset 2–5 classes: background, 0, 1, 2, 3; balanced, augmented)
Best accuracy (4 classes, WSIs)— ResNet-101 applied to DS3 dataset with 7 classes (dataset 3–7 classes: background, 0,1, 1.5, 2, 2.5, 3)
| Previously labeled dataset; ground truth: at least 2 pathologists’ scoring; 70%:20%:10% WSI for training, validation, and as a hold-out test set. | No external validation; no clinical dataset. |
Kabakçı 2021 [51] | Hybrid Cell Detection and Membrane Intensity Histogram Extraction methods were sequentially used for HER2 scoring, with testing on public and clinical datasets, and results were compared with ImmunoMembrane. | Mixed (clinical: ITU-MED-1, ITU-MED-2; Warwick). | ITU-MED-1: 13 cases/191 tissue images:
ITU-MED-2: 10 cases/148 tissue images:
| DL (LSTM); ML (k-Nearest Neighbors (kNN), Decision tree classifiers) for classification. | Accuracy (4 classes, patch-based) 91.43% ITU-MED-1, Best validation accuracy: 88.01% (LSTM), Best tissue-based scoring accuracy: 91.43% (Ensemble Boosted Trees); Compared with 74.07%. (ImmunoMembrane); ITU-MED-2, Best validation accuracy: 88.88% (LSTM), Best tissue-based scoring accuracy: 90.19% (Ensemble Boosted Trees, Ensemble Bagged Trees. Weighted kNN); Compared with 80.39% (ImmunoMembrane). | Ground truth: labeling by expert pathologists + FISH; The ITU-MED-1: 105 images for training, 86 for testing; the ITU-MED-2: 96 images for training, 52 for testing; Warwick: 51 WSI for training, 25 for testing. | |
Rashid, 2024 [52] | A combination of a transfer learning model (ResNet50) for feature extraction, a metaheuristic optimizer (NSGA-II) for selecting the most relevant features, and a machine learning algorithm (SVM) for classification was applied and tested on two datasets. | Mixed (Warwick, clinical) | Warwick (HER2SC) 79WSI clinical—126 individuals (HER2GAN):
| Transfer Learning Model (an ML strategy)—Resnet50; NSGA-II algorithm; SVM classifier. | Best accuracy (four classes, patch level) (Resnet50 + NSGA-II + SVM): 94.4% on HER2SC; 90.75% on HER2GAN. | 5-fold CV, 80%:20%. | No external validation for either dataset. |
Roshan, 2020 [26] | Digital image analysis using a free web application. | Clinical | (2) 60 samples/307 images. | ImmunoMembrane. | Accuracy (class “2+”, patch level) 86%. | Ground truth: manual IHC scoring by 2 pathologists and ISH. | No external validation. |
Marcuzzo, 2016 [53] | Surgical samples and core biopsies were prepared for digital analysis by VISIA Imaging, and results were compared with FISH results. | Clinical | 176 cases: 132 (75%) surgical specimens 44 (25%) biopsies. Negative (1+/0): 23 Equivocal (2+): 85 Positive (3+): 44 Inadequate: 24. | Specific software package: VISIA Imaging s.r.l. software (version 2.5.0.1, San Giovanni Valdarno, Italy). | Sensitivity/specificity (3 classes, WSI level) 100%/82%. | Comparison of results between types of specimen, staining distribution was done. | Combined classes of HER2 (0/1); no external validation. |
Shovon, 2023 [54] | Several popular deep-learning architectures were employed for feature extraction and classification. Various activation functions were utilized to achieve better results. The classification results of the model trained on H&E and IHC images were compared. | Public | BCI dataset: 4870 image pairs with a resolution of 1024×1024 of H&E and IHC, equal number images of HER2 0, 1+, 2+, 3+. | DL, ML: DenseNet201- Xception-SIE (single instance evaluation) (with the best performance); InceptionResNetV2; VGG16; VGG19; ResNet101; ResNet152V2; EfficientNetB7; InceptionV3. | Best accuracy (4 classes, patch level) DenseNet201- Xception-SIE: 97.56% (on IHC data) 97.12% (on H&E data). | 3896/977 images of H&E and IHC for training and validation. | No external validation; no clinical dataset. |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Dunenova, G.; Kalmataeva, Z.; Kaidarova, D.; Dauletbaev, N.; Semenova, Y.; Mansurova, M.; Grjibovski, A.; Kassymbekova, F.; Sarsembayev, A.; Semenov, D.; et al. The Performance and Clinical Applicability of HER2 Digital Image Analysis in Breast Cancer: A Systematic Review. Cancers 2024, 16, 2761. https://doi.org/10.3390/cancers16152761
Dunenova G, Kalmataeva Z, Kaidarova D, Dauletbaev N, Semenova Y, Mansurova M, Grjibovski A, Kassymbekova F, Sarsembayev A, Semenov D, et al. The Performance and Clinical Applicability of HER2 Digital Image Analysis in Breast Cancer: A Systematic Review. Cancers. 2024; 16(15):2761. https://doi.org/10.3390/cancers16152761
Chicago/Turabian StyleDunenova, Gauhar, Zhanna Kalmataeva, Dilyara Kaidarova, Nurlan Dauletbaev, Yuliya Semenova, Madina Mansurova, Andrej Grjibovski, Fatima Kassymbekova, Aidos Sarsembayev, Daniil Semenov, and et al. 2024. "The Performance and Clinical Applicability of HER2 Digital Image Analysis in Breast Cancer: A Systematic Review" Cancers 16, no. 15: 2761. https://doi.org/10.3390/cancers16152761
APA StyleDunenova, G., Kalmataeva, Z., Kaidarova, D., Dauletbaev, N., Semenova, Y., Mansurova, M., Grjibovski, A., Kassymbekova, F., Sarsembayev, A., Semenov, D., & Glushkova, N. (2024). The Performance and Clinical Applicability of HER2 Digital Image Analysis in Breast Cancer: A Systematic Review. Cancers, 16(15), 2761. https://doi.org/10.3390/cancers16152761