Next Article in Journal
A Survey on Explainable Artificial Intelligence (XAI) Techniques for Visualizing Deep Learning Models in Medical Imaging
Next Article in Special Issue
Joint Image Processing with Learning-Driven Data Representation and Model Behavior for Non-Intrusive Anemia Diagnosis in Pediatric Patients
Previous Article in Journal
Efficient End-to-End Convolutional Architecture for Point-of-Gaze Estimation
Previous Article in Special Issue
Enhancing Deep Learning Model Explainability in Brain Tumor Datasets Using Post-Heuristic Approaches
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Comparison of Visual and Quantra Software Mammographic Density Assessment According to BI-RADS® in 2D and 3D Images

1
Private Practice, 44123 Ferrara, Italy
2
AULSS 9 Scaligera, Breast Unit, Department of Radiology, 37122 Verona, Italy
3
ASL ROMA 3, Department of Radiology, 00125 Rome, Italy
4
UOC di Radiologia Toracica e Cardiovascolare, Dipartimento di Diagnostica per Immagini, Radioterapia Oncologica ed Ematologia, Fondazione Policlinico Universitario Agostino Gemelli IRCCS, 00168 Rome, Italy
5
Facoltà di Medicina e Chirurgia, Università Cattolica Sacro Cuore, Largo F. Vito 1, 00168 Rome, Italy
6
CIMEC-Center for Mind/Brain, University of Trento, 38068 Rovereto, Italy
7
Private Practice, 37122 Verona, Italy
*
Author to whom correspondence should be addressed.
J. Imaging 2024, 10(9), 238; https://doi.org/10.3390/jimaging10090238
Submission received: 5 July 2024 / Revised: 16 September 2024 / Accepted: 20 September 2024 / Published: 23 September 2024

Abstract

:
Mammographic density (MD) assessment is subject to inter- and intra-observer variability. An automated method, such as Quantra software, could be a useful tool for an objective and reproducible MD assessment. Our purpose was to evaluate the performance of Quantra software in assessing MD, according to BI-RADS® Atlas Fifth Edition recommendations, verifying the degree of agreement with the gold standard, given by the consensus of two breast radiologists. A total of 5009 screening examinations were evaluated by two radiologists and analysed by Quantra software to assess MD. The agreement between the three assigned values was expressed as intraclass correlation coefficients (ICCs). The agreement between the software and the two readers (R1 and R2) was moderate with ICC values of 0.725 and 0.713, respectively. A better agreement was demonstrated between the software’s assessment and the average score of the values assigned by the two radiologists, with an index of 0.793, which reflects a good correlation. Quantra software appears a promising tool in supporting radiologists in the MD assessment and could be part of a personalised screening protocol soon. However, some fine-tuning is needed to improve its accuracy, reduce its tendency to overestimate, and ensure it excludes high-density structures from its assessment.

1. Introduction

Mammographic density (MD) refers to the changes in the radiological appearance of breast parenchyma linked to the variability in its composition, specifically in the amount of fatty and fibroglandular tissues. MD represents a double risk factor: firstly, high MD promotes the onset of certain types of breast cancer (BC) and, secondly, it produces a masking effect on screening mammograms that can hinder an early diagnosis of the disease [1].
It has been estimated that women with very dense breasts present a higher risk of developing BC of 4–6 times those with low BD [2,3,4,5,6]. The parenchyma components vary throughout a woman’s life, especially in the postmenopausal period, in which there is an involution of fibroglandular tissue [7]. Several studies showed that a great number of women when screened have an MD that belongs to higher categories [7,8].
It is, therefore, necessary that MD assessment is accurate and reproducible over time as this would enable clinicians to tailor the screening process to suit the patient’s needs and to add, if appropriate, complementary screening tests such as ultrasound (US) [9], magnetic resonance imaging (MRI) [10], or digital breast tomosynthesis (DBT) [11] (when not already integrated into the screening test) to improve the evaluation of the dense breasts and the detection of cancer. In this regard, it is emblematic that the United States affirmed the importance of this assessment with the introduction of breast density notification laws, by which physicians are required to inform patients about their MD and its implications.
In addition, breast density might be a surrogate indicator of treatment response: a decrease in MD after tamoxifen therapy is linked to a lower risk of primary BC, recurrence, and mortality [12,13,14].
Various methods based on qualitative, semi-automated, or fully automated approaches are available to measure MD on mammograms [15].
The first approach is mainly based on the American College of Radiology’s Breast Imaging Reporting and Data System (BI-RADS) Fifth Edition guidelines. These recommendations are routinely used by breast radiologists, who conduct a visual assessment to assign a density category.
According to the BI-RADS guidelines, MD is classified into four groups based on the relative amounts of adipose versus fibroglandular tissue, which are: (a) almost entirely fatty, (b) scattered fibroglandular density, (c) heterogeneously dense, and (d) extremely dense [16].
In semi-automated methods, the radiologist outlines the dense area, and a computer program calculates the MD.
Regarding the last approach, there are two main types of fully automated techniques: those that measure the percentage of MD based on the area (i.e., MedDensity and Autodensity) and those that perform a volumetric assessment of breast density, such as QuantraTM (Hologic Inc., Bedford, MA, USA) and VolparaTM (Volpara Solutions, Mãkatina Company, Wellington, New Zealand) [15]. Specifically, Quantra is based on a calculation algorithm, powered by artificial intelligence, that allows for a volumetric estimate of MD as a function of the density of the fibroglandular tissue and its distribution [17,18,19].
In all of the above, it has been demonstrated that MD is a risk factor for BC [5,20].
Nevertheless, visual assessment is affected by a subjective component in the interpretation, which implicates a significant variability between different readers (inter-observer variability) and with respect to the same reader over time (intra-observer variability) that could result in a discrepancy in the MD categorization [21,22,23]. The reader’s ability to accurately assess MD is probably influenced by two key factors: their experience and their ongoing commitment to professional development. A study has shown that readers who are actively engaged in continuing education and specialise in breast imaging exhibit minimal to no differences in their measurements, both in terms of consistency inter- and intra-observer, having instead only a moderate concordance with the software [16]. Conversely, another study pointed out that the greatest inter-observer agreement was among radiologists with the longest experience in breast imaging, underlining its importance, and, also, that the greatest disagreement was on the assignment of the most dense categories (C and D) [24].
Also, semi-automated techniques involve user interaction that still introduces subjectivity besides requiring additional complex software, further burdening the already overloaded radiology reporting workflow [25].
Moreover, knowledge and awareness of MD positively affect women’s approach to screening programs: women who are informed for the first time about their MD category intend to continue screening and conduct the necessary complementary examinations. Additionally, even women who have been informed several times about their MD continue to prioritise regular screening and healthy habits [26].
The points mentioned above underscore the necessity for standardised MD assessment, thereby enabling its inclusion as a parameter in clinical decision-making. It is precisely for this purpose that fully automatic systems have been devised.
This paper aims to evaluate the performance of breast density calculation software, verifying the degree of agreement with the gold standard, given by the consensus of two screening readers.
To this end, the density classes assigned by the two breast radiologists and the automatic software will be considered.

2. Materials and Methods

2.1. Study Population and Data Collection

The examinations were collected from women aged 50–74 years during the screening setting. A total of 5009 exams performed from February to April 2022 in two departments of UOSD Breast Unit ULSS 9 Scaligera, Ospedale Fracastoro in San Bonifacio and Ospedale di Marzana, were revised and included in this study. Participating women gave written informed consent.
The average age of the patients included in the study was 61 years (SD 7.3). The average age of patients with non-dense breast classification “A/B” was around 62 years (specifically, 61.6 years for category “A” and 62.6 years for category “B”) and with dense breast classification “C/D” was around 60 years (specifically, 60.6 years for category “C” and 57.2 years for category “D”).
Three MD assessments were assigned to these screening tests: two by breast radiologists, with 3 and 10 years of experience, respectively, and one by Quantra software. MD assessment data were collected in three groups, referred to as follows: Reader 1, Reader 2, and Software.

2.2. Image Acquisition

The examinations were acquired with a Selenia Dimensions Hologic mammography system and were composed of digital breast tomosynthesis acquisitions combined with two-dimensional synthetic mammography (2DSM). This system sent the raw images to the software server and the processed images to PACS for radiologists’ reporting. The server transmitted the images to the algorithm, which analysed them and returned the results in .xml format to the software’s server. Finally, the software’s server generated the results in DICOM format.

2.3. Software for Automatic Breast Density Assessment

Quantra™ 2.2.3 breast density assessment software (Hologic Inc.) was used for the automatic analysis of the images.
This software analysed synthetic 2D and 3D images in 4 projections in DICOM format: right cranio-caudal, oblique mid-lateral, left cranio-caudal, and oblique mid-lateral. If the examination included multiple identical projections, e.g., two right cranio-caudal, just the second projection was considered for the assessment (Figure 1). Projections not recognized by the software were excluded.
Quantra 2.2.3 software produces an estimate of the MD based on the distribution and density of the parenchymal tissue using calculation algorithms provided by artificial intelligence.
It objectifies the MD assessment according to BI-RADS Fifth Edition recommendations. The system provides real-time visualisation of the assessment on the acquisition monitor, alongside the processed images, enabling customised diagnostic protocols.
Quantra analyses pixel data to quantify the characteristics of density and distribution of fibroglandular tissue. The characteristics are based on the variation in pixel values within the breast. The algorithm evaluates the density through a volumetric calculation of the amount of fibroglandular tissue present in each pixel. First, the software produces an estimate of the thickness of glandular tissue for each pixel in the image and calculates the total volume of fibroglandular tissue in the breast. Using a similar process, Quantra considers the contour of the breast on the image to compensate for unevenly compressed breast regions and to calculate the entire breast volume. The volumetric percentage of fibroglandular tissue is given by the ratio [27]:
V o l u m e t r i c d e n s i t y = V o l u m e f i b r o g h i a n d u l a r t i s s u e V o l u m e w h o l e b r e a s t
Following BI-RADS Fifth Edition guidelines, the software evaluates the distribution of the fibroglandular tissue present. Subsequently, this information is used by a classification scheme to assess the density category of each breast.
The classification model is programmed using images previously classified by experienced radiologists to recognize densities correctly (Figure 2, Figure 3, Figure 4 and Figure 5).
Quantra assigns density categories A, B, C, or D (Figure 6) to each breast and gives an estimate of breast density based on the distribution of fibroglandular tissue.
The final classification is given by the densest breast: for example, if the software has assigned a “B” to the right breast and a “C” to the left one, the final result would be a “C”.

2.4. Intraclass Correlation Coefficients

An analysis of the values assigned by breast radiologists and Quantra was conducted to extract the intraclass correlation coefficients (ICCs), which compare the different groups.
ICC represents a reliability index that indicates both correlation and agreement/reproducibility’ degree between measurements. It can be expressed as a value ranging between 0 and 1, where 1 represents a perfect match between measurements, and the different subranges suggest various levels of reliability [28]:
<0.50: poor correlation;
0.50–0.75: moderate correlation;
0.75–0.90: good correlation;
>0.90: excellent correlation.

3. Results

Quantra software is intended as a useful tool in supporting breast radiologists in the assessment of breast density, by eliminating humans’ subjective interpretation errors.
The MD values assigned by Reader 1 (R1), Reader 2 (R2), the mean of the two readers (R1-R2), and the Quantra Software (R3) are reported in Table 1 and Chart 1.
Firstly, a comparison between the density assessment of reader 1 (R1) and the Quantra software (R3) for the same mammographic images was made and, subsequently, the correlation index was calculated. The resulting ICC was 0.725, which represents a moderate correlation between measurements. The same method was used to compare the assessment between reader 2 (R2) and the software, and, similarly, there was a moderate correlation with an index of 0.713. The correlation was higher when the comparison was made between the assessment of the software and the consensus of the two readers, indeed, the obtained index was 0.793, which reflects a good correlation. Nevertheless, some differences have also been observed in the visual assessment of the two radiologists: the ICC was 0.712, reflecting a moderate correlation. The ICC for each comparison is reported in Table 2.
As reported in Table 1 and Table 2, when considering the “B” category of breast composition (scattered fibroglandular density), there was a high correspondence between the Quantra software and human readers, as confirmed by the ICC. Conversely, the software assigned the non-dense “A” category (almost entirely fatty) to considerably fewer cases than the radiologists. This trend was reversed in the denser categories, i.e., “C” (heterogeneously dense) and “D” (extremely dense), with Quantra having assigned them more than the human readers.
To elucidate the discrepancies in density assessment between the radiologists and software for the assignment of the “D” category, a thorough review of the relevant cases was conducted. The analysis revealed that, in some instances, the discrepancy was attributed to the presence of high-density structures, such as calcifications or implantable devices (e.g., breast implants, pacemakers), overlapped with the breast parenchyma, that the software was unable to accurately recognize. In these cases, these structures acquired a greater value than the breast tissue in Quantra’s calculation formula, and the resulting assigned category was erroneously the highest, i.e., “D” (Figure 7 and Figure 8).

4. Discussion and Conclusions

An artificial intelligence tool, such as Quantra software, could be a useful tool for an objective and reproducible density assessment compared with the visual assessment provided by human readers, as it is not linked to experience or updating skills.
One potential drawback of introducing this software in clinical practice concerns the possibility that it can give a different assessment to human readers, potentially leading to overtesting and, therefore, higher healthcare costs, and probably, limited insurance coverage in some countries. Nevertheless, our results are promising. In fact, this study showed that the Quantra 2.2.3 software has a moderate agreement with individual readers and a good agreement with the consensus of two breast radiologists, as can be inferred from the intraclass correlation indices, on the analysis of 5009 examinations performed during the screening.
To the best of our knowledge, most of the studies in the literature have investigated software from different manufacturers or different versions of one software, older versions of the BI-RADS system for MD classification, or different statistical analysis methods, making it difficult to compare their findings to our results directly. For instance, in 2012, Ciatto et al. studied version 1.3 of Quantra software, suggesting the value of ≤22% as cut off to best reproduce visual classification, represented by a reading panel of 11 radiologists, in discriminating dense and non-dense breasts [29]. Similarly, in 2015, Pahwa et al. reported a cut-off of 19.5% for the same purpose [30]. A study was conducted to compare radiologists and Quantra 2.2 software assessment. This research was part of the evidence submitted to the FDA for approval of the software in the United States and the results showed that the overall agreement in the analysis of tomosynthesis images was 72%, lower than for conventional 2D images, but in the range of inter-reader variability [31]. In another published study, the reported agreement between human readers was substantial to almost perfect, whilst between the readers or the consensus of the readers and Quantra software was found to be moderate to substantial [32].
Additionally, our study has evidenced that the correspondence between the visual assessment and Quantra software assessment is greater in some categories rather than in others: in fact, the greater concordance was only for category “B”, while the software tends to overestimate the “dense” classes “C” and “D” when compared with radiologists.
Therefore, in a personalised screening setting in which patients require additional workup with ultrasound if the software automatically assigns a “D”, this overestimation could lead to an increased number of unnecessary recalls, raising concerns about its sustainability in clinical practice.
We encountered some limitations in this study. During the exams’ revision, we observed some errors in the assessment; in particular, the software was affected by the presence of high-density structures such as calcifications, breast implants, pacemakers, and other devices overlapped with the breast parenchyma. Indeed, these structures attenuate X-rays significantly more than the breast tissue. The algorithm was unable to distinguish between the different densities and their distribution. For this reason, if these structures were present, the software assigned the highest density, “D”. Consequently, we hope that before its introduction into clinical practice, Quantra can be updated to recognize and exclude such structures from the assessment. Alternatively, the presence of these structures in the breast parenchyma should be an exclusion criterion from software analysis.
Furthermore, ICC takes into account the correlation between multiple measures and, also, expresses a degree of reliability of the measurement, which indicates the level of confidence in the replicability of the measurement itself [28]. However, in the present dataset, given the subjective nature of the classification, which is grounded on human assessment, the rating may be misleading as potentially biased by misclassification. In this sense, computing the distance proposed assumes a certain degree of confidence in a ground-truth classification, which is not possible to define by design. To have a better insight into the results here proposed, additional metrics should be considered and integrated into future research.
A further limitation found during data collection was the lack of some density assessments by radiologists; nevertheless, as these data gaps were limited, their weight was negligible.
Finally, this was a monocentric study: further research in other radiology departments would be useful to confirm the accuracy of our data.
Overall, Quantra software appears to be a promising tool that could be part of a personalised screening protocol in the near future. However, before we can fully rely on it, it needs some fine-tuning. Specifically, its accuracy needs to be improved, its tendency to overestimate needs to be reduced, and its ability to correctly identify and exclude high-density structures from the assessment needs to be ensured.

Author Contributions

Conceptualization, G.R.; Data curation, C.R., F.F. and G.R.; Writing—original draft, F.M., C.M., A.S., S.T. and A.G.T.; Writing—review & editing, F.M., R.R., M.C. and G.R.; Supervision, O.T., P.B. and A.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by the Ethics Committee of Verona and Rovigo (protocol code 55803, 27 September 2021).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author/s.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Lian, J.; Li, K. A Review of Breast Density Implications and Breast Cancer Screening. Clin. Breast Cancer 2020, 20, 283–290. [Google Scholar] [CrossRef] [PubMed]
  2. Harvey, J.A.; Bovbjerg, V.E. Quantitative Assessment of Mammographic Breast Density: Relationship with Breast Cancer Risk. Radiology 2004, 230, 29–41. [Google Scholar] [CrossRef] [PubMed]
  3. Vacek, P.M.; Geller, B.M. A Prospective Study of Breast Cancer Risk Using Routine Mammographic Breast Density Measurements. Cancer Epidemiol. Biomark. Prev. Publ. Am. Assoc. Cancer Res. Cosponsored Am. Soc. Prev. Oncol. 2004, 13, 715–722. [Google Scholar]
  4. Boyd, N.F.; Guo, H.; Martin, L.J.; Sun, L.; Stone, J.; Fishell, E.; Jong, R.A.; Hislop, G.; Chiarelli, A.; Minkin, S.; et al. Mammographic Density and the Risk and Detection of Breast Cancer. N. Engl. J. Med. 2007, 356, 227–236. [Google Scholar] [CrossRef]
  5. McCormack, V.A.; dos Santos Silva, I. Breast Density and Parenchymal Patterns as Markers of Breast Cancer Risk: A Meta-Analysis. Cancer Epidemiol. Biomark. Prev. Publ. Am. Assoc. Cancer Res. Cosponsored Am. Soc. Prev. Oncol. 2006, 15, 1159–1169. [Google Scholar] [CrossRef] [PubMed]
  6. Boyd, N.F.; Rommens, J.M.; Vogt, K.; Lee, V.; Hopper, J.L.; Yaffe, M.J.; Paterson, A.D. Mammographic Breast Density as an Intermediate Phenotype for Breast Cancer. Lancet Oncol. 2005, 6, 798–808. [Google Scholar] [CrossRef]
  7. Checka, C.M.; Chun, J.E.; Schnabel, F.R.; Lee, J.; Toth, H. The Relationship of Mammographic Density and Age: Implications for Breast Cancer Screening. AJR Am. J. Roentgenol. 2012, 198, W292–W295. [Google Scholar] [CrossRef]
  8. Liao, Y.-S.; Zhang, J.-Y.; Hsu, Y.-C.; Hong, M.-X.; Lee, L.-W. Age-Specific Breast Density Changes in Taiwanese Women: A Cross-Sectional Study. Int. J. Environ. Res. Public. Health 2020, 17, 3186. [Google Scholar] [CrossRef]
  9. Berg, W.A.; Blume, J.D.; Cormack, J.B.; Mendelson, E.B.; Lehrer, D.; Böhm-Vélez, M.; Pisano, E.D.; Jong, R.A.; Evans, W.P.; Morton, M.J.; et al. Combined Screening with Ultrasound and Mammography vs Mammography Alone in Women at Elevated Risk of Breast Cancer. JAMA 2008, 299, 2151–2163. [Google Scholar] [CrossRef]
  10. Berg, W.A.; Zhang, Z.; Lehrer, D.; Jong, R.A.; Pisano, E.D.; Barr, R.G.; Böhm-Vélez, M.; Mahoney, M.C.; Evans, W.P.; Larsen, L.H.; et al. Detection of Breast Cancer With Addition of Annual Screening Ultrasound or a Single Screening MRI to Mammography in Women With Elevated Breast Cancer Risk. JAMA 2012, 307, 1394–1404. [Google Scholar] [CrossRef]
  11. Rafferty, E.A.; Park, J.M.; Philpotts, L.E.; Poplack, S.P.; Sumkin, J.H.; Halpern, E.F.; Niklason, L.T. Assessing Radiologist Performance Using Combined Digital Mammography and Breast Tomosynthesis Compared with Digital Mammography Alone: Results of a Multicenter, Multireader Trial. Radiology 2013, 266, 104–113. [Google Scholar] [CrossRef] [PubMed]
  12. Cuzick, J.; Warwick, J.; Pinney, E.; Warren, R.M.L.; Duffy, S.W. Tamoxifen and Breast Density in Women at Increased Risk of Breast Cancer. J. Natl. Cancer Inst. 2004, 96, 621–628. [Google Scholar] [CrossRef] [PubMed]
  13. Nyante, S.J.; Sherman, M.E.; Pfeiffer, R.M.; Berrington de Gonzalez, A.; Brinton, L.A.; Aiello Bowles, E.J.; Hoover, R.N.; Glass, A.; Gierach, G.L. Prognostic Significance of Mammographic Density Change after Initiation of Tamoxifen for ER-Positive Breast Cancer. JNCI J. Natl. Cancer Inst. 2015, 107, dju425. [Google Scholar] [CrossRef] [PubMed]
  14. Kim, J.; Han, W.; Moon, H.-G.; Ahn, S.; Shin, H.-C.; You, J.-M.; Han, S.-W.; Im, S.-A.; Kim, T.-Y.; Koo, H.; et al. Breast Density Change as a Predictive Surrogate for Response to Adjuvant Endocrine Therapy in Hormone Receptor Positive Breast Cancer. Breast Cancer Res. BCR 2012, 14, R102. [Google Scholar] [CrossRef] [PubMed]
  15. Ekpo, E.U.; McEntee, M.F. Measurement of Breast Density with Digital Breast Tomosynthesis—A Systematic Review. Br. J. Radiol. 2014, 87, 20140460. [Google Scholar] [CrossRef]
  16. Breast Imaging Reporting & Data System. Available online: https://www.acr.org/Clinical-Resources/Reporting-and-Data-Systems/Bi-Rads (accessed on 5 March 2024).
  17. Ng, K.-H.; Yip, C.-H.; Taib, N.A.M. Standardisation of Clinical Breast-Density Measurement. Lancet Oncol. 2012, 13, 334–336. [Google Scholar] [CrossRef]
  18. Chen, J.-H.; Gulsen, G.; Su, M.-Y. Imaging Breast Density: Established and Emerging Modalities. Transl. Oncol. 2015, 8, 435–445. [Google Scholar] [CrossRef]
  19. Yaffe, M.J. Mammographic Density. Measurement of Mammographic Density. Breast Cancer Res. BCR 2008, 10, 209. [Google Scholar] [CrossRef]
  20. Eng, A.; Gallant, Z.; Shepherd, J.; McCormack, V.; Li, J.; Dowsett, M.; Vinnicombe, S.; Allen, S.; dos-Santos-Silva, I. Digital Mammographic Density and Breast Cancer Risk: A Case–Control Study of Six Alternative Density Assessment Methods. Breast Cancer Res. 2014, 16, 439. [Google Scholar] [CrossRef]
  21. Ciatto, S.; Houssami, N.; Apruzzese, A.; Bassetti, E.; Brancato, B.; Carozzi, F.; Catarzi, S.; Lamberini, M.P.; Marcelli, G.; Pellizzoni, R.; et al. Categorizing Breast Mammographic Density: Intra- and Interobserver Reproducibility of BI-RADS Density Categories. Breast Edinb. Scotl. 2005, 14, 269–275. [Google Scholar] [CrossRef]
  22. Lobbes, M.B.I.; Cleutjens, J.P.M.; Lima Passos, V.; Frotscher, C.; Lahaye, M.J.; Keymeulen, K.B.M.I.; Beets-Tan, R.G.; Wildberger, J.; Boetes, C. Density Is in the Eye of the Beholder: Visual versus Semi-Automated Assessment of Breast Density on Standard Mammograms. Insights Imaging 2012, 3, 91–99. [Google Scholar] [CrossRef] [PubMed]
  23. Sprague, B.L.; Conant, E.F.; Onega, T.; Garcia, M.P.; Beaber, E.F.; Herschorn, S.D.; Lehman, C.D.; Tosteson, A.N.A.; Lacson, R.; Schnall, M.D.; et al. Variation in Mammographic Breast Density Assessments Among Radiologists in Clinical Practice: A Multicenter Observational Study. Ann. Intern. Med. 2016, 165, 457–464. [Google Scholar] [CrossRef] [PubMed]
  24. Gard, C.C.; Aiello Bowles, E.J.; Miglioretti, D.L.; Taplin, S.H.; Rutter, C.M. Misclassification of Breast Imaging Reporting and Data System (BI-RADS) Mammographic Density and Implications for Breast Density Reporting Legislation. Breast J. 2015, 21, 481–489. [Google Scholar] [CrossRef] [PubMed]
  25. Diffey, J.; Berks, M.; Hufton, A.; Chung, C.; Verow, R.; Morrison, J.; Wilson, M.; Boggis, C.; Morris, J.; Maxwell, A.; et al. A Stepwedge-Based Method for Measuring Breast Density: Observer Variability and Comparison with Human Reading. In Proceedings of the Medical Imaging 2010: Physics of Medical Imaging, San Diego, CA, USA, 18 March 2010; SPIE: San Diego, CA, USA, 2010; Volume 7622, pp. 107–114. [Google Scholar]
  26. Santiago-Rivas, M.; Benjamin, S.; Andrews, J.Z.; Jandorf, L. Breast Density Awareness and Knowledge, and Intentions for Breast Cancer Screening in a Diverse Sample of Women Age Eligible for Mammography. J. Cancer Educ. Off. J. Am. Assoc. Cancer Educ. 2019, 34, 90–97. [Google Scholar] [CrossRef]
  27. Kontos, D.; Bakic, P.R.; Acciavatti, R.J.; Conant, E.F.; Maidment, A.D.A. A Comparative Study of Volumetric and Area-Based Breast Density Estimation in Digital Mammography: Results from a Screening Population. In Digital Mammography; Martí, J., Oliver, A., Freixenet, J., Martí, R., Eds.; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2010; Volume 6136, pp. 378–385. ISBN 978-3-642-13665-8. [Google Scholar]
  28. Koo, T.K.; Li, M.Y. A Guideline of Selecting and Reporting Intraclass Correlation Coefficients for Reliability Research. J. Chiropr. Med. 2016, 15, 155–163. [Google Scholar] [CrossRef]
  29. Ciatto, S.; Bernardi, D.; Calabrese, M.; Durando, M.; Gentilini, M.A.; Mariscotti, G.; Monetti, F.; Moriconi, E.; Pesce, B.; Roselli, A.; et al. A First Evaluation of Breast Radiological Density Assessment by QUANTRA Software as Compared to Visual Classification. Breast Edinb. Scotl. 2012, 21, 503–506. [Google Scholar] [CrossRef]
  30. Pahwa, S.; Hari, S.; Thulkar, S.; Angraal, S. Evaluation of Breast Parenchymal Density with QUANTRA Software. Indian J. Radiol. Imaging 2015, 25, 391–396. [Google Scholar] [CrossRef]
  31. Kshirsagar, A. QuantraTM 2.2 Software Design Intent and Clinical Performance. Available online: https://www.hologic.co.uk/sites/default/files/2023-05/Design%20Intent%20%26%20Performance%20White%20Paper%20GBR%20EN.pdf (accessed on 4 July 2024).
  32. Tari, D.U.; Santonastaso, R.; De Lucia, D.R.; Santarsiere, M.; Pinto, F. Breast Density Evaluation According to BI-RADS 5th Edition on Digital Breast Tomosynthesis: AI Automated Assessment Versus Human Visual Assessment. J. Pers. Med. 2023, 13, 609. [Google Scholar] [CrossRef]
Figure 1. Raw images for MD assessment by Quantra Software.
Figure 1. Raw images for MD assessment by Quantra Software.
Jimaging 10 00238 g001
Figure 2. RCC, LCC, RMLO, and LMLO 2D-synthetic mammograms assigned to category A, “almost entirely fat”, by the Quantra 2.2.3 software according to BI-RADS lexicon.
Figure 2. RCC, LCC, RMLO, and LMLO 2D-synthetic mammograms assigned to category A, “almost entirely fat”, by the Quantra 2.2.3 software according to BI-RADS lexicon.
Jimaging 10 00238 g002
Figure 3. RCC, LCC, RMLO, and LMLO 2D-synthetic mammograms assigned to category B, “scattered fibroglandular”, by the Quantra 2.2.3 software according to BI-RADS lexicon.
Figure 3. RCC, LCC, RMLO, and LMLO 2D-synthetic mammograms assigned to category B, “scattered fibroglandular”, by the Quantra 2.2.3 software according to BI-RADS lexicon.
Jimaging 10 00238 g003
Figure 4. RCC, LCC, RMLO, and LMLO 2D-synthetic mammograms assigned to category C, “heterogeneously dense”, by the Quantra 2.2.3 software according to BI-RADS lexicon.
Figure 4. RCC, LCC, RMLO, and LMLO 2D-synthetic mammograms assigned to category C, “heterogeneously dense”, by the Quantra 2.2.3 software according to BI-RADS lexicon.
Jimaging 10 00238 g004
Figure 5. RCC, LCC, RMLO, and LMLO 2D-synthetic mammograms assigned to category D, “extremely dense”, by the Quantra 2.2.3 software according to BI-RADS lexicon.
Figure 5. RCC, LCC, RMLO, and LMLO 2D-synthetic mammograms assigned to category D, “extremely dense”, by the Quantra 2.2.3 software according to BI-RADS lexicon.
Jimaging 10 00238 g005
Figure 6. Density categories assigned by Quantra Software according to BI-RADS lexicon, fifth edition.
Figure 6. Density categories assigned by Quantra Software according to BI-RADS lexicon, fifth edition.
Jimaging 10 00238 g006
Chart 1. Graphic illustration of the data reported in Table 1 on the frequency of breast density categories assigned by each reader (R1 and R2), the mean of the two readers (R1-R2), and the Quantra software (R3).
Chart 1. Graphic illustration of the data reported in Table 1 on the frequency of breast density categories assigned by each reader (R1 and R2), the mean of the two readers (R1-R2), and the Quantra software (R3).
Jimaging 10 00238 ch001
Figure 7. RCC, LCC, RMLO, and LMLO 2D-synthetic mammograms of the same patient. The visual assessment made by radiologists for this exam was “B”. Quantra software assigned the highest category, “D”, due to the presence of breast implants, resulting in an incorrect assessment.
Figure 7. RCC, LCC, RMLO, and LMLO 2D-synthetic mammograms of the same patient. The visual assessment made by radiologists for this exam was “B”. Quantra software assigned the highest category, “D”, due to the presence of breast implants, resulting in an incorrect assessment.
Jimaging 10 00238 g007
Figure 8. RCC, LCC, RMLO, and LMLO 2D-synthetic mammograms of the same patient. The visual assessment made by radiologists for this exam was “B”. Quantra software assigned the highest category, “D”, due to the presence of a loop recorder and some macrocalcifications in the left breast, resulting in an incorrect assessment.
Figure 8. RCC, LCC, RMLO, and LMLO 2D-synthetic mammograms of the same patient. The visual assessment made by radiologists for this exam was “B”. Quantra software assigned the highest category, “D”, due to the presence of a loop recorder and some macrocalcifications in the left breast, resulting in an incorrect assessment.
Jimaging 10 00238 g008
Table 1. The frequency of breast density categories assigned by each reader, the mean of the two readers, and Quantra software.
Table 1. The frequency of breast density categories assigned by each reader, the mean of the two readers, and Quantra software.
BI-RADS Category:Reader 1 (R1)Reader 2 (R2)Mean R1-R2Quantra Software (R3)
A1646 (33%)1538 (30%)1592 (32%)673 (13%)
B2241 (45%)2364 (48%)2303 (46%)2131 (43%)
C915 (18%)891 (18%)903 (18%)1728 (34%)
D178 (4%)179 (4%)179 (4%)477 (10%)
Table 2. Intraclass Correlation Coefficients.
Table 2. Intraclass Correlation Coefficients.
Intraclass Correlation CoefficientsResults
ICC (R1 vs. R3)0.725
ICC (R2 vs. R3)0.713
ICC (R1 vs. R2)0.712
ICC (meanR1-R2 vs. R3)0.793
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Morciano, F.; Marcazzan, C.; Rella, R.; Tommasini, O.; Conti, M.; Belli, P.; Spagnolo, A.; Quaglia, A.; Tambalo, S.; Trisca, A.G.; et al. Comparison of Visual and Quantra Software Mammographic Density Assessment According to BI-RADS® in 2D and 3D Images. J. Imaging 2024, 10, 238. https://doi.org/10.3390/jimaging10090238

AMA Style

Morciano F, Marcazzan C, Rella R, Tommasini O, Conti M, Belli P, Spagnolo A, Quaglia A, Tambalo S, Trisca AG, et al. Comparison of Visual and Quantra Software Mammographic Density Assessment According to BI-RADS® in 2D and 3D Images. Journal of Imaging. 2024; 10(9):238. https://doi.org/10.3390/jimaging10090238

Chicago/Turabian Style

Morciano, Francesca, Cristina Marcazzan, Rossella Rella, Oscar Tommasini, Marco Conti, Paolo Belli, Andrea Spagnolo, Andrea Quaglia, Stefano Tambalo, Andreea Georgiana Trisca, and et al. 2024. "Comparison of Visual and Quantra Software Mammographic Density Assessment According to BI-RADS® in 2D and 3D Images" Journal of Imaging 10, no. 9: 238. https://doi.org/10.3390/jimaging10090238

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop