Skip to main content
  • Original Article
  • Open access
  • Published:

Generalizable attention U-Net for segmentation of fibroglandular tissue and background parenchymal enhancement in breast DCE-MRI

Abstract

Objectives

Development of automated segmentation models enabling standardized volumetric quantification of fibroglandular tissue (FGT) from native volumes and background parenchymal enhancement (BPE) from subtraction volumes of dynamic contrast-enhanced breast MRI. Subsequent assessment of the developed models in the context of FGT and BPE Breast Imaging Reporting and Data System (BI-RADS)-compliant classification.

Methods

For the training and validation of attention U-Net models, data coming from a single 3.0-T scanner was used. For testing, additional data from 1.5-T scanner and data acquired in a different institution with a 3.0-T scanner was utilized. The developed models were used to quantify the amount of FGT and BPE in 80 DCE-MRI examinations, and a correlation between these volumetric measures and the classes assigned by radiologists was performed.

Results

To assess the model performance using application-relevant metrics, the correlation between the volumes of breast, FGT, and BPE calculated from ground truth masks and predicted masks was checked. Pearson correlation coefficients ranging from 0.963 ± 0.004 to 0.999 ± 0.001 were achieved. The Spearman correlation coefficient for the quantitative and qualitative assessment, i.e., classification by radiologist, of FGT amounted to 0.70 (p < 0.0001), whereas BPE amounted to 0.37 (p = 0.0006).

Conclusions

Generalizable algorithms for FGT and BPE segmentation were developed and tested. Our results suggest that when assessing FGT, it is sufficient to use volumetric measures alone. However, for the evaluation of BPE, additional models considering voxels’ intensity distribution and morphology are required.

Critical relevance statement

A standardized assessment of FGT density can rely on volumetric measures, whereas in the case of BPE, the volumetric measures constitute, along with voxels’ intensity distribution and morphology, an important factor.

Key points

• Our work contributes to the standardization of FGT and BPE assessment.

• Attention U-Net can reliably segment intricately shaped FGT and BPE structures.

• The developed models were robust to domain shift.

Graphical Abstract

Introduction

Breast cancer is the most common cancer in the female population worldwide [1]. Screening and early detection followed by treatment are indispensable to improving survival rates. Due to the limitations of mammography, MRI examinations are recommended for high-risk patients and for patients with extremely dense breasts [2]. The amount of fibroglandular tissue (FGT) and the contrast uptake of the healthy FGT, i.e., background parenchymal enhancement (BPE), are important factors related to breast cancer risk, diagnosis, and management [3,4,5,6,7]. For example, the study performed by Ray et al. demonstrated that high BPE was linked to higher rates of abnormal interpretation and biopsy procedures as well as lower specificity [7]. The findings of the study by Hu et al. indicate that elevated BPE is correlated with an increased risk of breast cancer [4]. This is why the American College of Radiology (ACR) in the fifth edition of the Breast Imaging Reporting and Data System (BI-RADS) for MRI reporting recommends visual estimation of the FGT density followed by its classification into four categories: (a) almost entirely fat, (b) scattered, (c) heterogeneous, and (d) extreme. The BPE, after visual assessment, should be classified into (a) minimal, (b) mild, (c) moderate, and (d) marked categories [8]. As shown by different studies, such a qualitative assessment is prone to inter-reader variability. The reported Cohen’s kappa values for BPE assessment range from 0.3 (fair agreement) to 0.95 (almost perfect agreement) [9]. This is the reason why currently there are ongoing efforts to standardize the FGT and BPE assessment by quantitative measures. The most commonly used approaches are based on a region of interest selected by the user, which is prone to inter-reader variability, and on segmentation of the whole structures followed by their quantification [10]. Currently, the ACR is awaiting more robust data for the recommendation of quantitative assessments in clinical practice [8]. To achieve this goal, highly reliable and robust algorithms segmenting FGT and BPE are indispensable.

The main challenge is that breast MRI acquisition is not standardized: different institutions and doctors use different equipment, magnetic field strength, protocols (including the timing and the amount of the post-contrast acquisitions), different contrast agents, and patient positioning, to acquire and process the data. As a result, MRI data exhibit different resolution, contrast, noise level, and artifacts leading to high variability in the quality and reliability of the diagnosis.

Over the past few years, the field of CNN-based segmentation models for medical imaging data has witnessed remarkable advancements. Among these, the U-Net architecture [11] and its various adaptations [12,13,14,15,16,17,18] have been developed. A particularly noteworthy enhancement to the U-Net family is the attention U-Net [19], which integrates attention gates into the architecture. This enables the model to efficiently concentrate on essential features while ignoring irrelevant regions.

It has been shown that the CNN-based segmentation architectures can reliably segment the FGT tissue from native MRI sequences, with the DSC values ranging from 0.81 ± 0.11 to 0.87 ± 0.08 [20,21,22,23,24,25]. Regarding the BPE segmentation, the reported approaches rely on the FGT segmentation from native images with subsequent transfer of the resulting mask to subtraction images, followed by the BPE segmentation based on mean and standard deviation of the voxels’ intensity values [22, 26] or on a predefined threshold [27, 28]. It should be noted that voxels’ intensity values in the MRI data are dependent on the magnetic field strength and the scanner hardware as well as on many adapted parameters during data acquisition and post-processing [29]. Moreover, the voxels’ intensity values in the subtraction data depend on the amount of the contrast agent and the timing of the acquisition. Additionally, the slices acquired at the breast periphery feature higher intensity than slices acquired in the middle of the breast. Thus, BPE segmentation directly from subtraction volumes is desired. In our previous work [30], we showed promising results using U-Net architecture and data from a single scanner for training, validation, and testing. However, segmentation of FGT and testing on an external dataset was not included in this study.

The primary objective of the first part of this study was to develop two automated and generalizable segmentation models: one segmenting the FGT from native volumes, the other segmenting the BPE from subtraction volumes. In the second part of the study, the resulting models were applied to 80 DCE-MRI examinations of 75 patients from our institution, with the aim to assess the correlation between the volumes of FGT and BPE and classes assigned visually by radiologists.

Materials and methods

Patient data

This retrospective study has been approved by the local ethics committee. The main parameters of the datasets utilized in this study are provided in Table 1. Datasets 1–3 originate from our institution, whereas dataset 4 is a small subset of public Duke-Breast-Cancer Dataset [31]. All the data was acquired in a transverse plane in a prone position with fat-saturation of the DCE T1 sequences.

Table 1 Main parameters of the three datasets used in this work

To curate datasets 1–3, we conducted a search in the Picture Archiving and Communication System (PACS) of our institution. The search aimed to identify DCE-MRI examinations fulfilling following inclusion criteria: (a) age above 18 years, (b) absence of implants, (c) availability of assessments for FGT and BPE classes in the corresponding radiological report, and (d) BI-RADS Assessment Category indicating a likelihood of malignancy of 1 or 2, corresponding to a 0% likelihood [8]. Examinations lacking proper fat saturation or suffering from motion blur were excluded. The FGT and BPE classes, determined via a board-certified radiologist, were extracted from the radiological report.

Dataset 1 was curated by searching through examinations acquired between September 2013 and October 2015 using a 3.0-T scanner. This dataset was used for model training, validation, and testing. For volumetric analysis, the search through examinations acquired between September 2020 and October 2022 continued until a total of 80 eligible examinations resulting in possibly well-equalized distributions of FGT and BPE classes were identified. In this way, dataset 2, containing examinations acquired with 1.5-T scanner, and dataset 3, containing examinations acquired with 3.0-T scanner, were obtained. Examinations from dataset 2 were additionally used for testing models’ performance.

AI model development design

The design of AI model development differed from previous works by utilizing two separate models: one for FGT segmentation from the native volume, the other one for BPE segmentation from the subtraction volumes. This approach allows for an investigation of all subtraction images, without the need for precise registration. This is important, as even a small misalignment between sequences (cf. Fig. S1) can have a significant impact on volumetric analysis, especially in cases with almost entirely fat FGT and minimal BPE. Another advantage of using two separate models is that potential errors in FGT segmentation do not affect BPE segmentation. For instance, artifacts or extreme superior and inferior regions with higher intensity on native volume may be mistakenly included in the FGT segmentation.

As the FGT and BPE are very complex and fine structures, it is important that the predicted mask has similar resolution to the resolution with which the MRI data was acquired. Hence, we opted for the highest resolution of our data, i.e., 448 × 448, as the input and output size. To accommodate a reasonable batch size for training with NVIDIA GeForce RTX 3090 (24 GB), we chose to train the models slice by slice (Fig. 1). Our approach utilized a 2D implementation of the attention U-Net from the repository of Yingkai Sha [32] with spatial 2D dropout layers added in each convolution stack. The additional advantage of using a 2D model is that it can effectively incorporate volumes with varying numbers of slices and any necessary rescaling is performed solely in 2D.

Fig. 1
figure 1

Schematic representation of the model development pipeline. Two independent attention U-Net models are trained: the first one is trained to segment the fibroglandular tissue (FGT) and the fatty tissue from native DCE data; the second one is trained to segment BPE and non-enhancing tissue from the subtraction data. This separation ensures accurate segmentation even for not well-registered cases. In both cases, the segmentation is performed slice-wise ensuring that with the chosen hardware, the predicted mask has high resolution able to accurately capture the intricate details of the FGT and BPE structures (Icons made by Freepik and Netscript from flaticon.com)

Ground truth masks

The ground truth masks were created in 3D Slicer [33] by S.N.. Firstly, the breast was segmented without skin by the use of Grow from Seeds algorithm, Gaussian smoothing, and fine-tuning with Paint and Erase. Afterwards, the FGT and BPE were segmented by thresholding, customized to each volume followed by the fine-tuning allowing also for artifacts’ removal. A sample of final segmentation masks was verified by A.L. (resident in radiology with more than 3 years of experience in breast imaging) and A.B. (board-certified radiologist with over 15 years of experience in breast imaging). Due to the time and resource constrains, intra- and inter-reader variability investigation was not performed.

Dataset splitting

Dataset 1 was split into a patient-stratified train-validation-test sets, and the FGT model was trained using 2112 slices from 20 patients for training, 416 slices from 4 patients for validation, and 520 slices from 5 patients for testing. The model was additionally evaluated on dataset 2 and dataset 4, which comprised 1004 slices from 6 patients. Importantly, the subtraction volumes exhibit lower contrast and signal-to-noise ratio compared to native volumes. To account for this, a larger amount of data was utilized for the BPE model. The total training set for the BPE model comprised 11,829 slices from 54 patients, with a validation set of 2469 slices from 12 patients, and a testing set of 3095 slices from 16 patients. Subsequently, the BPE segmentation model was tested on dataset 2 and dataset 4, which collectively contained 2672 slices from 6 patients.

Model training

All the data were rescaled to 0–1 range prior to training. A subset of the dataset was used for hyperparameter tuning using fivefold cross validation. The best hyperparameters obtained in this way were then fine-tuned during training using the entire dataset. Five rounds of training of native and subtraction models were then performed using best fine-tuned hyperparameters. Noteworthy, the best performance was achieved with focal Tversky loss [34] function harshly penalizing the false negatives by setting the α parameter of the loss to 0.99 and the β parameter to 0.01. Additionally, during training, brightness augmentation in the 0.2–1.8 range delivered best performance on the test set. All final hyperparameters together with average inference runtimes are reported in Table S1.

Model evaluation

The model obtained in each training run was evaluated on the test data coming from three datasets: dataset 1, 2, and 4 (cf. Table 1). Firstly, the evaluation was centered around application-relevant metrics, i.e., breastvol, FGT(%) (1)/BPE(%) (2), derived from ground truth and predicted masks.

$${FGT}_{(\%)}=\frac{{FGT}_{vol}}{{breast}_{vol}} 100\%$$
(1)
$${BPE}_{(\%)}=\frac{{BPE}_{vol}}{{breast}_{vol}} 100\%$$
(2)

Their correlation was plotted, followed by linear fit and calculation of Pearson correlation coefficient (r). Secondly, the volumetric DSC was computed for the breast and the FGT/BPE masks. Additionally, a weighted DSC was calculated, with the weights proportional to FGT(%)/BPE(%). This adjustment was made to account for the higher penalization of small shifts in case of lower FGT(%)/BPE(%). The models were additionally evaluated with Bland–Altman plots. Lastly, the overlays of the ground truth and the predicted masks were assessed visually.

Volumetric analysis

The best performing models were used to quantify the density of the healthy breast tissue and its percentage taking up the contrast using datasets 2 and 3. FGT(%) according to Eq. (1) and BPE/FGT (%) according to Eq. (3) were calculated from the predicted masks.

$${BPE/FGT}_{(\%)}=\frac{{BPE}_{vol}}{{FGT}_{vol}} 100\%$$
(3)

Next, the correlation between those quantitative measures and qualitative assessment by radiologists was assessed by using Spearman correlation coefficient (ρ), taking into account errors in the calculation of FGT(%) and BPE/FGT(%). These were calculated by propagation of uncertainty from mean absolute errors of the native and subtraction model using the test set.

All of the evaluation was performed by S.N..

Results

Evaluation of the FGT and BPE segmentation models

The performance metrics of the models trained with the best fine-tuned hyperparameters is reported in Table 2. The r for the native models amounted to 0.999 ± 0.001 for the breast and 0.985 ± 0.001 for the FGT(%). These models were characterized by DSC of 0.950 ± 0.002 for the breast segmentation and by DSC and weighted DSC of 0.820 ± 0.005 and 0.864 ± 0.004 for the FGT segmentation, respectively. The r for the subtraction models amounted to 0.992 ± 0.001 for the breast and 0.963 ± 0.004 for the BPE(%). These models featured DSC of 0.927 ± 0.001 for the breast segmentation and DSC and weighted DSC of 0.628 ± 0.018 and 0.715 ± 0.015 for the BPE segmentation, respectively.

Table 2 The evaluation of native and subtraction models; the mean and standard deviation calculated from 5 training rounds is reported for each metrics in each case

For further analysis, one native model and one subtraction model was chosen. The choice was made based on r and DSC values as well as the visual assessment. Analysis of the volumetric correlations between ground truth and predicted masks for the test set for the chosen models is displayed in Fig. 2. The fit regression lines are close to lines describing a perfect correlation. The chosen native model featured DSC of 0.950 (0.937–0.963, 95% confidence intervals (CI)) for the breast, and 0.824 (0.760–0.888, 95% CI) for the FGT. The chosen subtraction model was characterized by DSC of 0.923 (0.924–0.935, 95% CI) for the breast and 0.655 (0.614–0.696, 95% CI) for BPE. The Bland–Altman plots revealed a slight bias of − 0.39% in the estimation of FGT% with the limits of agreements (LoA) of − 7.3% and 6.5%. The bias in the estimation of BPE% amounted to − 0.98% with LoA of − 7.2% and 5.2%. Figure 3 showcases the overlays obtained for three sample native and corresponding subtraction slices, each originating from a different dataset. These models were then used for segmenting volumes from dataset 3.

Fig. 2
figure 2

Analysis of the volumetric correlations between ground truth and predicted masks for the test set. a 1st row: correlation plots for the chosen native model for breastvol and FGT(%) (for 11 volumes). 2nd row: correlation plots for the chosen subtraction model for breastvol and BPE(%) (for 49 volumes). b Bland-Altmann plots for the chosen native and subtraction models

Fig. 3
figure 3

Visual comparison of the overlay between ground truth and predicted masks for a chosen native slice and its corresponding subtraction slice from each dataset

Results of volumetric analysis

Figure 4 depicts the comparison of quantitative measures, i.e., FGT(%) and BPE/FGT(%), calculated from the predicted masks with the qualitative classification into four classes by the radiologists. In the case of FGT(%) distribution, the mean values increase with the increasing FGT class, which is not the case for the mean BPE/FGT(%) values and the increasing BPE class. The FGT(%) distribution of almost entirely fat and scattered FGT classes fully overlap. This is also the case for BPE/FGT(%) in case of minimal, mild, and moderate BPE classes. The coefficient ρ for FGT(%) amounted to 0.70 (p < 0.0001), whereas for BPE/FGT(%) amounted to 0.37 (p = 0.0006).

Fig. 4
figure 4

Comparison of the quantitative measures and qualitative assessments by the radiologists. a The distributions of the BPE and FGT classes. b Plot of calculated FGT(%) and BPE/FGT(%) from the predicted masks for each class. c The corresponding box plots

Discussion

The attention U-Net models developed in this study accurately segmented FGT and BPE structures from breast DCE-MRI data. Table 3 presents a summary of the chosen native and BPE models’ performance compared to other studies focused on BPE segmentation using CNN-based models. However, due to the variability in approaches and datasets utilized across studies, a direct comparison is not possible. Owing to the separation of segmentation from native and subtraction volumes, the models developed in this study are applicable to DCE examinations, in which the series are not perfectly registered, opening the possibility to study the volumetric changes also in the further subtraction sequences. Additional advantage is that errors in the FGT segmentation, such as inclusion of high-intensity artifacts, do not affect the BPE segmentation.

Table 3 Comparison of our work with other studies concerning with BPE segmentation with CNN-based models

The comparison of quantitative measures and qualitative assessment of breast tissue density revealed a correlation characterized by ρ = 0.70 (p < 0.0001), which is statistically strong. The box plots summarizing the FGT(%) distribution for each class (cf. Figure 4c, upper row) reveal the possibility of FGT class assessment based solely on the volumetric measures. This implies that breast DCE-MRI data from different institutions could be merged, and subsequently distinct and non-overlapping FGT(%) ranges for each class could be defined, thus allowing reproducible and standardized breast density assessment. This could be further explored in the investigation of breast cancer risk factors enabling triaging of the patients.

It is crucial to note that the statistically low correlation between quantitative and qualitative assessment of contrast uptake (ρ = 0.37, p = 0.0006, cf. Figure 4c lower row) highlights the complexity of its evaluation and underlines the need for an additional model, taking into account the voxels’ intensity values distribution and morphology [10] as well as the sensitivity of radiologists to a potential masking effect for the standardized assessment.

In the context of the four-class BI-RADS-compliant classification of the BPE, few approaches have been explored. Borkowski et al. reported a direct classification of the entire MRI slice by CNN [35]. Nam et al. extracted radiomic features from the BPE, segmented on the basis of mean and standard deviation of voxels’ intensity values in the subtraction image within the FGT mask, followed by the classification by a tree-based model [22]. The classification accuracy in both of these approaches could potentially be improved by implementing the BPE segmentation model developed in this work, thus bringing the models closer to implementation into clinical practice.

We are also aware of the limitations of our study. All the MRI scanners used in our study were from a single manufacturer, and dataset 2 and dataset 4 had a limited number of MRI scans. We are currently working to expand our test dataset to address these limitations.

Conclusion

The findings presented in this study highlight that a standardized assessment of FGT can rely solely on volumetric measures, while standardized assessment of BPE requires additional models that consider the distribution of intensity and morphology within enhancing voxels.

Due to the impracticality of annotating vast amounts of data from various MRI scanners, it is essential to develop segmentation models trained with limited data yet robust to domain shifts. Our work presents an end-to-end pipeline that creates generalizable models capable of accurately segmenting intricately shaped FGT and BPE structures from native and low-contrast subtraction volumes. Our approach can be extended to other MRI protocols.

Availability of data and materials

The dataset generated and analyzed during the current study is not publicly available due to the confidentiality of the patient data as stated in the Study Plan approved by the Ethics Committee of Kanton of Zürich. Part of the dataset can be made available from the corresponding author to bona fide researchers for non-commercial purposes on reasonable request.

Abbreviations

ACR:

American College of Radiology

BI-RADS:

Breast Imaging Reporting and Data System

BPE:

Background parenchymal enhancement

CI:

Confidence intervals

CNN:

Convolutional neural networks

DCE-MRI:

Dynamic contrast-enhanced magnetic resonance imaging

DSC:

Dice similarity coefficient

FGT:

Fibroglandular tissue

LoA:

Limits of agreement

PACS:

Picture Archiving and Communication System

ρ:

Spearman correlation coefficient

r:

Pearson correlation coefficient

References

  1. Breast cancer statistics | World Cancer Research Fund International. In: WCRF Int. https://www.wcrf.org/cancer-trends/breast-cancer-statistics/. Accessed 2 Mar 2023

  2. Mann RM, Athanasiou A, Baltzer PAT et al (2022) Breast cancer screening in women with extremely dense breasts recommendations of the European Society of Breast Imaging (EUSOBI). Eur Radiol. https://doi.org/10.1007/s00330-022-08617-6

    Article  PubMed  PubMed Central  Google Scholar 

  3. Thompson CM, Mallawaarachchi I, Dwivedi DK et al (2019) The association of background parenchymal enhancement at breast MRI with breast cancer: a systematic review and meta-analysis. Radiology 292:552–561. https://doi.org/10.1148/radiol.2019182441

    Article  PubMed  Google Scholar 

  4. Hu X, Jiang L, You C, Gu Y (2021) Fibroglandular tissue and background parenchymal enhancement on breast MR imaging correlates with breast cancer. Front Oncol 11:4025. https://doi.org/10.3389/fonc.2021.616716

    Article  Google Scholar 

  5. Liao GJ, Bancroft LH, Strigel RM et al (2020) Background parenchymal enhancement on breast MRI: a comprehensive review. J Magn Reson Imaging JMRI 51:43–61. https://doi.org/10.1002/jmri.26762

    Article  PubMed  Google Scholar 

  6. (2022) Background parenchymal enhancement and uptake as breast cancer imaging biomarkers: a state-of-the-art review. Clin Imaging 83:41–50. https://doi.org/10.1016/j.clinimag.2021.11.021

  7. Ray KM, Kerlikowske K, Lobach IV et al (2018) Effect of background parenchymal enhancement on breast MR imaging interpretive performance in community-based practices. Radiology 286:822–829. https://doi.org/10.1148/radiol.2017170811

    Article  PubMed  Google Scholar 

  8. ACR BI-RADS MRI -Reporting. https://www.acr.org/-/media/ACR/Files/RADS/BI-RADS/MRI-Reporting.pdf. Accessed 15 Dec 2021

  9. Bignotti B, Signori A, Valdora F et al (2017) Evaluation of background parenchymal enhancement on breast MRI: a systematic review. Br J Radiol 90:20160542. https://doi.org/10.1259/bjr.20160542

    Article  PubMed  PubMed Central  Google Scholar 

  10. Chalfant JS, Mortazavi S, Lee-Felker SA (2021) Background parenchymal enhancement on breast MRI: assessment and clinical implications. Curr Radiol Rep 9:10. https://doi.org/10.1007/s40134-021-00386-2

    Article  Google Scholar 

  11. Ronneberger O, Fischer P, Brox T (2015) U-Net: convolutional networks for biomedical image segmentation. In: Navab N, Hornegger J, Wells WM, Frangi AF (eds) Medical Image Computing and Computer-Assisted Intervention – MICCAI 2015. Springer International Publishing, Cham, pp 234–241

    Chapter  Google Scholar 

  12. Zuo Q, Chen S, Wang Z (2021) R2AU-Net: attention recurrent residual convolutional neural network for multimodal medical image segmentation. Secur Commun Netw 2021:e6625688. https://doi.org/10.1155/2021/6625688

    Article  Google Scholar 

  13. Diakogiannis FI, Waldner F, Caccetta P, Wu C (2020) ResUNet-a: a deep learning framework for semantic segmentation of remotely sensed data. ISPRS J Photogramm Remote Sens 162:94–114. https://doi.org/10.1016/j.isprsjprs.2020.01.013

    Article  Google Scholar 

  14. Huang H, Lin L, Tong R et al (2020) UNet 3+: a full-scale connected UNet for medical image segmentation. https://doi.org/10.1109/ICASSP40776.2020.9053405

    Book  Google Scholar 

  15. Chen J, Lu Y, Yu Q et al (2021) TransUNet: transformers make strong encoders for medical image segmentation. Available via https://www.cs.jhu.edu/~alanlab/Pubs21/chen2021transunet.pdf

  16. Cao H, Wang Y, Chen J et al (2021) Swin-Unet: Unet-like pure transformer for medical image segmentation. arXiv:2105.05537v1. https://doi.org/10.48550/arXiv.2105.0553

  17. Isensee F, Jaeger PF, Kohl SAA et al (2021) nnU-Net: a self-configuring method for deep learning-based biomedical image segmentation. Nat Methods 18:203–211. https://doi.org/10.1038/s41592-020-01008-z

    Article  CAS  PubMed  Google Scholar 

  18. Milletari F, Navab N, Ahmadi S-A (2016) V-Net: fully convolutional neural networks for volumetric medical image segmentation. 2016 Fourth International Conference on 3D Vision (3DV). pp 565–571

    Chapter  Google Scholar 

  19. Oktay O, Schlemper J, Folgoc LL et al (2018) Attention U-Net: learning where to look for the pancreas. arXiv:1804.03999v3. https://doi.org/10.48550/arXiv.1804.03999

  20. Dalmış MU, Litjens G, Holland K et al (2017) Using deep learning to segment breast and fibroglandular tissue in MRI volumes. Med Phys 44:533–546. https://doi.org/10.1002/mp.12079

    Article  CAS  PubMed  Google Scholar 

  21. Huo L, Hu X, Xiao Q et al (2021) Segmentation of whole breast and fibroglandular tissue using nnU-Net in dynamic contrast enhanced MR images. Magn Reson Imaging 82:31–41. https://doi.org/10.1016/j.mri.2021.06.017

    Article  CAS  Google Scholar 

  22. Nam Y, Park GE, Kang J, Kim SH (2021) Fully automatic assessment of background parenchymal enhancement on breast MRI using machine-learning models. J Magn Reson Imaging 53:818–826. https://doi.org/10.1002/jmri.27429

    Article  PubMed  Google Scholar 

  23. Zhang Y, Chen J-H, Chang K-T et al (2019) Automatic breast and fibroglandular tissue segmentation in breast MRI using deep learning by a fully-convolutional residual neural network U-Net. Acad Radiol 26:1526–1535. https://doi.org/10.1016/j.acra.2019.01.012

    Article  Google Scholar 

  24. Ma X, Wang J, Zheng X et al (2020) Automated fibroglandular tissue segmentation in breast MRI using generative adversarial networks. Phys Med Biol 65:105006. https://doi.org/10.1088/1361-6560/ab7e7f

    Article  PubMed  Google Scholar 

  25. Müller-Franzes G, Müller-Franzes F, Huck L et al (2023) Fibroglandular tissue segmentation in breast MRI using vision transformers – a multiinstitutional evaluation. Sci Rep 13:14207. https://doi.org/10.1038/s41598-023-41331-x

  26. Nguyen AA-T, Arasu VA, Strand F et al (2020) Comparison of segmentation methods in assessing background parenchymal enhancement as a biomarker for response to neoadjuvant therapy. Tomography. 6:101–110. https://doi.org/10.18383/j.tom.2020.00009

    Article  PubMed  Google Scholar 

  27. Wei D, Jahani N, Cohen E et al (2021) Fully automatic quantification of fibroglandular tissue and background parenchymal enhancement with accurate implementation for axial and sagittal breast MRI protocols. Med Phys 48:238–252. https://doi.org/10.1002/mp.14581

    Article  CAS  PubMed  Google Scholar 

  28. Wu S, Zuley ML, Berg WA et al (2017) DCE-MRI background parenchymal enhancement quantified from an early versus delayed post-contrast sequence: association with breast cancer presence. Sci Rep 7:2115. https://doi.org/10.1038/s41598-017-02341-8

    Article  CAS  PubMed Central  Google Scholar 

  29. Bloem JL, Reijnierse M, Huizinga TWJ, van der Helm-van Mil AHM (2018) MR signal intensity: staying on the bright side in MR image interpretation. RMD Open 4:e000728. https://doi.org/10.1136/rmdopen-2018-000728

    Article  PubMed  PubMed Central  Google Scholar 

  30. Nowakowska S, Borkowski K, Ruppert C et al (2022) Deep learning for automatic segmentation of background parenchymal enhancement in breast MRI. Medical Imaging with Deep Learning 2022.

  31. Saha A, Harowicz MR, Grimm LJ et al (2018) A machine learning approach to radiogenomics of breast cancer: a study of 922 subjects and 529 DCE-MRI features. Br J Cancer 119:508–516. https://doi.org/10.1038/s41416-018-0185-8

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  32. Sha Y (Kyle) (2023) keras-unet-collection

    Google Scholar 

  33. Fedorov A, Beichel R, Kalpathy-Cramer J et al (2012) 3D Slicer as an image computing platform for the Quantitative Imaging Network. Magn Reson Imaging 30:1323–1341. https://doi.org/10.1016/j.mri.2012.05.001

    Article  PubMed  PubMed Central  Google Scholar 

  34. Abraham N, Khan NM (2018) A novel focal Tversky loss function with improved attention U-Net for lesion segmentation. arXiv:1810.07842v1. https://doi.org/10.48550/arXiv.1810.07842

  35. Borkowski K, Rossi C, Ciritsis A et al (2020) Fully automatic classification of breast MRI background parenchymal enhancement using a transfer learning approach. Medicine (Baltimore) 99:e21243. https://doi.org/10.1097/MD.0000000000021243

    Article  PubMed  Google Scholar 

Download references

Funding

We would like to acknowledge the financial support from the Clinical Research Priority Program (CRPP) Artificial Intelligence in oncological Imaging of the University of Zurich.

Author information

Authors and Affiliations

Authors

Contributions

The concept and methodology of this work was developed by C. Rossi, S.N, and A.C. The project administration was done by C. Rossi and A.B. The data analysis and algorithms were developed by S.N. with the support of K.B. and C. Ruppert. A.L. and A.B. verified the segmentation masks. S.N. wrote the original draft, which was reviewed and reedited by all co-authors. All authors read and approved the final manuscript.

Corresponding author

Correspondence to Sylwia Nowakowska.

Ethics declarations

Ethics approval and consent to participate

The local Ethics Committee of Kanton of Zürich (Kantonale Ethikkommission) approved this retrospective study. As a general informed consent at University Hospital Zürich was introduced in 2016, all female patients above the age of 18 undergoing breast MRI examination from September 2013 to October 2015, not explicitly refusing the use of their data for scientific purposes, were considered for this study. MRI examinations of patients, who had follow-up examinations after 2016 and did not agree for scientific use of their data, were excluded. The local Ethics Committee of Kanton Zürich waived the informed consent for patients having the MRI examination before 2016 and not having the follow-up examination after 2016, as it would be disproportionally difficult to obtain it.

Consent for publication

Not applicable.

Competing interests

The authors declare that they have no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Additional file 1: Figure S1.

The registration issue between native and four contrast-enhanced sequences in DCE-MRI: with the progress of the data acquisition the subtraction slices are progressively shifted. As a result, the FGT mask does not include all the enhancing tissue. This can lead to underestimation of BPE(%), which might be particularly significant in case of entirely fat FGT and minimal BPE. Table S1. Overview of the Data, Hardware, Best Hyperparameters and Inference Time for the models reported in the main text.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Nowakowska, S., Borkowski, K., Ruppert, C.M. et al. Generalizable attention U-Net for segmentation of fibroglandular tissue and background parenchymal enhancement in breast DCE-MRI. Insights Imaging 14, 185 (2023). https://doi.org/10.1186/s13244-023-01531-5

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s13244-023-01531-5

Keywords