Skip to main content
  • Original Article
  • Open access
  • Published:

Deep learning-based thigh muscle segmentation for reproducible fat fraction quantification using fat–water decomposition MRI

Abstract

Background

Time-efficient and accurate whole volume thigh muscle segmentation is a major challenge in moving from qualitative assessment of thigh muscle MRI to more quantitative methods. This study developed an automated whole thigh muscle segmentation method using deep learning for reproducible fat fraction quantification on fat–water decomposition MRI.

Results

This study was performed using a public reference database (Dataset 1, 25 scans) and a local clinical dataset (Dataset 2, 21 scans). A U-net was trained using 23 scans (16 from Dataset 1, seven from Dataset 2) to automatically segment four functional muscle groups: quadriceps femoris, sartorius, gracilis and hamstring. The segmentation accuracy was evaluated on an independent testing set (3 × 3 repeated scans in Dataset 1 and four scans in Dataset 2). The average Dice coefficients between manual and automated segmentation were > 0.85. The average percent difference (absolute) in volume was 7.57%, and the average difference (absolute) in mean fat fraction (meanFF) was 0.17%. The reproducibility in meanFF was calculated using intraclass correlation coefficients (ICCs) for the repeated scans, and automated segmentation produced overall higher ICCs than manual segmentation (0.921 vs. 0.902). A preliminary quantitative analysis was performed using two-sample t test to detect possible differences in meanFF between 14 normal and 14 abnormal (with fat infiltration) thighs in Dataset 2 using automated segmentation, and significantly higher meanFF was detected in abnormal thighs.

Conclusions

This automated thigh muscle segmentation exhibits excellent accuracy and higher reproducibility in fat fraction estimation compared to manual segmentation, which can be further used for quantifying fat infiltration in thigh muscles.

Key points

  • This fully automated deep learning-based thigh muscle segmentation exhibits excellent accuracy.

  • It can delineate four clinically relevant thigh muscle groups in seconds.

  • It provides higher reproducibility in fat fraction estimations compared to manual segmentation.

Background

Previous studies have shown that fat infiltration can be observed in thigh muscles due to diseases such as neuromuscular and metabolic disorders or age-related muscle atrophy (sarcopenia) [1,2,3,4]. More importantly, the degree of intramuscular fat infiltration serves as a marker for disease severity and progression [5,6,7,8,9,10].

Quantitative magnetic resonance imaging (MRI) provides an excellent approach for noninvasively measuring the fat composition in various organs, including muscle tissue. Fat–water decomposition MRI, one of the quantitative MRI techniques, is highly suitable for fat fraction quantification due to its ability to separate MRI signals from water and fat protons based on their chemical shift difference. This technique was first proposed by Dixon [11] as the original two-point Dixon method, which acquires two sets of images at echo times when water and fat protons are in phase and out of phase [11]. With the advancement of MRI techniques, multipoint Dixon methods [12, 13] have been developed to acquire images at three or more echo times with different phase shifts between fat and water protons for fat–water separation. Compared to two-point Dixon method, three-point Dixon sequences [12] are able to correct the magnetic field inhomogeneities, which offer separated fat and water images for quantitative analysis. Some advanced multipoint Dixon techniques can address other confounding factors and provide more accurate fat and water images and the corresponding proton density fat fraction (PDFF) maps [14, 15]. Currently, multipoint Dixon sequences have been clinically implemented and also widely used in research studies for quantifying intramuscular fat infiltration in the thigh [2,3,4,5,6,7,8,9, 15,16,17,18,19,20].

However, the need for segmentation severely limits the application of fat–water decomposition MRI in quantitative fat assessment in clinical practice. Image segmentation is the first and crucial step in order to quantitatively evaluate the intramuscular fat fraction in the thigh. The conventional methods for segmenting thigh muscles in the current literature [3,4,5, 7,8,9, 15,16,17,18,19,20] rely on manually drawn regions-of-interest (ROIs), which is not only an extremely time-consuming and cumbersome process but also prone to subjective bias and inter-reader variation. A previous study [20] showed the average time for segmenting four entire muscle regions was approximately 6 h for each scan. Hence, most previous studies performed manual segmentation on one or a few representative slices [3,4,5, 7,8,9, 15,16,17]. Such a process can lower the accuracy and reproducibility, which is also not feasible for clinical adoption. Although recent studies developed semiautomated or automated algorithms [21,22,23,24,25], they are either based on nonquantitative MRI or difficult to apply in clinical settings [26]. A faster, simpler and fully automated thigh muscle segmentation method for reproducible fat fraction quantification is highly desired.

In recent years, deep learning techniques, in particular convolutional neural networks (CNNs), have been successfully applied in medical image analysis field, notably in image segmentation [27]. CNNs [28, 29] are trainable models established by multiple layers with operations including convolution, pooling and activation, which can capture the highly nonlinear mappings between inputs and outputs. This approach provides a powerful tool for image segmentation, as CNNs can automatically learn and leverage the representative patterns of the training images and then make accurate predictions on prospective images after optimization. U-net [30], one of the most well-known CNN architectures, is designed particularly for semantic segmentation of biomedical images. This network utilizes an encoder–decoder structure with skip connections and is able to extract higher resolution features more efficiently. U-net has been applied in plenty of studies for medical image segmentation and achieved promising and robust results [27]. However, the performance of such technique on thigh muscle segmentation has not yet been investigated and evaluated in the current literature.

The purpose of this study was to train and validate a CNN with sufficient accuracy for automatically segmenting four functional thigh muscle groups using fat–water decomposition MRI. We further evaluated its reproducibility in fat fraction estimation by comparing with manual segmentation and performed a preliminary quantitative analysis showing its ability in differentiating the abnormal thighs with fat infiltration.

Materials and methods

This study was approved by the institutional review board. Owing to the fact that this is a retrospective study, informed consent was waived.

Datasets

Figure 1 provides an overview of the datasets used in this study. Detailed demographic and clinical characteristics are provided in Additional file 1: Table S1.

Fig. 1
figure 1

An overview of the two datasets and workflow of this study

Dataset 1: a public reference database MyoSegmenTUM [20]

This database consists of 25 fat–water decomposition MRI scans collected from 19 subjects (15 healthy volunteers and 4 patients with neuromuscular diseases) using 6-echo 3D spoiled gradient echo (GRE) sequences. Details on MRI acquisitions were described in the reference [20]. Three healthy subjects were scanned three times with repositioning for reproducibility assessment. The manual segmentation masks of all scans were provided in the database as ground truth. The manual ROIs delineated four clinically relevant muscle groups (whole volume in 3D): quadriceps femoris muscle (ROI1), sartorius muscle (ROI2), gracilis muscle (ROI3) and hamstring muscles (ROI4), with an average segmentation time of ~ 6 h per scan.

Dataset 2: local clinical data

Fat–water decomposition MRI images of 21 subjects were collected between 2018 and 2020 for clinical purpose in our site on a 1.5-T GE Signa scanner, using a 3D three-point Dixon method and an iterative algorithm IDEAL [13] for fat–water decomposition. Images were acquired with the following parameters: repetition time (TR) = 7.608 ms, effective echo time (TE) = 3.576 ms, flip angle = 8°, reconstructed matrix size = 512 × 512,voxel size = 0.7422 × 0.7422 × 6mm3, the band width of 244 Hz/voxel. Scanning was performed with two consecutive stacks (upper and lower portions of the thigh) in the axial plane to cover the whole bilateral thigh volumes using GE HD 12-channel body array coil, and each stack had approximately 40 slices. The manual segmentation masks of 11 scans were drawn by an operator and examined and approved by a radiologist with 12 years’ experience. The manual masks contained the same four ROIs as Dataset 1.

MRI preprocessing

For all scans, the left and right sides of the thighs were first separated and treated as two thigh volumes for analysis, which also doubled the dataset size. As a preprocessing step, the subcutaneous fat and skin were removed automatically on water images to potentially improve the segmentation performance: (1) A K-means clustering was first performed on water images and the cluster with higher intensity was kept (subcutaneous fat with lower intensity was removed); (2) the skin region was then removed by applying an order-statistic filtering on the mask generated in the last step; and (3) the final mask was generated after filling the holes and a dilation (in case muscle regions were incorrectly removed in the previous steps). The MRI preprocessing was performed in MATLAB R2019a. The water and fat images covering only the muscle regions for each thigh volume were used for the CNN training and evaluation.

Convolutional neural network training

A CNN was trained to automatically segment the four thigh muscle groups (ROI1–ROI4) using a U-net architecture [30] in Keras [31] with TensorFlow backend [32] on a Windows workstation with an Nvidia 1080Ti GPU. The network architecture is shown in Fig. 2. The whole dataset with manual ROIs for developing the CNN consisted of 25 scans from Dataset 1 and 11 scans from Dataset 2. The network was trained in 2D using a total of 4968 slices from 46 thigh volumes (23 scans, 16 from Dataset 1 and seven from Dataset 2). The preprocessed water and fat images were used as two input channels, and both channels were independently normalized to [0, 1]. The manual segmentation masks were provided as ground truth. The network was trained using a batch size of 6 for 1000 epochs on ~ 3/4 of the training set (3940 slices from 36 thigh volumes of 18 scans, 13 from Dataset 1 and five from Dataset 2). Validation was performed after each epoch using on the rest ~ 1/4 of the training set (1028 slices from ten thigh volumes of five scans, three from Dataset 1 and two from Dataset 2). The network was trained using Adam optimizer with a learning rate of 10−6 [33]. The training objective was to minimize the categorical cross-entropy loss between the CNN outputs and the manual segmentations. The CNN outputs were compressed to the range (0,1) using a sigmoid function. A threshold of 0.5 was applied to binarize the CNN outputs. The binarized CNN outputs were stacked back to thigh volumes, and for each ROI, only the largest region in 3D was kept as the final automated mask.

Fig. 2
figure 2

U-net architecture for segmentation of the four thigh muscle regions

Convolutional neural network evaluation

The CNN segmentation was evaluated using an independent testing set consisting of those MRI scans that were not used for CNN training but had manual ROIs. The testing set included the MRI images acquired from the three subjects with three repeated scans in Dataset 1 (3 × 3 scans, 6 × 3 thigh volumes) and four scans in Dataset 2 (eight thigh volumes). The performance of the CNN segmentation was assessed by comparing with manual segmentation in three ways: (1) Dice coefficient at the volume level; (2) differences in volume and fat fraction estimation; the percent differences in volume between manual and CNN segmentation were assessed; for each scan, a fat fraction map was generated as the ratio of fat image over the sum of water and fat images, and the mean fat fraction (meanFF) values in each muscle group were calculated. (3) Reproducibility in meanFF determined by the intraclass correlation coefficients (ICCs) in each ROI using the repeated scans in Dataset 1. For (1) and (2), in order to avoid bias from the repeated scans, the values of each thigh volume with repeated scans were averaged for assessment.

A quantitative analysis using CNN segmentation

A preliminary quantitative analysis was performed using the clinical scans of 14 subjects in Dataset 2 that were not used for CNN training. As few patients received biopsy-based pathological examination, radiological assessment was used as a reference standard. Among the 14 subjects, seven of them had no radiological evidence of abnormal fat infiltration in their thigh muscles, while the remaining seven subjects showed abnormal fat replacement of muscle (14 normal vs. 14 abnormal thigh volumes). The meanFF values in each ROI were calculated using the CNN segmentation masks. A one-tailed, two-sample t test assuming unequal variances was performed to detect possible differences in meanFF between normal and abnormal thighs. A p value of < 0.05 considered statistically significant.

Results

Performance of the CNN segmentation

Figures 3 and 4 show the MRI images and the segmentation results of a representative slice in the independent testing set from Dataset 1 and Dataset 2, respectively. The manual segmentation and automated segmentation show excellent agreement visually. The CNN generated the ROIs within 10–30 s for each thigh volume.

Fig. 3
figure 3

A scheme of this study, with MRI images and the segmentation results of an example from Dataset 1 in the testing set. ac the original water image, fat image and PDFF map provided by Dataset 1, d the manual segmentation masks of four ROIs, e, f the preprocessed water image and fat image covering the whole thigh muscles as CNN input channels, g the automated CNN-derived segmentation masks. For d, g, segmentations are superimposed on the water image, with ROI1 (quadriceps femoris) in red, ROI2 (sartorius) in green, ROI3 (gracilis) in blue and ROI4 (hamstrings) in yellow

Fig. 4
figure 4

MRI images and the segmentation results of an example from Dataset 2 in the testing set. ac the original water image, fat image and the calculated fat fraction map, d the manual segmentation masks of four ROIs, e, f the preprocessed water image and fat image covering the whole thigh muscles as CNN input channels, g the automated CNN-derived segmentation masks. For d, g, segmentations are superimposed on the water image, with ROI1 (quadriceps femoris) in red, ROI2 (sartorius) in green, ROI3 (gracilis) in blue and ROI4 (hamstrings) in yellow

Table 1 summarizes the mean and standard deviation Dice coefficients between manual and automated segmentation results for ROI1 ~ ROI4 of each thigh volume in the testing set. To avoid bias, the Dice coefficients of each thigh volume with repeated scans were first averaged (their mean ± standard deviation was calculated as shown in Table 1). It is noted that some ROIs are very small, or some regions (particularly at either end of muscle insertion/attachment) can be more inaccurate, which impacts the Dice coefficient. Our mean Dice coefficients were > 0.85 for all ROIs, even for the smallest ROI2. The Dice coefficients of ROI2 and ROI3 were relatively lower than those of ROI1 and ROI4 because ROI2 and ROI3 are smaller muscle groups.

Table 1 Segmentation accuracy assessed by Dice coefficients between manual and automated segmentation for each ROI of each thigh volume in the testing set

Figure 5 demonstrates the percent differences in volume and meanFF differences between manual and automated segmentation for each thigh volume in the testing set (the values of each thigh volume with repeated scans were averaged). The average percent difference in volume (absolute values) was 7.57%. As ROI2 and ROI3 are relatively small muscle regions, the percent difference in volume could be large even if the volume difference was small. The differences in meanFF of all thigh volumes were within an acceptable range (− 0.5 to 1.5%). The average of the absolute differences in meanFF was 0.17%.

Fig. 5
figure 5

a Percent differences in volume between manual and automated segmentation for the 26 thigh volumes in the testing set. The average of the absolute differences was 7.57%. b The differences in meanFF between manual and automated segmentation for the 26 thigh volumes in the testing set. The average of the absolute differences was 0.17%

Table 2 demonstrates the ICCs of meanFF in the four ROIs based on the repeated scans. As shown, CNN-based segmentation method produced more reproducible fat fraction measures than manual segmentation in all ROIs.

Table 2 Reproducibility analyses calculated by ICCs on meanFF of the repeated scans

Quantitative analysis results

Figure 6 illustrates the meanFF differences between the 14 normal and 14 abnormal thigh volumes in Dataset 2 calculated using the CNN-based automated ROIs. For all ROIs, abnormal thighs showed significantly higher fat fraction than normal thighs, although a clear threshold could not be defined based on this preliminary analysis.

Fig. 6
figure 6

A quantitative analysis on meanFF using the CNN segmentation based on 14 normal and 14 abnormal thigh volumes of the clinical data. The averages of meanFF values (indicated by the red lines) and the two-sample t test p values are shown in each plot

Discussion

In this work, we adopt the U-net architecture to train a CNN using fat–water decomposition MRI for thigh muscle segmentation. Time-efficient thigh muscle segmentation is a major challenge in moving from a primarily qualitative assessment of thigh muscle MRI in clinical practice, to potentially more accurate and quantitative methods. Compared to the time-consuming and labor-intensive manual annotation, which ordinarily takes at least 5–6 h, the proposed fully automated method provided sufficiently accurate segmentation taking between 10–30 s for each thigh volume. In terms of fat fraction quantification, the automated segmentation demonstrated higher reproducibility in fat fraction estimation compared to manual segmentation (overall ICCs: 0.921 vs. 0.902). A reproducible measurement allows serial monitoring, which can be useful, particularly if there is an intervention to allow for the detection of disease progression or for evaluation of treatment efficacy. Furthermore, the automated segmentation enabled the quantitative evaluation of fat infiltration in thigh muscles and showed potential for assisting diagnosis or surveillance when applied to clinical data.

This CNN segmentation method was trained using images from two datasets that were acquired by research used six-point Dixon (Dataset 1) and also clinically implemented three-point Dixon (Dataset 2). Although the two datasets were obtained with different sequences and scanners, our generalized network demonstrates satisfying segmentation accuracy in both datasets. This an important strength of our segmentation network as it can be applied to both clinical and research data for further studies. Both three-point and the more advanced six-point Dixon methods have been widely used for quantifying muscle fat content [2]. The six-point Dixon method used in Dataset 1 [20] not only addresses magnetic field inhomogeneities but also corrects for confounding factors including T1 bias, T2* decay and multispectral complexity of fat protons, in order to perform more accurate fat–water separation and obtain the PDFF maps. PDFF is a standardized biomarker of tissue fat concentration that reflects the fundamental property of tissue [14]. If available, such a method is preferred for more accurate muscle fat quantification, as suggested by previous studies [2, 15]. For three-point Dixon method, although some confounding factors exist, it has also been widely used to measure the fat replacement of skeletal muscle [2] due to its higher availability in clinical practice, and its inter- and intra-rater reliability has also been validated [5, 15, 16].

This automated, fast and reproducible thigh muscle segmentation is expected to be useful for quantitative assessment of the intramuscular fat infiltration in clinical practice. We thereby performed a preliminary analysis using our local clinical data, which demonstrates the feasibility of using this CNN-based thigh muscle segmentation method in differentiating the normal thighs and the abnormal thighs with fat infiltration. Significant differences between normal and abnormal thighs were detected in all four ROIs using the CNN-based automated segmentation.

Our method shows advantages compared to other studies that tackled this challenging task in recent years [21,22,23,24,25]. Kemnitz et al. [21] developed a semiautomated thigh muscle segmentation method using an active shape model. In order to achieve high agreement with manual segmentation, this method required manual interaction to modify the individual thigh component masks as the first step before applying the segmentation. Later in 2019, they also applied a U-net deep learning architecture for fully automated thigh muscle segmentation [22]. The segmentation was performed only for a particular anatomical location (33% of the femoral bone), which might not be used for volumetric analysis. In addition, these two studies used nonquantitative T1-weighted MRI, which limited their applications in quantitative studies. Rodrigues et al. [23] developed an automated thigh muscle segmentation method based on local texture analysis using three-point Dixon MRI data. They segmented the entire muscle region without distinguishing any muscle groups. Mesbah et al. [24] segmented three thigh muscle groups on the fat and water images utilizing a 3-D Joint Markov Gibbs Random Field model. The approach was performed on the preselected 50 central slices in a total of seven steps, which might make it difficult to apply in clinical settings. Moreover, the above studies did not perform reproducibility analysis. Ogier et al. [25] designed a semiautomatic segmentation pipeline of individual leg muscles using two-point Dixon and three-point Dixon images based on automatic propagation through nonlinear registrations of initial delineation. The reproducibility was also assessed on four subjects scanned twice on the same day. However, the initial manual thigh muscle segmentation of three slices was required. It is also noted that the two-point Dixon technique can suffer from phase errors due to static field inhomogeneity, which should be used with caution in quantitative measurement [34]. Our CNN-based automated method for thigh muscle segmentation on the whole volumes is designed for quantitative fat–water decomposition MRI (three-point and six-point Dixon sequences), which could be further used to identify potential image-based biomarkers. It produces sufficiently accurate segmentation of four muscle groups in seconds and exhibits high reproducibility compared with manual segmentation.

The current study has some limitations. First, the number of subjects was relatively small. In order to reduce the effect of the small dataset, the sample size was doubled after separating the left and right thigh volumes, and our network was trained in 2D using a total of 4968 slices. However, more data are needed to further improve the segmentation performance and also validate its reproducibility. Second, the current CNN segmentation was not suitable for patients with markedly severe fat infiltration (i.e., whole muscle group involvement obscuring the muscle boundary), as we had limited data of such cases to train the network. Third, for the preliminary quantitative analysis, the pathology results, usually used as a gold standard, were missing for our dataset. In addition, only meanFF was calculated, and no clear thresholds were found to distinguish the normal and abnormal thighs in our preliminary results. Fourth, it is noted that both segmentations yielded only moderate reproducibility in ROI2, which suggests ROI2, the sartorius muscle, might not be suitable for quantitative determination of fat infiltration. This is probably because ROI2 is the smallest muscle region among these four ROIs. For a very small ROI, a little difference in ROI delineation can lead to a large difference in fat fraction estimation. Finally, the segmentation network was trained and tested only based on these two datasets. Although satisfying results were achieved for these two datasets, external validation using an independent dataset acquired with a different scanner is further needed to ensure the reproducibility and feasibility of this segmentation method in clinical practice. Further studies are also needed to establish age-specific reference ranges as well as to improve and validate the segmentation performance across the spectrum of diseases. In addition, more image-based features or biomarkers can be investigated in future studies with a larger dataset, and other factors, including age, sex and body mass index, need to be considered.

Conclusions

In conclusion, we present a deep learning-based thigh muscle segmentation method using fat–water decomposition MRI. This fast and automated method exhibits excellent segmentation accuracy and higher reproducibility in fat fraction estimation compared to manual segmentation, which would be beneficial to clinical practice such as quantifying fat infiltration in thigh muscles associated with ageing, or in conditions such as neuromuscular disorders. Moreover, reproducible segmentation allows serial monitoring which can be particularly useful for detection of disease progression or for evaluation of treatment efficacy.

Availability of data and materials

The datasets used and/or analyzed during the current study are available from the corresponding author on reasonable request.

Abbreviations

CNN:

Convolutional neural network

FF:

Fat fraction

GRE:

Gradient echo

ICC:

Intraclass correlation coefficient

PDFF:

Proton density fat fraction

ROI:

Region-of-interest

TE:

Echo time

TR:

Repetition time

References

  1. Yoshiko A, Kaji T, Sugiyama H, Koike T, Oshida Y, Akima H (2018) Muscle quality characteristics of muscles in the thigh, upper arm and lower back in elderly men and women. Eur J Appl Physiol 118(7):1385–1395

    Article  Google Scholar 

  2. Burakiewicz J, Sinclair CD, Fischer D, Walter GA, Kan HE, Hollingsworth KG (2017) Quantifying fat replacement of muscle by quantitative MRI in muscular dystrophy. J Neurol 264(10):2053–2067

    Article  Google Scholar 

  3. Kumar D, Karampinos DC, MacLeod TD et al (2014) Quadriceps intramuscular fat fraction rather than muscle size is associated with knee osteoarthritis. Osteoarthritis Cartilage 22(2):226–234

    Article  CAS  Google Scholar 

  4. Inhuber S, Sollmann N, Schlaeger S et al (2019) Associations of thigh muscle fat infiltration with isometric strength measurements based on chemical shift encoding-based water-fat magnetic resonance imaging. Eur Radiol Exp 3(1):1–10

    Article  Google Scholar 

  5. Wren TA, Bluml S, Tseng-Ong L, Gilsanz V (2008) Three-point technique of fat quantification of muscle tissue as a marker of disease progression in Duchenne muscular dystrophy: preliminary study. AJR Am J Roentgenol 190(1):W8–W12

    Article  Google Scholar 

  6. Leung DG (2019) Advancements in magnetic resonance imaging-based biomarkers for muscular dystrophy. Muscle Nerve 60(4):347–360

    Article  Google Scholar 

  7. Willis TA, Hollingsworth KG, Coombs A et al (2013) Quantitative muscle MRI as an assessment tool for monitoring disease progression in LGMD2I: a multicentre longitudinal study. PLoS One 8(8):e70993

    Article  CAS  Google Scholar 

  8. Willis TA, Hollingsworth KG, Coombs A et al (2014) Quantitative magnetic resonance imaging in limb-girdle muscular dystrophy 2I: a multinational cross-sectional study. PLoS One 9(2):e903777

    Article  Google Scholar 

  9. Morrow JM, Sinclair CD, Fischmann A et al (2016) MRI biomarker assessment of neuromuscular disease progression: a prospective observational cohort study. Lancet Neurol 15(1):65–77

    Article  Google Scholar 

  10. Li W, Zheng Y, Zhang W, Wang Z, Xiao J, Yuan Y (2015) Progression and variation of fatty infiltration of the thigh muscles in Duchenne muscular dystrophy, a muscle magnetic resonance imaging study. Neuromuscul Disord 25(5):375–380

    Article  Google Scholar 

  11. Dixon WT (1984) Simple proton spectroscopic imaging. Radiology 153(1):189–194

    Article  CAS  Google Scholar 

  12. Glover GH (1991) Multipoint Dixon technique for water and fat proton and susceptibility imaging. J Magn Reson Imaging 1(5):521–530

    Article  CAS  Google Scholar 

  13. Reeder SB, Pineda AR, Wen Z et al (2005) Iterative decomposition of water and fat with echo asymmetry and least-squares estimation (IDEAL): application with fast spin-echo imaging. Magn Reson Med 54(3):636–644

    Article  Google Scholar 

  14. Reeder SB, Hu HH, Sirlin CB (2012) Proton density fat-fraction: a standardized MR-based biomarker of tissue fat concentration. J Magn Reson Imaging 36(5):1011–1014

    Article  Google Scholar 

  15. Grimm A, Meyer H, Nickel MD et al (2018a) Evaluation of 2-point, 3-point, and 6-point Dixon magnetic resonance imaging with flexible echo timing for muscle fat quantification. Eur J Radiol 103:57–64

    Article  Google Scholar 

  16. Schroeder J, Tobler P, Stalder A et al (2019) Intra-rater and inter-rater reliability of quantitative thigh muscle magnetic resonance imaging. Imaging Med 11(2):3–12

    Google Scholar 

  17. Grimm A, Meyer H, Nickel MD et al (2018b) Repeatability of Dixon magnetic resonance imaging and magnetic resonance spectroscopy for quantitative muscle fat assessments in the thigh. J Cachexia Sarcopenia Muscle 9(6):1093–1100

    Article  Google Scholar 

  18. Li K, Dortch RD, Welch EB et al (2014) Multi-parametric MRI characterization of healthy human thigh muscles at 3.0 T—relaxation, magnetization transfer, fat/water, and diffusion tensor imaging. NMR Biomed 27(9):1070–1084

    Article  Google Scholar 

  19. Baum T, Inhuber S, Dieckmeyer M et al (2016) Association of quadriceps muscle fat with isometric strength measurements in healthy males using chemical shift encoding-based water-fat magnetic resonance imaging. J Comput Assist Tomogr 40(3):447

    Article  Google Scholar 

  20. Schlaeger S, Freitag F, Klupp E et al (2018) Thigh muscle segmentation of chemical shift encoding-based water-fat magnetic resonance images: the reference database MyoSegmenTUM. PLoS One 13(6):e0198200

    Article  Google Scholar 

  21. Kemnitz J, Eckstein F, Culvenor AG et al (2017) Validation of an active shape model-based semi-automated segmentation algorithm for the analysis of thigh muscle and adipose tissue cross-sectional areas. MAGMA 30(5):489–503

    Article  CAS  Google Scholar 

  22. Kemnitz J, Baumgartner CF, Eckstein F et al (2020) Clinical evaluation of fully automated thigh muscle and adipose tissue segmentation using a U-Net deep learning architecture in context of osteoarthritic knee pain. MAGMA 33:483–493

    Article  CAS  Google Scholar 

  23. Rodrigues R, Pinheiro AMG (2019) Segmentation of skeletal muscle in thigh dixon MRI based on texture analysis. arXiv:1904.04747

  24. Mesbah S, Shalaby AM, Stills S et al (2019) Novel stochastic framework for automatic segmentation of human thigh MRI volumes and its applications in spinal cord injured individuals. PLoS One 14(5):e0216487

    Article  Google Scholar 

  25. Ogier AC, Heskamp L, Michel CP et al (2020) A novel segmentation framework dedicated to the follow-up of fat infiltration in individual muscles of patients with neuromuscular disorders. Magn Reson Med 83(5):1825–1836

    Article  CAS  Google Scholar 

  26. de Mello R, Ma Y, Ji Y, Du J, Chang EY (2019) Quantitative MRI musculoskeletal techniques: an update. AJR Am J Roentgenol 213(3):524–533

    Article  Google Scholar 

  27. Hesamian MH, Jia W, He X, Kennedy P (2019) Deep learning techniques for medical image segmentation: achievements and challenges. J Digit Imaging 32(4):582–596

    Article  Google Scholar 

  28. LeCun Y, Bottou L, Bengio Y, Haffner P et al (1998) Gradient-based learning applied to document recognition. Proc IEEE 86(11):2278–2324

    Article  Google Scholar 

  29. Krizhevsky A, Sutskever I, Hinton GE. Imagenet classification with deep convolutional neural networks. In: Advances in neural information processing systems. 2012.

  30. Ronneberger O, Fischer P, Brox T. U-net: Convolutional networks for biomedical image segmentation. In: International Conference on Medical image computing and computer-assisted intervention. Springer; 2015.

  31. Chollet F. Keras. github.com. 2015. https://keras.io.

  32. Abadi M, Barham P, Chen J, et al. Tensorflow: a system for large-scale machine learning. In: 12th {USENIX} Symposium on Operating Systems Design and Implementation ({OSDI} 16). 2016.

  33. Kingma DP, Ba J (2014) Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980

  34. Grimm A, Meyer H, Nittka M et al (2017) Magnetic resonance methods for quantitative evaluation of intramuscular adipose tissue. In: Proceedings International Society for Magnetic Resonance in Medicine

Download references

Funding

None.

Author information

Authors and Affiliations

Authors

Contributions

JD and VV performed literature search, figures production, study design, data collection and data interpretation. PC, CHC, GY and SC had inputs in study design, data collection, data interpretation. All authors read and approved the final manuscript.

Corresponding author

Correspondence to Varut Vardhanabhuti.

Ethics declarations

Ethics approval and consent to participate

This study was approved by the institutional review board. Owing to the fact that this is a retrospective study, informed consent was waived.

Consent for publication

Not applicable.

Competing interests

The authors declare that they have no competing interests.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary information

Additional file 1: Table S1.

Subjects’ demographic and clinical characteristics.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Ding, J., Cao, P., Chang, HC. et al. Deep learning-based thigh muscle segmentation for reproducible fat fraction quantification using fat–water decomposition MRI. Insights Imaging 11, 128 (2020). https://doi.org/10.1186/s13244-020-00946-8

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s13244-020-00946-8

Keywords