Skip to main content
  • Original Article
  • Open access
  • Published:

Deep learning for differentiation of osteolytic osteosarcoma and giant cell tumor around the knee joint on radiographs: a multicenter study

Abstract

Objectives

To develop a deep learning (DL) model for differentiating between osteolytic osteosarcoma (OS) and giant cell tumor (GCT) on radiographs.

Methods

Patients with osteolytic OS and GCT proven by postoperative pathology were retrospectively recruited from four centers (center A, training and internal testing; centers B, C, and D, external testing). Sixteen radiologists with different experiences in musculoskeletal imaging diagnosis were divided into three groups and participated with or without the DL model’s assistance. DL model was generated using EfficientNet-B6 architecture, and the clinical model was trained using clinical variables. The performance of various models was compared using McNemar’s test.

Results

Three hundred thirty-three patients were included (mean age, 27 years ± 12 [SD]; 186 men). Compared to the clinical model, the DL model achieved a higher area under the curve (AUC) in both the internal (0.97 vs. 0.77, p = 0.008) and external test set (0.97 vs. 0.64, p < 0.001). In the total test set (including the internal and external test sets), the DL model achieved higher accuracy than the junior expert committee (93.1% vs. 72.4%; p < 0.001) and was comparable to the intermediate and senior expert committee (93.1% vs. 88.8%, p = 0.25; 87.1%, p = 0.35). With DL model assistance, the accuracy of the junior expert committee was improved from 72.4% to 91.4% (p = 0.051).

Conclusion

The DL model accurately distinguished osteolytic OS and GCT with better performance than the junior radiologists, whose own diagnostic performances were significantly improved with the aid of the model, indicating the potential for the differential diagnosis of the two bone tumors on radiographs.

Critical relevance statement

The deep learning model can accurately distinguish osteolytic osteosarcoma and giant cell tumor on radiographs, which may help radiologists improve the diagnostic accuracy of two types of tumors.

Key points

• The DL model shows robust performance in distinguishing osteolytic osteosarcoma and giant cell tumor.

• The diagnosis performance of the DL model is better than junior radiologists’.

• The DL model shows potential for differentiating osteolytic osteosarcoma and giant cell tumor.

Graphical Abstract

Introduction

Bone tumors are a group of primary or secondary neoplastic lesions of bone with various pathological types and biological behaviors [1]. According to the 5th edition of the World Health Organization (WHO) classification of bone tumors published in 2020, bone tumors are classified as benign, intermediate, or malignant [2]. Primary bone tumors of the extremities are commonly found in the bones around the knee joint [3,4,5], with the most common types including osteochondroma, osteosarcoma (OS), and giant cell tumor (GCT) of bone [6,7,8]. The imaging manifestation of osteochondroma is a benign bone tumor growing outside the bone, with diagnostic characteristics including cartilage cap coverage, and it is therefore not difficult to make a correct diagnosis according to medical imaging [9, 10]. OS and GCT of bone both show strong local aggressiveness on imaging [7, 8], but they are classified as malignant bone tumors and intermediate bone tumors according to the WHO classification of bone tumors, and their clinical treatment plans and prognosis are substantially different. Therefore, it is necessary to obtain an accurate differential diagnosis between OS and GCT before clinical treatment [1, 7, 8].

Digital radiography is widely acknowledged as the primary imaging method for diagnosing bone lesions and is extensively utilized in clinical practice [11]. Radiographs can display the overall image of bone tumors as a whole and reflect their biological behaviors, providing diagnostic and differential diagnostic information [12]. Tumoral bone formation on radiographs is a characteristic feature in the diagnosis of OS [11]. OS can be categorized into osteolytic, osteoblastic, and mixed subtypes according to the osteogenic quantity within the OS on radiographs [13]. However, distinguishing osteolytic OS from GCT becomes more challenging when tumoral bone formation is absent, especially for junior radiologist, as both exhibit localized, aggressive bone destruction [14]. Patient age holds diagnostic value in distinguishing between the two tumor types, but similar ages between patients with these two different tumors may confuse the radiologist’s judgment [11]. Therefore, we constructed a clinical model to investigate the role of clinical variables.

As an emerging machine learning technology, deep learning (DL) has been widely applied to medical image analysis of bone tumors [15,16,17], which can distinguish benign and malignant bone tumors [11, 12, 14, 18,19,20,21]. A recent study has highlighted DL’s potential in effectively classifying healthy and pathological X-rays in children [22]. However, previous studies have primarily focused on the benign and malignant classification of various bone tumors at different sites throughout the body [15, 18, 21], rather than specifically addressing the differential diagnosis of osteolytic OS and GCT around the knee joint. To our knowledge, there have been no reported studies utilizing DL for this specific purpose.

The purpose of this study was to develop a DL model for the differential diagnosis of osteolytic OS and GCT of bone on knee radiographs and to compare its diagnostic performance with that of radiologists with and without model assistance.

Materials and methods

Subjects

The retrospective study adhered to the principles outlined in the Helsinki Declaration and received institutional review board approval, including a waiver for written informed consent. This multicenter study collected patients with OS and GCT of bone around the knee joint obtained from four tertiary referral centers, from 2013 to 2022. The training set and internal test set were obtained from the First Affiliated Hospital of Sun Yat-Sen University (center A), while the external test set was obtained from Foshan Hospital of Traditional Chinese Medicine (center B), People’s Hospital of Huizhou City Center (center C) and the People’s Hospital of Guangxi Zhuang Autonomous Region (center D). Radiographs were obtained from different digital X-ray imaging devices with automatic examination parameters set in each of the four centers. Detailed information on the digital X-ray imaging devices is provided in Supplementary Table S5.

There were 333 patients (osteolytic OS:136, GCT:197) in this study. The inclusion and exclusion criteria are shown in Fig. 1. The osteosarcoma, in which tumoral bone formation is not observed on both anteroposterior and lateral radiographs, is considered as osteolytic osteosarcoma. According to the inclusion and exclusion criteria (Fig. 1), two radiologists (both unknown to the study, with more than 10 years of experience in reading musculoskeletal radiographs) independently reviewed all radiographs and selected the patient included in the study. The clinical data of all patients are summarized in Table 1. All included bone tumors were pathologically confirmed. Both anteroposterior and lateral radiographs were available for each patient.

Fig. 1
figure 1

Inclusion and exclusion criteria. DR, digital radiography; OS, osteosarcoma; GCT, giant cell tumor; DL, deep learning

Table 1 Demographic data of patients with osteolytic OS and GCT from four centers

Radiologists’ interpretations

In the internal and external test sets, sixteen radiologists from 6 tertiary referral hospitals were selected using the Multi-Reader Multi-Case (MRMC) reading method [23, 24] and Digital Imaging and Communication in Medicine (DICOM) Viewer 2.2.9 software (Medical Company, Poland) to independently evaluate the radiographs. All observers were blinded to the pathologic diagnosis results. The time taken by radiologists to evaluate radiographs was recorded.

All radiologists were categorized into three “expert committee” groups (expert decision by majority rule) based on their experience in reading musculoskeletal radiographs. There were 5 radiologists (> 2 but ≤ 5 years) in group A, 7 radiologists (> 5 but ≤ 8 years) in group B, and 4 radiologists (> 8 but < 13 years) in group C. All radiologists participated in the evaluation of performance with and without assistance from the DL model. The washout period between the two evaluations was more than 4 weeks (4–6 weeks, with an average of 5.8 weeks).

Preprocessing

All images were downloaded from the picture archiving and communication system in DICOM format at their original dimensions and resolution, at which point the basic information of the patients on the images was removed. All images were converted from DICOM to 8-bit JPEG using MicroDicom software (Version 3.8.1.422, MicroDicom Ltd, Bulgarian). To facilitate input to the neural networks, the pixel size of each image was adjusted to 1080 by 1080, and the pixel values were scaled into the range [0,1].

DL model building

Model training was performed in Python 3.6 (https://www.python.org) and PyTorch 1.6.0 (https://pytorch.org). We adopted the EfficientNet-B6 architecture [25] with weights pretrained on ImageNet. Five-fold cross-validation was used for model training and selecting. The model took preprocessed images with a resolution of 1080 × 1080 as inputs and output the predicted probability for each image. It employed Binary Cross Entropy (BCE) loss as the loss function and incorporated data augmentation operations such as horizontal/vertical flips, rotation, and contrast adjustments. sThe network was optimized with the following parameters: basic learning rate = 1 × 10-4; batch size = 2; and epoch = 150. The model with the minimum validation loss was selected for testing on both the internal and external test sets.

Clinical model and integrated model building

A clinical model based on clinical variables, including patient age, sex and tumor site (distal femur, proximal tibia and proximal fibula), was developed using logistic regression. Sex was coded using binary coding, and the tumor site was coded using one-hot coding. The clinical model was trained using a cross-validation strategy. The model with the best validation performance was selected for testing on both the internal and external test sets. Additionally, we incorporated the predictions generated by the DL model as a variable into the clinical model to establish an integrated model. Similarly, the training and testing strategy for the integrated model remains consistent with that of the clinical model.

Feature visualization and analysis

For feature visualization, gradient-weighted class activation mapping (Grad-CAM) [26] was used to generate the Grad-CAM maps with the last convolution layer of the DL model, representing the model’s attention to different portions of an input radiograph. For the feature analysis, the t-distributed stochastic neighbor embedding (t-SNE) [27] algorithm was used to reduce the 2304-dimensional features extracted by the DL model to three dimensions and then visualize them in 3D space to display the difference in the distribution of the features of GCT and OS.

Statistical analysis

In comparisons of clinical variables between two groups and to assess the impact of different digital X-ray imaging devices on DL model performance, the chi-square test was used for categorical variables (sex, tumor site, digital X-ray imaging device), and the independent samples t test was used for continuous variables (age). Additionally, we grouped patients according to patient age and divided into quartiles to analyze diagnostic performance of different radiologists in different age groups. The accuracy, sensitivity, specificity, area under the ROC curve (AUC), and corresponding 95% confidence interval were calculated. The optimal cutoff value was determined by the maximum Youden index. The ROC curves of various models were compared using the DeLong test. The accuracy of the various models was compared using McNemar’s test. All analyses were conducted using R (Version 4.0.4), SPSS (Version 24.0), and MedCalc (Version 15.8) statistical software. p < 0.05 was considered statistically significant.

Results

Patient characteristics

A total of 333 patients with bone tumors were included in this study, with an average age of 27 years, including 186 males and 147 females. Among them, there were 136 patients with osteolytic OS (male: 72, female: 64), aged 3–61 years, with an average age of 19 years, and 197 patients with GCT of bone (male: 117, female: 80), aged 12–70 years, with an average age of 32 years (as shown in Table 1). There was no significant difference in sex between patients with osteolytic OS and GCT; however, there was a significant difference in age and tumor site between the two types of bone tumors, as shown in Table S1.

Model performance

The DL model achieved an AUC of 0.94 (0.90–0.97) in the training set, 0.97 (0.90–1.00) in the internal test set, and 0.97 (0.88–1.0) in the external test set (all p < 0.001; as shown in Table 2, Fig. 2a). According to the DeLong test, we found no evidence of a significant difference in the performance of the DL model between the internal test set and external test set (p = 0.79). Additionally, there was no significant difference in the DL model performance among digital X-ray imaging devices (p = 0.43) according to the chi-square test (as shown in Table S5).

Table 2 Diagnostic performance of the deep learning model, clinical model, and integrated model in the training set, internal test set, and external test set
Fig. 2
figure 2

ROC curves of the three models in the internal test set (a) and in the external test set (b). CLI, clinical model; DL, DL model

The logistic regression clinical model based on clinical variables achieved an AUC of 0.77 (0.65–0.87, p = 0.001) in the internal test set and 0.64 (0.50–0.76, p = 0.17) in the external test set (as shown in Table 2, Fig. 2b).

According to the DeLong test, the diagnostic performance of the DL model was superior to that of the clinical model in both the internal (p = 0.008) and external test sets (p < 0.001) and superior to that of the integrated model in both the internal (p = 0.28) and external test sets (p = 0.11).

Diagnostic performance of radiologists’ evaluation with and without the DL model

The comparative results of the diagnostic performance among the DL model, radiologists without model assistance, and radiologists with model assistance are shown in Table 3, S2, S3, and S4. In the total test set composed of the internal and external test sets, there was a significant difference in the diagnostic performance between the DL model and the expert committee “A” (junior radiologists) (p < 0.001). The diagnostic accuracy of the expert committee “A” was significantly improved with the help of the DL model (p = 0.05). Specifically, among the five junior radiologists (> 2 but ≤ 5 years), the accuracy of the DL model was better than that of the radiologists (93.1% vs. 52.6%, 55.2%, 81.0%, 75.0%, 78.5%; all p < 0.05), and the accuracy of all the junior radiologists was improved with the aid of the model (52.6% to 86.2%, p < 0.001; 55.2% to 87.9%, p < 0.001; 81.0% to 90.5%, p = 0.49; 75.0% to 80.2%, p = 0.35; 78.5% to 92.2%, p = 0.003; as shown in Table S2). Additionally, the diagnostic performance of the DL model was superior to that of the expert committee “C” (senior radiologists), and the diagnostic accuracy of the expert committee “C” was improved with the help of the DL model but the difference was insignificant. Specifically, among the four senior radiologists (> 8 years), the accuracy of the DL model was better than that of the radiologists (93.1% vs. 85.3%, 85.3%, 88.8%, 86.2%; p = 0.06, p = 0.06, p = 0.25, p = 0.09), and the accuracy of three of the senior radiologists was improved with the aid of the model (85.3% to 87.1%, p = 0.70; 85.3% to 90.5%, p = 0.23; 88.8% to 92.2%, p = 0.37; as shown in Table S4).

Table 3 Diagnostic performance comparison among the DL model, radiologist evaluation without model assistance, and radiologist evaluation with model assistance

Among 16 radiologists in reading radiographs from 116 patients, 9 had shorter diagnostic time with DL assistance (decrease of 1 to 15 min), while 4 had longer diagnostic time (increase of 5 to 10 min) and 3 remained unchanged (Tables S2–S4). Patients in the total test set were further grouped according to patient age, divided into quartiles: 27 patients (OS 21 and GCT 4) in group A; 30 patients (OS 4 and GCT 26) in group B; 31 patients (OS 1 and GCT 30) in group C; and 28 patients (OS 7 and GCT 22) in group D. The statistical results are shown in Table 4. In the 21–30-year age group, the diagnostic performance of the DL model was better than that of the junior radiologist group (p = 0.003).

Table 4 Comparison of diagnostic performance between the DL model and radiologist evaluation in the patient subgroups

Feature visualization and analysis

For the 116 patients in the total test set, the main features of the DL model concentrated in the area of bone destruction caused by the tumor on the radiographs and their corresponding overlapping Grad-CAM heatmaps. The red areas in the heatmaps were simultaneously present in 95 patients’ bone tumor areas on the anteroposterior and lateral radiographs (shown in Figs. 3 and 4), but only in 19 patients’ bone tumor areas on the anteroposterior radiographs and 2 patients’ bone tumor areas on the lateral radiographs.

Fig. 3
figure 3

Anteroposterior (a) and lateral (b) radiographs of GCT of the distal femur and their overlapping Grad-CAM maps

Fig. 4
figure 4

Anteroposterior (a) and lateral (b) radiographs of OS of the proximal tibia and their overlapping Grad-CAM maps

The t-SNE feature analysis results from the total test set showed that the spatial distribution of extracted image features for GCT and OS were different (shown in Fig. 5), indicating that the spatial distribution of the features extracted by the DL model was involved in the differential diagnosis of the two diseases.

Fig. 5
figure 5

3D display of features extracted by the DL model in the total test set: red crosses represent osteolytic OS patients; green dots represent GCT patients

Discussion

In this multicenter study, DL models were developed and validated to distinguish between osteolytic OS and GCT on radiographs and compared to radiologists. Overall, the DL model showed a higher diagnostic accuracy than the clinical model or integrated model (0.97 vs. 0.72, 0.91). In the test set, the DL model achieved higher accuracy than the junior expert committee (93.1% vs. 72.4%; p < 0.001) and was comparable to the intermediate and senior expert committee (93.1% vs. 88.8%, 87.1%; p = 0.25, p = 0.35, respectively). With DL model assistance, the performance of junior expert committee improved (72.4% to 91.4%, p = 0.051).

This study revealed a significant difference in age between the two types of bone tumors, which is consistent with previous literature, indicating that patient age holds some diagnostic value in distinguishing between the two tumor types [11]. Regarding the tumor site, this study revealed a statistically significant difference between osteolytic OS and GCT, which is inconsistent with previous research about OS and GCT [6,7,8]. We believe that this may be attributed to sample bias due to the limited number of osteolytic OS patients included. The logistic regression model based on clinical characteristics exhibited poor performance in the differential diagnosis of the two tumor sites in the overall test set (AUC = 0.72).

In clinical practice, radiologists may misdiagnose osteolytic OS as GCT of bone and vice versa due to the similar appearances on radiographs. The diagnostic accuracy of radiologists in distinguishing between osteolytic OS and GCT of bone is greatly influenced by their clinical experience in reading musculoskeletal radiographs (accuracy ranging from 52.6% to 90.5%, Tables S2, S3, S4). This study used the MRMC method and voting rules (minority follows majority) to demonstrate that radiologists with different levels of clinical experience exhibit different diagnostic performances (accuracy: 76.7%, 88.8%, and 87.1%). Greater clinical experience in reading musculoskeletal radiographs led to improved accuracy of the radiologists in differentiating between osteolytic OS and GCT of bone, which aligns with the learning growth curve characteristics [7]. We conducted a stratified analysis of patients with either type of tumor at different ages. The study revealed that within the patient group with age range of 21 to 30 years, the radiologists’ diagnostic accuracy was lower, while the DL model demonstrated significantly higher accuracy (Table 4). This indicates that when faced with patients of similar age presenting with these aggressive bone tumors, radiologists may experience reduced diagnostic accuracy due to the similarities in radiographic appearance. When 16 radiologists evaluated all radiographs with AI, the evaluation time varied compared to that without AI assistance (Tables S2–S4), which may be due to the different experiences. Of these, 11 radiologists’ diagnostic accuracy with AI was improved, suggesting that artificial intelligence can provide valuable information. Our study demonstrated that the DL model outperformed both the clinical model and the combined model. This suggests that although patient age has a certain reference value in the differential diagnosis between tumor types, the overlap among the patient ages may have reduced the diagnostic accuracy. In contrast, the radiographic appearance can reflect the underlying pathology and biological behavior of the bone tumors, highlighting the superior diagnostic value of the DL model based on radiograph features.

Our proposed DL model identified the tumor area and its surrounding normal bone in the full field of the radiograph, rather than only choosing the tumor area [18, 21], avoiding the need for tumor area labeling. Additionally, the DL model takes into account not only the bone tumor itself but also the shape and size of the bone in which the tumor is located, thereby capturing important variations associated with age. The DL model proposed in this study may include clinical information about age and tumor site, resulting in better diagnostic performance. According to the Grad-CAM heatmaps and t-SNE feature analysis, the features extracted by the DL model concentrated in the bone destruction area caused by the tumors and were spatially able to differentiate between the two types of tumors to some extent.

This study was subject to several limitations. First, the research only focused on two primary aggressive bone tumors around the knee joint, OS and GCT of bone, without considering other bone tumors. In our opinion, the region around the knee joint is the most common site for primary bone tumors, and the most common aggressive bone tumors among them are OS and GCT of bone [3,4,5,6,7,8]. The radiographic manifestations of osteolytic OS and GCT are similar, which can easily confuse the diagnosis [10]. In the future, we intend to incorporate other types of bone tumors to expand the application scope of our model in real clinical scenarios. Second, this study only utilized radiographic information for the establishment of the DL model without incorporating other imaging modalities, such as CT or MRI. This is because digital radiography is recognized as a first-line imaging modality for bone lesion evaluation [11], facilitating the widespread implementation of the DL model in clinical practice. Third, we used the MRMC method to evaluate the radiologist’s diagnostic performance, but the sample size of the patients in the external test set was limited. To further evaluate the DL model, we may continue to collect more data in future studies.

In conclusion, we developed a DL model for the differential diagnosis of osteolytic OS and GCT of bone on knee radiographs. Our model outperformed junior radiologists in terms of diagnostic accuracy, required less time, and enhanced the diagnostic performance of radiologists as an assistive tool, which demonstrated its potential for accurate differential diagnosis in clinical applications.

Availability of data and materials

The data underlying this article will be shared on reasonable request to the corresponding author.

Abbreviations

DL:

Deep learning

GCT:

Giant cell tumor

OS:

Osteosarcoma

References

  1. Choi JH, Ro JY (2021) The 2020 WHO classification of tumors of bone: an updated review. Adv Anat Pathol 28:119–138

    Article  CAS  PubMed  Google Scholar 

  2. Anderson WJ, Doyle LA (2021) Updates from the 2020 World Health Organization classification of soft tissue and bone tumours. Histopathology 78:644–657

    Article  PubMed  Google Scholar 

  3. Do NT, Jung ST, Yang HJ, Kim SH (2021) Multi-level seg-unet model with global and patch-based X-ray images for knee bone tumor detection. Diagnostics (Basel) 11:691

    Article  CAS  PubMed  Google Scholar 

  4. Bahamonde L, Catalan J (2006) Bone tumors around the knee: risks and benefits of arthroscopic procedures. Arthroscopy 22:558–564

    Article  PubMed  Google Scholar 

  5. Voskuil R, Evenski AJ, Montgomery C, Emory C (2019) Malignant bone tumors of the knee: how to identify and treat. J Knee Surg 32:305–314

    Article  PubMed  Google Scholar 

  6. Tepelenis K, Papathanakos G, Kitsouli A et al (2021) Osteochondromas: an updated review of epidemiology, pathogenesis, clinical presentation, radiological features and treatment options. In Vivo 35:681–691

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  7. Ritter J, Bielack S (2010) Osteosarcoma. Ann Oncol 21:320–325

    Article  Google Scholar 

  8. BasuMallick A, Chawla SP (2021) Giant cell tumor of bone: an update. Curr Oncol Rep 23:51

    Article  CAS  Google Scholar 

  9. Garcia RA, Inwards CY, Unni KK (2011) Benign bone tumors–recent developments. Semin Diagn Pathol 28:73–85

    Article  PubMed  Google Scholar 

  10. Ariga A, Kanda H, Ando Y, Gokita T (2020) Giant cell tumor with atypical imaging implying osteosarcoma. Clin Nucl Med 45:892–893

    Article  PubMed  Google Scholar 

  11. von Schacky CE, Wilhelm NJ, Schäfer VS et al (2021) Multitask deep learning for segmentation and classification of primary bone tumors on radiographs. Radiology 301:398–406

    Article  Google Scholar 

  12. Miller TT (2008) Bone tumors and tumorlike conditions: analysis with conventional radiography. Radiology 246:662–674

    Article  PubMed  Google Scholar 

  13. Nguyen JC, Baghdadi S, Pogoriler J, Guariento A, Rajapakse CS, Arkader A (2022) Pediatric osteosarcoma: correlation of imaging findings with histopathologic features, treatment, and outcome. Radiographics 42:1196–1213

    Article  PubMed  Google Scholar 

  14. He Y, Pan I, Bao B et al (2020) Deep learning-based classification of primary bone tumors on radiographs: a preliminary study. EBioMedicine 62:103121

    Article  PubMed  PubMed Central  Google Scholar 

  15. Zhou X, Wang H, Feng C et al (2022) Emerging applications of deep learning in bone tumors: current advances and challenges. Front Oncol 12:908873

    Article  PubMed  PubMed Central  Google Scholar 

  16. Eweje FR, Bao B, Wu J et al (2021) Deep learning for classification of bone lesions on routine MRI. EBioMedicine 68:103402

    Article  PubMed  PubMed Central  Google Scholar 

  17. Noguchi S, Nishio M, Sakamoto R et al (2022) Deep learning-based algorithm improved radiologists’ performance in bone metastases detection on CT. Eur Radiol 32:7976–7987

    Article  PubMed  Google Scholar 

  18. Li J, Li S, Li X et al (2022) Primary bone tumor detection and classification in full-field bone radiographs via YOLO deep learning model. Eur Radiol. https://doi.org/10.1007/s00330-022-09289-y

    Article  PubMed  PubMed Central  Google Scholar 

  19. Murphey MD, Kransdorf MJ (2021) Staging and classification of primary musculoskeletal bone and soft-tissue tumors according to the 2020 WHO update, from the AJR Special Series on Cancer Staging. AJR Am J Roentgenol 217:1038–1052

    Article  PubMed  Google Scholar 

  20. Park CW, Oh SJ, Kim KS (2022) Artificial intelligence-based classification of bone tumors in the proximal femur on plain radiographs: system development and validation. PLoS One 17:e0264140

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  21. von Schacky CE, Wilhelm NJ, Schäfer VS (2022) Development and evaluation of machine learning models based on X-ray radiomics for the classification and differentiation of malignant and benign bone tumors. Eur Radiol 32:6247–6257

    Article  Google Scholar 

  22. Breden S, Hinterwimmer F, Consalvo S et al (2023) Deep learning-based detection of bone tumors around the knee in X-rays of children. J Clin Med 12(18):5960

    Article  PubMed  PubMed Central  Google Scholar 

  23. Obuchowski NA, Bullen J (2022) Multireader diagnostic accuracy imaging studies: fundamentals of design and analysis. Radiology 303(1):26–34

    Article  PubMed  Google Scholar 

  24. Liu R, Pan D, Xu Y (2022) A deep learning-machine learning fusion approach for the classification of benign, malignant, and intermediate bone tumors. Eur Radiol 32:1371–1383

    Article  PubMed  Google Scholar 

  25. Tan M, Le Q (2019) Efficientnet: rethinking model scaling for convolutional neural networks. Int Conf Mach Learn PMLR 2019:6105–6114

    ADS  Google Scholar 

  26. Selvaraju RR, Cogswell M, Das A, Vedantam R, Parikh D, Batra D (2020) Grad-CAM: visual explanations from deep networks via gradient-based localization. Int J Comput Vis 128:336–359

    Article  Google Scholar 

  27. Van Der Maaten L, Hinton G (2008) Visualizing data using t-SNE. J Mach Learn Res 9:27

    Google Scholar 

Download references

Acknowledgements

We thanked 16 radiologists (Dr. Wenyao Fan, Dr. Zedi Huang, Dr. Zifeng Li, Dr. Sicong Li, Dr. Rui Mi, Dr. Xiaping Xia, Dr. Chunyang Yang, Dr. Xingyan Xie, Dr. Shuanghao Ke, Dr. Tianming Guan, Dr. Dan Wei, Dr. Haoqin Zhang, Dr. Ruodai Wu, Dr. Ximei Peng, Dr. Shuifang Wei, Dr. Yaocheng Wen) for their help in evaluating the images.

Funding

This work was funded by Natural Science Foundation of Guangdong Province, China (Grant No.2022A1515011593) and Medical Science, Technology Foundation of Guangdong Province, China (Grant No.A2021010) and Shenzhen Municipal Scheme for Basic Research (JCYJ20210324100208022).

Author information

Authors and Affiliations

Authors

Contributions

XFD, STF, and ZHG were responsible for the overall study design. JJS, HXL, LO, DYX, YS, TMG, HYO, DMD, and ZHG supervised the data collection. JJS, HXL, BL, and RHL performed data analysis. JJS and HXL completed manuscript drafting. LO, DYX, YS, TMG, HYO, DMD, BSH, STF, XFD, and ZHG were responsible for manuscript editing. All authors read, discussed, and approved the final version of the manuscript. All authors had full access to the data in the study and took responsibility for the integrity of the data and the accuracy of the data analysis as well as the decision to submit for publication.

Corresponding authors

Correspondence to Shiting Feng, Xianfen Diao or Zhenhua Gao.

Ethics declarations

Ethics approval and consent to participate

This study was conducted with the approval of the Ethics Committee of the First Affiliated Hospital, Sun Yat-Sen University.

Consent for publication

Written informed consent was waived by the Institutional Review Board.

Competing interests

The authors of this manuscript declare no relationships with any companies, whose products or services may be related to the subject matter of the article.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Additional file 1: Table S1.

Comparison of the clinical characteristics of the patients with osteolytic OS and GCT. Table S2. Diagnostic performance comparison among the DL model, radiologist’s evaluation without model assistance, and radiologist’s evaluation with model assistance in group A. Table S3. Diagnostic performance comparison of the DL model, radiologist’s evaluation without model assistance, and radiologist’s evaluation with model assistance in group B. Table S4. Diagnostic performance comparison of the DL model, radiologist’s evaluation without model assistance, and radiologist’s evaluation with model assistance in group C. Table S5. Detailed information on digital X-ray imaging devices.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Shao, J., Lin, H., Ding, L. et al. Deep learning for differentiation of osteolytic osteosarcoma and giant cell tumor around the knee joint on radiographs: a multicenter study. Insights Imaging 15, 35 (2024). https://doi.org/10.1186/s13244-024-01610-1

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s13244-024-01610-1

Keywords