Skip to main content

Table 3 TRIPOD adherence of included studies

From: A systematic review of radiomics in pancreatitis: applying the evidence level rating tool for promoting clinical transferability

35 Selected Items in 20 Criteria According to 6 Sections (N = 30)

Study, n (%)

Overall (excluding 5c, 11, 14b, 10c, 10e, 12, 13, 17, and 19a)

478 (61)

Section 1: Title and Abstract

14 (23)

 1. Title—identify developing/validating a model, target population, and the outcome

2 (7)

 2. Abstract—provide a summary of objectives, study design, setting, participants, sample size, predictors, outcome, statistical analysis, results, and conclusions

12 (40)

Section 2: Introduction

37 (62)

 3a. Background—Explain the medical context and rationale for developing/validating the model

30 (100)

 3b. Objective—Specify the objectives, including whether the study describes the development/validation of the model or both

7 (23)

Section 3: Methods

252 (65)

 4a. Source of data—describe the study design or source of data (randomized trial, cohort, or registry data)

30 (100)

 4b. Source of data—specify the key dates

26 (87)

 5a. Participants—specify key elements of the study setting including number and location of centers

30 (100)

 5b. Participants—describe eligibility criteria for participants (inclusion and exclusion criteria)

21 (70)

 5c. Participants—give details of treatment received, if relevant (N = 6)

0 (0)

 6a. Outcome—clearly define the outcome, including how and when assessed

30 (100)

 6b. Outcome—report any actions to blind assessment of the outcome

0 (0)

 7a. Predictors—clearly define all predictors, including how and when assessed

23 (77)

 7b. Predictors—report any actions to blind assessment of predictors for the outcome and other predictors

15 (50)

 8. Sample size—explain how the study size was arrived at

0 (0)

 9. Missing data—describe how missing data were handled with details of any imputation method

0 (0)

 10a. Statistical analysis methods—describe how predictors were handled

24 (80)

 10b. Statistical analysis methods—specify type of model, all model-building procedures (any predictor selection), and method for internal validation

23 (77)

 10d. Statistical analysis methods—specify all measures used to assess model performance and if relevant, to compare multiple models (discrimination and calibration)

30 (100)

 11. Risk groups—provide details on how risk groups were created, if done (N = 0)

0 (0)

Section 4: Results

94 (52)

 13a. Participants—describe the flow of participants, including the number of participants with and without the outcome. A diagram may be helpful

16 (53)

 13b. Participants—describe the characteristics of the participants, including the number of participants with missing data for predictors and outcome

24 (80)

 14a. Model development—specify the number of participants and outcome events in each analysis

25 (83)

 14b. Model development—report the unadjusted association between each candidate predictor and outcome, if done (N = 5)

1 (20)

 15a. Model specification—present the full prediction model to allow predictions for individuals (regression coefficients, intercept)

5 (17)

 15b. Model specification—explain how to the use the prediction model (nomogram, calculator, etc.)

2 (7)

 16. Model performance—report performance measures (with confidence intervals) for the prediction model

22 (73)

Section 5: Discussion

81 (90)

 18. Limitations—Discuss any limitations of the study

30 (100)

 19b. Interpretation—Give an overall interpretation of the results

30 (100)

 20. Implications—Discuss the potential clinical use of the model and implications for future research

21 (70)

Section 6: Validation for Model type 2a, 2b, 3, and 4 (N = 13)

9 (17)

 10c. Statistical analysis methods—describe how the predictions were calculated

0 (0)

 10e. Statistical analysis methods—describe any model updating (recalibration), if done (N = 0)

0 (0)

 12. Development versus validation—Identify any differences from the development data in setting, eligibility criteria, outcome, and predictors

5 (38)

 13c. Participants (for validation)—show a comparison with the development data of the distribution of important variables

4 (31)

 17. Model updating—report the results from any model updating, if done (N = 0)

0 (0)

 19a. Interpretation (for validation)—discuss the results with reference to performance in the development data and any other validation data

0 (0)

  1. TRIPOD Transparent Reporting of a Multivariable Prediction Model for Individual Prognosis or Diagnosis