Skip to main content

Has the quality of reporting improved since it became mandatory to use the Standards for Reporting Diagnostic Accuracy?

Abstract

Objectives

To investigate whether making the Standards for Reporting Diagnostic Accuracy (STARD) mandatory by the leading journal ‘Radiology’ in 2016 improved the quality of reporting of diagnostic accuracy studies.

Methods

A validated search term was used to identify diagnostic accuracy studies published in Radiology in 2015 and 2019. STARD adherence was assessed by two independent reviewers. Each item was scored as yes (1 point) if adequately reported or as no (0 points) if not. The total STARD score per article was calculated. Wilcoxon–Mann–Whitney tests were used to evaluate differences of the total STARD scores between 2015 and 2019. In addition, the total STARD score was compared between studies stratified by study design, citation rate, and data collection.

Results

The median number of reported STARD items for the total of 66 diagnostic accuracy studies from 2015 to 2019 was 18.5 (interquartile range [IQR] 17.5–20.0) of 29. Adherence to the STARD checklist significantly improved the STARD score from a median of 18.0 (IQR 15.5–19.5) in 2015 to a median of 19.5 (IQR 18.5–21.5) in 2019 (p < 0.001). No significant differences were found between studies stratified by mode of data collection (prospective vs. retrospective studies, p = 0.68), study design (cohort vs. case–control studies, p = 0.81), and citation rate (two groups divided by median split [< 0.56 citations/month vs. ≥ 0.56 citations/month], p = 0.54).

Conclusions

Making use of the STARD checklist mandatory significantly increased the adherence with reporting standards for diagnostic accuracy studies and should be considered by editors and publishers for widespread implementation.

Critical relevance statement

Editors may consider making reporting guidelines mandatory to improve the scientific quality.

Graphical Abstract

Key points

  • The median number of reported STARD items for all studies was 18.5.

  • Adherence to the STARD checklist significantly improved from 2015 to 2019.

  • Editors may consider making reporting guidelines mandatory to improve the scientific quality.

Introduction

Diagnostic accuracy studies play an important role in introducing a new diagnostic test into clinical practice [1] because diagnostic test accuracy compared with an established reference standard provides information about how well diagnostic tests may improve clinical decision making [2]. Diagnostic accuracy studies are at risk of bias [3, 4] because measures of diagnostic accuracy, such as sensitivity and specificity, are not fixed values but reflect the performance of the index test under certain study and test circumstances [2, 4,5,6]. Therefore, a detailed description of the methodology, setting, and subjects is crucial for readers to judge the trustworthiness of the results (internal validity) and appraise the applicability of the medical test in clinical practice (external validity, i.e., generalizability) [5].

In the past, studies published in journals with high impact factors had shortcomings in reporting diagnostic accuracy, leading to overestimation of test performance and improper recommendations with disadvantages for patient outcomes [7]. Furthermore, “incomplete reporting has been identified as a major source of avoidable waste in biomedical research” [8] and growing health care costs [9, 10]. Following the successful CONSORT (Consolidated Standards of Reporting Trials) initiative [11], the Standards for Reporting Diagnostic Accuracy (STARD) statement was published in 2003 [12] and updated in 2015 [8]. It consists of a checklist of 30 essential items to guide authors in planning and reporting diagnostic accuracy studies [8]. Since then, STARD has been endorsed by more than 200 biomedical journals [13].

In February 2016, the use of reporting guideline checklists became mandatory for all original research manuscripts submitted to Radiology, which had endorsed STARD since its publication [14, 15]. We used this as an opportunity to investigate the reporting quality of diagnostic accuracy studies published in Radiology before and after guideline implementation and to evaluate whether reporting quality improved after mandating reporting guideline use. Further, we analyzed whether the total STARD score differed between studies stratified by study design, citation rate, and data collection.

Methods

This analysis, even not fulfilling all criteria of a meta-analysis, complied with the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) guidelines [16]. Our analysis was therefore not eligible for registration in the international prospective register of systematic reviews (PROSPERO) [17].

Literature search

To identify diagnostic accuracy studies published in Radiology in 2015 and 2019, we performed a systematic literature review in MEDLINE (using PubMed) using a validated search strategy proposed by Devillé et al. [18], which served as the basis (Additional file 1: Table S1) for our search strategy. The search strategy is detailed in Additional file 1: Table S2. Additionally, we manually searched the website of Radiology for additionally eligible studies which were not identified in MEDLINE. PubMed was last searched on April 8, 2020; the website of Radiology on June 23, 2020.

Study selection

Articles were included if (1) there was at least one measure of diagnostic accuracy (sensitivity, specificity, likelihood ratios, predictive values, area under the receiver operator curve, accuracy), (2) the results of at least one medical imaging test were compared against a reference standard, and (3) the study was conducted in human subjects. Articles dealing with predictive or prognostic accuracy as well as commentaries, editorials, letters, reviews and the development of models were excluded. Two reviewers (A.S., an advanced medical student, with 3 years of experience in performing literature reviews of diagnostic accuracy studies, and A.T., a dentist, with 1 year of experience in this field) independently reviewed all studies for inclusion; discrepancies were resolved in consensus meetings with a third reviewer (B.K., a physician with 8 years of experience in radiological research). First, we went through all titles, keywords, and abstracts to identify potentially eligible articles. Finally, the full texts of the articles remaining after this step were assessed for eligibility. The following information was extracted from each included article: publication date (2015 vs. 2019), mode of data collection (prospective vs. retrospective), and study design (cohort vs. case–control study).

Adherence to STARD

Although two studies reported good reproducibility of the STARD checklist [19, 20], two reviewers (A.S., A.T.) independently pilot-tested the STARD checklist on four articles from 2014 TO 2020. Uncertainties regarding the explanation and elaboration of each item were discussed to make sure that the reviewers agreed about the interpretation of the STARD criteria. For the purpose of our analysis, we excluded item 11 (rationale for choosing the reference standard (if alternatives exist)) from the STARD checklist following the approach of Wilczynski [5, 21] because in case that no information regarding this item was found in an article, it was not possible to reliably determine whether the authors simply forgot to mention it in the manuscript or ignored it in their study because no alternatives existed. Thus, the finally used checklist consisted of 29 items. Each adequately reported item was scored yes (1 point) or no (0 points). As items 10, 12, 13, and 21 refer to both the index test and the reference standard, we split these items and counted each of the two modalities as ½ item (0.5 points). Both reviewers (A.S., A.T.) evaluated independently all included articles according to the 29-item checklist. Discrepancies were resolved in consensus meetings. If no consensus could be reached, a third reviewer (B.K.) helped to make the final decision. Reviewers were not blinded to journal, publication year, and authors. The reviewers did not evaluate the methodological quality [22] of the study but the quality of reporting [8].

Data and statistical analysis

We calculated the total STARD score for each included article by adding the number of reported STARD items (range, 0–29). The median and interquartile range (IQR) for the total STARD scores were calculated. Assuming that each item is of equal weight, a higher score suggests a better reporting quality. Wilcoxon–Mann–Whitney’s test was used to compare the STARD score between papers published in 2015 and papers published in 2019. This comparison was performed with inclusion of all studies as well as with inclusion of the following subgroups: prospective studies, retrospective studies, cohort studies, case control studies, studies with a citation rate above median, and studies with citation rate below median. In addition, Wilcoxon–Mann–Whitney’s test was applied to analyze whether the total STARD score differed between studies stratified by study design (cohort vs. case control studies), citation rate (equal or above vs. below median citation rate), and data collection (prospective vs. retrospective). Vargha and Delaney’s A was used as effect size measure.

The citation rate was calculated by dividing the total number of times each article had been cited by April 30, 2021, by the total number of months since publication (print version). These numbers were provided by the citation index reported in Web of Science (Thomson Reuters, New York, NY, USA).

Cohen´s κ statistics was used to calculate interrater reliability. According to Landis and Koch [23], a κ value of 0.4–0.60 indicates moderate; a κ value of 0.61–0.80, substantial; and a κ value of 0.81–1.00, (almost) perfect agreement between the reviewers. p values less than < 0.05 were considered statistically significant. The code for the statistical analysis was written in R language, version 4.2.0.

Results

Search results and study characteristics

The systematic literature search identified 289 publications, the manual search 354. Independent assessment of title, abstract, and keywords according to our inclusion criteria by two reviewers (A.S., A.T.) identified 75 potentially relevant articles from the literature search and 77 from the manual search. After exclusion of 63 duplicates, two readers (A.S., A.T.) independently examined the full texts of 89 articles. The multilevel selection process finally led to the identification of 66 eligible articles. The PRISMA 2020 flow diagram (Fig. 1) [16] provides detailed information on the study selection process. The median citation rate was 0.56 citations per month (range: 0.1–2.35). Baseline characteristics of all studies included are compiled in Table 1.

Fig. 1
figure 1

PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) 2020 flow diagram of selected diagnostic accuracy studies published in Radiology in 2015 and 2019 with detailed description for exclusion criteria. N/A, not applicable

Table 1 Characteristics of included studies published in Radiology

Adherence to STARD

The median number of reported STARD items for the 66 diagnostic accuracy studies analyzed was 18.5 (IQR 17.5–20.0) of 29, with a range of 13 to 24.5. A list of all included studies with individual total STARD scores is provided in Additional file 1: Table S3.

Overall agreement in scoring the 33 (sub-)items was 85% (1859/2178). Based on a Cohen´s κ statistic of 0.70 (95% CI 0.66, 0.73), interrater reliability was substantial.

Studies published in 2019 showed a 2.2 points higher (95% CI 1.2–3.3, p < 0.001, Vargha and Delaney’s A = 0.24) total STARD score than those published in 2015 (19.5 [IQR 18.5–21.5] vs. 18.0 [15.5–19.5]), indicating a better overall reporting quality. A significant difference in the median STARD score between 2015 and 2019 was also found for subgroups consisting of prospective studies (3.3, 95% CI 1.6–5.0, p < 0.001, Vargha and Delaney’s A = 0.14), cohort studies (2.0, 95% CI 0.8–3.1, p = 0.002, Vargha and Delaney’s A = 0.28), case control studies (4.15, 95% CI 1.0–7.3, p = 0.017, Vargha and Delaney’s A = 0.0), studies with a citation rate above median (2.5, 95% CI 0.9–4.1, p = 0.003, Vargha and Delaney’s A = 0.23), and studies with a citation rate below median (2.2, 95% CI 0.6–3.8, p = 0.008, Vargha and Delaney’s A = 0.24). Retrospective studies did not show a significant difference in the reporting quality between 2015 and 2019 (1.4, 95% CI − 0.1 to 2.9, p = 0.065, Vargha and Delaney’s A = 0.36).

No difference in the total STARD score was found between studies stratified by mode of data collection (p = 0.68, Vargha and Delaney’s A = 0.47), study design (p = 0.81, Vargha and Delaney’s A = 0.53), and citation rate (p = 0.54, Vargha and Delaney’s A = 0.54). Detailed results are provided in Table 2.

Table 2 Summary of Wilcoxon–Mann–Whitney tests

Item-specific adherence to STARD

The results for adherence to individual STARD items and comparisons of reporting frequencies between studies published in 2015 and 2019 are shown in Table 3 and Fig. 2. Although STARD has been mandatory in Radiology since 2016, seven relevant items of the STARD checklist (item 12b (prespecified definition of test positivity cutoffs of the reference standard), items 15 and 16 (handling of missing and indeterminate results), 18 (sample size calculation), 23 (a cross-tabulation), 25 (adverse events), and 28 (registration number) were infrequently reported (< 33%) in the 27 diagnostic accuracy studies published in Radiology in 2019 with very poor reporting rates for items 18 and 23 (11%, 3/27). Providing a registration number, however, notably improved by 16%-points between 2015 (3%, 1/39) and 2019 (19%, 5/27). Six of the 33 (sub-)items were moderately (33–66% of studies) reported in 2019. Especially reporting of study objectives and hypotheses (item 4) nearly doubled between 2015 (31%, 12/39) and 2019 (59%, 16/27). Last, 20 of 33 items were frequently (> 66%) reported by studies published in 2019. Particularly item 9 (sample selection) improved by 37%-points (56%, 22/39 vs. 93%, 25/27), item 19 (flow diagram) by 58%-points (38%, 15/39 vs. 96%, 26/27), and item 20 (baseline demographics) by 31%-points (62%, 24/39 vs. 93%, 25/27) between 2015 and 2019. For all items (items 10, 12, 13, 21) concerning both the index test and reference standard, the information was more frequently reported for the index test.

Table 3 Reporting frequencies of individual STARD items for all studies and comparison of reporting frequencies with STARD being recommended (2015) vs. STARD being mandatory (2019) in Radiology
Fig. 2
figure 2

Graph shows results of comparison of overall adherence (%) to 29 STARD 2015 checklist items and reporting of individual items between studies published in 2015, with STARD being recommended, and those published in 2019, with STARD being mandatory. Studies published in 2019 adhered in general to more items of the STARD 2015 checklist. However, items 12b (prespecified definition of test positivity cutoffs of the reference standard), 14 (methods for estimating diagnostic accuracy measures), 15 and 16 (handling of missing and indeterminate results) and item 22 (time interval between tests) were more often reported in 2015. When referring to the split items 10, 12, 13 and 21, the graph reveals that information about the index test (a items) was more frequently reported than for the reference standard (b items). STARD, Standards for Reporting Diagnostic Accuracy

Discussion

Shortcomings in reporting diagnostic accuracy studies hamper an objective assessment of the clinical performance of diagnostic tests [24]. To improve reporting quality, the STARD statement was developed [12]. In our analysis, we assessed the reporting quality of 66 diagnostic accuracy studies published before and after using the STARD guidelines became mandatory. We found that (1) adherence to the STARD 2015 checklist was moderate (median 18.5 of 29 items), (2) mandating guideline use had a significant effect on the total STARD score (p < 0.001), and (3) that further improvement is especially necessary to ensure adequate reporting of items that are prone to bias and variation [3, 8], such as prespecified definitions of test positivity cutoffs, handling of indeterminate and missing results, providing sample size calculations, and cross-tabulations.

Compared with a previous study by this author group, we found a higher average number of reported items than in diagnostic accuracy studies published in European Radiology [25]. This could be due to the fact that European Radiology is a STARD-endorsing journal, while the use of the STARD checklist is mandatory for studies submitted to Radiology. Therefore making STARD and other checklists mandatory may be considered by Insights into Imaging and other journals of the European Society of Radiology Journal Family to improve the scientific quality. Nevertheless, we found no significant differences in the STARD score between studies stratified by study design, data collection, and citation rate within the studies published in European Radiology as well. This is in-line with the results of this assessment. Choi et al. analyzed 63 studies published from 2011 to 2015 in the Korean Journal of Radiology, a STARD-endorsing journal [26]. The mean total STARD score of their analysis was 20 of 27 items (74%), indicating a relatively high overall reporting quality. This could be due to the fact that the authors excluded item 28 (providing a registration number). In our study, we found the lowest adherence rate for this item (9%, 6/66), which might have affected our total scores. Furthermore, Choi et al. also found no effect of the citation rate on STARD adherence. This is in-line with the results reported by Hogan et al. [27] in 2020 and in contrast with the results of the large assessment by Dilauro et al. [28] who found a weak positive correlation between the total STARD score and the citation rate. Most of the above-mentioned studies additionally compared the reporting quality of diagnostic accuracy studies in journals that had endorsed STARD with those that did not. Their results revealed that STARD endorsement had a relevant impact on the total STARD score [26, 27, 29]. To the best of our knowledge, ours is the first investigation explicitly assessing the impact of mandatory guideline use on reporting quality over time.

A summary of the relevant literature on STARD adherence is provided in Table 4.

Table 4 Summary of results of relevant studies dealing with the reporting quality in diagnostic accuracy studies using the STARD Checklist

Our study has some potential limitations. First, we searched MEDLINE using a validated search strategy to identify relevant diagnostic accuracy studies. Since the search strategy has 80.0% sensitivity and 97.3% specificity [18], some studies may not have been recognized by our search filter. We minimized this risk by additionally identifying further studies by a manual search of the website. Second, we excluded item 11 with the qualifier “if alternatives exist” from the original STARD 2015 checklist for reasons mentioned above. This may have affected the results of our analysis depending on the performance of item 11. Additionally, we focused on a single journal to be able to draw direct comparisons after a policy change in 2016. Due to these two points, the generalizability of our results may be limited and further studies in journals making such policy change are warranted. Also, by choosing articles published in 2019 instead of 2020 or 2021, the immediacy of our data might be affected. We made this decision due to the ongoing COVID-19 crisis since 2020, which brought a great increase in submissions about this single topic with reductions in diagnostic accuracy studies. Third, we were rather strict in assigning scores. For example, baseline characteristics (item 20) were only judged as being satisfactorily reported when some information other than sex and age, such as underlying conditions, was also provided. In addition, several items are prone to subjective assessment. To reduce rater bias, we explicitly defined each item, did pilot exercises, and resolved discrepancies in consensus meetings. Finally, the update of STARD was released in October 2015. Consequently, some authors of studies published in 2015 may not yet have had access to the revised checklist. Nevertheless, we decided to use this list for all studies because the update was intended to facilitate the use of STARD and to highlight items prone to bias and variation, as suggested by recent evidence [8]. Interestingly, five of nine new checklist items were already frequently reported in our study sample: Item 2 (structured summary), 3 (clinical background), 26 (study limitations), 27 (implications for practice), and 30 (sources of funding), which may suggest that reporting these items has already been adopted.

In conclusion, our results showed overall adherence to reporting guidelines in diagnostic accuracy studies to be moderate to good. With the STARD guidelines being mandatory since 2016, studies published in 2019 had a relevantly higher total STARD score than those published in 2015. Making the STARD guidelines mandatory may thus positively affect the reporting quality of diagnostic accuracy studies. This should encourage journals and publishers to add mandatory reporting guidelines to their author instructions.

Availability of data and materials

Used data sets analyzed during the study are available from the corresponding author upon reasonable request.

Abbreviations

CI:

Confidence interval

IQR:

Interquartile range

STARD:

Standards for Reporting Diagnostic Accuracy

References

  1. Sackett DL, Haynes RB (2002) The architecture of diagnostic research. BMJ 7336:539

    Article  Google Scholar 

  2. van Stralen KJ, Stel VS, Reitsma JB, Dekker FW, Zoccali C, Jager KJ (2009) Diagnostic methods I: sensitivity, specificity, and other measures of accuracy. Kidney Int 12:1257–1263

    Article  Google Scholar 

  3. Whiting P, Rutjes AWS, Reitsma JB, Glas AS, Bossuyt PMM, Kleijnen J (2004) Sources of variation and bias in studies of diagnostic accuracy: a systematic review. Ann Intern Med 3:189–202

    Article  Google Scholar 

  4. Rutjes AWS, Reitsma JB, Di Nisio M, Smidt N, van Rijn JC, Bossuyt PMM (2006) Evidence of bias and variation in diagnostic accuracy studies. CMAJ 4:469–476

    Article  Google Scholar 

  5. Cohen JF, Korevaar DA, Altman DG et al (2016) STARD 2015 guidelines for reporting diagnostic accuracy studies: explanation and elaboration. BMJ Open 11:e012799–e012799

    Article  Google Scholar 

  6. Irwig L, Bossuyt P, Glasziou P, Gatsonis C, Lijmer J (2002) Designing studies to ensure that estimates of test accuracy are transferable. BMJ 7338:669–671

    Article  Google Scholar 

  7. Ochodo EA, de Haan MC, Reitsma JB, Hooft L, Bossuyt PM, Leeflang MM (2013) Overinterpretation and misreporting of diagnostic accuracy studies: evidence of “spin.” Radiology 2:581–588

    Article  Google Scholar 

  8. Bossuyt PM, Reitsma JB, Bruns DE et al (2015) STARD 2015: an updated list of essential items for reporting diagnostic accuracy studies. BMJ Br Med J. https://doi.org/10.1373/clinchem.2015.246280

    Article  Google Scholar 

  9. Chan A-W, Song F, Vickers A et al (2014) Increasing value and reducing waste: addressing inaccessible research. Lancet 9913:257–266

    Article  Google Scholar 

  10. Sailer AM, van Zwam WH, Wildberger JE, Grutters JPC (2015) Cost-effectiveness modelling in diagnostic imaging: a stepwise approach. Eur Radiol 12:3629–3637

    Article  Google Scholar 

  11. Schulz KF, Altman DG, Moher D (2010) CONSORT 2010 Statement: updated guidelines for reporting parallel group randomised trials. BMC Med. https://doi.org/10.1371/journal.pmed.1000251

    Article  PubMed  PubMed Central  Google Scholar 

  12. Bossuyt PM, Reitsma JB, Bruns DE et al (2003) Towards complete and accurate reporting of studies of diagnostic accuracy: the STARD initiative. BMJ 7379:41–44

    Article  Google Scholar 

  13. Cohen J, Korevaar DA, Gatsonis C et al (2017) STARD for abstracts: essential items for reporting diagnostic accuracy studies in journal or conference abstracts. BMJ j3751

  14. Levine D, Kressel HY (2015) Radiology 2016: The care and scientific rigor used to process and evaluate original research manuscripts for publication. Radiology 1:6–10

    Google Scholar 

  15. Dewey M, Levine D, Bossuyt PM, Kressel HY (2019) Impact and perceived value of journal reporting guidelines among Radiology authors and reviewers. Eur Radiol 8:3986–3995

    Article  Google Scholar 

  16. Page MJ, McKenzie JE, Bossuyt PM et al (2021) The PRISMA 2020 statement: an updated guideline for reporting systematic reviews. BMJ n71

  17. Booth A, Clarke M, Dooley G et al (2012) The nuts and bolts of PROSPERO: an international prospective register of systematic reviews. Syst Rev 1:2. https://doi.org/10.1186/2046-4053-1-2

    Article  PubMed  PubMed Central  Google Scholar 

  18. Devillé WLJM, Bezemer PD, Bouter LM (2000) Publications on diagnostic test evaluation in family medicine journals: an optimal search strategy. J Clin Epidemiol 1:65–69

    Article  Google Scholar 

  19. Smidt N, Rutjes AWS, van der Windt DAWM et al (2006) Reproducibility of the STARD checklist: an instrument to assess the quality of reporting of diagnostic accuracy studies. BMC Med Res Methodol 1:12

    Article  Google Scholar 

  20. Virgili G, Michelessi M, Miele A et al (2017) STARD 2015 was reproducible in a large set of studies on glaucoma. PLoS One 10:e0186209

    Article  Google Scholar 

  21. Wilczynski NL (2008) Quality of reporting of diagnostic accuracy studies: no change since STARD statement publication–before-and-after study. Radiology 3:817–823

    Article  Google Scholar 

  22. Whiting PF, Rutjes AWS, Westwood ME et al (2011) QUADAS-2: a revised tool for the quality assessment of diagnostic accuracy studies. Ann Intern Med 8:529–536

    Article  Google Scholar 

  23. Landis JR, Koch GG (1977) The measurement of observer agreement for categorical data. Biometrics 1:159–174

    Article  Google Scholar 

  24. Korevaar DA, van Enst W, Spijker R, Bossuyt P, Hooft L (2013) Reporting quality of diagnostic accuracy studies: a systematic review and meta-analysis of investigations on adherence to STARD. Evid Based Med. https://doi.org/10.1136/eb-2013-101637

    Article  PubMed  Google Scholar 

  25. Stahl AC, Tietz AS, Kendziora B, Dewey M (2022) Has the STARD statement improved the quality of reporting of diagnostic accuracy studies published in European Radiology? Eur Radiol. https://doi.org/10.1007/s00330-022-09008-7

    Article  PubMed  PubMed Central  Google Scholar 

  26. Choi YJ, Chung MS, Koo HJ, Park JE, Yoon HM, Park SH (2016) Does the reporting quality of diagnostic test accuracy studies, as defined by STARD 2015, affect citation? Korean J Radiol 5:706–714

    Article  Google Scholar 

  27. Hogan KO, Fraga GR (2020) Compliance with standards for STARD 2015 reporting recommendations in pathology. Am J Clin Pathol 6:828–836

    Article  Google Scholar 

  28. Dilauro M, McInnes MD, Korevaar DA et al (2016) Is There an association between STARD statement adherence and citation rate? Radiology 1:62–67

    Article  Google Scholar 

  29. Hong PJ, Korevaar DA, McGrath TA et al (2018) Reporting of imaging diagnostic accuracy studies with focus on MRI subgroup: adherence to STARD 2015. J Magn Reson Imaging 2:523–544

    Article  Google Scholar 

  30. Smidt N, Rutjes AWS, van der Windt DAWM et al (2006) The quality of diagnostic accuracy studies since the STARD statement. Neurology 5:792

    Article  Google Scholar 

Download references

Acknowledgements

The authors thank D. Levine and H. Kressel for discussion and support. We would also like to thank Bettina Herwig for copy editing.

Funding

Open Access funding enabled and organized by Projekt DEAL.

Author information

Authors and Affiliations

Authors

Contributions

AS and AT collected and interpreted the data. BK did the statistical analysis, while AS and AT were major contributors in writing the manuscript. MD had the idea for this research work and was the supervisor. BK was involved as second supervisor. All authors read and approved the final manuscript. The research was mainly conducted at Ludwig Maximilian University of Munich.

Corresponding author

Correspondence to Benjamin Kendziora.

Ethics declarations

Ethics approval and consent to participate

Not applicable.

Consent for publication

Not applicable.

Competing interests

MD has received grant support from the FP7 Program of the European Commission for the randomized multicenter DISCHARGE trial (EC-GA 603266 in HEALTH.2013.2.4.2-2). He also received grant support from German Research Foundation (DFG) in the Heisenberg Program (DE 1361/14-1, DFG project number: 213705389), graduate program on quantitative biomedical imaging (BIOQIC, GRK 2260/1, DFG project number: 289347353), for fractal analysis of myocardial perfusion (DE 1361/18-1, DFG project number: 392304398), the DFG Priority Programme Radiomics (DFG project number: 402688427) for the investigation of coronary plaque and coronary flow (DE 1361/19-1 [DFG project number: 428222922] and DE 1361/20-1 [DFG project number: 428223139] in SPP 2177/1). He also received funding from the Berlin University Alliance (GC_SC_PC 27) and from the Digital Health Accelerator of the Berlin Institute of Health. MD is the European Society of Radiology (ESR) Publications Chair (2022–2025); the opinions expressed in this article are the author’s own and do not represent the view of ESR. Per the guiding principles of ESR, the work as Publications Chair is on a voluntary basis and only remuneration of travel expenses occurs. MD is also the editor of Cardiac CT, published by Springer Nature, and offers hands-on courses on CT imaging (www.ct-kurs.de). Institutional master research agreements exist with Siemens, General Electric, Philips, and Canon. The terms of these arrangements are managed by the legal department of Charité—Universitätsmedizin Berlin. MD holds a joint patent with Florian Michallek on dynamic perfusion analysis using fractal analysis (PCT/EP2016/071551 and USPTO 2021 10,991,109 approved).

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Additional file 1

: Full search strategy and list of STARD scores for all included articles.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Stahl, AC., Tietz, AS., Dewey, M. et al. Has the quality of reporting improved since it became mandatory to use the Standards for Reporting Diagnostic Accuracy?. Insights Imaging 14, 85 (2023). https://doi.org/10.1186/s13244-023-01432-7

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s13244-023-01432-7

Keywords