Skip to main content

How does DICOM support big data management? Investigating its use in medical imaging community

Abstract

The diagnostic imaging field is experiencing considerable growth, followed by increasing production of massive amounts of data. The lack of standardization and privacy concerns are considered the main barriers to big data capitalization. This work aims to verify whether the advanced features of the DICOM standard, beyond imaging data storage, are effectively used in research practice. This issue will be analyzed by investigating the publicly shared medical imaging databases and assessing how much the most common medical imaging software tools support DICOM in all its potential. Therefore, 100 public databases and ten medical imaging software tools were selected and examined using a systematic approach. In particular, the DICOM fields related to privacy, segmentation and reporting have been assessed in the selected database; software tools have been evaluated for reading and writing the same DICOM fields. From our analysis, less than a third of the databases examined use the DICOM format to record meaningful information to manage the images. Regarding software, the vast majority does not allow the management, reading and writing of some or all the DICOM fields. Surprisingly, if we observe chest computed tomography data sharing to address the COVID-19 emergency, there are only two datasets out of 12 released in DICOM format. Our work shows how the DICOM can potentially fully support big data management; however, further efforts are still needed from the scientific and technological community to promote the use of the existing standard, encouraging data sharing and interoperability for a concrete development of big data analytics.

Key points

  • Standardization is crucial for big data capitalization.

  • DICOM supports key actions for imaging data management.

  • The majority of shared research databases does not fully exploit DICOM format.

  • Imaging software tools do not fully support DICOM advanced feature.

  • There is need to fully promote DICOM in data sharing and software development.

Introduction

The modern era is going through a rapid technological evolution, and we are witnessing the production of a huge amount of information that can be worthily enhanced with appropriate management and analysis. We are, in fact, in the so-called big data era, where the value of data can be the real engine of innovation [1].

The healthcare sector, where multi-sources patient information is routinely collected, is gaining volume and complexity. Among big data types, imaging data can be considered the largest in volume. In fact, it covers not only gigapixel images, such as tissues or organs at subcellular resolutions, but also metadata and quantitative measurements. For instance, neuroimaging is currently producing more than 10 petabytes of data every year with a staggering ninefold increase in data complexity (i.e., data acquisition modalities) over the last three decades [2, 3].

Therefore, the standardization of the medical imaging formats plays a crucial role in the effective exploitation of the data and subsequent clinical decision making [1].

The World Health Organization (WHO) recognizes the lack of standardization, together with privacy concerns, as the main barrier to big data exploitation [4]. The Digital Imaging and Communications in Medicine (DICOM) format is the current standard for storing and transmitting medical images, enabling the integration of medical imaging devices such as scanners, servers, workstations, printers, network hardware and picture archiving and communication systems (PACS) from multiple manufacturers [5]. DICOM encompasses raw imaging data and all the metadata related to the procedures of image acquisition and curation, including a series of processes such as de-identification of sensible data, annotation of regions of interest within the medical image, image enhancement or structured reporting.

In addition to DICOM, the most used formats designed for medical images are the analyzed format and its most recent version, Neuroimaging Informatics Technology Initiative (NIfTI), MetaImage (mhd) and nearly raw raster data (NRRD). They are mostly used for saving images for post-processing operations [6]. It is important to note, however, that DICOM remains the standard for the clinical management of images, which are routinely acquired in DICOM and eventually transformed into other formats. For this reason, our work does not focus on the already consolidated use of DICOM as a clinical standard but on the extension of the standard toward big data analytics, as a putative bridge between the clinic and research.

In fact, during the last years, the DICOM steering committee has carefully followed these needs by defining and extending the standard [7, 8]. Nevertheless, the simple definition of a standard is not enough to satisfy needs since the availability to users and the support with appropriate software tools are crucial.

This work aims to evaluate the effective implementation of the DICOM standard in the big data perspective and, in particular:

  • To recognize and introduce how the DICOM standard effectively supports the current challenges in diagnostic imaging management and analytics.

  • To verify whether the standard is implemented in clinical and research practice, investigating the public data shared by the research community and checking how much the standard is actually supported in the main software for management and processing of diagnostic images.

The scientific literature shows different initiatives that aim to evaluate the use of DICOM and related software tools for research purposes, in particular for the management of diagnostic data oriented to quantitative imaging [7, 9, 10] or for operations such as de-identification [11]; in this work, we intend to carry out a comprehensive evaluation that includes all the main data curation operations both for software tools and released datasets.

The next paragraphs deepen the introduction of this context and show the current support of the DICOM format in the diagnostic workflow.

Big data workflow

Despite large strides in the introduction of PACS over the past few decades and the acceptance of the international DICOM standard for the storage and transfer of medical imaging data, there still remain significant barriers for the effective implementation of big data analytics on diagnostic imaging data.

Diagnostic images constitute a huge amount of data circulating in health care.

Unfortunately, medical data are often stored in qualitative reports, which do not allow recovering the images of interest as well as the related details. The clinical decision, in fact, is usually made on information deriving from several sources which may vary from the patient to the diagnostic systems but could not be stored in an appropriate and standardized way. Each of these single attributes could be considered as a fundamental element for the consequent definition of algorithms or, e.g., supervised or unsupervised models.

Figure 1 shows a typical radiological workflow, in which the patient undergoes a diagnostic imaging examination. The results of the diagnostic procedure are elaborated by the instrumentation (e.g., computed tomography (CT) or magnetic resonance (MR) scan and stored in the DICOM format after the image formation procedure. Series, scans or reports are usually safely stored in the hospital's PACS and could be recalled by a radiologist or a clinician whenever needed. For instance, data retrieval is a fundamental step for reporting the result of the examination, identifying any potential pathologic condition and giving the patient or other practitioners the relevant indications. Many critical and significant elements may be lost during the conventional workflow because of the lack of standardization, the use of different formats or errors in the recording stage. We can recognize three fundamental actions, as shown in Fig. 1:

  1. 1.

    De-identification/anonymization: A DICOM file contains both the image and a large variety of data in the header. All of these elements can includeidentifiable information about the patient, the study and the institution. Sharing such sensitive data demands proper protection in order to ensure data safety and maintain patient privacy [12].

  2. 2.

    Annotation/segmentation: This phase includes all the operations of delineation, demarcation, localization and measurement of regions of interest (organs, lesions, suspicious or notable areas) within the diagnostic images [1]. As a final result, some quantitative data are reported with the images and could be used to build datasets for the development and validation of algorithms, models and the so-called semantic segmentation [13].

  3. 3.

    Clinical reporting: This operation includes the collection of data related to the patient's pathological state, before and after the diagnostic session. A report usually includes critical data such as the clinical outcomes, clinical characterization of the subjects and information about therapeutic treatments. Similar information has a huge potential in terms of data analysis but is often inaccessible due to the qualitative and descriptive nature of these documents.

Fig. 1
figure 1

Conventional radiological workflow and collection of information for big data analytics using specific DICOM tags. Blue arrows refer to the radiological workflow; green arrows refer to data collection for big data analytics

All the information processed and recorded during these steps should be prepared as efficiently as possible: The resulting datasets should be analyzed readily and should not require additional curation steps. In fact, any additional procedure would be infeasible, especially considering large amounts of data.

The FAIR Guiding Principles [14, 15] state that good data management is “the key conduit leading to knowledge discovery and innovation, and to subsequent data and knowledge integration and reuse by the community after the data publication process.” FAIR, in fact, stands for findability, accessibility, interoperability and reusability, which are perceived as the four key factors affecting data quality.

It is important to note that, in the age of machine learning, “reusability” refers to reuse not only by humans but also by machines. Consequently, it is important to consider how to make data readable by machines in order to make best use of modern technologies.

Privacy and de-identification

Privacy is one of the most discussed topics related to data collection, analysis and interpretation since, nowadays, these activities and their results are considered as a new business. Medical data may have several legitimate secondary uses, such as research projects or teaching, but it is strictly necessary to receive the informed consent from the patients. Another example is the development of decision support systems, which could be easily commercialized as soon as the results of the digital toolkit are considered as adequate. Additionally, most personal data should be removed securely, even if in many cases some data or a link to personal data must be kept. The increasing demand for predictive models, clinical decision support systems and data analysis has led to a situation where the research needs are often in conflict with privacy rules [16]. At this moment, there is no commonly accepted solution since the amount and type of personal data requested by researchers varies on a case-by-case basis and the choice of information to be kept, modified or deleted depends on the purposes and regulations to comply with.

To overcome these problems, governments and institutions have developed rules and regulations that bring together privacy, data and research purposes. To this aim, some of the most requested and used regulations that must be considered to use data for research purposes are:

  • HIPAA, 1996 Health Insurance Portability and Liability Act (HIPAA),

  • GDPR, the general data protection regulation (EU) 2016/679.

Specifically, the GDPR gives EU citizens or residents the right to request the deletion, modification or access of their data, while the HIPAA does not confer this right. The GDPR compliance affects all personal data, while the HIPAA is limited to the “Protected Health Information” (PHI), defined as information that can be used to directly or indirectly identify an individual in relation to his past, present or future health condition. In addition, the HIPAA contains the "Safe Harbor" method, which lists 18 identifiers of types to be removed or modified, whereas there are no explicit lists of the elements to be eliminated in the GDPR.

Pseudonymization and anonymization are different ways to perform de-identification. Specifically, following the ISO 25237:2017 standard, anonymization is defined as a “process by which personal data are irreversibly altered in such a way that a data subject can no longer be identified directly or indirectly, either by the data controller alone or in collaboration with any other party,” whereas pseudonymization is a “particular type of de-identification that both removes the association with a data subject and adds an association between a particular set of characteristics relating to the data subject and one or more pseudonyms.”

From these definitions, anonymization can be used in accordance with HIPAA and both anonymization and pseudonymization are considered as “data processing” under the GDPR. In the latter case, the right of access, modification and cancellation of the citizen's personal data must be guaranteed. However, as the anonymous data have no direct or indirect links to identify the original patient, any additional processing or processing performed on that dataset falls outside the scope of the GDPR.

The implementation of de-identification on clinical diagnostic examinations passes through the search, elimination and replacement of identifiers within the images and tags that describe them. Most DICOM objects contain demographic and associated medical images and information about the patient, which must be kept confidential or removed in case the tests are to be used for research purposes. As reported in the document “Security and Privacy in DICOM” [12], since 1999 the DICOM standard has included options to encrypt and protect data that move through the connections of network in response to the implementation of HIPAA and not in response to cybersecurity problems; furthermore, in 2001 DICOM extended the use of the CMS (Cryptographic Message Syntax) to encrypt DICOM data, allowing the encryption of the PHI of a DICOM object.

Profiles and options to address the removal and replacement of attributes within a DICOM Dataset are reported in the Annex E of DICOM PS3.15 2020d. Specifically, it contains a Basic Profile with several retain or clean options, and other implementations such as “In standard Compliance of IOD.” Each attribute can be either replaced, cleaned, removed or kept, depending on the confidentiality level and the importance of the identifier. The following tags should be added after the application of one of the de-identification profiles, to retain the history of metadata as in Table 1.

Table 1 The DICOM attributes to add after the application of the DICOM attribute confidentiality profiles

There are various software tools to perform de-identification on DICOM images. Aryanto et al. [11] offer a complete overview and critical analysis.

It is very important to mention that simply removing or modifying metadata in DICOM images may not be sufficient to prevent re-identification of the subject. In fact, topograms for CT data and ultrasound images may have patient information burned into the pixel data [17]. To manage this critical issue, some specific studies and tools are available to the scientific community [18,19,20,21]. In addition, it is demonstrated that humans or specific software could identify individual subjects by reconstructing facial images contained in cranial MR or CT [22,23,24]. Further efforts are needed to develop reliable de-identification methods for medical images containing identifiable anatomical details, such as facial features. Moreover, the creation of specific DICOM attributes that encode this operation is expected to fully standardize the de-identification procedures.

Annotation and segmentation

The key actions to extract fundamental features from medical images for both clinical decision making and research are annotation and segmentation. The annotation procedure focuses on labeling images with additional information useful for data detection, classification and grouping [25,26,27,28]. In particular, this routine allows to transform descriptive and qualitative image features in machine-readable data, thus making them suitable for automatic image analyses as supervised artificial intelligence (AI) methodologies [29]. Segmentation can be considered as a special case of annotation where one or more image areas are isolated in so-called regions of interest (ROI). Indeed, segmentation is based on drawing (manually, semi-automatically or fully automatically) a binary mask of pixels belonging to the ROI.

In the context of big data analytics, annotation and segmentation routines are fundamental processes for data description, sharing and analysis. Furthermore, the huge amount of data to be managed requires approaches able to codify and standardize the annotation data.

To address this issue, the National Electrical Manufacturers Association (NEMA) developed DICOM-RT, i.e., the first extension of the DICOM standard that could include information regarding annotations [30]. DICOM-RT was developed to specifically address the standardization of data deriving from radiotherapy (e.g., external beam, treatment planning, dose, radiotherapy images). In particular, five DICOM-RT objects were defined to manage: areas of significance (DICOM-RT Structure Set), transfer of treatment plans (DICOM-RT Plan), dose distribution of the radiation therapy plan (DICOM-RT Dose), radiotherapy images (DICOM-RT Image) and treatment session report (DICOM-RT Treatment Record). Segmentation information provided as ROIs was specifically included in DICOM-RT Structure Set object and, conditionally for data containing dose points or dose curves, in DICOM-RT Dose object.

More recently, DICOM-SEG [31] has been introduced. It is a dedicated modality in which the annotation routine is encoded as text (Table 2) and the positional information of the annotation is specified by a codified segmentation image (image data).

Table 2 Most pertinent/specific DICOM tags related to DICOM-SEG modality

It is worth mentioning that software to include segmentation/annotation information as DICOM-SEG modality was developed, thus allowing the management of such information in DICOM. In particular, dedicated software packages as DCMTK [32], ITK [33, 34], dcmqi [10] (built upon DCMTK and ITK) and pydicom-seg [35] are suitable for this purpose.

Structured report

With the recent technical advances, the need to achieve full interoperability with the increasing amount of multi-modal patient data is arising. Structured reporting (SR) is becoming essential for clinical decision-making and research applications, including big data and machine learning.

SR aims to standardize both the format and lexicon used in radiology reports [36]. A definition for SR is set by describing three increasing levels of SR according to Weiss and Bolos [37]: The first and basic level consists of a structured format with paragraph and subheadings; the second is marked by a consistent organization with items reported in a certain order; and the third and more complex is characterized by the consistent use of dedicated lexicon and ontology.

The main reasons prompting to move from traditional free text reporting to standardized and structured reporting are summarized below and encompass both clinical and research considerations [38,39,40]. First of all, the use of checklist-style SR and standardized SR templates ensures that all relevant items for a particular examination are addressed. This may reduce diagnostic error, improve report clarity and quality and ensure consistent use of terminology across practices. Secondly, the use of standardized lexicon and structure prevents ambiguity and facilitates comparability of disease states, treatments and any type of clinical results. Even if this constitutes guidance for referring physicians, it should be highlighted that a simple comparison of the results is crucial for clinical purposes.

Finally, the capability of SR to include quantitative imaging biomarkers (radiomics) and parameters (e.g., laboratory results) might make SR as “Big-data container” leading not only to an integrated and precise clinical decision (e.g., diagnosis, treatment option) but also to a substantial support for modern clinical research. The standardized structure and vocabulary typical of SR can be well suitable to be analyzed by computers, thus facilitating data sharing (e.g., registries and biobanks) or data mining in research [38,39,40,41].

In summary, a wide adoption of SR is critical not only for communicating results to physicians or patients, but also for making diagnostic imaging data suitable by AI algorithms. Clinical decision-making and research applications in AI and big data in medical imaging heavily depend on data and standardization. One of the main challenges for the development of AI solutions for health care and radiology remains the unstructured nature of the data stored in electronic health records. In particular, radiological report data are often available only as unstructured narrative text [41,42,43].

The implementation of SR is complex and still scarce in clinical routine for several reasons. One of the biggest challenges in SR implementation is resistance to switch from the traditional narrative reporting to SR. Another issue concerns the risk of errors in case of improper use in clinical routine. Moreover, including unnecessary or irrelevant information in a template report may negatively impact the coherence of the report and the subsequent comprehension by referring physicians. Finally, the SR checklist schema may interfere with the radiologist’s reasoning and ability with a negative impact on the search pattern and visual attention. The so-called eye-dwell phenomenon may happen in case radiologists are more focused on the report template rather than the images. This may not only increase reporting time, but generate errors or missed findings [38,39,40].

Several steps were made by healthcare providers in order to overcome the above-mentioned limitations and encourage the use of uniform language and structure in radiology reporting, which are the basis for successfully implementing SR in clinical practice. For instance, RSNA developed RadLex, a standardized ontology of radiological terms in constant updating and developed starting from SNOMED-CT. RadLex can be used together with popular medical lexicons such as SNOMED-CT, ICD-10, CPT and BrainInfo [44]. Moreover, RSNA started the so-called Reporting Initiative with the aim of developing and providing vendor-neutral reporting templates [45]. This led to the publication of the Management of Radiology Report Templates profile by IHE, which extensively describes the concepts and technical details for interoperable, standardized and structured report templates [46].

Since an essential requirement for the successful implementation of SR is to respect the current radiology workflow, the DICOM standard plays a key role [5]. Given its potentiality, NEMA introduced the DICOM-SR which defines the syntax and semantics of structured and standardized diagnostic reports. An exhaustive description of DICOM-SR can be found in Clunie’s work [47]. Briefly, like a DICOM image, the DICOM-SR has a header, which encodes the information of the patient and study identification, and a content, that instead is responsible for the coding of the report itself. The information elements in the report are hierarchically connected in a tree model, identifying the Sources and Targets Nodes and their relationships. Each element has a name and a value, forming the pairs Name-Value [48, 49]. DICOM-SR contains text with links to other data such as images, waveforms and spatial or temporal coordinates. Although DICOM-SR is not as widespread as DICOM for digital images, its use has many advantages [50, 51]. DICOM-SR documents can be stored and sent along with the images belonging to the same study in PACS. In addition, DICOM-SR supports unified lexicons such as RadLex, ICD-10 and SNOMED. Finally, DICOM-SR templates have been defined to constrain the possible structures and to provide the basic codes that can be used to encode specific reports [50]. Specifically, a DICOM-SR template is applied to the document content to harmonize its structure. Each template is assigned to an unique template identifier (TID) with a related name and is specified by a table where each line corresponds to a so-called node with defined content item or indicates another template to be included in the SR document [48]. A specific limitation of DICOM-SR is that even if it provides a data structure which embeds structured reports in a standard “container” that can be read across different software applications, it does not define how the content should be structured or standardized.

Table 3 and Fig. 2 show the definition of the DICOM-SR Template for a Measurement Report and its sub-templates (TID 1500 from DICOM PS3.16 and PS3.21).

Table 3 DICOM SR template for measurement report, template ID 1500 (from DICOM PS3.16)
Fig. 2
figure 2

DICOM SR measurement report template structure (template ID 1500) and its sub-templates (from DICOM PS3.21). TID: template ID, MR, magnetic resonance; CT, computed tomography; PET, positron emission tomography

Similar to DICOM-SEG, several efforts have been made in software development supporting the management of structured reports in DICOM standard [10, 32,33,34].

Materials and methods

To evaluate the degree of effective use of the DICOM standard in the previously detailed actions (i.e., de-identification/anonymization, segmentation/annotation and clinical reporting), two different experiments have been designed and will be described in the following paragraphs. The first explores the DICOM coding of information in the main databases shared by the scientific community, whereas the second investigates and evaluates the possibility of managing DICOM fields with a systematic selection of DICOM visualization software.

Imaging database evaluation

The databases included in our benchmark have been collected by a quasi-systematic internet survey, starting from four well-known lists of public repositories [52,53,54,55].

The study of databases aims to identify the type of information released and to evaluate its standardization. In particular, in this analysis we intend to verify whether the DICOM standard is used to collect, where possible, the information available.

Each database has been evaluated considering its modality of access and several key features, described hereafter. Furthermore, we performed a three-step review of the available repositories, in order to select which of them were suitable for this project. The datasets released up to March 1, 2021, were included. The detailed description of the database selection criteria is as follows:

Step 1:

  • Access: We looked for open or public access. In particular, repositories which required an application with a project, or private datasets were not considered.

  • Format: Only datasets providing imaging scans in DICOM format were selected.

  • Species: We considered only humans, whereas there were few databases with rodents, small animals or phantom images.

Step 2:

Among selected databases providing images in DICOM format, we excluded those not including any additional information than the acquired images. Specifically, we considered the repositories having at least one of the following additional resources, not necessarily provided in DICOM format:

  • annotations/segmentations;

  • radiologist reports;

  • any clinical information.

In this step, the databases that provided clinical information, in any format, were included. The latter criterion was adopted since, in this step, we aimed to verify how many databases had sufficient information to reconstruct dedicated DICOM-SR. Specifically, we considered this condition verified in the following cases:

  1. 1.

    traceability of the referenced images;

  2. 2.

    clinical information related to single imaging modality.

Step 3:

From the repositories identified in step 2, we selected databases that provided additional information according to the DICOM standard. In particular, we excluded all databases not including at least DICOM-SEG or DICOM-SR. In the case of TCIA repository, we used the TCIA portal (https://nbia.cancerimagingarchive.net/nbia-search) for selection and filtering of the collections. Moreover, the collections containing the TCIA third-party analysis results provided in DICOM-SEG or DICOM-SR modalities were also included. In particular, the aforementioned derived data corresponded to contributions generated by researchers who were not part of the group which originally submitted the related TCIA collection.

In order to test the actual presence of available DICOM images, and any associated additional information in DICOM format, each dataset was inspected in its entire content. Specifically, the first patient folder containing DICOM images was opened by using PostDicom software (https://www.postdicom.com/), which allows the reading of DICOM series. Exploiting PostDicom's ability of reading information encoded in both the DICOM-SEG and DICOM-SR, the same procedure was performed for the associated additional information in DICOM format (if present).

Furthermore, to complement the analysis, it was considered appropriate to investigate the sharing of imaging data during the spread of coronavirus disease 2019 (COVID-19). The COVID-19 pandemic required a joint effort for the urgent development of tools and data sharing to better face the emergency. In this situation, diagnostic imaging also plays a crucial role since the virus can generate a viral lung infection with a typical pattern in the chest computed tomography (CT), such as ground glass opacity, crazy-paving pattern and consolidation [56].

In particular, initiatives for the development of automatic diagnostic support tools based on AI techniques for diagnosis and clinical prediction have proliferated [57]. For the development of these tools, it is essential to make available large and curated datasets that include, in addition to imaging scans, information relating to the segmentation of regions of interest (i.e., lesions generated by viral pneumonia) and information related to the clinical status of the subject and the clinical outcome [58,59,60,61]. It is clear that, in this context, the use of standards plays a key role and, therefore, the analysis of CT COVID-19 datasets sharing deserves particular attention.

For this purpose, publicly accessible datasets including chest CT images with COVID-19 lesions, available at the first pandemic year (up to March 2021), have been identified in leading scientific research repositories. For each dataset, a sample of the data was downloaded to verify the format used for the release of CT scans and associated segmentations and clinical information.

Imaging software evaluation

In this section, we specifically consider the analysis of the software tools suitable for the diagnostic workflow. For this reason, the tools allowing at least the direct visualization of the DICOM images have been focused on, and their ability to manage the other data curation actions has been analyzed.

The software and tools included in our benchmark have been collected by a quasi-systematic Internet survey using the I Do Imaging (IDI) [62] initiative. Moreover, all software tools listed in the DICOM/PACS viewer webpage of Radiology Cafè [63] have been included to account for relevant software tools intended for clinical use.

The first includes all free software tools released and reviewed for research purposes, whereas the latter includes all software tested by a consultant radiologist and thought to represent the best currently available online, including core functionalities required by a radiologist for reviewing studies and/or teaching. In order to meet our evaluation criteria, the IDI database has been filtered for DICOM support, display of images and rank greater than four stars. Note that we have not included software or tools with no rating. Each software has been evaluated according to the following inclusion criteria:

  • DICOM format: We have considered only software tools that could at least read a DICOM series.

  • Display: This is a mandatory requirement, each software should, at least, display the image contained in a DICOM series.

  • License: We considered free or open-source software. Commercial software or tools were included only if a free trial version was available.

After selecting software tools that fulfilled the inclusion criteria, we considered the following features for the software evaluation procedure:

  • De-identification: Software that could modify a DICOM header, e.g., editing or removing tag values and save back all the de-identified personal information;

  • Segmentation and annotation: Software and tools that could read and/or write a DICOM file with segmentations or annotations, using the dedicated DICOM tags;

  • Structured report: Reading and writing DICOM structured reports, i.e., enclosed documents or data.

To evaluate if a specific software was able to read a DICOM-SR file, we checked if DICOM-SR content was successfully read by the software and displayed to the operator.

In order to test the functionality of the software selected in the management of the DICOM fields, specific “probe” datasets have been created. Two exemplary DICOM publicly available folders, specifically downloaded from TCIA Prostate-diagnosis [64] and CMET-MRhead [65], served for this scope. For the segmentation and annotation functionalities, the data were already embedded as DICOM-SEG modality in CMET-MRhead repository. Conversely, since NRRD format was used for these data types in Prostate-Diagnosis repository, dcmqi tool [10] was used to convert NRRD segmentations to DICOM-SEG. To generate DICOM-SR from the probe datasets, the template TID1500 “Measurement Report” was used (from DICOM PS3.16, http://dicom.nema.org/medical/dicom/current/output/html/part16.html#sect_TID_1500). In particular, for both probe datasets, basic measurements (e.g., mean, standard deviation, median, range, volume) were evaluated on the original DICOM images by means of the associated DICOM-SEG and encapsulated in DICOM-SR by using the dcmqi package [10].

All software and tools ran on a Windows V10 operating system, except for the Horos DICOM Viewer that was tested on Mac OS Catalina V10.15.6.

Results

Imaging database evaluation

Regarding the selection of databases to be analyzed, 210 databases were identified after the quasi-systematic web survey. 100 repositories were collected after the first selection criteria (step 1), among which 83 databases were selected by applying the criteria of step 2.

In step 3, we identified which of the 83 detected databases provided additional information according to the DICOM standard. In particular, we excluded all databases not including at least DICOM-SEG or DICOM-SR. After the exclusion of 49 items, 34 datasets were finally selected (Fig. 3).

Fig. 3
figure 3

Flow diagram describing the process of imaging database evaluation and selection. DICOM, Digital Imaging and Communications in Medicine; DICOM-SEG, DICOM segmentation object; DICOM-SR, DICOM structured report object

Table 4 shows the 83 databases identified after step 2, with lines colored in bold corresponding to the 34 datasets selected after applying step 3 criteria, including their characteristics (e.g., dataset name, collection, pathology) and additional information other than DICOM images (e.g., CI, clinical info; A/S, annotations/segmentations; CR, clinical report). Focusing on the seventh, eighth and ninth columns (CI, A/S, CR), it is evident that, respectively:

  • No clinical information, when available, is reported in DICOM format. Of the 67 databases including clinical information, only 29 provided clinical information that are sufficient to reconstruct DICOM-SR;

  • Twenty-seven out of 56 databases that contain the definition of regions of interest use the DICOM-SEG format.

  • Twenty-four datasets providing clinical reports, out of 83, use the DICOM-SR format. Furthermore, it was verified that all datasets comply with the correct DICOM format for data de-identification.

Table 4 Selected databases (n = 83) providing additional information other than DICOM images

Of note, the highest result variability was found for A/S data (eighth column in Table 4). Indeed, excluding databases that provided segmentation or annotation in DICOM-SEG, ten databases released segmentation data in DICOM-RT, ten databases in image format (eight in NIfTI, one in mhd and one in NRRD data type), nine stored in tabular format (XLS and CSV data type) and nine stored in structured data format (four in XML and five in JSON data type). Except for image format, the remaining data formats require dedicated routines to manage A/S information by associating file content to positional information.

Focusing on the COVID-19 databases, 12 datasets have been identified [66,67,68,69,70,71,72,73,74,75,76,77]. Of note, we found that only two of the reviewed COVID-19 datasets were released in DICOM format [76, 77].

Five datasets use a non-specific image format for the medical domain as portable network graphics [66,67,68], tagged image file format [69] and hierarchical data format [70]. The remaining datasets [71,72,73,74,75] use the NIfTI format for CT scans and, when available, segmentation images.

Imaging software evaluation

Although the following software met the inclusion criteria, CollectiveMinds (www.cmrad.com) was excluded since it resulted a Web platform for collaborative reporting, accessible only to licensed medical doctors and Papaya (http://mangoviewer.com/papaya.html) was excluded since it was based on the same API of Mango. Finally, ten software tools resulted from the quasi-systematic selection, as listed in Table 5.

Table 5 Evaluation of DICOM viewers included in the study

All software tools proved to be suitable in basic operations such as loading, selecting, viewing and manipulating the raw images from the probe datasets.

The third and fourth columns of Table 5 show the results, in dichotomous representation, of the software evaluation with the respect to annotation/segmentation and structured report, respectively.

For each selected software tool, the aforementioned procedures were tested for both reading and writing operations. The following results summarize the software evaluation procedure:

  • Four out of ten analyzed software tools were able to modify DICOM header by editing DICOM tag values.

  • None of the selected software tools allows the writing of regions of interest in the DICOM-SEG format; at the same time, two software tools (PostDicom and 3D Slicer) allow DICOM-SEG reading.

  • Four software tools (Radiant, PostDicom and Horos Viewer) allow the reading of information encoded in the DICOM-SR format, and none of the analyzed software tools allows DICOM-SR writing.

Of note, 3D Slicer software allows both writing DICOM-SEG and reading DICOM-SR images by the use of “QuantitativeReporting” extension (https://qiicr.gitbook.io/quantitativereporting-guide/).

In addition, only two software tools (Horos Viewer and MricROgl) allow the de-identification of DICOM folders in full compliance with the DICOM standard.

Discussion

In the era of big data, it is increasingly important to pay attention and care to data management in order to fully exploit the potential of modern analytical techniques. The definition and the proper use of standards certainly play a leading role in this perspective.

The DICOM standard has been described and evaluated for a series of key actions involved in the radiological workflow. The results of our research first show that the DICOM format fully supports them, allowing to encapsulate in a single format much of the information necessary for subsequent analytical phases.

Considering a typical radiomic workflow [98, 99], an artificial system can find appropriately de-identified data, information related to the patient's clinical status (DICOM-SR) and information on the localization of the region of interest within a single DICOM folder. For example, the details concerning the ROI (DICOM-SEG) can be useful to calculate the radiomic descriptors, thus favoring the aggregation of suitable data to develop reliable systems for classification or prediction of clinical outcomes.

It is interesting to note that although the DICOM format has considerable potential to foster big data analytics, it is only partially exploited in the sharing of imaging data for research purposes.

The results of the imaging database evaluation show that the DICOM-SR format is rarely used to contain clinical information. It should be pointed out that all the 24 datasets providing clinical reports released such information in DICOM-SR format according to guidelines drawn up for challenge tasks or specific initiatives [83, 91, 92, 96]. Moreover, regarding DICOM-SEG, one-third of the analyzed datasets use different and not fully standardized formats to share information on regions of interest. On this topic, the “DICOM4QI” initiative is worthy of note; it aims at evaluating interoperability of the image analysis tools and workstations, applied to exchange of the quantitative image analysis results using DICOM standard [7, 9, 10].

The imaging software evaluation shows that the support provided by software to fully exploit the potential of the DICOM format is still considerably limited, reducing the possibility for researchers and clinicians to create and make available suitable DICOM datasets. Therefore, there is a need to spur the development of initiatives that increase the attention on radiological software not only for the visualization reporting, but also for preparing datasets suitable for big data analytics. Interesting and helpful initiatives, such as OHIF viewer initiative [103] and NCI Imaging Data Commons (IDC), should be highlighted. Although not included in the present study, the first aimed to deliver an extensible platform to support site-specific workflows and accommodate evolving research requirements, according to DICOM specifications. Instead, IDC highlighted the role of the DICOM format as a cornerstone for sharing data and harmonizing analyses [104].

The need to promote the DICOM standard is additionally demonstrated by the analysis of the CT COVID-19 datasets. Indeed, only the two most recent COVID-19 collections released CT data in DICOM format. Although DICOM is the format used in the clinical acquisition routine, its limited adoption may indicate that the actual emergency conditions enhance the difficulty in finding adequate tools to manage the standard and, therefore, to promptly proceed to the de-identification, sharing and care of data. This difficulty may therefore lead researchers to use alternative formats easier to manage and disseminate, such as textual tables and non-medical image data.

It is important to note that this work includes databases mainly oriented to oncological studies, although this was not a prerequisite. Indeed, neurological studies, in which diagnostic images play a decisive role [105,106,107], have not been evaluated due to the definition of ad hoc standards and tools to manage and share neuroimaging data. In this field, the dedicated standard named Brain Imaging Data Structure (BIDS) [108] has been developed to organize and describe neuroimaging data with the use of different file formats than DICOM (e.g., NIfTI, JSON, text files). Similar to the DICOM, the BIDS standard allows to manage both metadata and derived quantitative measurements, opening to automated data analysis workflows.

It should be considered that, to enhance the feasibility of this work, narrow inclusion criteria that could not allow a comprehensive analysis were chosen and, therefore, it is not possible to exclude that some useful data have been neglected. For example, software tools and databases whose release requires the submission and approval of specific research projects were not considered.

Conclusions

In conclusion, this work provides an overview of the potential, not always exploited, of the DICOM format for capitalizing the radiological workflow from a big data perspective. The analysis of both the databases and the software shows that further efforts are needed by researchers, clinicians and companies to promote and facilitate the use of standards to increase the value of imaging data, according to FAIR principles.

Availability of data and materials

This work is based on data and information already available with open access.

Abbreviations

A/S:

Annotations/segmentations

AI:

Artificial intelligence

CI:

Clinical info

COVID-19:

Coronavirus disease 2019

CR:

Clinical report

CR:

Computed radiography

CSV:

Comma separated value

CT:

Computed tomography

DICOM:

Digital Imaging and Communications in Medicine

DICOM-RT:

DICOM radiation therapy

JSON:

JavaScript Object Notation

mhd:

MetaImage

MR:

Magnetic resonance

NEMA:

National Electrical Manufacturers Association

NIfTI:

Neuroimaging Informatics Technology Initiative

NRRD:

Nearly raw raster data

PT (or PET):

Positron emission tomography

ROI:

Region of interest

RTDOSE:

DICOM-RT dose

RTPLAN:

DICOM-RT plan

SR:

Structured report

TCIA:

The Cancer Imaging Archive

TID:

Template identifier

XLS:

Excel spreadsheet

XML:

Extensible markup language

References

  1. Aiello M, Cavaliere C, D’Albore A, Salvatore M (2019) The challenges of diagnostic imaging in the era of big data. J Clin Med 8:316

    Article  Google Scholar 

  2. Cirillo D, Valencia A (2019) Big data analytics for personalized medicine. Curr Opin Biotechnol 58:161–167. https://doi.org/10.1016/j.copbio.2019.03.004

    Article  CAS  PubMed  Google Scholar 

  3. Dinov ID (2016) Volume and value of big healthcare data. J Med Stat Inform. https://doi.org/10.7243/2053-7662-4-3

    Article  PubMed  PubMed Central  Google Scholar 

  4. Peterson CB, Hamilton C, Hasvold P (2016) From innovation to implementation: eHealth in the WHO European region. WHO Regional Office for Europe, Copenhagen

    Google Scholar 

  5. Mildenberger P, Eichelberg M, Martin E (2002) Introduction to the DICOM standard. Eur Radiol 12:920–927. https://doi.org/10.1007/s003300101100

    Article  PubMed  Google Scholar 

  6. Larobina M, Murino L (2014) Medical image file formats. J Digit Imaging 27:200–206. https://doi.org/10.1007/s10278-013-9657-9

    Article  PubMed  Google Scholar 

  7. Fedorov A, Clunie D, Ulrich E et al (2016) DICOM for quantitative imaging biomarker development: a standards based approach to sharing clinical data and structured PET/CT analysis results in head and neck cancer research. PeerJ 4:e2057. https://doi.org/10.7717/peerj.2057

    Article  PubMed  PubMed Central  Google Scholar 

  8. Fedorov A, Schwier M, Clunie D et al (2018) An annotated test-retest collection of prostate multiparametric MRI. Sci Data 5:1–13. https://doi.org/10.1038/sdata.2018.281

    Article  Google Scholar 

  9. Fedorov A, Rubin D, Kalpathy-Cramer J et al (2015) Interoperable communication of quantitative image analysis results using DICOM standard. figshare

  10. Herz C, Fillion-Robin J-C, Onken M et al (2017) DCMQI: an open source library for standardized communication of quantitative image analysis results using DICOM. Cancer Res 77:e87–e90. https://doi.org/10.1158/0008-5472.CAN-17-0336

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  11. Aryanto KYE, Oudkerk M, van Ooijen PMA (2015) Free DICOM de-identification tools in clinical research: functioning and safety of patient privacy. Eur Radiol 25:3685–3695. https://doi.org/10.1007/s00330-015-3794-0

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  12. Medema J Security and Privacy in DICOM. 2

  13. Simpson AL, Antonelli M, Bakas S et al (2019) A large annotated medical image dataset for the development and evaluation of segmentation algorithms. arXiv:190209063 [cs, eess]

  14. Hodson S, Jones S, Collins S et al (2018) Turning FAIR data into reality: interim report from the European Commission Expert Group on FAIR data. https://doi.org/10.5281/zenodo.1285272

  15. Vesteghem C, Brøndum RF, Sønderkær M et al (2019) Implementing the FAIR Data Principles in precision oncology: review of supporting initiatives. Brief Bioinform. https://doi.org/10.1093/bib/bbz044

    Article  PubMed Central  Google Scholar 

  16. Caspers J (2021) Translation of predictive modeling and AI into clinics: a question of trust. Eur Radiol 31:4947–4948. https://doi.org/10.1007/s00330-021-07977-9

    Article  PubMed  PubMed Central  Google Scholar 

  17. Moore SM, Maffitt DR, Smith KE et al (2015) De-identification of medical images with retention of scientific research value. Radiographics 35:727–735. https://doi.org/10.1148/rg.2015140244

    Article  PubMed  Google Scholar 

  18. Vcelak P, Kryl M, Kratochvil M, Kleckova J (2019) Identification and classification of DICOM files with burned-in text content. Int J Med Inform 126:128–137. https://doi.org/10.1016/j.ijmedinf.2019.02.011

    Article  PubMed  Google Scholar 

  19. Monteiro E, Costa C, Oliveira JL (2017) A de-identification pipeline for ultrasound medical images in DICOM format. J Med Syst 41:89. https://doi.org/10.1007/s10916-017-0736-1

    Article  PubMed  Google Scholar 

  20. DicomCleanerTM. http://www.dclunie.com/pixelmed/software/webstart/DicomCleanerUsage.html. Accessed 6 Aug 2021

  21. PyDicom. DICOM in Python

  22. Parks CL, Monson KL (2017) Automated facial recognition of computed tomography-derived facial images: patient privacy implications. J Digit Imaging 30:204–214. https://doi.org/10.1007/s10278-016-9932-7

    Article  PubMed  Google Scholar 

  23. Schwarz CG, Kremers WK, Therneau TM et al (2019) Identification of anonymous MRI research participants with face-recognition software. N Engl J Med 381:1684–1686. https://doi.org/10.1056/NEJMc1908881

    Article  PubMed  PubMed Central  Google Scholar 

  24. Chen JJ-S, Juluru K, Morgan T et al (2014) Implications of surface-rendered facial CT images in patient privacy. AJR Am J Roentgenol 202:1267–1271. https://doi.org/10.2214/AJR.13.10608

    Article  PubMed  Google Scholar 

  25. Temal L, Dojat M, Kassel G, Gibaud B (2008) Towards an ontology for sharing medical images and regions of interest in neuroimaging. J Biomed Inform 41:766–778. https://doi.org/10.1016/j.jbi.2008.03.002

    Article  PubMed  Google Scholar 

  26. Kohli MD, Summers RM, Geis JR (2017) Medical image data and datasets in the era of machine learning—whitepaper from the 2016 C-MIMI meeting dataset session. J Digit Imaging 30:392–399. https://doi.org/10.1007/s10278-017-9976-3

    Article  PubMed  PubMed Central  Google Scholar 

  27. Channin DS, Mongkolwat P, Kleper V, Rubin DL (2009) The annotation and image mark-up project. Radiology 253:590–592. https://doi.org/10.1148/radiol.2533090135

    Article  PubMed  Google Scholar 

  28. Tommasi T, Orabona F, Caputo B (2008) Discriminative cue integration for medical image annotation. Pattern Recogn Lett 29:1996–2002. https://doi.org/10.1016/j.patrec.2008.03.009

    Article  Google Scholar 

  29. Philbrick KA, Weston AD, Akkus Z et al (2019) RIL-contour: a medical imaging dataset annotation tool for and with deep learning. J Digit Imaging. https://doi.org/10.1007/s10278-019-00232-0

    Article  PubMed  PubMed Central  Google Scholar 

  30. Law MYY, Liu B (2009) DICOM-RT and its utilization in radiation therapy. Radiographics 29:655–667. https://doi.org/10.1148/rg.293075172

    Article  PubMed  Google Scholar 

  31. A.51 Segmentation IOD. http://dicom.nema.org/medical/dicom/current/output/chtml/part03/sect_A.51.html

  32. DCMTK - DICOM Toolkit. https://dicom.offis.de/dcmtk.php.en

  33. Insight Toolkit: ITK. https://itk.org/

  34. McCormick MM, Liu X, Ibanez L, Jomier J, Marion C (2014) ITK: enabling reproducible research and open science. Front Neuroinform. https://doi.org/10.3389/fninf.2014.00013

    Article  PubMed  PubMed Central  Google Scholar 

  35. pydicom-seg. https://github.com/razorx89/pydicom-seg

  36. Marcovici PA, Taylor GA (2014) JOURNAL CLUB: structured radiology reports are more complete and more effective than unstructured reports. AJR Am J Roentgenol 203:1265–1271. https://doi.org/10.2214/AJR.14.12636

    Article  PubMed  Google Scholar 

  37. Weiss DL, Langlotz CP (2008) Structured reporting: patient care enhancement or productivity nightmare? Radiology 249:739–747. https://doi.org/10.1148/radiol.2493080988

    Article  PubMed  Google Scholar 

  38. Ganeshan D, Duong P-AT, Probyn L et al (2018) Structured reporting in radiology. Acad Radiol 25:66–73. https://doi.org/10.1016/j.acra.2017.08.005

    Article  PubMed  Google Scholar 

  39. Gul P, Gul P (2019) Education in radiology structured reporting in radiology. Are we ready to implement it? PJR 29(1):49–53

    Google Scholar 

  40. European Society of Radiology (ESR) (2018) ESR paper on structured reporting in radiology. Insights Imaging 9:1–7. https://doi.org/10.1007/s13244-017-0588-8

    Article  Google Scholar 

  41. Pinto dos Santos D, Baeßler B (2018) Big data, artificial intelligence, and structured reporting. Eur Radiol Exp 2:42. https://doi.org/10.1186/s41747-018-0071-4

    Article  PubMed  PubMed Central  Google Scholar 

  42. Marcheschi P (2017) Relevance of eHealth standards for big data interoperability in radiology and beyond. Radiol Med 122:437–443. https://doi.org/10.1007/s11547-016-0691-9

    Article  PubMed  Google Scholar 

  43. Pinto dos Santos D, Kotter E (2018) Structured radiology reporting on an institutional level—benefit or new administrative burden? Ann NY Acad Sci 1434:274–281. https://doi.org/10.1111/nyas.13741

    Article  PubMed  Google Scholar 

  44. Langlotz CP (2006) RadLex: a new method for indexing online educational materials. Radiographics 26:1595–1597. https://doi.org/10.1148/rg.266065168

    Article  PubMed  Google Scholar 

  45. Morgan TA, Helibrun ME, Kahn CE (2014) Reporting initiative of the radiological society of North America: progress and new directions. Radiology 273:642–645. https://doi.org/10.1148/radiol.14141227

    Article  PubMed  Google Scholar 

  46. Kahn CE, Genereaux B, Langlotz CP (2015) Conversion of radiology reporting templates to the MRRT standard. J Digit Imaging 28:528–536. https://doi.org/10.1007/s10278-015-9787-3

    Article  PubMed  PubMed Central  Google Scholar 

  47. Clunie DA (2000) DICOM structured reporting. PixelMed Pub, Bangor

    Google Scholar 

  48. Hussein R, Engelmann U, Schroeter A, Meinzer H-P (2004) DICOM structured reporting: part 1. Overview and characteristics. Radiographics 24:891–896. https://doi.org/10.1148/rg.243035710

    Article  PubMed  Google Scholar 

  49. Hussein R, Engelmann U, Schroeter A, Meinzer H-P (2004) DICOM structured reporting: part 2. Problems and challenges in implementation for PACS workstations. Radiographics 24:897–909. https://doi.org/10.1148/rg.243035722

    Article  PubMed  Google Scholar 

  50. Torres JS, Damian SegrellesQuilis J, Espert IB, García VH (2012) Improving knowledge management through the support of image examination and data annotation using DICOM structured reporting. J Biomed Inform 45:1066–1074. https://doi.org/10.1016/j.jbi.2012.07.004

    Article  PubMed  Google Scholar 

  51. Noumeir R (2003) DICOM structured report document type definition. IEEE Trans Inf Technol Biomed 7:318–328. https://doi.org/10.1109/TITB.2003.821334

    Article  PubMed  Google Scholar 

  52. Open-Access Medical Image Repositories-aylward.org (2020). http://www.aylward.org/notes/open-access-medical-image-repositories. Accessed 30 Sept 2020

  53. Sfikas G (2020) sfikas/medical-imaging-datasets

  54. FAIRsharing. https://fairsharing.org/biodbcore/?q=dicom

  55. Medical data for machine learning. https://github.com/beamandrew/medical-data

  56. Pan F, Ye T, Sun P et al (2020) Time course of lung changes at chest CT during recovery from coronavirus disease 2019 (COVID-19). Radiology 295:715–721. https://doi.org/10.1148/radiol.2020200370

    Article  PubMed  Google Scholar 

  57. Computer‐aided diagnosis in the era of deep learning-Chan-2020-Medical Physics-Wiley Online Library (2020). https://aapm.onlinelibrary.wiley.com/doi/10.1002/mp.13764. Accessed 21 July 2020

  58. Gieraerts C, Dangis A, Janssen L et al (2020) Prognostic value and reproducibility of AI-assisted analysis of lung involvement in COVID-19 on low-dose submillisievert chest CT: sample size implications for clinical trials. Radiol Cardiothorac Imaging 2:e200441. https://doi.org/10.1148/ryct.2020200441

    Article  PubMed  PubMed Central  Google Scholar 

  59. Wang M, Xia C, Huang L et al (2020) Deep learning-based triage and analysis of lesion burden for COVID-19: a retrospective study with external validation. Lancet Digital Health 2:e506–e515. https://doi.org/10.1016/S2589-7500(20)30199-0

    Article  PubMed  Google Scholar 

  60. Lessmann N, Sánchez CI, Beenen L et al (2021) Automated assessment of COVID-19 reporting and data system and chest CT severity scores in patients suspected of having COVID-19 using artificial intelligence. Radiology 298:E18–E28. https://doi.org/10.1148/radiol.2020202439

    Article  PubMed  Google Scholar 

  61. Harmon SA, Sanford TH, Xu S et al (2020) Artificial intelligence for the detection of COVID-19 pneumonia on chest CT using multinational datasets. Nat Commun 11:4080. https://doi.org/10.1038/s41467-020-17971-2

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  62. Free medical imaging software: I do imaging (2020). https://idoimaging.com/. Accessed 30 Sept 2020

  63. Clarke C (2020) DICOM viewers. In: Radiology cafe. https://www.radiologycafe.com/radiology-trainees/dicom-viewers. Accessed 30 Sept 2020

  64. PROSTATE-DIAGNOSIS-The Cancer Imaging Archive (TCIA) Public Access-Cancer Imaging Archive Wiki. https://wiki.cancerimagingarchive.net/display/Public/PROSTATE-DIAGNOSIS

  65. Slicer Server by Kitware-CMET-MRhead (2020). http://slicer.kitware.com/midas3/folder/4433. Accessed 30 Sept 2020

  66. Peng Y, Tang Y, Lee S, Zhu Y, Summers RM, Lu Z (2020) COVID-19-CT-CXR: a freely accessible and weakly labeled chest X-ray and CT image collection on COVID-19 from biomedical literature. arXiv

  67. Zhao J, Zhang Y, He X, Xie P (2020) COVID-CT-dataset: a CT scan dataset about COVID-19. arXiv:200313865 [cs, eess, stat]

  68. Angelov P, Soares E (2020) Explainable-by-design approach for COVID-19 classification via CT-scan. Health Informatics

  69. Rahimzadeh M, Attar A, Sakhaei SM (2020) A fully automated deep learning-based network for detecting COVID-19 from a new and large lung CT scan dataset. Biomed Signal Process Control 68:102588

    Article  Google Scholar 

  70. Song J, Wang H, Liu Y et al (2020) End-to-end automatic differentiation of the coronavirus disease 2019 (COVID-19) from viral pneumonia based on chest CT. Eur J Nucl Med Mol Imaging 47:2516–2524. https://doi.org/10.1007/s00259-020-04929-1

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  71. Vayá M de la I, Saborit JM, Montell JA et al (2020) BIMCV COVID-19+: a large annotated dataset of RX and CT images from COVID-19 patients. arXiv:200601174 [cs, eess]

  72. Zhang K, Liu X, Shen J et al (2020) Clinically applicable AI system for accurate diagnosis, quantitative measurements, and prognosis of COVID-19 pneumonia using computed tomography. Cell 181:1423-1433.e11. https://doi.org/10.1016/j.cell.2020.04.045

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  73. An P, Xu S, Harmon SA et al (2020) CT images in COVID-19

  74. Jun M, Cheng G, Yixin W et al (2020) COVID-19 CT lung and infection segmentation dataset

  75. Morozov SP, Andreychenko AE, Pavlov NA et al (2020) MosMedData: chest CT scans with COVID-19 related findings dataset. arXiv:200506465 [cs, eess]

  76. Desai S, Baghal A, Wongsurawat T et al (2020) Chest imaging representing a COVID-19 positive rural U.S. population. Sci Data 7:414. https://doi.org/10.1038/s41597-020-00741-6

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  77. Tsai EB, Simpson S, Lungren MP et al (2021) The RSNA international COVID-19 open radiology database (RICORD). Radiology 299:E204–E213. https://doi.org/10.1148/radiol.2021203957

    Article  PubMed  Google Scholar 

  78. Kalpathy-Cramer J, Zhao B, Goldgof D et al (2016) A comparison of lung nodule segmentation algorithms: methods and results from a multi-institutional study. J Digit Imaging 29:476–487. https://doi.org/10.1007/s10278-016-9859-z

    Article  PubMed  PubMed Central  Google Scholar 

  79. Jayashree Kalpathy-Cramer SN (2015) Multi-site collection of lung CT data with nodule segmentations. The Cancer Imaging Archive

  80. Goldgof D, Hall L, Hawkins S et al (2017) Long and short survival in adenocarcinoma lung CTs DICOM. Available via https://wiki.cancerimagingarchive.net/pages/viewpage.action?pageId=24284406

  81. Paul R, Hawkins S, Yoganand B, Goldgof D (2016) Deep feature transfer learning in combination with traditional features predicts survival among patients with lung adenocarcinoma. Tomography 2:388–395. https://doi.org/10.18383/j.tom.2016.00211

    Article  PubMed  PubMed Central  Google Scholar 

  82. Hawkins SH, Korecki JN, Balagurunathan Y et al (2014) Predicting outcomes of nonsmall cell lung cancer using CT image features. IEEE Access 2:1418–1426. https://doi.org/10.1109/ACCESS.2014.2373335

    Article  Google Scholar 

  83. Urban T, Ziegler E, Pieper S et al (2019) Crowds cure cancer: crowdsourced data collected at the RSNA 2018 annual meeting. Available via https://wiki.cancerimagingarchive.net/pages/viewpage.action?pageId=52757630

  84. Kiser K, Ahmed S, Stieb SM et al (2021) Thoracic volume and pleural effusion segmentations in diseased lungs for benchmarking chest CT processing pipelines. Available via https://wiki.cancerimagingarchive.net/pages/viewpage.action?pageId=68551327

  85. Kiser KJ, Barman A, Stieb S, Fuller CD, Giancardo L (2020) Novel autosegmentation spatial similarity metrics capture the time required to correct segmentations better than traditional metrics in a thoracic cavity segmentation workflow. J Digit Imaging 34(3):541–553

    Article  Google Scholar 

  86. Aerts HJWL, Wee L, Rios Velazquez E et al (2019) Data from NSCLC-radiomics

  87. Gevaert O, Mitchell LA, Achrol AS et al (2014) Glioblastoma multiforme: exploratory radiogenomic analysis by using quantitative image features. Radiology 273:168–174. https://doi.org/10.1148/radiol.14131731

    Article  PubMed  Google Scholar 

  88. Bakas S, Akbari H, Sotiras A et al (2017) Segmentation labels for the preoperative scans of the TCGA-GBM collection. Available via https://wiki.cancerimagingarchive.net/pages/viewpage.action?pageId=24282666

  89. Bakas S, Akbari H, Sotiras A et al (2017) Advancing the cancer genome atlas glioma MRI collections with expert segmentation labels and radiomic features. Sci Data 4:170117. https://doi.org/10.1038/sdata.2017.117

    Article  PubMed  PubMed Central  Google Scholar 

  90. Beers A, Gerstner E, Rosen B et al (2018) DICOM-SEG conversions for TCGA-LGG and TCGA-GBM segmentation datasets. Available via https://wiki.cancerimagingarchive.net/pages/viewpage.action?pageId=41517733

  91. Clunie DA, Hickman H, Ver Hoef W et al (2019) DICOM SR of clinical data and measurement for breast cancer collections to TCIA. Available via https://wiki.cancerimagingarchive.net/pages/viewpage.action?pageId=50135479

  92. Fedorov A, Hancock M, Clunie D et al (2018) Standardized representation of the TCIA LIDC-IDRI annotations using DICOM. Available via https://peerj.com/preprints/27378.pdf

  93. Fedorov A, Hancock M, Clunie D et al (2019) Standardized representation of the LIDC annotations using DICOM. Available via https://peerj.com/preprints/27378/

  94. Armato SG, McLennan G, Bidaut L et al (2011) The Lung Image Database Consortium (LIDC) and Image Database Resource Initiative (IDRI): a completed reference database of lung nodules on CT scans: the LIDC/IDRI thoracic CT database of lung nodules. Med Phys 38:915–931. https://doi.org/10.1118/1.3528204

    Article  PubMed  PubMed Central  Google Scholar 

  95. Armato SG, McLennan G, Bidaut L et al (2015) Data From LIDC-IDRI

  96. Kalpathy-Cramer J, Beers A, Mamonov A et al (2019) Crowds cure cancer: crowdsourced data collected at the RSNA 2017 annual meeting. Available via https://wiki.cancerimagingarchive.net/pages/viewpage.action?pageId=33948774

  97. Burnside ES, Drukker K, Li H et al (2016) Using computer-extracted image phenotypes from tumors on breast magnetic resonance imaging to predict breast cancer pathologic stage: breast MRI phenotypes predict stage. Cancer 122:748–757. https://doi.org/10.1002/cncr.29791

    Article  PubMed  Google Scholar 

  98. Guo W, Li H, Zhu Y et al (2015) Prediction of clinical phenotypes in invasive breast carcinomas from the integration of radiomics and genomics data. J Med Imag 2:041007. https://doi.org/10.1117/1.JMI.2.4.041007

    Article  Google Scholar 

  99. Zhu Y, Li H, Guo W et al (2015) Deciphering genomic underpinnings of quantitative MRI-based radiomic phenotypes of invasive breast carcinoma. Sci Rep 5:17787. https://doi.org/10.1038/srep17787

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  100. Li H, Zhu Y, Burnside ES et al (2016) MR imaging radiomics signatures for predicting the risk of breast cancer recurrence as given by research versions of MammaPrint, Oncotype DX, and PAM50 gene assays. Radiology 281:382–391. https://doi.org/10.1148/radiol.2016152110

    Article  PubMed  Google Scholar 

  101. Li H, Zhu Y, Burnside ES et al (2016) Quantitative MRI radiomics in the prediction of molecular classifications of breast cancer subtypes in the TCGA/TCIA data set. NPJ Breast Cancer. https://doi.org/10.1038/npjbcancer.2016.12

    Article  PubMed  PubMed Central  Google Scholar 

  102. Morris E, Burnside E, Whitman G et al (2014) Using computer-extracted image phenotypes from tumors on breast MRI to predict stage

  103. Ziegler E, Urban T, Brown D et al (2020) Open health imaging foundation viewer: an extensible open-source framework for building web-based imaging applications to support cancer research. JCO Clin Cancer Inform. https://doi.org/10.1200/CCI.19.00131

    Article  PubMed  PubMed Central  Google Scholar 

  104. Fedorov A, Beichel R, Kalpathy-Cramer J et al (2020) Quantitative imaging informatics for cancer research. JCO Clin Cancer Inform 4:444–453. https://doi.org/10.1200/CCI.19.00165

    Article  PubMed  Google Scholar 

  105. Aiello M, Cavaliere C, Salvatore M (2016) Hybrid PET/MR imaging and brain connectivity. Front Neurosci 10:64

    Article  Google Scholar 

  106. Marchitelli R, Aiello M, Cachia A et al (2018) Simultaneous resting-state FDG-PET/fMRI in Alzheimer disease: relationship between glucose metabolism and intrinsic activity. Neuroimage 176:246–258. https://doi.org/10.1016/j.neuroimage.2018.04.048

    Article  CAS  PubMed  Google Scholar 

  107. Aiello M, Cavaliere C, Fiorenza D, Duggento A, Passamonti L, Toschi N (2018) Neuroinflammation in neurodegenerative diseases: current multi-modal imaging studies and future opportunities for hybrid PET/MRI. Neuroscience. https://doi.org/10.1016/j.neuroscience.2018.07.033

    Article  PubMed  Google Scholar 

  108. Gorgolewski KJ, Alfaro-Almagro F, Auer T et al (2017) BIDS apps: improving ease of use, accessibility, and reproducibility of neuroimaging data analysis methods. PLoS Comput Biol 13:e1005209. https://doi.org/10.1371/journal.pcbi.1005209

    Article  CAS  PubMed  PubMed Central  Google Scholar 

Download references

Funding

This work was partially funded by the Italian Ministry for Education, University and Research—Project “MOLIM ONCOBRAIN LAB—Metodi innovativi di imaging molecolare per lo studio di malattie oncologiche e neurodegenerative” ARS0100144 Prot. U.001378531-08-2018 and by the Italian Ministry of Health: “Ricerca Corrente” project and by POR CAMPANIA FESR 2014–2020, AP1-OS1.3 (DGR no. 140/2020) for project “Protocolli TC del torace a bassissima dose e tecniche di intelligenza artificiale per la diagnosi precoce e quantificazione della malattia da COVID-19” CUP D54I20001410002.

Author information

Authors and Affiliations

Authors

Contributions

Guarantor of integrity of the entire study was MA. Study ideation and design were contributed by MA and MS. Databases and software evaluation were contributed by VB, GE and PB. Manuscript editing was contributed by MA, GP and MS. All authors contributed to the writing and read and approved the final manuscript.

Corresponding author

Correspondence to Marco Aiello.

Ethics declarations

Ethics approval and consent to participate

Not applicable.

Consent for publication

Not applicable.

Competing interests

The authors declare that they have no competing interests.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Aiello, M., Esposito, G., Pagliari, G. et al. How does DICOM support big data management? Investigating its use in medical imaging community. Insights Imaging 12, 164 (2021). https://doi.org/10.1186/s13244-021-01081-8

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s13244-021-01081-8

Keywords