You are viewing an old version of this page. View the current version.

Compare with Current View Page History

« Previous Version 10 Next »

Finding labeled datasets on TCIA

Many collections on TCIA contain "labels" which can be used for training and testing artificial intelligence models.  However, users who are not familiar with medical imaging may benefit from assistance learning how to identify and utilize data types suitable for common deep learning tasks such as image classification, object detection, and object segmentations.  This page seeks to summarize key information about finding image labels that may not be obvious to researchers who do not have a background in radiology or histopathology.  It also provides a host of useful links from third party sources which could be useful to data scientists who are new to working with medical images.

Object Segmentation

Many collections on TCIA contain some kind of "ground truth" labels which provide information about object location and boundaries in the images.  In radiology images these are typically created by one or more radiologists hand-drawing boundaries around objects such as the patient's tumor(s) or organs on each image.  These kinds of data can be shared using a few different file formats. 

  1. DICOM provides support for these kinds of data using SEG and RTSTRUCT modalities.  
  2. Many popular open source tools export these labels in other formats.  Popular formats include NIFTI, NRRD, and MHA.

On TCIA you can find these data in a couple of ways.  

  1. For Collections datasets you can look for SEG / RTSTRUCT in the modality column to determine where DICOM segmentations or contours are available.  You can also filter for "Image Analyses" in the supporting data column.  If a collection says "Image Analyses" but does not include SEG or RTSTRUCT in the modality this is typically because the analysis was in some other format.  This could be segmentation data in NIFTI/NRRD/MHA formats, but it might also represent some other kind of analysis such as image classification.
  2. For Analysis Results of existing TCIA collections it is a bit more straightforward.  Simply use the filter above the table to search for "segmentations" which will find any instance of this in the Analysis Artifacts column.

Image classification

TCIA includes a wealth of non-image data which could be utilized for image classification purposes.  

  1. Clinical data (outcomes, stage)
  2. Distinguishing between cancer types (lgg vs gbm)
  3. Genomic/Proteomic subtypes

Suggested Deep Learning Parameters for TCIA Result Submission

Deep Learning parameters are critical for researchers to reproduce Deep Learning experiments. However, it is usually the author's discretion for the completeness and format of reported parameters in manuscripts. This document proposes a list of essential Deep Learning parameters to be included in TCIA results submission process. The goal is to explicitly capture these parameters in a common format such that TCIA users can easily reproduce and compare analysis results from TCIA data.

List of Deep Learning Parameters

  1. Deep Neural Network (DNN) Name - for example, VGG16, ResNet-101, UNet, etc., or a link to GitHub repository or manuscript for customized DNNs if applicable.

  2. Data Augmentation Methods - for example, color augmentation (HSV or RGB color space), transformation, noise, GAN, patch generation, downsizing parameters, etc.

  3. Training, Validation, and Testing Set Configuration - for example number of samples per each set, total number of samples, etc.

  4. Hyperparameters - for example, learning rate, early stopping, batch size, number of epochs, etc.

  5. Training Statistics - for example, wall time spent in training, accuracy metrics such as if average score or best score is reported, etc.

  6. Training Environment - for example, GPU type, Deep Learning framework used such as TensorFlow/PyTorch, number of GPUs, number of nodes, etc

Other useful stuff

  1. https://www.youtube.com/watch?v=-XUKq3B4sdw - how a radiologist interprets lung CTs?
  2. https://www.kaggle.com/gzuidhof/full-preprocessing-tutorial - how to pre-process images for deep learning
  3. https://theaisummer.com/medical-image-coordinates/ - DICOM deep learning for medical imaging novices
  4. https://developer.nvidia.com/clara-medical-imaging - NVIDIA package for simplifying deep learning tasks in medical imaging
  5. https://forums.fast.ai/t/fastai-v2-has-a-medical-imaging-submodule/56117 - FastAI package for simplifying deep learning in medical imaging
  6. "TCIA as a Centralized Data Resource for Development of AI" from RSNA 2019
  7. https://www.kaggle.com/marcovasquez/basic-eda-data-visualization - RSNA intracranial hemorrhaging guide  
  8. https://github.com/RSNA/AI-Deep-Learning-Lab - RSNA 2019 deep learning course
  9. https://github.com/RSNA/MagiciansCorner - Notebooks, datasets, other content for the Radiology:AI series known as Magicians Corner by Brad Erickson
  • No labels