You are viewing an old version of this page. View the current version.

Compare with Current View Page History

« Previous Version 5 Next »

Finding labeled datasets on TCIA

Many collections on TCIA contain "labels" which can be used for training and testing artificial intelligence models.  However, users who are not familiar with medical imaging may benefit from assistance learning how to identify and utilize data types suitable for common deep learning tasks such as image classification, object detection, and object segmentations.  This page seeks to summarize key information about finding image labels that may not be obvious to researchers who do not have a background in radiology or histopathology.

Object Segmentation

Many collections on TCIA contain some kind of "ground truth" labels which provide information about object location and boundaries in the images.  In radiology images these are typically created by one or more radiologists hand-drawing boundaries around objects such as the patient's tumor(s) or organs on each image.  These kinds of data can be shared using a few different file formats. 

  1. DICOM provides support for these kinds of data using SEG and RTSTRUCT modalities.  
  2. Many popular open source tools export these labels in other formats.  Popular formats include NIFTI, NRRD, and MHA.

Image classification

TCIA includes a wealth of non-image data which could be utilized for image classification purposes.  

  1. Clinical data (outcomes, stage)
  2. Distinguishing between cancer types (lgg vs gbm)
  3. Genomic/Proteomic subtypes

Suggested Deep Learning Parameters for TCIA Result Submission

Deep Learning parameters are critical for researchers to reproduce Deep Learning experiments. However, it is usually the author's discretion for the completeness and format of reported parameters in manuscripts. This document proposes a list of essential Deep Learning parameters to be included in TCIA results submission process. The goal is to explicitly capture these parameters in a common format such that TCIA users can easily reproduce and compare analysis results from TCIA data.

List of Deep Learning Parameters

  1. Deep Neural Network (DNN) Name - for example, VGG16, ResNet-101, UNet, etc., or a link to GitHub repository or manuscript for customized DNNs if applicable.

  2. Data Augmentation Methods - for example, color augmentation (HSV or RGB color space), transformation, noise, GAN, patch generation, downsizing parameters, etc.

  3. Training, Validation, and Testing Set Configuration - for example number of samples per each set, total number of samples, etc.

  4. Hyperparameters - for example, learning rate, early stopping, batch size, number of epochs, etc.

  5. Training Statistics - for example, wall time spent in training, accuracy metrics such as if average score or best score is reported, etc.

  6. Training Environment - for example, GPU type, Deep Learning framework used such as TensorFlow/PyTorch, number of GPUs, number of nodes, etc

Other useful stuff

  1. https://www.youtube.com/watch?v=-XUKq3B4sdw - how a radiologist interprets lung CTs?
  2. https://www.kaggle.com/gzuidhof/full-preprocessing-tutorial - how to pre-process images for deep learning
  3. https://developer.nvidia.com/clara-medical-imaging
  4. https://forums.fast.ai/t/fastai-v2-has-a-medical-imaging-submodule/56117
  • No labels