Cross-modality (CT-MRI) prior augmented deep learning for robust lung tumor segmentation from small MR datasets Journal Article


Authors: Jiang, J.; Hu, Y. C.; Tyagi, N.; Zhang, P.; Rimner, A.; Deasy, J. O.; Veeraraghavan, H.
Article Title: Cross-modality (CT-MRI) prior augmented deep learning for robust lung tumor segmentation from small MR datasets
Abstract: Purpose: Accurate tumor segmentation is a requirement for magnetic resonance (MR)-based radiotherapy. Lack of large expert annotated MR datasets makes training deep learning models difficult. Therefore, a cross-modality (MR-CT) deep learning segmentation approach that augments training data using pseudo MR images produced by transforming expert-segmented CT images was developed. Methods: Eighty-one T2-weighted MRI scans from 28 patients with non-small cell lung cancers (nine with pretreatment and weekly MRI and the remainder with pre-treatment MRI scans) were analyzed. Cross-modality model encoding the transformation of CT to pseudo MR images resembling T2w MRI was learned as a generative adversarial deep learning network. This model was used to translate 377 expert segmented non-small cell lung cancer CT scans from the Cancer Imaging Archive into pseudo MRI that served as additional training set. This method was benchmarked against shallow learning using random forest, standard data augmentation, and three state-of-the art adversarial learning-based cross-modality data (pseudo MR) augmentation methods. Segmentation accuracy was computed using Dice similarity coefficient (DSC), Hausdorff distance metrics, and volume ratio. Results: The proposed approach produced the lowest statistical variability in the intensity distribution between pseudo and T2w MR images measured as Kullback–Leibler divergence of 0.069. This method produced the highest segmentation accuracy with a DSC of (0.75 ± 0.12) and the lowest Hausdorff distance of (9.36 mm ± 6.00 mm) on the test dataset using a U-Net structure. This approach produced highly similar estimations of tumor growth as an expert (P = 0.37). Conclusions: A novel deep learning MR segmentation was developed that overcomes the limitation of learning robust models from small datasets by leveraging learned cross-modality information using a model that explicitly incorporates knowledge of tumors in modality translation to augment segmentation training. The results show the feasibility of the approach and the corresponding improvement over the state-of-the-art methods. © 2019 American Association of Physicists in Medicine
Keywords: clinical article; cancer radiotherapy; nuclear magnetic resonance imaging; magnetic resonance imaging; prospective study; benchmarking; external beam radiotherapy; tumor growth; non small cell lung cancer; image segmentation; tumor segmentation; measurement accuracy; human; article; random forest; x-ray computed tomography; deep learning; data augmentation; cross-modality learning; generative adversarial networks; cross modality learning
Journal Title: Medical Physics
Volume: 46
Issue: 10
ISSN: 0094-2405
Publisher: American Association of Physicists in Medicine  
Date Published: 2019-10-01
Start Page: 4392
End Page: 4404
Language: English
DOI: 10.1002/mp.13695
PUBMED: 31274206
PROVIDER: scopus
PMCID: PMC6800584
DOI/URL:
Notes: Source: Scopus
Altmetric Score
MSK Authors
  1. Andreas Rimner
    291 Rimner
  2. Pengpeng Zhang
    93 Zhang
  3. Joseph Owen Deasy
    320 Deasy
  4. Yu-Chi Hu
    68 Hu
  5. Neelam Tyagi
    67 Tyagi
  6. Jue Jiang
    16 Jiang