Skip to main content

Unpaired Multi-contrast MR Image Synthesis Using Generative Adversarial Networks

  • Conference paper
  • First Online:
Simulation and Synthesis in Medical Imaging (SASHIMI 2019)

Part of the book series: Lecture Notes in Computer Science ((LNIP,volume 11827))

Included in the following conference series:

Abstract

Magnetic Resonance Imaging (MRI) has been established as an important diagnostic tool for research and clinical purposes. Multi-contrast scans can enhance the accuracy for many deep learning algorithms. However, these scans may not be available in some situations. Thus, it is valuable to synthetically generate non-existent contrasts from the available one. Existing methods based on Generative Adversarial Networks (GANs) lack the freedom to map one image to multiple contrasts using only a single generator and discriminator, hence, requiring training of multiple models for multi-contrast MR synthesis. We present a novel method for multi-contrast MR image synthesis with unpaired data using GANs. Our method leverages the strength of Star-GAN to translate a given image to n contrasts using a single generator and discriminator. We also introduce a new generation loss function, which enforces the generator to produce high-quality images which are perceptually closer to the real ones and exhibit high structural similarity as well. We experiment on IXI dataset to learn all possible mappings among T\(_1\)-weighted, T\(_2\)-weighted, Proton Density (PD) weighted and Magnetic Resonance Angiography (MRA) images. Qualitative and quantitative comparison against baseline method shows the superiority of our approach.

This work was supported by National Natural Science Foundation (NNSF) of China under Grant 61873166, 61673275 and 61473184.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    https://brain-development.org/ixi-dataset/.

  2. 2.

    https://github.com/ANTsX/ANTsPy.

References

  1. Katti, G., Ara, S.A., Shireen, A.: Magnetic resonance imaging (MRI) - a review. Int. J. Dent. Clin. 3, 65–70 (2011)

    Google Scholar 

  2. Joyce, T., Chartsias, A., Tsaftaris, S.A.: Robust Multi-modal MR image synthesis. In: Descoteaux, M., Maier-Hein, L., Franz, A., Jannin, P., Collins, D.L., Duchesne, S. (eds.) MICCAI 2017. LNCS, vol. 10435, pp. 347–355. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-66179-7_40

    Chapter  Google Scholar 

  3. Beers, A., et al.: High-resolution medical image synthesis using progressively grown generative adversarial networks (2018)

    Google Scholar 

  4. Liu, J., et al.: Applications of deep learning to MRI images: a survey. Big Data Min. Anal. 1(1), 1–18 (2018). https://doi.org/10.26599/BDMA.2018.9020001

    Article  Google Scholar 

  5. Han, C., et al.: GAN-based synthetic brain MR image generation. In: 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), pp. 734–738 (2018). https://doi.org/10.1109/ISBI.2018.8363678

  6. Choi, Y., Choi, M., Kim, M., Ha, J.W., Kim, S., Choo, J.: StarGAN: unified generative adversarial networks for multi-domain image-to-image translation. In: 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (2018)

    Google Scholar 

  7. Zhu, J.Y., Park, T., Isola, P., Efros, A.A.: Unpaired image-to-image translation using cycle-consistent adversarial networks. In: 2017 IEEE International Conference on Computer Vision (ICCV) (2017)

    Google Scholar 

  8. Mirza, M., Osindero, S.: Conditional generative adversarial nets. arXiv e-prints arXiv:1411.1784 (2014)

  9. Isola, P., Zhu, J.Y., Zhou, T., Efros, A.A.: Image-to-image translation with conditional adversarial networks. In: CVPR (2017)

    Google Scholar 

  10. Ronneberger, O., Fischer, P., Brox, T.: U-Net: convolutional networks for biomedical image segmentation. In: Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F. (eds.) MICCAI 2015. LNCS, vol. 9351, pp. 234–241. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-24574-4_28

    Chapter  Google Scholar 

  11. Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 13(4), 600–612 (2004). https://doi.org/10.1109/TIP.2003.819861

    Article  Google Scholar 

  12. Zhang, R., Isola, P., Efros, A.A., Shechtman, E., Wang, O.: The unreasonable effectiveness of deep features as a perceptual metric. In: 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (2018)

    Google Scholar 

  13. Wei, X., Gong, B., Liu, Z., Lu, W., Wang, L.: Improving the improved training of Wasserstein GANs: a consistency term and its dual effect. arXiv e-prints arXiv:1803.01541 (2018)

  14. Goodfellow, I.J., et al.: Generative adversarial nets. In: Proceedings of the 27th International Conference on Neural Information Processing Systems, NIPS 2014, vol. 2, pp. 2672–2680. MIT Press, Cambridge (2014). http://dl.acm.org/citation.cfm?id=2969033.2969125

  15. Xiang, L., Li, Y., Lin, W., Wang, Q., Shen, D.: Unpaired deep cross-modality synthesis with fast training. In: Stoyanov, D., et al. (eds.) DLMIA/ML-CDS -2018. LNCS, vol. 11045, pp. 155–164. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-00889-5_18

    Chapter  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Jing Wu .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2019 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Sohail, M., Riaz, M.N., Wu, J., Long, C., Li, S. (2019). Unpaired Multi-contrast MR Image Synthesis Using Generative Adversarial Networks. In: Burgos, N., Gooya, A., Svoboda, D. (eds) Simulation and Synthesis in Medical Imaging. SASHIMI 2019. Lecture Notes in Computer Science(), vol 11827. Springer, Cham. https://doi.org/10.1007/978-3-030-32778-1_3

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-32778-1_3

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-32777-4

  • Online ISBN: 978-3-030-32778-1

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics