Abstract
In this paper, a novel, multi-task fully convolutional network (FCN) architecture is proposed for automatic segmentation of brain tumour. The proposed network builds on the hierarchical relationship between tumour substructures with branch and leaf losses imposed and optimised simultaneously. The network takes multimodal MR images along with their symmetric-difference images as input and extracts multi-level contextual information, firstly by the branch losses which are then fed to the leaf loss in a combination stage. The model was evaluated on BRATS13 and BRATS15 datasets and results show that the proposed multi-task FCN outperforms single-task FCN on all sub-tasks. The method is among the most accurate available and its computational cost is relatively low at test time.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Menze, B.H., et al.: The multimodal brain tumour image segmentation benchmark (BRATS). Med. Imaging 34(10), 1993–2024 (2015)
Tustison, N.J., et al.: Optimal symmetric multimodal templates and concatenated random forests for supervised brain tumour segmentation (simplified) with ANTsR. Neuroinformatics 13(2), 209–225 (2015)
Pereira, S., et al.: Brain tumour segmentation using convolutional neural networks in MRI images. Med. Imaging 35(5), 1240–1251 (2016)
Havaei, M., et al.: Brain tumour segmentation with deep neural networks. Med. Image Anal. 35, 18–31 (2017)
Kamnitsas, K., et al.: Efficient multi-scale 3D CNN with fully connected CRF for accurate brain lesion segmentation. Med. Image Anal. 36, 61–78 (2017)
Simonyan, K., et al.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)
Shen, H., et al.: Efficient symmetry-driven fully convolutional network for multimodal brain tumour segmentation (2017). Submitted to ICIP
Kwon, D., Shinohara, R.T., Akbari, H., Davatzikos, C.: Combining generative models for multifocal glioma segmentation and registration. In: Golland, P., Hata, N., Barillot, C., Hornegger, J., Howe, R. (eds.) MICCAI 2014. LNCS, vol. 8673, pp. 763–770. Springer, Cham (2014). doi:10.1007/978-3-319-10404-1_95
Long, J., et al.: Fully convolutional networks for semantic segmentation. In: CVPR 2015 (2015)
Chen, L.-C., et al.: Semantic image segmentation with deep convolutional nets and fully connected CRFs. arXiv preprint arXiv:1412.7062 (2014)
Ronneberger, O., Fischer, P., Brox, T.: U-Net: convolutional networks for biomedical image segmentation. In: Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F. (eds.) MICCAI 2015. LNCS, vol. 9351, pp. 234–241. Springer, Cham (2015). doi:10.1007/978-3-319-24574-4_28
Chen, H., et al.: Deep contextual networks for neuronal structure segmentation. In: AAAI 2016 (2016)
Chen, H., et al.: DCAN: deep contour-aware networks for accurate gland segmentation. In: CVPR 2016 (2016)
He, K., et al.: Delving deep into rectifiers: surpassing human-level performance on imagenet classification. In: ICCV 2015 (2015)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2017 Springer International Publishing AG
About this paper
Cite this paper
Shen, H., Wang, R., Zhang, J., McKenna, S. (2017). Multi-task Fully Convolutional Network for Brain Tumour Segmentation. In: Valdés Hernández, M., González-Castro, V. (eds) Medical Image Understanding and Analysis. MIUA 2017. Communications in Computer and Information Science, vol 723. Springer, Cham. https://doi.org/10.1007/978-3-319-60964-5_21
Download citation
DOI: https://doi.org/10.1007/978-3-319-60964-5_21
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-60963-8
Online ISBN: 978-3-319-60964-5
eBook Packages: Computer ScienceComputer Science (R0)