Skip to main content

Overfitting of Neural Nets Under Class Imbalance: Analysis and Improvements for Segmentation

  • Conference paper
  • First Online:
Medical Image Computing and Computer Assisted Intervention – MICCAI 2019 (MICCAI 2019)

Part of the book series: Lecture Notes in Computer Science ((LNIP,volume 11766))

Abstract

Overfitting in deep learning has been the focus of a number of recent works, yet its exact impact on the behavior of neural networks is not well understood. This study analyzes overfitting by examining how the distribution of logits alters in relation to how much the model overfits. Specifically, we find that when training with few data samples, the distribution of logit activations when processing unseen test samples of an under-represented class tends to shift towards and even across the decision boundary, while the over-represented class seems unaffected. In image segmentation, foreground samples are often heavily under-represented. We observe that sensitivity of the model drops as a result of overfitting, while precision remains mostly stable. Based on our analysis, we derive asymmetric modifications of existing loss functions and regularizers including a large margin loss, focal loss, adversarial training and mixup, which specifically aim at reducing the shift observed when embedding unseen samples of the under-represented class. We study the case of binary segmentation of brain tumor core and show that our proposed simple modifications lead to significantly improved segmentation performance over the symmetric variants.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Change history

  • 10 October 2019

    For chapter 45:

    The original version of this chapter was revised. The value of the last column of the fifth row in Table 1 was corrected from “0.93” to “0.83.”

    For chapter 51:

    The original version of this chapter was revised. The given name and family name of an author were mixed up. The given name is Antonio, and the family name is García-Uceda Juárez.

Notes

  1. 1.

    The dot product between a filter and a signal is highest when these match perfectly.

References

  1. Bakas, S., et al.: Advancing the cancer genome atlas glioma MRI collections with expert segmentation labels and radiomic features. Sci. Data 4, 170117 (2017)

    Article  Google Scholar 

  2. Goodfellow, I.J., Shlens, J., Szegedy, C.: Explaining and harnessing adversarial examples. In: International Conference on Learning Representations (ICLR) (2015)

    Google Scholar 

  3. Kamnitsas, K., et al.: Efficient multi-scale 3D CNN with fully connected CRF for accurate brain lesion segmentation. Med. Image Anal. 36, 61–78 (2017)

    Article  Google Scholar 

  4. Landman, B.A., Xu, Z., Igelsias, J.E., Styner, M., Langerak, T.R., Klein, A.: 2015 MICCAI Multi-atlas Labeling Beyond The Cranial Vault - Workshop and Challenge (2015)

    Google Scholar 

  5. Lin, T.Y., Goyal, P., Girshick, R., He, K., Dollár, P.: Focal loss for dense object detection. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2980–2988 (2017)

    Google Scholar 

  6. Liu, W., Wen, Y., Yu, Z., Yang, M.: Large-margin softmax loss for convolutional neural networks. In: International Conference on Machine Leanring (ICML), pp. 507–516 (2016)

    Google Scholar 

  7. Valindria, V.V., et al.: Small organ segmentation in whole-body MRI using a two-stage FCN and weighting schemes. In: Shi, Y., Suk, H.-I., Liu, M. (eds.) MLMI 2018. LNCS, vol. 11046, pp. 346–354. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-00919-9_40

    Chapter  Google Scholar 

  8. Zhang, H., Cisse, M., Dauphin, Y.N., Lopez-Paz, D.: mixup: beyond empirical risk minimization. In: International Conference on Learning Representations (ICLR) (2018)

    Google Scholar 

Download references

Acknowledgements

ZL is grateful for a China Scholarship Council (CSC) Imperial Scholarship. This project has received funding from the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation programme (grant No 757173, project MIRA, ERC-2017-STG) and EPSRC (EP/R511547/1).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Zeju Li .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2019 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Li, Z., Kamnitsas, K., Glocker, B. (2019). Overfitting of Neural Nets Under Class Imbalance: Analysis and Improvements for Segmentation. In: Shen, D., et al. Medical Image Computing and Computer Assisted Intervention – MICCAI 2019. MICCAI 2019. Lecture Notes in Computer Science(), vol 11766. Springer, Cham. https://doi.org/10.1007/978-3-030-32248-9_45

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-32248-9_45

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-32247-2

  • Online ISBN: 978-3-030-32248-9

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics