A Multi-label Scene Categorization Model Based on Deep Convolutional Neural Network

  • Gaofeng Zhao
  • Wang LuoEmail author
  • Yang Cui
  • Qiang Fan
  • Qiwei Peng
  • Zhen Kong
  • Liang Zhu
  • Tai Zhang
Conference paper
Part of the Lecture Notes in Electrical Engineering book series (LNEE, volume 517)


Being one of the most fundamental embranchments of deep learning theory, scene categorization technology has been extensively researched because of its great value in engineering application, especially in the field of remote monitoring and intelligent fault detection. To bridge the gap between theoretical accuracy and practical performance of relevant classification models which is mainly caused by nonstandard labeling information, this paper builds a normative dataset composed of 10,000 high-quality manual labeled images from the power sector, and proposes a high-performance multi-label classification model utilizing deep convolutional neural network (CNN) inspired by Inception-v4 [1] on this basis. Experiments demonstrate that the model proposed achieves an accuracy of 94.125% on the test set and thus can be deployed into practical intelligent surveillance scenarios.


Multi-label Scene categorization CNN 



This research was supported by the project as follows: Science and Technology Project of SGCC “Research on feature recognition and prediction of typical ice and wind disaster for transmission lines based on small sample machine learning method”.


  1. 1.
    Szegedy, C., Ioffe, S., Vanhoucke, V., Alemi, A.A.: Inception-v4, inception-ResNet and the impact of residual connections on learning. In: Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, pp. 4278–4284 (2017)Google Scholar
  2. 2.
    Lecun, Y., Bottou, L., Bengio, Y., Ha, P.: Gradient-based learning applied to document recognition. Proc. IEEE, 1–46 (1998)Google Scholar
  3. 3.
    Hinton, G.E., Osindero, S., Teh, Y.-W.: A fast learning algorithm for deep belief nets. Neural Comput. 18(7), 1527–1554 (2006)MathSciNetCrossRefGoogle Scholar
  4. 4.
    Zhu, J., Liao, S., Lei, Z., Li, S.Z.: Multi-label convolutional neural network based pedestrian attribute classification. Image Vis. Comput. 58, 224–229 (2017)CrossRefGoogle Scholar
  5. 5.
    Yuan, L., Chen, F., Zhou, L., Hu, D.: Improve scene classification by using feature and kernel combination. Neurocomputing 170, 213–220 (2015)CrossRefGoogle Scholar
  6. 6.
    Bay, H., Ess, A., Tuytelaars, T., Van Gool, L.: SURF: speeded up robust features. Comput. Vis. Image Underst. 110, 346–359 (2008)CrossRefGoogle Scholar
  7. 7.
    Nogueira, K., Penatti, O.A.B., Jefersson, A.: Towards better exploiting convolutional neural networks for remote sensing scene classification. Pattern Recognit. 61, 539–556 (2017)CrossRefGoogle Scholar
  8. 8.
    Hubel, B.Y.D.H., Wiesel, A.D.T.N.: Receptive fields, binocular interaction and functional architecture in the cat’s visual cortex. J. Physiol. 106–154 (1962)CrossRefGoogle Scholar
  9. 9.
    Fukushima, K.: Neocognitron: a self-organizing neural network model for a mechanism of pattern recognition unaffected by shift in position. Biol. Cybern. 36(4), 193–202 (1980)CrossRefGoogle Scholar
  10. 10.
    Krizhevsky, A., Hinton, G.E.: ImageNet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 25(2), 1–9 (2012)Google Scholar
  11. 11.
    Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., Erhan, D., Vanhoucke, V., Rabinovich, A.: Going deeper with convolutions. Comput. Vis. Pattern Recognit. 1–9 (2015)Google Scholar
  12. 12.
    Lin, A., Chen, Q., Yan, S.: Network in network. Comput. Sci. 1–10 (2013)Google Scholar
  13. 13.
    Ioffe, S., Szegedy, C.: Batch normalization: accelerating deep network training by reducing internal covariate shift. 448–456 (2015)Google Scholar
  14. 14.
    Szegedy, C., Vanhoucke, V., Shlens, J., Wojna, Z.: Rethinking the inception architecture for computer vision. 2818–2826 (2015)Google Scholar
  15. 15.
    He, Z., Zhang, X., Ren, S., Sun, S.: Deep residual learning for image recognition. 770–778 (2015)Google Scholar

Copyright information

© Springer Nature Singapore Pte Ltd. 2020

Authors and Affiliations

  • Gaofeng Zhao
    • 1
  • Wang Luo
    • 1
    Email author
  • Yang Cui
    • 1
  • Qiang Fan
    • 1
  • Qiwei Peng
    • 1
  • Zhen Kong
    • 1
  • Liang Zhu
    • 2
  • Tai Zhang
    • 3
  1. 1.NARI Group Corporation (State Grid Electric Power Research Institute)NanjingChina
  2. 2.State Grid Hunan Electric Power Company LimitedChangshaChina
  3. 3.State Grid Sichuan Electric Power Company LimitedChengduChina

Personalised recommendations