Skip to main content

Improving the Lightweight Object Detection Method for YOLOv5

  • Conference paper
  • First Online:
Advanced Manufacturing and Automation XII (IWAMA 2022)

Part of the book series: Lecture Notes in Electrical Engineering ((LNEE,volume 994))

Included in the following conference series:

  • 1052 Accesses

Abstract

As an excellent algorithm, YOLOv5 is an object detection model with the advantages of high flexibility and fast speed, but it has problems with many network parameters, complex model structure, and low boundary regression accuracy for the target. For the above problems, this study improves on the YOLOv5s algorithm and proposes a new model YOLO-GC with lower hardware requirements, fewer network parameters and higher boundary regression accuracy. First, the upsampling module of YOLOv5s is replaced by the CARAFE upsampling module to increase the receptive field and semantic features, the Ghost module to reduce the number of parameters and calculation, and the CBAM attention mechanism to combine the spatial and channel attention map, the model pays more attention to the key areas to improve the model accuracy. The test results of this research method on the PASCAL VOC object detection benchmark dataset show that compared with YOLOv5s, the number of parameters is reduced by 44.2%, the model size is reduced by 42.8%, the is increased by 1.2%, and the :0.95 An increase of 5.5%.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 259.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 329.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 329.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Redmon, J., Farhadi, A.: YOLOv3: An incremental improvement (2018). arXiv preprint arXiv:1804.02767

  2. Bochkovskiy, A., Wang, C., Liao, H.M.: YOLOv4: Optimal Speed and Accuracy of Object Detection (2020). arXiv preprint arXiv:2004.10934

  3. Howard, A.G., Zhu, M., Chen, B., Kalenichenko, D., Wang, W., Weyand, T., et al.: MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications (2017). arXiv preprint arXiv:1704.04861

  4. Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., Chen, L.: MobileNetV2: inverted residuals and linear bottlenecks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4510–4520 (2018)

    Google Scholar 

  5. Howard, A., Sandler, M., Chu, G., Chen, L., Chen, B., Tan, M., et al.: Searching for MobileNetV3. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 1314–1324 (2019)

    Google Scholar 

  6. Zhang, X., Zhou, X., Lin, M., Sun, J.: ShuffleNet: an extremely efficient convolutional neural network for mobile devices. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6848–6856 (2018)

    Google Scholar 

  7. Ma, N., Zhang, X., Zheng, H.-T., Sun, J.: ShuffleNet V2: practical guidelines for efficient CNN architecture design. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) Computer Vision – ECCV 2018. LNCS, vol. 11218, pp. 122–138. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01264-9_8

    Chapter  Google Scholar 

  8. Han, K., Wang, Y., Tian, Q., Guo, J., Xu, C., Xu, C.: GhostNet: more features from cheap operations. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp.1580–1589 (2020)

    Google Scholar 

  9. Woo, S., Park, J., Lee, J.-Y., Kweon, I.S.: CBAM: convolutional block attention module. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11211, pp. 3–19. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01234-2_1

    Chapter  Google Scholar 

  10. Wang, J., Chen, K., Xu, R., Liu, Z., Loy, C.C., Lin, D.: CARAFE: content-aware reassembly of features. IEEE Trans. Pattern Anal. Mach. Intell. 3007–3016 (2019)

    Google Scholar 

  11. Tsung-Yi, L., Dollar, P., Girshick, R., He, K., Hariharan, B., Belongie, S.: Feature pyramid networks for object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR):2117–2125 (2017)

    Google Scholar 

  12. Liu, S., Qi, L., Qin, H., Shi, J., Jia, J.: path aggregation network for instance segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp.8759–8768 (2018)

    Google Scholar 

Download references

Acknowledgment

This work is supported by Fujian Provincial Natural Science Foundation (Grant No.: 2021J1851) and Xiamen Winjoin Technology Co. (Contract No.: S21228).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Ning Chen .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Chen, N., Li, Q., Ning, J., Wang, Q., Liao, N. (2023). Improving the Lightweight Object Detection Method for YOLOv5. In: Wang, Y., Yu, T., Wang, K. (eds) Advanced Manufacturing and Automation XII. IWAMA 2022. Lecture Notes in Electrical Engineering, vol 994. Springer, Singapore. https://doi.org/10.1007/978-981-19-9338-1_4

Download citation

Publish with us

Policies and ethics