Abstract
With the outbreak of COVID-19 and various influenza diseases, it is necessary to wear masks properly in crowded public places to prevent the spread of the virus. Therefore, detecting mask-wearing efficiently and accurately is essential for people’s physical health and safety. In this paper, we present a novel one-stage mask detection method, named attention-guided neural network (AGNN) that can efficiently detect non-mask-wearing faces in public. Specifically, we started with YOLOv5 as a baseline and integrated the coordinate attention mechanism module into YOLOv5 to guide the holistic model for improving the ability of feature extraction. Furthermore, we explored utilizing the focal loss to solve the problem of class imbalance. The experiment is conducted on the face mask detection dataset of real-life scenes with twenty different categories. Experimental results demonstrate that the proposed AGNN method achieves higher precision and recall than the original YOLOv5 in multi-classification mask detection.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Centers for Disease Control and Prevention (CDC): Interim Infection Prevention and Control Reccommendations for patients with suspected or confirmed Coronavirus Disease 2019 (COVID-19) in Healthcare settings. https://www.cdc.gov/coronavirus/2019-nconv/hcp/infection-control-recommendations.html. Accessed Feb 2022
World Health Organization (WHO): Coronavirus disease (COVID-19) pandemic. Accessed Feb 2022. https://www.who.int/emergencies/diseases/novel-coronavirus-2019. Accessed Feb 2022
Girshick, R., Donahue, J., Darrell, T., Malik, J.: Rich feature hierarchies for accurate object detection and semantic segmentation. In: 2014 IEEE Conference on Computer Vision and Pattern Recognition, pp. 580–587 (2014)
Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: unified, real-time object detection. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 779–788 (2016)
Zhang, X., Fan, H., Zhu, H., Huang, X., Wu, T., Zhou, H.: Improvement of YOLOV5 model based on the structure of multiscale domain adaptive network for crowdscape. In: 2021 IEEE 7th International Conference on Cloud Computing and Intelligent Systems (CCIS), pp. 171–175 (2021)
Wang, B., Zheng, J., Chen, C.L.P.: A survey on masked facial detection methods and datasets for fighting against COVID-19. IEEE Trans. Artif. Intell. 3, 323–343 (2022)
Sethi, S., Kathuria, M., Kaushik, T.: Face mask detection using deep learning: an approach to reduce risk of coronavirus spread. J. Biomed. Inform. 120, 103848 (2021)
Wang, B., Zhao, Y., Chen, C.L.P.: Hybrid transfer learning and broad learning system for wearing mask detection in the COVID-19 Era. IEEE Trans. Instrum. Meas. 70, 1–12 (2021)
Wang, Y., Pan, L.: YOLOV5s-Face face detection algorithm. In: 2022 China Automation Congress (CAC), pp. 1107–1112 (2022)
Redmon, J., Farhadi, A.: YOLO9000: better, faster, stronger. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 6517–6525 (2017)
Redmon, J., Farhadi, A.: YOLOv3: An Incremental Improvement. ArXiv (2018)
Bochkovskiy, A., Wang, C.-Y., Liao, H.: YOLOv4: Optimal Speed and Accuracy of Object Detection. ArXiv (2020)
Sudars, K., et al.: YOLOv5 deep neural network for quince and raspberry detection on RGB images. In: 2022 Workshop on Microwave Theory and Techniques in Wireless Communications (MTTW), pp. 19–22 (2022)
Hu, J., Shen, L., Albanie, S., Sun, G., Wu, E.: Squeeze-and-excitation networks. IEEE Trans. Pattern Anal. Mach. Intell. 42, 2011–2023 (2020)
Lao, M., Guo, Y., Liu, Y., Lew, M.S.: A language prior based focal loss for visual question answering. In: 2021 IEEE International Conference on Multimedia and Expo (ICME), pp. 1–6 (2021)
Woo, S., Park, J., Lee, J.-Y., Kweon, I.S.: CBAM: Convolutional Block Attention Module (2018)
Li, Y., Cheng, R., Zhang, C., Chen, M., Ma, J., Shi, X.: Sign language letters recognition model based on improved YOLOv5. In: 2022 9th International Conference on Digital Home (ICDH), pp. 188–193 (2022)
Hou, Q., Zhou, D., Feng, J.: Coordinate attention for efficient mobile network design. In: 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 13708–13717 (2021)
Chen, J., Wei, Y., Zhou, Y.: Dense crowd detection algorithm for YOLOv5 based on coordinate attention mechanism. In: 2022 2nd International Conference on Algorithms, High Performance Computing and Artificial Intelligence (AHPCAI), pp. 187–190 (2022)
Lin, T.-Y., Dollar, P., Girshick, R., He, K., Hariharan, B., Belongie, S.: Feature pyramid networks for object detection. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 936–944 (2017)
Liu, S., Qi, L., Qin, H., Shi, J., Jia, J.: Path aggregation network for instance segmentation. In: 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 8759–8768 (2018)
Zheng, Z., Wang, P., Liu, W., Li, J., Ye, R., Ren, D.: Distance-IoU loss: faster and better learning for bounding box regression. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp. 12993–13000 (2020)
Neubeck, A., Van Gool, L.: Efficient non-maximum suppression. In: 18th International Conference on Pattern Recognition (ICPR 2006), pp. 850–855 (2006)
Liu, R., Cui, B., Fang, X., Guo, B., Ma, Y., An, J.: Super-resolution of GF-1 multispectral wide field of view images via a very deep residual coordinate attention network. IEEE Geosci. Remote Sens. Lett. 19, 1–5 (2022)
Lin, T.-Y., Goyal, P., Girshick, R., He, K., Dollar, P.: Focal loss for dense object detection. IEEE Trans. Pattern Anal. Mach. Intell. 42, 318–327 (2020)
Cheng, Z., Chai, S.: A cyber intrusion detection method based on focal loss neural network. In: 2020 39th Chinese Control Conference (CCC), pp. 7379–7383 (2020)
Extreme Mart platform. https://cvmart.net
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Zhang, B., Li, S., Wang, Z., Wu, L. (2023). Attention-Guided Neural Network for Face Mask Detection. In: Yongtian, W., Lifang, W. (eds) Image and Graphics Technologies and Applications. IGTA 2023. Communications in Computer and Information Science, vol 1910. Springer, Singapore. https://doi.org/10.1007/978-981-99-7549-5_15
Download citation
DOI: https://doi.org/10.1007/978-981-99-7549-5_15
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-99-7548-8
Online ISBN: 978-981-99-7549-5
eBook Packages: Computer ScienceComputer Science (R0)