Abstract
In this research we focus on using deep learning for the training of real time detection of defected Nails and Nuts on a high speed production line using You Only Look Once (YOLO) algorithm for real time object detection and trying to increase the precision of detection and decrease the problems facing real time object detection models like Object occlusion, different orientation for objects, lighting conditions, undetermined moving objects and noise. A series of experiments have been done to achieve high prediction accuracy, the experimental results made on our costumed pascal visual object classes (VOC) dataset demonstrated that the mean Average Precision (mAP) could reach 85%. The proposed model showed very good prediction accuracy on the test dataset.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Sanchez-Lopez, J.R., Marin-Hernandez, A., Palacios, E.: Visual detection, tracking and pose estimation of a robotic arm end effector, April 2011. https://www.researchgate.net/publication/239918179
Tsarouchi, P., Michalos, G., Makris, S., Chryssolouris, G.: Vision system for robotic handling of randomly placed objects. Procedia CIRP 9, 61–66 (2013). https://doi.org/10.1016/j.procir.2013.06.169
Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: unified, real-time object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, June 2016, pp. 779–788 (2016)
Girshick, R., Donahue, J., Darrell, T., Malik, J.: Rich feature hierarchies for accurate object detection and semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 580–587 (2014). https://doi.org/10.1109/CVPR.2014.81
Girshick, R.: Fast R-CNN. In: 2015 IEEE International Conference on Computer Vision (ICCV), pp. 1440–1448 (2015). https://doi.org/10.1109/ICCV.2015.169
Ren, S., He, K., Girshick, R., Sun, J.: Faster R-CNN: towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 39 (2015). https://doi.org/10.1109/TPAMI.2016.2577031
He, K., Gkioxari, G., Dollár, P., Girshick, R.: Mask R-CNN. In: IEEE International Conference on Computer Vision (ICCV), pp. 2980–2988 (2017). https://doi.org/10.1109/ICCV.2017.322
Redmon, J., Farhadi, A.: YOLO9000: better, faster, stronger. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, July 2017, pp. 6517–6525 (2017)
Ioffe, S., Szegedy, C.: Batch normalization: accelerating deep network training by reducing internal covariate shift. In: International Conference on Machine Learning, Lille, France, July 2005, pp. 448–456 (2005)
Sang, J., et al.: An improved YOLOv2 for vehicle detection. Sensors 18 (2018). http://www.mdpi.com/1424-8220/18/12/4272. https://doi.org/10.3390/s18124272
Wang, Y., Ewert, D., Vossen, R., Jeschke, S.: A visual servoing system for interactive human-robot object transfer. Autom. Control Eng. J. 3 (2015). https://doi.org/10.12720/joace.3.4.277-283
Ahlin, K., Joffe, B., Hu, A.-P., Mcmurray, G., Sadegh, N.: Autonomous leaf picking using deep learning and visual-servoing. IFAC-PapersOnLine 49, 177–183 (2016). https://doi.org/10.1016/j.ifacol.2016.10.033
Ye, R., Pan, C.-S., Chang, M., Yu, Q.: Intelligent defect classification system based on deep learning. Adv. Mech. Eng. 10(03) (2018). https://doi.org/10.1177/1687814018766682
Biresaw, T.A., Nawaz, T., Ferryman, J., Dell, A.I.: ViTBAT: video tracking and behavior annotation tool. In: 2016 13th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), pp. 295–301 (2016). https://doi.org/10.1109/AVSS.2016.7738055
Maher, A., Taha, H., Zhang, B.: Realtime multi-aircraft tracking in aerial scene with deep orientation network. J. Real-Time Image Proc. 15(3), 495–507 (2018)
Everingham, M., Van Gool, L., Williams, C.K., Winn, J., Zisserman, A.: The pascal visual object classes (VOC) challenge. Int. J. Comput. Vis. 88(2), 303–338 (2010)
Maher, A., Li, C., Hu, H., Zhang, B.: Realtime human-UAV interaction using deep learning. In: Zhou, J., et al. (eds.) CCBR 2017. LNCS, vol. 10568, pp. 511–519. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-69923-3_55
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this paper
Cite this paper
Teama, T., Ma, H., Maher, A., Kassab, M.A. (2019). Real Time Object Detection Based on Deep Neural Network. In: Yu, H., Liu, J., Liu, L., Ju, Z., Liu, Y., Zhou, D. (eds) Intelligent Robotics and Applications. ICIRA 2019. Lecture Notes in Computer Science(), vol 11743. Springer, Cham. https://doi.org/10.1007/978-3-030-27538-9_42
Download citation
DOI: https://doi.org/10.1007/978-3-030-27538-9_42
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-27537-2
Online ISBN: 978-3-030-27538-9
eBook Packages: Computer ScienceComputer Science (R0)