Abstract
Wildlife monitoring and analysis are an active research field since last many decades. In this paper, we focus on wildlife monitoring and analysis through animal detection from natural scenes acquired by camera-trap networks. The image sequences obtained from camera-trap consist of highly cluttered images that hinder the detection of animal resulting in low-detection rates and high false discovery rates. To handle this problem, we have used a camera-trap database that has candidate animal proposals using multilevel graph cut in the spatiotemporal domain. These proposals are used to create a verification phase that identifies whether a given patch is animal or background. We have designed animal detection model using self-learned Deep Convolutional Neural Network (DCNN) features. This efficient feature set is then used for classification using state-of-the-art machine learning algorithms, namely support vector machine, k-nearest neighbor, and ensemble tree. Our intensive results show that our detection model using DCNN features provides accuracy of 91.4% on standard camera-trap dataset.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
S. Tilak et al., Monitoring wild animal communities with arrays of motion sensitive camera, in Int. J. Res. Rev. Wireless Sensor Netw., vol. 1, pp. 1929, 2011.
R. Kays et al., eMammalCitizen science camera trapping as a solution for broad-scale, long-term monitoring of wildlife populations, in Proc. North Am. Conservation Biol., 2014, pp. 8086.
Zhi Zhang, Zhihai He, Guitao Cao, and Wenming Cao, Animal Detection From Highly Cluttered Natural Scenes Using Spatiotemporal Object Region Proposals and Patch Verification, in IEEE Transactions on Multimedia, vol. 18, no. 10, October 2016.
Mario I. Chacon-Murguia, and Sergio Gonzalez-Duarte, An Adaptive Neural-Fuzzy Approach for Object Detection in Dynamic Backgrounds for Surveillance Systems, in IEEE Transaction on industrial electronics, vol. 59, no. 8, August 2012.
Wei Liu, Hongfei Yu, Huai Yuan, Hong Zhao, Xiaowei Xu, Effective background modelling and subtraction approach for moving object detection, in IET Computer Vision, April 2014.
V. Mahadevan and N. Vasconcelos, Background subtraction in highly dynamic scenes, in Proc. IEEE Conf. Comput. Vis. Pattern Recog., Jun. 2008, pp. 16.
R. Girshick, J. Donahue, T. Darrell, and J. Malik, Rich feature hierar-chies for accurate object detection and semantic segmentation, in Proc. IEEE Conf. Comput. Vis. Pattern Recog., Jun. 2014, pp. 580587.
R. Girshick,Fast r-CNN, in Proc. Int. Conf. Comput. Vis., pp. 14401448, 2015.
K. H. Shaoqing Ren and J. S. Ross Girshick, Faster R-CNN: Towards real-time object detection with region proposal networks, in Adv. Neural Inf. Process. Syst., 2015, pp. 9199.
J. R. Uijlings, K. E. van de Sande, T. Gevers, and A. W. Smeulders, Selective search for object recognition, in Int. J. Comput. Vis., vol. 104, no. 2, pp. 154171, 2013.
M. M. Cheng, Z. Zhang, W. Y. Lin, and P. Torr, BING: Binarized normed gradients for objectness estimation at 300fps, in Proc. IEEE Conf. Comput. Vis. Pattern Recog., Jun. 2014, pp. 32863293.
P.L. St-Charles, G.-A. Bilodeau, and R. Bergevin, Flexible background subtraction with self-balanced local sensitivity, in Proc. IEEE Conf. Comput. Vis. Pattern Recog. Workshops, Jun. 2014, pp. 414419.
X. Ren, T. X. Han, and Z. He, Ensemble video object cut in highly dynamic scenes, in Proc. IEEE Conf. Comput. Vis. Pattern Recog., Jun. 2013, pp. 19471954.
M. Oquab, L. Bottou, I. Laptev, and J. Sivic, Learning and transferring mid-level image representations using convolutional neural networks, in Proc. IEEE Conf. Comput. Vis. Pattern Recog., Jun. 2014, pp. 17171724.
A. S. Razavian, H. Azizpour, J. Sullivan, and S. Carlsson, CNN features off-the-shelf: An astounding baseline for recognition, in Proc. IEEE Conf. Comput. Vis. Pattern Recog. Workshops, Jun. 2014, pp. 512519.
C. Szegedy, A. Toshev, and D. Erhan,Deep neural networks for object detection, in Proc. Adv. Neural Inf. Process. Syst., 2013, pp. 25532561.
D. Erhan, C. Szegedy, A. Toshev, and D. Anguelov, Scalable object detection using deep neural networks, in Proc. IEEE Conf. Comput. Vis. Pattern Recog., Jun. 2014, pp. 21552162.
P. Sermanet et al., Overfeat: Integrated recognition, localization and detection using convolutional networks, in Proc. Int. Conf. Learn. Represent., Dec. 2013. [Online]. Available: http://adsabs.harvard.edu/abs/2014arXiv1412.1441S.
C. Szegedy, S. Reed, D. Erhan, and D. Anguelov, Scalable, high-quality object detection, Dec. 2014.
A. Krizhevsky, I. Sutskever, and G. E. Hinton, Imagenet classification with deep convolutional neural networks, in Proc. Adv. Neural Inf. Process. Syst., 2012, pp. 10971105.
Ken Chatfield, Karen Simonyan, Andrea Vedaldi, Andrew Zisserman, Return of the Devil in the Details: Delving Deep into Convolutional Nets, in proceedings of BMVC 2014.
Zhang, Z., Han, T.X., He, Z.: Coupled ensemble graph cuts and object verification for animal segmentation from highly cluttered videos. In: 2015 IEEE International Conference on Image Processing (ICIP), pp. 2830–2834. IEEE (2015).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Verma, G.K., Gupta, P. (2018). Wild Animal Detection Using Deep Convolutional Neural Network. In: Chaudhuri, B., Kankanhalli, M., Raman, B. (eds) Proceedings of 2nd International Conference on Computer Vision & Image Processing . Advances in Intelligent Systems and Computing, vol 704. Springer, Singapore. https://doi.org/10.1007/978-981-10-7898-9_27
Download citation
DOI: https://doi.org/10.1007/978-981-10-7898-9_27
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-10-7897-2
Online ISBN: 978-981-10-7898-9
eBook Packages: EngineeringEngineering (R0)