Abstract
An end-to-end neural network model based on DenseNet was designed to estimate the pose of the camera in this paper. The picture frame captured by the camera and the camera position (3-dimensional space coordinates) and pose (quaternion) corresponding to the picture frame are the inputs to the network model. Through the neural network model, the spatial structure information and the higher-layer features in the image are trained and learned, so that the network model finally outputs the 7-dimensional vector representing the camera position (3-dimensional space coordinates) and the pose (quaternion). Due to the pose estimation constraint of the network, the training effect of the model is guaranteed, and the pose estimation ability of the network is improved. The trained model is validated on the StMarysChurch Dataset. The experimental results show that the network model has good performances in accuracy and shorter training time.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsReferences
Kendall A, Grimes M, Cipolla R (2015) PoseNet: a convolutional network for real-time 6-DOF camera relocalization. In: CONFERENCE 2015, ICCV, Santiago, Chile
Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov D, Erhan D, Vanhoucke V, Rabinovich A (2015) Going deeper with convolutions. In: CONFERENCE 2015, CVPR, Boston, USA
Qiao Q (2018) Research on visual ego-motion estimation with convolutional neural network. University of Science and Technology of China, Hefei (2018)
Valada A, Radwan N, Burgard W (2018) Deep auxiliary learning for visual localization and odometry. In: CONFERENCE 2018, ICRA, Brisbance, Australia
Mayer N, llg E, Häusser P, Fischer P (2015) A large dataset to train convolutional networks for disparity, optical flow, and scene flow estimation. In: CONFERENCE 2015, CVPR, Boston, USA
Dosovitskiy A, Fischer P, Ilg E, Hausser P, Hazırbas C, Golkov V (2015) FlowNet: learning optical flow with convolutional networks. In: CONFERENCE 2015, ICCV, Santiago, Chile
Wu C (2015) Towards linear-time incremental structure from motion. In: CONFERENCE 2015, 3DV, Seattle WA, USA
Acknowledgments
This work was supported by National Natural Science Foundation of China (grant numbers 61520106010, 61741302).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Lu, L., Zhang, W. (2020). Visual Pose Estimation Based on the DenseNet Network. In: Jia, Y., Du, J., Zhang, W. (eds) Proceedings of 2019 Chinese Intelligent Systems Conference. CISC 2019. Lecture Notes in Electrical Engineering, vol 593. Springer, Singapore. https://doi.org/10.1007/978-981-32-9686-2_18
Download citation
DOI: https://doi.org/10.1007/978-981-32-9686-2_18
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-32-9685-5
Online ISBN: 978-981-32-9686-2
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)