Abstract
Light field (LF) images provide rich information and are suitable for high-level computer vision applications. To acquire capabilities of modeling the correlated information of LF, most of the previous methods have to stack several convolutional layers to improve the feature representation and result in heavy computation and large model sizes. In this paper, we propose channel-wise and SAI-wise attention modules to enhance the feature representation at a low cost. The channel-wise attention module helps to focus on important channels while the SAI-wise attention module guides the network to pay more attention to informative SAIs. The experimental results demonstrate that the baseline network can achieve better performance with the aid of the attention modules.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsReferences
Stanford Lytro Light Field Archive. http://lightfields.stanford.edu/LF2016.html
Abadi, M., et al.: TensorFlow: a system for large-scale machine learning. In: 12th \(USENIX\) Symposium on Operating Systems Design and Implementation OSDI 2016, pp. 265–283 (2016)
Chollet, F., et al.: Keras (2015). https://keras.io
Fiss, J., Curless, B., Szeliski, R.: Refocusing plenoptic images using depth-adaptive splatting. In: 2014 IEEE International Conference on Computational Photography (ICCP), pp. 1–9. IEEE (2014)
Heber, S., Yu, W., Pock, T.: U-shaped networks for shape from light field. In: Proceedings of the British Machine Vision Conference 2016, vol. 1, pp. 37.1–37.12 (2016)
Heber, S., Yu, W., Pock, T.: Neural EPI-volume networks for shape from light field. In: Proceedings of the IEEE International Conference on Computer Vision, vol. 2017-October, pp. 2271–2279, October 2017
Hu, J., Shen, L., Sun, G.: Squeeze-and-excitation networks. In: 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 7132–7141. IEEE, June 2018
Jeon, H.G., et al.: Depth from a light field image with learning-based matching costs. IEEE Trans. Pattern Anal. Mach. Intell. 41(2), 297–310 (2019)
Kalantari, N.K., Wang, T.C., Ramamoorthi, R.: Learning-based view synthesis for light field cameras. ACM Trans. Graph. 35(6), 193 (2016). (Proceedings of SIGGRAPH Asia 2016)
Lu, Z., Yeung, H.W.F., Qu, Q., Chung, Y.Y., Chen, X., Chen, Z.: Improved image classification with 4D light-field and interleaved convolutional neural network. Multimedia Tools Appl. 78(20), 29211–29227 (2019)
Nair, V., Hinton, G.E.: Rectified linear units improve restricted Boltzmann machines. In: Proceedings of the 27th International Conference on Machine Learning (ICML-10), pp. 807–814 (2010)
Ronneberger, O., Fischer, P., Brox, T.: U-net: convolutional networks for biomedical image segmentation. In: Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F. (eds.) MICCAI 2015. LNCS, vol. 9351, pp. 234–241. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-24574-4_28
Shin, C., Jeon, H.G., Yoon, Y., So Kweon, I., Joo Kim, S.: EPINET: a fully-convolutional neural network using epipolar geometry for depth from light field images. In: The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2018
Wang, T.C., Efros, A.A., Ramamoorthi, R.: Occlusion-aware depth estimation using light-field cameras. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 3487–3495 (2015)
Wang, T.-C., Zhu, J.-Y., Hiroaki, E., Chandraker, M., Efros, A.A., Ramamoorthi, R.: A 4D light-field dataset and CNN architectures for material recognition. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9907, pp. 121–138. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-46487-9_8
Yeung, H.W.F., Hou, J., Chen, J., Chung, Y.Y., Chen, X.: Fast light field reconstruction with deep coarse-to-fine modeling of spatial-angular clues. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11210, pp. 138–154. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01231-1_9
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this paper
Cite this paper
Hu, Z., Chung, Y.Y., Zandavi, S.M., Ouyang, W., He, X., Gao, Y. (2019). High-Performance Light Field Reconstruction with Channel-wise and SAI-wise Attention. In: Gedeon, T., Wong, K., Lee, M. (eds) Neural Information Processing. ICONIP 2019. Communications in Computer and Information Science, vol 1143. Springer, Cham. https://doi.org/10.1007/978-3-030-36802-9_14
Download citation
DOI: https://doi.org/10.1007/978-3-030-36802-9_14
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-36801-2
Online ISBN: 978-3-030-36802-9
eBook Packages: Computer ScienceComputer Science (R0)