Local Binary Pattern Metric-Based Multi-focus Image Fusion

  • Wenda ZhaoEmail author
  • Weiling Yin
  • Di You
  • Dong Wang
Part of the Studies in Computational Intelligence book series (SCI, volume 810)


Multi-focus image fusion is to integrate the partially focused images into one single image which is focused everywhere. Nowadays, it has become an important research topic due to the applications in more and more scientific fields. However, preserving more information of the low-contrast area in the focus area and maintaining the edge information are two challenges for existing approaches. In this paper, we address these two challenges with presenting a simple yet efficient multi-focus fusion method based on local binary pattern (LBP). In our algorithm, we measure the clarity using the LBP metric and construct the initial weight map. And then we use the connected area judgment strategy (CAJS) to reduce the noise in the initial map. Afterwards, the two source images are fused together by weighted arranging. The experimental results validate that the proposed algorithm outperforms state-of-the-art image fusion algorithms in both qualitative and quantitative evaluations, especially when dealing with low contrast regions and edge information.


Multi-focus image fusion Local binary pattern Connected area judgment strategy Preserving the low-contrast area information Maintaining the edge information 


  1. 1.
    Kaur, G., Kaur, P.: Survey on multifocus image fusion techniques. In: International Conference on Electrical, Electronics, and Optimization Techniques (ICEEOT), pp. 1420–1424 (2016)Google Scholar
  2. 2.
    Qu, Y., Yang, H.: Optical microscopy with flexible axial capabilities using a vari-focus liquid lens. J. Microsc. 258(3), 212–222 (2015)CrossRefGoogle Scholar
  3. 3.
    Abdipour, M., Nooshyar, M.: Multi-focus image fusion using sharpness criteria for visual sensor networks in wavelet domain. Comput. Electr. Eng. 51, 74-88 (2016) (ISSN)Google Scholar
  4. 4.
    Liu, Y., Wang, Z.: Simultaneous image fusion and denoising with adaptive sparse representation. IET Image Proc. 9(5), 347–357 (2015)CrossRefGoogle Scholar
  5. 5.
    Yang, Y., Tong, S., Huang, S., Lin, P., Fang, Y.: A hybrid method for multi-focus image fusion based on fast discrete curvelet transform. IEEE Access 5, 14898–14913 (2017)CrossRefGoogle Scholar
  6. 6.
    Yang, Y., Tong, S., Huang, S., Lin, P.: Multifocus image fusion based on NSCT and focused area detection. IEEE Sens. J. 15(5), 2824–2838 (2015)Google Scholar
  7. 7.
    Yang, Y., et al.: Multi-focus image fusion via clustering PCA based joint dictionary learning. IEEE Access 5, 16985–16997 (2017)CrossRefGoogle Scholar
  8. 8.
    Zhang, Q., Levine, M.D.: Robust multi-focus image fusion using multi-task sparse representation and spatial context. IEEE Trans. Image Process. 25(5), 2045–2058 (2016)MathSciNetCrossRefGoogle Scholar
  9. 9.
    Cao, L., Jin, L., Tao, H., Li, G., Zhuang, Z., Zhang, Y.: Multi-focus image fusion based on spatial frequency in discrete cosine transform domain. IEEE Signal Process. Lett. 22(2), 220–224 (2015)Google Scholar
  10. 10.
    Jiang, Q., Jin, X., Lee, S.J., Yao, S.: A novel multi-focus image fusion method based on stationary wavelet transform and local features of fuzzy sets. IEEE Access 5, 20286–20302 (2017)CrossRefGoogle Scholar
  11. 11.
    Shreyamsha Kumar, B.K.: Multi-focus and multispectral image fusion based on pixel significance using discrete cosine harmonic wavelet transform. Signal Image Video Process. 7(6), 125–1143 (2013)Google Scholar
  12. 12.
    Shreyamsha Kumar, B.K.: Image fusion based on pixel significance using cross bilateral filter. Signal Image Video Process. 9(5), 1193–1204 (2015)CrossRefGoogle Scholar
  13. 13.
    Liu, Y., Chen, X., Ward, R.K., Jane Wang, Z.: Image fusion with convolutional sparse representation. IEEE Signal Process. Lett. 23(12), 1882–1886 (2016)CrossRefGoogle Scholar
  14. 14.
    Guo, D., Yan, J., Xiaobo, Q.: High quality multi-focus image fusion using self-similarity and depth information. Opt. Commun. 338, 138–144 (2015)CrossRefGoogle Scholar
  15. 15.
    Gangapure, V.N., Banerjee, S., Chowdhury, A.S.: Steerable local frequency based multispectral multifocus image fusion. Inf. Fusion 23, 99–115 (2015)CrossRefGoogle Scholar
  16. 16.
    Farid, M.S., Mahmood, A., Al-Maadeed, S.A.: Multi-focus image fusion using content adaptive blurring. Inf. Fusion (2018)Google Scholar
  17. 17.
    Chen, Y., Guan, J., Cham, W.K.: Robust multi-focus image fusion using edge model and multi-matting. IEEE Trans. Image Process. 27(3), 1526–1541 (2018)MathSciNetCrossRefGoogle Scholar
  18. 18.
    Liu, Y., Chen, X., Peng, H., Wang, Z.: Multi-focus image fusion with a deep convolutional neural network. Inf. Fusion 36, 191–207 (2017)Google Scholar
  19. 19.
    Du, C., Gao, S.: Image segmentation-based multi-focus image fusion through multi-scale convolutional neural network. IEEE Access 5, 15750–15761 (2017)CrossRefGoogle Scholar
  20. 20.
    Zhang, Y., Bai, X., Wang, T.: Boundary finding based multi-focus image fusion through multi-scale morphological focus-measure. Inf. Fusion 35, 81–101 (2017)Google Scholar
  21. 21.
    Nejati, M., Samavi, S., Karimi, N., Reza Soroushmehr, S.M., Shirani, S., Roosta, I., Najarian, K.: Surface area-based focus criterion for multi-focus image fusion. Inf. Fusion 36, 284–295 (2017)Google Scholar
  22. 22.
    Li, S., Kang, X., Hu, J.: Image Fusion With Guided Filtering. IEEE Trans. Image Process. 22(7), 2864–2875 (2013)CrossRefGoogle Scholar
  23. 23.
    Zhao, J., Feng, H., Xu, Z., Li, Q., Tao, X.: Automatic blur region segmentation approach using image matting. Signal Image Video Process. 7(6), 1173–1181 (2013)CrossRefGoogle Scholar
  24. 24.
    Shi, J., Xu, L., Jia, J.: Discriminative blur detection features. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2965–2972 (2014)Google Scholar
  25. 25.
    Liu, R., Li, Z., Jia, J.: Image partial blur detection and classification. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1–8 (2008)Google Scholar
  26. 26.
    Vu, C.T., Phan, T.D., Chandler, D.M.: S3: a spectral and spatial measure of local perceived sharpness in natural images. IEEE Trans. Image Process. 21(3), 934–945 (2012)Google Scholar
  27. 27.
    Ojala, T., Pietikäinen, M., Harwood, D.: A comparative study of texture measures with classification based on featured distributions. Pattern Recognit. 29(1), 51–59 (1996)CrossRefGoogle Scholar
  28. 28.
    Jun, Z., Jizhao, H., Zhenglan, T., Feng, W.: Face detection based on LBP. In: IEEE International Conference on Electronic Measurement & Instruments (ICEMI), pp. 421–425 (2017)Google Scholar
  29. 29.
    Shu, Z., Liu, G., Xie, Z.: Real time target tracking scale adaptive based on LBP operator and nonlinear meanshift. In: International Conference on Cyber-Enabled Distributed Computing and Knowledge Discovery (CyberC), pp. 130–133 (2017)Google Scholar
  30. 30.
    Guo, Z., Zhang, L., Zhang, D.: A completed modeling of local binary pattern operator for texture classification. IEEE Trans. Image Process. 19(6), 1657–1663 (2010)MathSciNetCrossRefGoogle Scholar
  31. 31.
    Yi, X., Eramian, M.: LBP-based segmentation of defocus blur [J]. IEEE transactions on image processing. 25(4), 1626–1638 (2016)Google Scholar
  32. 32.
    Ojala, T., Pietikainen, M., Maenpaa, T.: Multiresolution gray-scale and rotation invariant texture classification with local binary patterns. IEEE Trans. Pattern Anal. Mach. Intell. 24(7), 971–987 (2002)CrossRefGoogle Scholar
  33. 33.
    Seiichi, S., Lu, H.: Underwater image dehazing using joint trilateral filter. Comput. Electr. Eng. 40(1), 41–50 (2014)Google Scholar
  34. 34.
  35. 35.
    Zhang, Q., Guo, B.: Multifocus image fusion using the non-subsampled contourlet transform. Signal Process. 89(7), 1334–1346 (2009)CrossRefGoogle Scholar
  36. 36.
    Hossny, M., Nahavandi, S., Creighton, D.: Comments on ‘information measure for performance of image fusion’. Electron. Lett. 44(18), 1066–1067 (2008)CrossRefGoogle Scholar

Copyright information

© Springer Nature Switzerland AG 2020

Authors and Affiliations

  1. 1.School of Information and Communication EngineeringDalian University of TechnologyDalianChina

Personalised recommendations