Skip to main content

A Collective Neurodynamic Optimization Approach to Nonnegative Tensor Decomposition

  • Conference paper
  • First Online:
Advances in Neural Networks - ISNN 2017 (ISNN 2017)

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 10262))

Included in the following conference series:

Abstract

In this paper, a collective neurodynamic optimization approach is proposed to nonnegative tensor factorization. Tensor decompositions are often applied in the data analysis. However, it is often a nonconvex optimization problem, which would cost much time and usually trap into the local minima. To solve this problem, a novel collective neurodynamic optimization approach is proposed by combining recurrent neural networks (RNN) and particle swarm optimization (PSO) algorithm. Each RNN still carries out local search. And then the best solution of each RNN improves through PSO framework. In the end, the global optimal solutions of nonnegative tensor factorization are obtained. Experiments results demonstrate the effectiveness for the nonconvex optimization with constraints.

J. Fan—The work described in the paper was supported by the National Key Research and Development Program of China under project 2016YFC1401007, the Foundation of High Resolution Special Research under 41-Y30B12-9001-14/16, and the National Natural Science Foundation of China under project 61273307.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Chen, Y., Han, D., Qi, L.: New ALS methods with extrapolating search directions and optimal step size for complex-valued tensor decompositions. IEEE Trans. Sig. Process. 59, 5888–5898 (2011)

    Article  MathSciNet  Google Scholar 

  2. Cichocki, A., Mandic, D., Lathauwer, L., Zhou, G., Zhao, Q., Caiafa, C.: Tensor decompositions for signal processing applications: from two-way to multiway component analysis. IEEE Sig. Process. Mag. 32, 145–163 (2015)

    Article  Google Scholar 

  3. Dauwels, J., Srinivasan, K., Reddy, M., Cichocki, A.: Near-lossless multichannel EEG compression based on matrix and tensor decompositionss. IEEE J. Biomed. Health Inform. 17, 708–714 (2013)

    Article  Google Scholar 

  4. Kolda, T., Bader, B.: Tensor decompositions and applications. SIAM Rev. 51, 455–500 (2009)

    Article  MathSciNet  MATH  Google Scholar 

  5. Chen, Y., Hsu, C., Hsu, H., Liao, H.: Simultaneous tensor decomposition and completion using factor priors. IEEE Trans. Pattern Anal. Mach. Intell. 36, 577–591 (2014)

    Article  Google Scholar 

  6. Lee, D., Seung, H.: Algorithms for nonnegative matrix factorization. In: Advances in Neural Information Processing Systems, pp. 556–562 (2000)

    Google Scholar 

  7. Cichocki, A., Zdunek, R., Phan, A., Amari, S.: Nonegative Matrix and Tensor Factorizations: Application to Exploratory Multi-way Data Analysis and Blind Source Separation. Wiley, Chichester (2009)

    Book  Google Scholar 

  8. Phan, A., Cichocki, A.: Extended HALS algorithm for nonnegative Tucker decomposition and its applications for multiway analysis and classification. Neurocomputing 74, 1956–1969 (2011)

    Article  Google Scholar 

  9. Wang, J., Wu, G.: Recurrent neural networks for LU decomposition and Cholesky factorization. Math. Comput. Model. 18, 1–8 (1993)

    Article  MathSciNet  MATH  Google Scholar 

  10. Xia, Y.S., Wang, J.: A recurrent neural network for nonlinear convex optimization subject to nonlinear inequality constraints. IEEE Trans. Circ. Syst. I Regul. Pap. 51, 1385–1394 (2004)

    Article  MathSciNet  Google Scholar 

  11. Yan, Z., Wang, J., Li, G.C.: A collective neurodynamic optimization approach to bound-constrained nonconvex optimization. Neural Netw. 55, 20–29 (2014)

    Article  MATH  Google Scholar 

  12. Kennedy, J., Eberhart, R.: Particle swarm optimization. In: IEEE International Conference on Neural Networks, pp. 1942–1948 (1995)

    Google Scholar 

  13. Han, M., Fan, J.C., Wang, J.: A dynamic feedforward neural network based on gaussian particle swarm optimization and its application for predictive control. IEEE Trans. Neural Netw. 22, 1457–1468 (2011)

    Article  Google Scholar 

  14. Lathauwer, L.D., Castaing, J., Cardoso, J.F.: Fourth-order cumulantbased blind identification of underdetermined mixtures. IEEE Trans. Sig. Process. 55, 2965–2973 (2007)

    Article  Google Scholar 

  15. Liu, Q.S., Wang, J.: A one-layer recurrent neural network with a discontinuous hard-limiting activation function for quadratic programming. IEEE Trans. Neural Netw. 19, 558–570 (2008)

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Jianchao Fan .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2017 Springer International Publishing AG

About this paper

Cite this paper

Fan, J., Wang, J. (2017). A Collective Neurodynamic Optimization Approach to Nonnegative Tensor Decomposition. In: Cong, F., Leung, A., Wei, Q. (eds) Advances in Neural Networks - ISNN 2017. ISNN 2017. Lecture Notes in Computer Science(), vol 10262. Springer, Cham. https://doi.org/10.1007/978-3-319-59081-3_25

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-59081-3_25

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-59080-6

  • Online ISBN: 978-3-319-59081-3

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics