Abstract
Tensor decomposition (TD) is a natural extension of matrix factorization (MF), introduced for matrices in the previous chapter, when tensors instead of matrices are considered. In contrast to the MF that is usually represented as a product of two matrices, TD has various forms. In contrast to the matrices that were extensively studied over long period, tensor has much shorter history of extensive investigations, especially from the application point of views. Thus, there are no de facto standards to be used for the specific application. Similar to the aim of MF, that of TD is also to reduce the degrees of freedoms. Nevertheless, how the degrees of freedom can be reduced has many variations for TD. In this chapter, we introduce three principal realizations of TD: sum of outer product of vectors, product summation of (smaller) tensor and matrices, and product summation of (smaller) tensors. These three methods have their own unique pros and cons. In addition to the algorithm to perform each of TDs, we will also discuss about these pros and cons of three methods introduced.
I painted her as an unapproachable enigma and never even tried to see her for who she was.Ichigo, Darling in the FranXX, Season 1, Episode 16
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsNotes
- 1.
Although the detailed algorithms of individual TDs will be presented in the later sections, readers might feel that they would like to try them in advance with reading prior sections that demonstrate examples. In that case, see Appendix A where I list some of the implementations on various platforms.
- 2.
See Appendix for more details about Moore-Penrose pseudoinverse. Alternatively, one can simply execute linear regression analysis , Eq. (3.35).
References
Barata, J.C.A., Hussein, M.S.: The Moore–Penrose pseudoinverse: a tutorial review of the theory. Braz. J. Phys. 42(1), 146–165 (2012). https://doi.org/10.1007/s13538-011-0052-z
Kolda, T., Bader, B.: Tensor decompositions and applications. SIAM Rev. 51(3), 455–500 (2009). https://doi.org/10.1137/07070111X
Oseledets, I.: Tensor-train decomposition. SIAM J. Sci. Comput. 33(5), 2295–2317 (2011). https://doi.org/10.1137/090752286
Author information
Authors and Affiliations
Appendix
Appendix
1.1 I Moore-Penrose Pseudoinverse
Moore-Penrose pseudoinverse [1] , which is denoted as A †, of matrix A satisfies the following conditions:
-
AA † A = A
-
A † AA † = A †
-
(A † A)T = A † A
-
(AA †)T = AA †
Suppose we need to find \(\boldsymbol {x} \in \mathbb {R}^M\) that satisfies
where \(A \in \mathbb {R}^{N \times M}\) and \(\boldsymbol {b} \in \mathbb {R}^N\). It is known that there is a unique solution only when N = M.
Moore-Penrose pseudoinverse can solve Eq. (3.72) because
gives
-
the unique solution of Eq. (3.72) when N = M.
-
the x that satisfies Eq. (3.72) with minimum |x| when N < M (i.e., when no unique solutions are available).
-
the x with minimum |A x −b| when N > M (equivalent to the so-called linear regression analysis ).
When N < M, there are infinitely large number of solutions that satisfy Eq. (3.72). Moore-Penrose pseudoinverse allows us to select one of them, which has minimum |x|. On the other hand, when N > M, there are not always solutions that satisfy Eq. (3.72). Moore-Penrose pseudoinverse allows us to select the solution having the minimum |A x −b|, i.e., the smallest residuals. Thus, by computing Moore-Penrose pseudoinverse, we can always compute x that satisfies Eq. (3.72) as much as possible in some sense.
How to compute A † is as follows. Apply SVD to A as
\(U \in \mathbb {R}^{N \times M}, \varSigma , V \in \mathbb {R}^{M \times M}\) for N > M and \(U, \varSigma \in \mathbb {R}^{N \times N}, V \in \mathbb {R}^{M \times N}\) for N < M. When U or V is not a square matrix, U T U = V T V = I, but UU T ≠ I and V V T ≠ I. When U and V are square matrices, U T U = UU T = V T V = V V T = I.
Then A † can be defined as
It is not difficult to show that A † = V Σ −1 U T satisfies the required conditions because
and
where V T V = I for N > M and U T U = I for N < M are used.
Then when N > M,
On the other hand, when N < M,
When N = M, these are obvious because AA † = A † A = I.
The reason why we can treat Eq. (3.72) using Moore-Penrose pseudoinverse as mentioned in the above is as follows. Define
with arbitrary vector w. Then because
when AA † = I, i.e., N ≤ M, A x 0 = b, x 0 is a solution of Eq. (3.72). This corresponds to the cases where there are no unique solutions because the number of variables, M, is larger than the number of equations, N. x 0 can be a unique solution only when A † A = I as well, i.e., N = M because of Eq. (3.86). This corresponds to the cases where there is a unique solution because the number of variables, M, is equal to the number of equations, N.
Here one should notice that \(A^\dagger \boldsymbol {b} \perp \left (I - A^\dagger A \right ) \boldsymbol {w}\) because
Thus from Eq. (3.86)
This means \(|\boldsymbol {x}_0| > \left |A^\dagger \boldsymbol {b} \right |\). Therefore, A † b is the solution that satisfies Eq. (3.72) and has the smallest |x 0| (in other words, the solution with the L2 regulation term).
When AA †≠ I, i.e., N > M, there are no solutions. This corresponds to the cases where there are no solutions because the number of variables, M, is smaller than the number of equations, N. In this case, x = A † b is known to be optimal (i.e., the solution with minimum |A x −b|). In order to prove this, first we need to compute A T(AA † b −b) as
With taking transposition of the above, we can also get
Using these, we can show
This means that x = A † b is an optimal solution of Eq. (3.72).
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this chapter
Cite this chapter
Taguchi, Yh. (2020). Tensor Decomposition. In: Unsupervised Feature Extraction Applied to Bioinformatics. Unsupervised and Semi-Supervised Learning. Springer, Cham. https://doi.org/10.1007/978-3-030-22456-1_3
Download citation
DOI: https://doi.org/10.1007/978-3-030-22456-1_3
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-22455-4
Online ISBN: 978-3-030-22456-1
eBook Packages: EngineeringEngineering (R0)