Local convergence of the alternating least squares algorithm for canonical tensor approximation (Q2910973)
From MaRDI portal
| This is the item page for this Wikibase entity, intended for internal use and editing purposes. Please use this page instead for the normal view: Local convergence of the alternating least squares algorithm for canonical tensor approximation |
scientific article; zbMATH DE number 6081317
| Language | Label | Description | Also known as |
|---|---|---|---|
| English | Local convergence of the alternating least squares algorithm for canonical tensor approximation |
scientific article; zbMATH DE number 6081317 |
Statements
12 September 2012
0 references
low-rank approximation
0 references
nonlinear Gauss-Seidel method, PARAFAC
0 references
local convergence
0 references
alternating least square algorithm
0 references
tensor
0 references
global minimization
0 references
Lagrange multipliers
0 references
Local convergence of the alternating least squares algorithm for canonical tensor approximation (English)
0 references
The author presents a local convergence result for the alternating least square (ALS) algorithm under the assumption that the Hessian of the loss function at the solution is essentially positive definite, except on a trivial null space caused by the scaling indeterminacy. This assumption necessarily requires the local essential uniqueness of the CP decomposition (that is, uniqueness up to scaling), which is reasonable for tensors of order higher than three. The proof shows that an inbuilt normalization procedure in the ALS algorithm acts (in first order) as a projection onto a subspace complementary to the null space of the Hessian. On this subspace the linear Gauss-Seidel method is known to be contractive.NEWLINENEWLINEThe main ideas are not specifically related to the least squares error as the loss function to be minimized. It turns out that the global minimization in one ALS direction can be replaced by a single Newton step to obtain an approximate scheme which is still convergent (under the same assumptions). One convenient aspect of this approach is that it avoids the explicit use of Lagrange multipliers. It is therefore easily accessible and, in principle, applicable to more delicate types of redundancies as they, for instance, occur in the Tucker format or in the newly developed TT format.
0 references