Efficient Low-Tubal-Rank Tensor Estimation via Alternating Preconditioned Gradient Descent
By: Zhiyu Liu , Zhi Han , Yandong Tang and more
Potential Business Impact:
Makes computer math problems solve much faster.
The problem of low-tubal-rank tensor estimation is a fundamental task with wide applications across high-dimensional signal processing, machine learning, and image science. Traditional approaches tackle such a problem by performing tensor singular value decomposition, which is computationally expensive and becomes infeasible for large-scale tensors. Recent approaches address this issue by factorizing the tensor into two smaller factor tensors and solving the resulting problem using gradient descent. However, this kind of approach requires an accurate estimate of the tensor rank, and when the rank is overestimated, the convergence of gradient descent and its variants slows down significantly or even diverges. To address this problem, we propose an Alternating Preconditioned Gradient Descent (APGD) algorithm, which accelerates convergence in the over-parameterized setting by adding a preconditioning term to the original gradient and updating these two factors alternately. Based on certain geometric assumptions on the objective function, we establish linear convergence guarantees for more general low-tubal-rank tensor estimation problems. Then we further analyze the specific cases of low-tubal-rank tensor factorization and low-tubal-rank tensor recovery. Our theoretical results show that APGD achieves linear convergence even under over-parameterization, and the convergence rate is independent of the tensor condition number. Extensive simulations on synthetic data are carried out to validate our theoretical assertions.
Similar Papers
Guaranteed Nonconvex Low-Rank Tensor Estimation via Scaled Gradient Descent
Machine Learning (Stat)
Cleans messy data to find hidden patterns faster.
Dynamic Low-rank Approximation of Full-Matrix Preconditioner for Training Generalized Linear Models
Machine Learning (CS)
Makes computer learning faster with smarter math.
Guaranteed Noisy CP Tensor Recovery via Riemannian Optimization on the Segre Manifold
Machine Learning (Stat)
Finds hidden patterns in messy data.