Tensor train decomposition for solving large-scale linear equations

Hengnu Chen, Lei Deng, Zheng Qu, Ling Liang, Tianyi Yan, Yuan Xie, Guoqi Li*

*此作品的通讯作者

科研成果: 期刊稿件文章同行评审

2 引用 (Scopus)

摘要

Solving large-scale linear equations is an important problem in signal processing, machine vision, financial mathematics, quantum physics, chemistry, and many other areas. Due to the curse of dimensionality, classic numerical treatments of such problems are difficult and inefficient, which are usually addressed by low rank approximation methods. Tensor train decomposition (TTD) is one of these methods that can reduce the execution cost theoretically. In this work, we present a TTD based batch alternating least squares (BALS) method, termed as TTD-BALS, for solving large-scale linear equations. TTD-BALS makes the computation efficient and stable, since it can convert the large-scale optimization problem into sequential optimization subproblems with smaller scale. We further uncover that all existing TTD-based methods together with our TTD-BALS are not universal and can be troubled by local minima, and therefore they are not suitable for general and arbitrary linear equations. Fortunately, based on our analysis, if some specific requirements we identified can be satisfied, TTD-BALS is able to provide a very good solution. In fact these requirements can be satisfied in a wide range of domains such as signal processing and scientific computing. Furthermore, we prove that the proposed method is able to reduce the computational complexity, and conduct numerical experiments to confirm its effectiveness.

源语言英语
页(从-至)203-217
页数15
期刊Neurocomputing
464
DOI
出版状态已出版 - 13 11月 2021

指纹

探究 'Tensor train decomposition for solving large-scale linear equations' 的科研主题。它们共同构成独一无二的指纹。

引用此

Chen, H., Deng, L., Qu, Z., Liang, L., Yan, T., Xie, Y., & Li, G. (2021). Tensor train decomposition for solving large-scale linear equations. Neurocomputing, 464, 203-217. https://doi.org/10.1016/j.neucom.2021.08.034