TY - JOUR
T1 - Large-scale Riemannian meta-optimization via subspace adaptation
AU - Yu, Peilin
AU - Wu, Yuwei
AU - Gao, Zhi
AU - Fan, Xiaomeng
AU - Jia, Yunde
N1 - Publisher Copyright:
© 2025 Elsevier Inc.
PY - 2025/3
Y1 - 2025/3
N2 - Riemannian meta-optimization provides a promising approach to solving non-linear constrained optimization problems, which trains neural networks as optimizers to perform optimization on Riemannian manifolds. However, existing Riemannian meta-optimization methods take up huge memory footprints in large-scale optimization settings, as the learned optimizer can only adapt gradients of a fixed size and thus cannot be shared across different Riemannian parameters. In this paper, we propose an efficient Riemannian meta-optimization method that significantly reduces the memory burden for large-scale optimization via a subspace adaptation scheme. Our method trains neural networks to individually adapt the row and column subspaces of Riemannian gradients, instead of directly adapting the full gradient matrices in existing Riemannian meta-optimization methods. In this case, our learned optimizer can be shared across Riemannian parameters with different sizes. Our method reduces the model memory consumption by six orders of magnitude when optimizing an orthogonal mainstream deep neural network (e.g. ResNet50). Experiments on multiple Riemannian tasks show that our method can not only reduce the memory consumption but also improve the performance of Riemannian meta-optimization.
AB - Riemannian meta-optimization provides a promising approach to solving non-linear constrained optimization problems, which trains neural networks as optimizers to perform optimization on Riemannian manifolds. However, existing Riemannian meta-optimization methods take up huge memory footprints in large-scale optimization settings, as the learned optimizer can only adapt gradients of a fixed size and thus cannot be shared across different Riemannian parameters. In this paper, we propose an efficient Riemannian meta-optimization method that significantly reduces the memory burden for large-scale optimization via a subspace adaptation scheme. Our method trains neural networks to individually adapt the row and column subspaces of Riemannian gradients, instead of directly adapting the full gradient matrices in existing Riemannian meta-optimization methods. In this case, our learned optimizer can be shared across Riemannian parameters with different sizes. Our method reduces the model memory consumption by six orders of magnitude when optimizing an orthogonal mainstream deep neural network (e.g. ResNet50). Experiments on multiple Riemannian tasks show that our method can not only reduce the memory consumption but also improve the performance of Riemannian meta-optimization.
KW - Large-scale optimization
KW - Riemannian manifolds
KW - Riemannian meta-optimization
KW - Subspace adaptation
UR - http://www.scopus.com/inward/record.url?scp=85216846906&partnerID=8YFLogxK
U2 - 10.1016/j.cviu.2025.104306
DO - 10.1016/j.cviu.2025.104306
M3 - Article
AN - SCOPUS:85216846906
SN - 1077-3142
VL - 253
JO - Computer Vision and Image Understanding
JF - Computer Vision and Image Understanding
M1 - 104306
ER -