ACMo: Angle-Calibrated Moment Methods for Stochastic Optimization

Xunpeng Huang, Runxin Xu*, Hao Zhou, Zhe Wang*, Zhengyang Liu, Lei Li

*此作品的通讯作者

科研成果: 书/报告/会议事项章节会议稿件同行评审

摘要

Stochastic gradient descent (SGD) is a widely used method for its outstanding generalization ability and simplicity. Adaptive gradient methods have been proposed to further accelerate the optimization process. In this paper, we revisit existing adaptive gradient optimization methods with a new interpretation. Such new perspective leads to a refreshed understanding of the roles of second moments in stochastic optimization. Based on this, we propose Angle-Calibration Moment method (ACMo), a novel stochastic optimization method. It enjoys the benefits of second moments with only first moment updates. Theoretical analysis shows that ACMo is able to achieve the same convergence rate as mainstream adaptive methods. Experiments on a variety of CV and NLP tasks demonstrate that ACMo has a comparable convergence to state-of-the-art Adam-type optimizers, and even a better generalization performance in most cases. The code is available at https://github.com/Xunpeng746/ACMo.

源语言英语
主期刊名35th AAAI Conference on Artificial Intelligence, AAAI 2021
出版商Association for the Advancement of Artificial Intelligence
7857-7864
页数8
ISBN(电子版)9781713835974
出版状态已出版 - 2021
活动35th AAAI Conference on Artificial Intelligence, AAAI 2021 - Virtual, Online
期限: 2 2月 20219 2月 2021

出版系列

姓名35th AAAI Conference on Artificial Intelligence, AAAI 2021
9A

会议

会议35th AAAI Conference on Artificial Intelligence, AAAI 2021
Virtual, Online
时期2/02/219/02/21

指纹

探究 'ACMo: Angle-Calibrated Moment Methods for Stochastic Optimization' 的科研主题。它们共同构成独一无二的指纹。

引用此