TY - JOUR
T1 - LightAdam
T2 - Towards a Fast and Accurate Adaptive Momentum Online Algorithm
AU - Zhou, Yangfan
AU - Huang, Kaizhu
AU - Cheng, Cheng
AU - Wang, Xuguang
AU - Liu, Xin
N1 - Funding Information:
This work was partially supported by Chinese Academy of Sciences under grant No. Y9BEJ11001 and the innovation workstation of Suzhou Institute of Nano-Tech and Nano-Bionics (SINANO) under grant No. E010210101. This work was also partially supported by National Natural Science Foundation of China under no.61876155 and Jiangsu Science and Technology Programme under no. BE2020006-4.
Publisher Copyright:
© 2022, The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature.
PY - 2022/3
Y1 - 2022/3
N2 - Adaptive optimization algorithms enjoy fast convergence and have been widely exploited in pattern recognition and cognitively-inspired machine learning. These algorithms may however be of high computational cost and low generalization ability due to their projection steps. Such limitations make them difficult to be applied in big data analytics, which may typically be seen in cognitively inspired learning, e.g. deep learning tasks. In this paper, we propose a fast and accurate adaptive momentum online algorithm, called LightAdam, to alleviate the drawbacks of projection steps for the adaptive algorithms. The proposed algorithm substantially reduces computational cost for each iteration step by replacing high-order projection operators with one-dimensional linear searches. Moreover, we introduce a novel second-order momentum and engage dynamic learning rate bounds in the proposed algorithm, thereby obtaining a higher generalization ability than other adaptive algorithms. We theoretically analyze that our proposed algorithm has a guaranteed convergence bound, and prove that our proposed algorithm has better generalization capability as compared to Adam.
AB - Adaptive optimization algorithms enjoy fast convergence and have been widely exploited in pattern recognition and cognitively-inspired machine learning. These algorithms may however be of high computational cost and low generalization ability due to their projection steps. Such limitations make them difficult to be applied in big data analytics, which may typically be seen in cognitively inspired learning, e.g. deep learning tasks. In this paper, we propose a fast and accurate adaptive momentum online algorithm, called LightAdam, to alleviate the drawbacks of projection steps for the adaptive algorithms. The proposed algorithm substantially reduces computational cost for each iteration step by replacing high-order projection operators with one-dimensional linear searches. Moreover, we introduce a novel second-order momentum and engage dynamic learning rate bounds in the proposed algorithm, thereby obtaining a higher generalization ability than other adaptive algorithms. We theoretically analyze that our proposed algorithm has a guaranteed convergence bound, and prove that our proposed algorithm has better generalization capability as compared to Adam.
KW - Adaptive training algorithm
KW - Convex optimization
KW - Online learning
KW - Projection-free
UR - http://www.scopus.com/inward/record.url?scp=85122690180&partnerID=8YFLogxK
U2 - 10.1007/s12559-021-09985-9
DO - 10.1007/s12559-021-09985-9
M3 - Article
AN - SCOPUS:85122690180
SN - 1866-9956
VL - 14
SP - 764
EP - 779
JO - Cognitive Computation
JF - Cognitive Computation
IS - 2
ER -