TY - GEN
T1 - Randomness and Interpolation Improve Gradient Descent
T2 - 16th International Conference on Cyber-Enabled Distributed Computing and Knowledge Discover, CyberC 2024
AU - Li, Jiawen
AU - Lefevre, Pascal
AU - PP Abdul Majeed, Anwar
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - Based on Stochastic Gradient Descent (SGD), the paper introduces two optimizers, named Interpolational Accelerating Gradient Descent (IAGD) as well as Noise-Regularized Stochastic Gradient Descent (NRSGD). IAGD leverages second-order Newton Interpolation to expedite the convergence process during training, assuming relevancy in gradients between iterations. To avoid over-fitting, NRSGD incorporates a noise regularization technique that introduces controlled noise to the gradients during the optimization process. Comparative experiments of this research are conducted on the CIFAR-10, and CIFAR-100 datasets, benchmarking different CNNs(Convolutional Neural Networks) with IAGD and NRSGD against classical optimizers in Keras Package. Results demonstrate the potential of those two viable improvement methods in SGD, implicating the effectiveness of the advancements.
AB - Based on Stochastic Gradient Descent (SGD), the paper introduces two optimizers, named Interpolational Accelerating Gradient Descent (IAGD) as well as Noise-Regularized Stochastic Gradient Descent (NRSGD). IAGD leverages second-order Newton Interpolation to expedite the convergence process during training, assuming relevancy in gradients between iterations. To avoid over-fitting, NRSGD incorporates a noise regularization technique that introduces controlled noise to the gradients during the optimization process. Comparative experiments of this research are conducted on the CIFAR-10, and CIFAR-100 datasets, benchmarking different CNNs(Convolutional Neural Networks) with IAGD and NRSGD against classical optimizers in Keras Package. Results demonstrate the potential of those two viable improvement methods in SGD, implicating the effectiveness of the advancements.
KW - Deep Learning
KW - Interpolation
KW - Optimization
KW - Stochastic Gradient Descent(SGD)
UR - http://www.scopus.com/inward/record.url?scp=85215130364&partnerID=8YFLogxK
U2 - 10.1109/CyberC62439.2024.00020
DO - 10.1109/CyberC62439.2024.00020
M3 - Conference Proceeding
AN - SCOPUS:85215130364
T3 - Proceedings - 2024 International Conference on Cyber-Enabled Distributed Computing and Knowledge Discover, CyberC 2024
SP - 56
EP - 59
BT - Proceedings - 2024 International Conference on Cyber-Enabled Distributed Computing and Knowledge Discover, CyberC 2024
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 24 October 2024 through 26 October 2024
ER -