Randomness and Interpolation Improve Gradient Descent: A Simple Exploration in CIFAR datasets

jiawen Li, Pascal LEFEVRE, Anwar PP Majeed

Research output: Chapter in Book or Report/Conference proceedingConference Proceedingpeer-review

Abstract

Based on Stochastic Gradient Descent (SGD), the paper introduces two optimizers, named Interpolational Accelerating Gradient Descent (IAGD) as well as Noise-Regularized Stochastic Gradient Descent (NRSGD). IAGD leverages second-order Newton Interpolation to expedite the convergence process during training, assuming relevancy in gradients between iterations. To avoid over-fitting, NRSGD incorporates a noise regularization technique that introduces controlled noise to the gradients during the optimization process. Comparative experiments of this research are conducted on the CIFAR-10, and CIFAR-100 datasets, benchmarking different CNNs(Convolutional Neural Networks) with IAGD and NRSGD against classical optimizers in Keras Package. Results demonstrate the potential of those two viable improvement methods in SGD, implicating the effectiveness of the advancements.
Original languageEnglish
Title of host publicationRandomness and Interpolation Improve Gradient Descent: A Simple Exploration in CIFAR datasets
PublisherIEEE TCCC CyberC
Publication statusPublished - Sept 2024

Fingerprint

Dive into the research topics of 'Randomness and Interpolation Improve Gradient Descent: A Simple Exploration in CIFAR datasets'. Together they form a unique fingerprint.

Cite this