Randomness and Interpolation Improve Gradient Descent: A Simple Exploration in CIFAR Datasets

Jiawen Li*, Pascal Lefevre, Anwar PP Abdul Majeed

*Corresponding author for this work

Research output: Chapter in Book or Report/Conference proceedingConference Proceedingpeer-review

Abstract

Based on Stochastic Gradient Descent (SGD), the paper introduces two optimizers, named Interpolational Accelerating Gradient Descent (IAGD) as well as Noise-Regularized Stochastic Gradient Descent (NRSGD). IAGD leverages second-order Newton Interpolation to expedite the convergence process during training, assuming relevancy in gradients between iterations. To avoid over-fitting, NRSGD incorporates a noise regularization technique that introduces controlled noise to the gradients during the optimization process. Comparative experiments of this research are conducted on the CIFAR-10, and CIFAR-100 datasets, benchmarking different CNNs(Convolutional Neural Networks) with IAGD and NRSGD against classical optimizers in Keras Package. Results demonstrate the potential of those two viable improvement methods in SGD, implicating the effectiveness of the advancements.

Original languageEnglish
Title of host publicationProceedings - 2024 International Conference on Cyber-Enabled Distributed Computing and Knowledge Discover, CyberC 2024
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages56-59
Number of pages4
ISBN (Electronic)9798331506896
DOIs
Publication statusPublished - 2024
Event16th International Conference on Cyber-Enabled Distributed Computing and Knowledge Discover, CyberC 2024 - Guangzhou, China
Duration: 24 Oct 202426 Oct 2024

Publication series

NameProceedings - 2024 International Conference on Cyber-Enabled Distributed Computing and Knowledge Discover, CyberC 2024

Conference

Conference16th International Conference on Cyber-Enabled Distributed Computing and Knowledge Discover, CyberC 2024
Country/TerritoryChina
CityGuangzhou
Period24/10/2426/10/24

Keywords

  • Deep Learning
  • Interpolation
  • Optimization
  • Stochastic Gradient Descent(SGD)

Fingerprint

Dive into the research topics of 'Randomness and Interpolation Improve Gradient Descent: A Simple Exploration in CIFAR Datasets'. Together they form a unique fingerprint.

Cite this