AdaNorm: Adaptive Gradient Norm Correction based Optimizer for CNNs

October 12, 2022 ยท Entered Twilight ยท ๐Ÿ› IEEE Workshop/Winter Conference on Applications of Computer Vision

๐Ÿ’ค TWILIGHT: Eternal Rest
Repo abandoned since publication

Repo contents: CIFAR10, CIFAR100, README.md, TinyImageNet

Authors Shiv Ram Dubey, Satish Kumar Singh, Bidyut Baran Chaudhuri arXiv ID 2210.06364 Category cs.CV: Computer Vision Citations 11 Venue IEEE Workshop/Winter Conference on Applications of Computer Vision Repository https://github.com/shivram1987/AdaNorm โญ 11 Last Checked 1 month ago
Abstract
The stochastic gradient descent (SGD) optimizers are generally used to train the convolutional neural networks (CNNs). In recent years, several adaptive momentum based SGD optimizers have been introduced, such as Adam, diffGrad, Radam and AdaBelief. However, the existing SGD optimizers do not exploit the gradient norm of past iterations and lead to poor convergence and performance. In this paper, we propose a novel AdaNorm based SGD optimizers by correcting the norm of gradient in each iteration based on the adaptive training history of gradient norm. By doing so, the proposed optimizers are able to maintain high and representive gradient throughout the training and solves the low and atypical gradient problems. The proposed concept is generic and can be used with any existing SGD optimizer. We show the efficacy of the proposed AdaNorm with four state-of-the-art optimizers, including Adam, diffGrad, Radam and AdaBelief. We depict the performance improvement due to the proposed optimizers using three CNN models, including VGG16, ResNet18 and ResNet50, on three benchmark object recognition datasets, including CIFAR10, CIFAR100 and TinyImageNet. Code: https://github.com/shivram1987/AdaNorm.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Computer Vision