HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

Perturbated Gradients Updating within Unit Space for Deep Learning

Ching-Hsun. Tseng Liu-Hsueh. Cheng Shin-Jye. Lee Xiaojun Zeng

Perturbated Gradients Updating within Unit Space for Deep Learning

Abstract

In deep learning, optimization plays a vital role. By focusing on image classification, this work investigates the pros and cons of the widely used optimizers, and proposes a new optimizer: Perturbated Unit Gradient Descent (PUGD) algorithm with extending normalized gradient operation in tensor within perturbation to update in unit space. Via a set of experiments and analyses, we show that PUGD is locally bounded updating, which means the updating from time to time is controlled. On the other hand, PUGD can push models to a flat minimum, where the error remains approximately constant, not only because of the nature of avoiding stationary points in gradient normalization but also by scanning sharpness in the unit ball. From a series of rigorous experiments, PUGD helps models to gain a state-of-the-art Top-1 accuracy in Tiny ImageNet and competitive performances in CIFAR- {10, 100}. We open-source our code at link: https://github.com/hanktseng131415go/PUGD.

Code Repositories

hanktseng131415go/pugd
Official
pytorch

Benchmarks

BenchmarkMethodologyMetrics
image-classification-on-cifar-10ViT-B/16 (PUGD)
Percentage correct: 99.13
image-classification-on-cifar-100ViT-B/16 (PUGD)
Percentage correct: 93.95
image-classification-on-tiny-imagenet-1ViT-B/16 (PUGD)
Validation Acc: 90.74%
image-classification-on-tiny-imagenet-1DeiT-B/16 (PUGD)
Validation Acc: 91.02%

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp