HyperAIHyperAI

Command Palette

Search for a command to run...

4 months ago

Regularizing cross entropy loss via minimum entropy and K-L divergence

Ibraheem Abdulrahman Oladipupo

Regularizing cross entropy loss via minimum entropy and K-L divergence

Abstract

I introduce two novel loss functions for classification in deep learning. Thetwo loss functions extend standard cross entropy loss by regularizing it withminimum entropy and Kullback-Leibler (K-L) divergence terms. The first of thetwo novel loss functions is termed mixed entropy loss (MIX-ENT for short),while the second one is termed minimum entropy regularized cross-entropy loss(MIN-ENT for short). The MIX-ENT function introduces a regularizer that can beshown to be equivalent to the sum of a minimum entropy term and a K-Ldivergence term. However, it should be noted that the K-L divergence term hereis different from that in the standard cross-entropy loss function, in thesense that it swaps the roles of the target probability and the hypothesisprobability. The MIN-ENT function simply adds a minimum entropy regularizer tothe standard cross entropy loss function. In both MIX-ENT and MIN-ENT, theminimum entropy regularizer minimizes the entropy of the hypothesis probabilitydistribution which is output by the neural network. Experiments on theEMNIST-Letters dataset shows that my implementation of MIX-ENT and MIN-ENT letsthe VGG model climb from its previous 3rd position on the paperswithcodeleaderboard to reach the 2nd position on the leaderboard, outperforming theSpinal-VGG model in so doing. Specifically, using standard cross-entropy, VGGachieves 95.86% while Spinal-VGG achieves 95.88% classification accuracies,whereas using VGG (without Spinal-VGG) our MIN-ENT achieved 95.933%, while ourMIX-ENT achieved 95.927% accuracies. The pre-trained models for both MIX-ENTand MIN-ENT are at https://github.com/rahmanoladi/minimum entropy project.

Code Repositories

Benchmarks

BenchmarkMethodologyMetrics
image-classification-on-emnist-lettersMIX-ENT + VGG-5
Accuracy: 95.928
image-classification-on-emnist-lettersMIN-ENT + VGG-5
Accuracy: 95.933

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
Regularizing cross entropy loss via minimum entropy and K-L divergence | Papers | HyperAI