HyperAIHyperAI

Command Palette

Search for a command to run...

5 months ago

IA-RED$^2$: Interpretability-Aware Redundancy Reduction for Vision Transformers

Pan Bowen ; Panda Rameswar ; Jiang Yifan ; Wang Zhangyang ; Feris Rogerio ; Oliva Aude

IA-RED$^2$: Interpretability-Aware Redundancy Reduction for Vision
  Transformers

Abstract

The self-attention-based model, transformer, is recently becoming the leadingbackbone in the field of computer vision. In spite of the impressive successmade by transformers in a variety of vision tasks, it still suffers from heavycomputation and intensive memory costs. To address this limitation, this paperpresents an Interpretability-Aware REDundancy REDuction framework (IA-RED$^2$).We start by observing a large amount of redundant computation, mainly spent onuncorrelated input patches, and then introduce an interpretable module todynamically and gracefully drop these redundant patches. This novel frameworkis then extended to a hierarchical structure, where uncorrelated tokens atdifferent stages are gradually removed, resulting in a considerable shrinkageof computational cost. We include extensive experiments on both image and videotasks, where our method could deliver up to 1.4x speed-up for state-of-the-artmodels like DeiT and TimeSformer, by only sacrificing less than 0.7% accuracy.More importantly, contrary to other acceleration approaches, our method isinherently interpretable with substantial visual evidence, making visiontransformer closer to a more human-understandable architecture while beinglighter. We demonstrate that the interpretability that naturally emerged in ourframework can outperform the raw attention learned by the original visualtransformer, as well as those generated by off-the-shelf interpretationmethods, with both qualitative and quantitative results. Project Page:http://people.csail.mit.edu/bpan/ia-red/.

Benchmarks

BenchmarkMethodologyMetrics
efficient-vits-on-imagenet-1k-with-deit-sIA-RED$^2$
GFLOPs: 3.2
Top 1 Accuracy: 79.1

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
IA-RED$^2$: Interpretability-Aware Redundancy Reduction for Vision Transformers | Papers | HyperAI