HyperAI超神经

Domain Generalization On Imagenet R

评估指标

Top-1 Error Rate

评测结果

各个模型在此基准测试上的表现结果

模型名称
Top-1 Error Rate
Paper TitleRepository
Model soups (ViT-G/14)4.54Model soups: averaging weights of multiple fine-tuned models improves accuracy without increasing inference time
GPaCo (ViT-L)39.7Generalized Parametric Contrastive Learning
SEER (RegNet10B)43.9Vision Models Are More Robust And Fair When Pretrained On Uncurated Images Without Supervision
Pyramid Adversarial Training Improves ViT (Im21k)42.16Pyramid Adversarial Training Improves ViT Performance
VOLO-D5+HAT40.3Improving Vision Transformers by Revisiting High-frequency Components
Stylized ImageNet (ResNet-50)58.5 ImageNet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness
Mixer-B/8-SAM76.5When Vision Transformers Outperform ResNets without Pre-training or Strong Data Augmentations
LLE (ViT-B/16, SWAG, Edge Aug)31.3A Whac-A-Mole Dilemma: Shortcuts Come in Multiples Where Mitigating One Amplifies Others
Model soups (BASIC-L)3.90Model soups: averaging weights of multiple fine-tuned models improves accuracy without increasing inference time
MAE (ViT-H, 448)33.5Masked Autoencoders Are Scalable Vision Learners
PRIME with JSD (ResNet-50)53.7PRIME: A few primitives can boost robustness to common corruptions
ConvFormer-B3648.9MetaFormer Baselines for Vision
CAR-FT (CLIP, ViT-L/14@336px)10.3Context-Aware Robust Fine-Tuning-
ConvFormer-B36 (384)47.8MetaFormer Baselines for Vision
ConvNeXt-XL (Im21k, 384)31.8A ConvNet for the 2020s
ResNet-152x2-SAM71.9When Vision Transformers Outperform ResNets without Pre-training or Strong Data Augmentations
FAN-Hybrid-L(IN-21K, 384))28.9Understanding The Robustness in Vision Transformers
RVT-Ti*56.1Towards Robust Vision Transformer
RVT-S*52.3Towards Robust Vision Transformer
ConvFormer-B36 (IN21K, 384)33.5MetaFormer Baselines for Vision
0 of 39 row(s) selected.