HyperAI
HyperAI超神经
首页
算力平台
文档
资讯
论文
教程
数据集
百科
SOTA
LLM 模型天梯
GPU 天梯
顶会
开源项目
全站搜索
关于
中文
HyperAI
HyperAI超神经
Toggle sidebar
全站搜索…
⌘
K
Command Palette
Search for a command to run...
首页
SOTA
自然语言推理
Natural Language Inference On Wnli
Natural Language Inference On Wnli
评估指标
Accuracy
评测结果
各个模型在此基准测试上的表现结果
Columns
模型名称
Accuracy
Paper Title
Repository
Turing NLR v5 XXL 5.4B (fine-tuned)
95.9
-
-
DeBERTa
94.5
DeBERTa: Decoding-enhanced BERT with Disentangled Attention
T5-XXL 11B
93.2
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
XLNet
92.5
XLNet: Generalized Autoregressive Pretraining for Language Understanding
ALBERT
91.8
ALBERT: A Lite BERT for Self-supervised Learning of Language Representations
StructBERTRoBERTa ensemble
89.7
StructBERT: Incorporating Language Structures into Pre-training for Deep Language Understanding
-
T5-XL 3B
89.7
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
HNNensemble
89
A Hybrid Neural Network Model for Commonsense Reasoning
RoBERTa (ensemble)
89
RoBERTa: A Robustly Optimized BERT Pretraining Approach
T5-Large 770M
85.6
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
HNN
83.6
A Hybrid Neural Network Model for Commonsense Reasoning
T5-Base 220M
78.8
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
BERTwiki 340M (fine-tuned on WSCR)
74.7
A Surprisingly Robust Trick for Winograd Schema Challenge
FLAN 137B (zero-shot)
74.6
Finetuned Language Models Are Zero-Shot Learners
BERT-large 340M (fine-tuned on WSCR)
71.9
A Surprisingly Robust Trick for Winograd Schema Challenge
BERT-base 110M (fine-tuned on WSCR)
70.5
A Surprisingly Robust Trick for Winograd Schema Challenge
FLAN 137B (few-shot, k=4)
70.4
Finetuned Language Models Are Zero-Shot Learners
T5-Small 60M
69.2
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
ERNIE 2.0 Large
67.8
ERNIE 2.0: A Continual Pre-training Framework for Language Understanding
SqueezeBERT
65.1
SqueezeBERT: What can computer vision teach NLP about efficient neural networks?
0 of 23 row(s) selected.
Previous
Next
Natural Language Inference On Wnli | SOTA | HyperAI超神经