HyperAI
HyperAI超神经
首页
算力平台
文档
资讯
论文
教程
数据集
百科
SOTA
LLM 模型天梯
GPU 天梯
顶会
开源项目
全站搜索
关于
中文
HyperAI
HyperAI超神经
Toggle sidebar
全站搜索…
⌘
K
Command Palette
Search for a command to run...
首页
SOTA
常识推理
Common Sense Reasoning On Commonsenseqa
Common Sense Reasoning On Commonsenseqa
评估指标
Accuracy
评测结果
各个模型在此基准测试上的表现结果
Columns
模型名称
Accuracy
Paper Title
Repository
GPT-4o (HPT)
92.54
Hierarchical Prompting Taxonomy: A Universal Evaluation Framework for Large Language Models
DeBERTaV3-large+KEAR
91.2
Human Parity on CommonsenseQA: Augmenting Self-Attention with External Attention
PaLM 2 (few‑shot, CoT, SC)
90.4
PaLM 2 Technical Report
KEAR
89.4
Human Parity on CommonsenseQA: Augmenting Self-Attention with External Attention
DEKCOR
83.3
Fusing Context Into Knowledge Graph for Commonsense Question Answering
Unicorn 11B (fine-tuned)
79.3
UNICORN on RAINBOW: A Universal Commonsense Reasoning Model on a New Multitask Benchmark
MUPPET Roberta Large
79.2
Muppet: Massive Multi-task Representations with Pre-Finetuning
UnifiedQA 11B (fine-tuned)
79.1
UnifiedQA: Crossing Format Boundaries With a Single QA System
DRAGON
78.2
Deep Bidirectional Language-Knowledge Graph Pretraining
T5-XXL 11B (fine-tuned)
78.1
UnifiedQA: Crossing Format Boundaries With a Single QA System
Albert Lan et al. (2020) (ensemble)
76.5
ALBERT: A Lite BERT for Self-supervised Learning of Language Representations
UnifiedQA 11B (zero-shot)
76.2
UnifiedQA: Crossing Format Boundaries With a Single QA System
QA-GNN
76.1
QA-GNN: Reasoning with Language Models and Knowledge Graphs for Question Answering
XLNet+GraphReason
75.3
Graph-Based Reasoning over Heterogeneous External Knowledge for Commonsense Question Answering
GrapeQA: PEGA
73.5
GrapeQA: GRaph Augmentation and Pruning to Enhance Question-Answering
-
RoBERTa+HyKAS Ma et al. (2019)
73.2
Towards Generalizable Neuro-Symbolic Systems for Commonsense Question Answering
-
GPT-3 Direct Finetuned
73.0
Human Parity on CommonsenseQA: Augmenting Self-Attention with External Attention
STaR (on GPT-J)
72.3
STaR: Bootstrapping Reasoning With Reasoning
RoBERTa-Large 355M
72.1
RoBERTa: A Robustly Optimized BERT Pretraining Approach
STaR without Rationalization (on GPT-J)
68.8
STaR: Bootstrapping Reasoning With Reasoning
0 of 38 row(s) selected.
Previous
Next
Common Sense Reasoning On Commonsenseqa | SOTA | HyperAI超神经