HyperAI
HyperAI超神经
首页
算力平台
文档
资讯
论文
教程
数据集
百科
SOTA
LLM 模型天梯
GPU 天梯
顶会
开源项目
全站搜索
关于
中文
HyperAI
HyperAI超神经
Toggle sidebar
全站搜索…
⌘
K
Command Palette
Search for a command to run...
首页
SOTA
问答
Question Answering On Strategyqa
Question Answering On Strategyqa
评估指标
Accuracy
评测结果
各个模型在此基准测试上的表现结果
Columns
模型名称
Accuracy
Paper Title
Repository
PaLM 2 (few-shot, CoT, SC)
90.4
PaLM 2 Technical Report
Rethinking with retrieval (GPT-3)
77.73
Rethinking with Retrieval: Faithful Large Language Model Inference
Self-Evaluation Guided Decoding (Codex, CoT, single reasoning chain, 6-shot gen, 4-shot eval)
77.2
-
-
U-PaLM 540B
76.6
Transcending Scaling Laws with 0.1% Extra Compute
-
PaLM 540B
76.4
Transcending Scaling Laws with 0.1% Extra Compute
-
Minerva 540B
61.9
Transcending Scaling Laws with 0.1% Extra Compute
-
SearchChain
-
Chain-of-Action: Faithful and Multimodal Question Answering through Large Language Models
Least-to-Most
-
Least-to-Most Prompting Enables Complex Reasoning in Large Language Models
SearchChain
-
Search-in-the-Chain: Interactively Enhancing Large Language Models with Search for Knowledge-intensive Tasks
CoA w/o actions
-
Chain-of-Action: Faithful and Multimodal Question Answering through Large Language Models
CoA
-
Chain-of-Action: Faithful and Multimodal Question Answering through Large Language Models
Least-to-Most
-
Chain-of-Action: Faithful and Multimodal Question Answering through Large Language Models
0 of 12 row(s) selected.
Previous
Next
Question Answering On Strategyqa | SOTA | HyperAI超神经