Question Answering On Squad20
评估指标
EM
F1
评测结果
各个模型在此基准测试上的表现结果
模型名称 | EM | F1 | Paper Title | Repository |
---|---|---|---|---|
BISAN-CC (single model) | 80.208 | 83.149 | - | - |
bert (single model) | 79.971 | 83.184 | - | - |
PMI-Masking Random Baseline (single model) | 80.038 | 82.796 | - | - |
PwP+BERT (single model) | 80.117 | 83.189 | - | - |
Tuned BERT Large Cased (single model) | 82.803 | 85.863 | - | - |
BERT-Base-DT (single model) | 74.769 | 77.706 | - | - |
ELECTRA+RL+EV (single model) | 89.021 | 91.765 | - | - |
ALBERT (single model) | 88.107 | 90.902 | ALBERT: A Lite BERT for Self-supervised Learning of Language Representations | |
PMI-Masking Additional Data Random Baseline (single model) | 80.377 | 83.262 | - | - |
XLNET-123 (single model) | 86.436 | 89.086 | - | - |
mgrc | 75.344 | 78.381 | - | - |
SemBERT (single model) | 84.800 | 87.864 | - | - |
batch2 (single model) | 73.742 | 76.858 | - | - |
Candi-Net+BERT (single model) | 80.388 | 82.908 | - | - |
Fusion Adapters TriviaQA NQ Singl | 78.933 | 81.863 | - | - |
BERT+AC(single model) | 78.052 | 81.174 | - | - |
electra+nlayers+kdav (ensemble) | 90.002 | 92.497 | - | - |
BERT + ConvLSTM + MTL + Verifier (single model) | 84.924 | 88.204 | - | - |
Ensemble ALBERT | - | 90.123 | Ensemble ALBERT on SQuAD 2.0 | |
L6Net + BERT (single model) | 79.181 | 82.259 | - | - |
0 of 286 row(s) selected.