HyperAI超神经

Question Answering On Danetqa

评估指标

Accuracy

评测结果

各个模型在此基准测试上的表现结果

模型名称
Accuracy
Paper TitleRepository
Human Benchmark0.915RussianSuperGLUE: A Russian Language Understanding Evaluation Benchmark
Multilingual Bert0.624--
ruRoberta-large finetune0.82--
MT5 Large0.657mT5: A massively multilingual pre-trained text-to-text transformer
majority_class0.503Unreasonable Effectiveness of Rule-Based Heuristics in Solving Russian SuperGLUE Tasks-
RuGPT3Medium0.634--
RuBERT plain0.639--
RuGPT3Small0.61--
ruBert-large finetune0.773--
Baseline TF-IDF1.10.621RussianSuperGLUE: A Russian Language Understanding Evaluation Benchmark
YaLM 1.0B few-shot0.637--
ruT5-large-finetune0.711--
Random weighted0.52Unreasonable Effectiveness of Rule-Based Heuristics in Solving Russian SuperGLUE Tasks-
RuGPT3XL few-shot0.59--
RuBERT conversational0.606--
SBERT_Large_mt_ru_finetuning0.697--
Golden Transformer0.917--
RuGPT3Large0.604--
SBERT_Large0.675--
ruT5-base-finetune0.732--
0 of 22 row(s) selected.