Cross Lingual Question Answering On Tydiqa
评估指标
EM
评测结果
各个模型在此基准测试上的表现结果
| Paper Title | Repository | ||
|---|---|---|---|
| ByT5 (fine-tuned) | 81.9 | ByT5: Towards a token-free future with pre-trained byte-to-byte models | |
| U-PaLM 62B (fine-tuned) | 78.4 | Transcending Scaling Laws with 0.1% Extra Compute | - |
| Flan-U-PaLM 540B (direct-prompting) | 68.3 | Scaling Instruction-Finetuned Language Models | |
| Flan-PaLM 540B (direct-prompting) | 67.8 | Scaling Instruction-Finetuned Language Models | |
| ByT5 XXL | 60.0 | ByT5: Towards a token-free future with pre-trained byte-to-byte models | |
| U-PaLM-540B (CoT) | 54.6 | Transcending Scaling Laws with 0.1% Extra Compute | - |
| PaLM-540B (CoT) | 52.9 | PaLM: Scaling Language Modeling with Pathways | |
| Decoupled | 42.8 | Rethinking embedding coupling in pre-trained language models | |
| PaLM 2-M (one-shot) | - | PaLM 2 Technical Report | |
| PaLM 2-S (one-shot) | - | PaLM 2 Technical Report | |
| PaLM 2-L (one-shot) | - | PaLM 2 Technical Report |
0 of 11 row(s) selected.