Common Sense Reasoning On Arc Easy

评估指标

Accuracy

评测结果

各个模型在此基准测试上的表现结果

Paper TitleRepository
ST-MoE-32B 269B (fine-tuned)95.2ST-MoE: Designing Stable and Transferable Sparse Expert Models
LLaMA 3 8B+MoSLoRA (fine-tuned)90.5Mixture-of-Subspaces in Low-Rank Adaptation
PaLM 2-L (1-shot)89.7PaLM 2 Technical Report
PaLM 2-M (1-shot)88.0PaLM 2 Technical Report
LLaMA-3 8B + MixLoRA86.5MixLoRA: Enhancing Large Language Models Fine-Tuning with LoRA-based Mixture of Experts
Camelidae-8×34B86.2Parameter-Efficient Sparsity Crafting from Dense to Mixture-of-Experts for Instruction Tuning on General Tasks
PaLM 2-S (1-shot)85.6PaLM 2 Technical Report
LLaMA 65B + CFG (0-shot)84.2Stay on topic with Classifier-Free Guidance-
GAL 120B (0-shot)83.8Galactica: A Large Language Model for Science
LLaMA-2 13B + MixLoRA83.5MixLoRA: Enhancing Large Language Models Fine-Tuning with LoRA-based Mixture of Experts
LLaMA 30B + CFG (0-shot)83.2Stay on topic with Classifier-Free Guidance-
Mixtral 8x7B (0-shot)83.1Mixtral of Experts
FLAN 137B (few-shot, k=14)80.7Finetuned Language Models Are Zero-Shot Learners
Mistral 7B (0-shot)80.5Mixtral of Experts
LLaMA 33B (0-shot)80.0LLaMA: Open and Efficient Foundation Language Models
Mistral 7B (0-shot)80.0Mistral 7B
FLAN 137B (0-shot)79.6Finetuned Language Models Are Zero-Shot Learners
LLaMA 13B + CFG (0-shot)79.1Stay on topic with Classifier-Free Guidance-
LLaMA 65B (0-shot)78.9LLaMA: Open and Efficient Foundation Language Models
LLaMA-2 7B + MixLoRA77.7MixLoRA: Enhancing Large Language Models Fine-Tuning with LoRA-based Mixture of Experts
0 of 47 row(s) selected.
Common Sense Reasoning On Arc Easy | SOTA | HyperAI超神经