Datasets:
Dataset Viewer
Model (CoT)
stringlengths 48
89
| TheoremQA
float64 7.8
48.4
| MATH
float64 5
69.2
| GSM
float64 17.4
94.5
| GPQA
float64 20.7
48.9
| MMLU-STEM
float64 29.7
79
|
|---|---|---|---|---|---|
[Mistral-7B-v0.2-base](https://huggingface.co/TIGER-Lab/Mistral-7B-Base-V0.2)
| 19.2
| 10.2
| 36.2
| 24.7
| 50.1
|
[Mixtral-8x7B-base](https://huggingface.co/mistralai/Mixtral-8x7B-v0.1)
| 23.2
| 28.4
| 74.4
| 29.2
| 59.7
|
[Mixtral-8x7B-instruct](https://huggingface.co/mistralai/Mixtral-8x7B-v0.1)
| 25.3
| 22.1
| 71.7
| 32.4
| 61.4
|
[Qwen-1.5-7B-base](https://huggingface.co/Qwen/Qwen1.5-7B)
| 14.2
| 13.3
| 54.1
| 26.7
| 53.7
|
[Qwen-1.5-14B-base](https://huggingface.co/Qwen/Qwen1.5-14B)
| 14
| 25.2
| 61.6
| 35.8
| 64.5
|
[Qwen-1.5-72B-base](https://huggingface.co/Qwen/Qwen1.5-72B)
| 29.3
| 35.1
| 77.6
| 36.3
| 68.5
|
[Yi-6B-base](https://huggingface.co/01-ai/Yi-6B)
| 12
| 5.8
| 32.6
| 20.7
| 46.9
|
[Yi-34B-base](https://huggingface.co/01-ai/Yi-34B)
| 23.2
| 15.9
| 67.9
| 29.7
| 62.6
|
[ChatGLM3-6B-base](https://huggingface.co/THUDM/chatglm3-6b)
| 11.3
| 25.7
| 72.3
| 28.7
| 42.4
|
[Gemma-7B-base](https://huggingface.co/google/gemma-7b)
| 21.5
| 24.3
| 46.4
| 25.7
| 53.3
|
[LLaMA-2-13B-base](https://huggingface.co/meta-llama/Llama-2-13b)
| 10.9
| 5
| 29.6
| 26.2
| 42.9
|
[LLeMMA-7B](https://huggingface.co/EleutherAI/llemma_7b)
| 17.2
| 18
| 36.4
| 23.2
| 45.2
|
[LLeMMA-34B](https://huggingface.co/EleutherAI/llemma_34b)
| 21.1
| 25
| 71.9
| 29.2
| 41.7
|
[InternLM2-7B](https://huggingface.co/internlm/internlm2-7b)
| 7.8
| 20.2
| 70.8
| 22.7
| 34.3
|
[InternLM2-20B](https://huggingface.co/internlm/internlm2-20b)
| 19.5
| 25.5
| 76.1
| 34.3
| 61.2
|
[Deepseek-7B-base](https://huggingface.co/deepseek-ai/deepseek-llm-7b-base)
| 15.7
| 6.4
| 17.4
| 25.7
| 43.1
|
[Deepseek-67B-base](https://huggingface.co/deepseek-ai/deepseek-llm-67b-base)
| 25.3
| 15.9
| 66.5
| 31.8
| 57.4
|
[GPT-4-turbo-0409](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4)
| 48.4
| 69.2
| 94.5
| 46.2
| 76.5
|
[InternLM-Math-7B](https://huggingface.co/internlm/internlm2-math-7b)
| 13.2
| 34.6
| 78.1
| 22.7
| 41.1
|
[InternLM-Math-20B](https://huggingface.co/internlm/internlm2-math-20b)
| 17.1
| 37.7
| 82.9
| 28.9
| 50.1
|
[Deepseek-Math-7B](https://huggingface.co/deepseek-ai/deepseek-math-7b-base)
| 25.3
| 36.2
| 64.2
| 29.7
| 56.4
|
[Deepseek-Math-7B-Instruct](https://huggingface.co/deepseek-ai/deepseek-math-7b-instruct)
| 23.7
| 46.8
| 82.9
| 31.8
| 59.3
|
[WizardMath-7B-1.1](https://huggingface.co/WizardLM/WizardMath-7B-V1.1)
| 11.7
| 33
| 83.2
| 28.7
| 52.7
|
[MetaMath-Mistral-7B](https://huggingface.co/meta-math/MetaMath-Mistral-7B)
| 16.5
| 28.2
| 77.7
| 30.8
| 51.3
|
[Abel-7B-002](https://huggingface.co/GAIR/Abel-7B-002)
| 19.3
| 29.5
| 83.2
| 30.3
| 29.7
|
[OpenMath-Mistral-7B](https://huggingface.co/nvidia/OpenMath-Mistral-7B-v0.1)
| 13.1
| 9.1
| 24.5
| 26.5
| 43.7
|
[Rho-1-Math-7B](https://huggingface.co/microsoft/rho-math-7b-v0.1)
| 21
| 31
| 66.9
| 29.2
| 53.1
|
[LLaMA-3-8B-base](https://huggingface.co/meta-llama/Meta-Llama-3-8B)
| 20.1
| 21.3
| 54.8
| 27.2
| 55.6
|
[LLaMA-3-8B-Instrct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct)
| 22.8
| 30
| 79.5
| 34.5
| 60.2
|
[LLaMA-3-70B-base](https://huggingface.co/meta-llama/Meta-Llama-3-70B)
| 32.3
| 42.5
| 77.6
| 36.3
| 73.7
|
[LLaMA-3-70B-instruct](https://huggingface.co/meta-llama/Meta-Llama-3-70B-Instruct)
| 42
| 47.4
| 92
| 48.9
| 79
|
[Qwen-1.5-110B-base](https://huggingface.co/Qwen/Qwen1.5-110B)
| 34.9
| 49.6
| 85.4
| 35.9
| 73.4
|
[MAmmoTH2-8x7B-base](https://huggingface.co/TIGER-Lab/MAmmoTH2-8x7B)
| 32.2
| 39
| 75.4
| 36.8
| 67.4
|
[MAmmoTH2-8x7B-plus](https://huggingface.co/TIGER-Lab/MAmmoTH2-8x7B-Plus)
| 34.1
| 47
| 86.4
| 37.4
| 72.4
|
This dataset contains the results used for Science Leaderboard
- Downloads last month
- 15