Adding Evaluation Results
Browse filesThis is an automated PR created with https://huggingface.co/spaces/Weyaxi/open-llm-leaderboard-results-pr
The purpose of this PR is to add evaluation results from the Open LLM Leaderboard to your model card.
If you encounter any issues, please report them to https://huggingface.co/spaces/Weyaxi/open-llm-leaderboard-results-pr/discussions
README.md
CHANGED
|
@@ -1,5 +1,108 @@
|
|
| 1 |
---
|
| 2 |
license: mit
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 3 |
---
|
| 4 |
|
| 5 |
This is a fine-tuned 13B chat model
|
|
@@ -74,4 +177,17 @@ This incident marked a turning point in Ava's crusade for environmental protecti
|
|
| 74 |
|
| 75 |
Today, Yosemite National Park remains a testament to the power of individual actions combined with collective effort. Its pristine landscapes continue inspiring countless visitors each year, reminding us all that we have a responsibility towards safeguarding our planet for future generations. And amidst these stunning vistas stands Ava, proudly carrying forth John Muir's legacy, ensuring that his dream of preserving nature lives on forever.</s>
|
| 76 |
|
| 77 |
-
```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
---
|
| 2 |
license: mit
|
| 3 |
+
model-index:
|
| 4 |
+
- name: Mixtral_13B_Chat
|
| 5 |
+
results:
|
| 6 |
+
- task:
|
| 7 |
+
type: text-generation
|
| 8 |
+
name: Text Generation
|
| 9 |
+
dataset:
|
| 10 |
+
name: AI2 Reasoning Challenge (25-Shot)
|
| 11 |
+
type: ai2_arc
|
| 12 |
+
config: ARC-Challenge
|
| 13 |
+
split: test
|
| 14 |
+
args:
|
| 15 |
+
num_few_shot: 25
|
| 16 |
+
metrics:
|
| 17 |
+
- type: acc_norm
|
| 18 |
+
value: 67.41
|
| 19 |
+
name: normalized accuracy
|
| 20 |
+
source:
|
| 21 |
+
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=cloudyu/Mixtral_13B_Chat
|
| 22 |
+
name: Open LLM Leaderboard
|
| 23 |
+
- task:
|
| 24 |
+
type: text-generation
|
| 25 |
+
name: Text Generation
|
| 26 |
+
dataset:
|
| 27 |
+
name: HellaSwag (10-Shot)
|
| 28 |
+
type: hellaswag
|
| 29 |
+
split: validation
|
| 30 |
+
args:
|
| 31 |
+
num_few_shot: 10
|
| 32 |
+
metrics:
|
| 33 |
+
- type: acc_norm
|
| 34 |
+
value: 85.87
|
| 35 |
+
name: normalized accuracy
|
| 36 |
+
source:
|
| 37 |
+
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=cloudyu/Mixtral_13B_Chat
|
| 38 |
+
name: Open LLM Leaderboard
|
| 39 |
+
- task:
|
| 40 |
+
type: text-generation
|
| 41 |
+
name: Text Generation
|
| 42 |
+
dataset:
|
| 43 |
+
name: MMLU (5-Shot)
|
| 44 |
+
type: cais/mmlu
|
| 45 |
+
config: all
|
| 46 |
+
split: test
|
| 47 |
+
args:
|
| 48 |
+
num_few_shot: 5
|
| 49 |
+
metrics:
|
| 50 |
+
- type: acc
|
| 51 |
+
value: 64.54
|
| 52 |
+
name: accuracy
|
| 53 |
+
source:
|
| 54 |
+
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=cloudyu/Mixtral_13B_Chat
|
| 55 |
+
name: Open LLM Leaderboard
|
| 56 |
+
- task:
|
| 57 |
+
type: text-generation
|
| 58 |
+
name: Text Generation
|
| 59 |
+
dataset:
|
| 60 |
+
name: TruthfulQA (0-shot)
|
| 61 |
+
type: truthful_qa
|
| 62 |
+
config: multiple_choice
|
| 63 |
+
split: validation
|
| 64 |
+
args:
|
| 65 |
+
num_few_shot: 0
|
| 66 |
+
metrics:
|
| 67 |
+
- type: mc2
|
| 68 |
+
value: 58.98
|
| 69 |
+
source:
|
| 70 |
+
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=cloudyu/Mixtral_13B_Chat
|
| 71 |
+
name: Open LLM Leaderboard
|
| 72 |
+
- task:
|
| 73 |
+
type: text-generation
|
| 74 |
+
name: Text Generation
|
| 75 |
+
dataset:
|
| 76 |
+
name: Winogrande (5-shot)
|
| 77 |
+
type: winogrande
|
| 78 |
+
config: winogrande_xl
|
| 79 |
+
split: validation
|
| 80 |
+
args:
|
| 81 |
+
num_few_shot: 5
|
| 82 |
+
metrics:
|
| 83 |
+
- type: acc
|
| 84 |
+
value: 80.43
|
| 85 |
+
name: accuracy
|
| 86 |
+
source:
|
| 87 |
+
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=cloudyu/Mixtral_13B_Chat
|
| 88 |
+
name: Open LLM Leaderboard
|
| 89 |
+
- task:
|
| 90 |
+
type: text-generation
|
| 91 |
+
name: Text Generation
|
| 92 |
+
dataset:
|
| 93 |
+
name: GSM8k (5-shot)
|
| 94 |
+
type: gsm8k
|
| 95 |
+
config: main
|
| 96 |
+
split: test
|
| 97 |
+
args:
|
| 98 |
+
num_few_shot: 5
|
| 99 |
+
metrics:
|
| 100 |
+
- type: acc
|
| 101 |
+
value: 56.63
|
| 102 |
+
name: accuracy
|
| 103 |
+
source:
|
| 104 |
+
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=cloudyu/Mixtral_13B_Chat
|
| 105 |
+
name: Open LLM Leaderboard
|
| 106 |
---
|
| 107 |
|
| 108 |
This is a fine-tuned 13B chat model
|
|
|
|
| 177 |
|
| 178 |
Today, Yosemite National Park remains a testament to the power of individual actions combined with collective effort. Its pristine landscapes continue inspiring countless visitors each year, reminding us all that we have a responsibility towards safeguarding our planet for future generations. And amidst these stunning vistas stands Ava, proudly carrying forth John Muir's legacy, ensuring that his dream of preserving nature lives on forever.</s>
|
| 179 |
|
| 180 |
+
```
|
| 181 |
+
# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
|
| 182 |
+
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_cloudyu__Mixtral_13B_Chat)
|
| 183 |
+
|
| 184 |
+
| Metric |Value|
|
| 185 |
+
|---------------------------------|----:|
|
| 186 |
+
|Avg. |68.98|
|
| 187 |
+
|AI2 Reasoning Challenge (25-Shot)|67.41|
|
| 188 |
+
|HellaSwag (10-Shot) |85.87|
|
| 189 |
+
|MMLU (5-Shot) |64.54|
|
| 190 |
+
|TruthfulQA (0-shot) |58.98|
|
| 191 |
+
|Winogrande (5-shot) |80.43|
|
| 192 |
+
|GSM8k (5-shot) |56.63|
|
| 193 |
+
|