Update README.md
Browse files
README.md
CHANGED
|
@@ -133,10 +133,10 @@ Meta-Llama-3-70B-Instruct 9.006250
|
|
| 133 |
|
| 134 |
### OpenLLM Leaderboard Manual Evaluation
|
| 135 |
|
| 136 |
-
| Model | ARC | Hellaswag | MMLU | TruthfulQA | Winogrande | GSM8K* |
|
| 137 |
-
| :---- | ---: | ------: | ---: | ---: | ---: | ---: |
|
| 138 |
-
| Smaug-Llama-3-70B-Instruct | 70.6 | 86.1 | 79.2 | 62.5 | 83.5 | 90.5 |
|
| 139 |
-
| Llama-3-70B-Instruct | 71.4 | 85.7 | 80.0 | 61.8 | 82.9 | 91.1 |
|
| 140 |
|
| 141 |
**GSM8K** The GSM8K numbers quoted here are computed using a recent release
|
| 142 |
of the [LM Evaluation Harness](https://github.com/EleutherAI/lm-evaluation-harness/).
|
|
|
|
| 133 |
|
| 134 |
### OpenLLM Leaderboard Manual Evaluation
|
| 135 |
|
| 136 |
+
| Model | ARC | Hellaswag | MMLU | TruthfulQA | Winogrande | GSM8K* | Average |
|
| 137 |
+
| :---- | ---: | ------: | ---: | ---: | ---: | ---: | ---: |
|
| 138 |
+
| Smaug-Llama-3-70B-Instruct | 70.6 | 86.1 | 79.2 | 62.5 | 83.5 | 90.5 | 78.7 |
|
| 139 |
+
| Llama-3-70B-Instruct | 71.4 | 85.7 | 80.0 | 61.8 | 82.9 | 91.1 | 78.8 |
|
| 140 |
|
| 141 |
**GSM8K** The GSM8K numbers quoted here are computed using a recent release
|
| 142 |
of the [LM Evaluation Harness](https://github.com/EleutherAI/lm-evaluation-harness/).
|