Update README.md
Browse files
README.md
CHANGED
|
@@ -193,13 +193,13 @@ As always, the training data, training code, and metrics are publicly available.
|
|
| 193 |
|
| 194 |
# CrystalChat Performance
|
| 195 |
|
| 196 |
-
| Model | Trained Tokens | Avg. of Avg. | Language Avg. | Coding Avg.
|
| 197 |
-
|
| 198 |
-
| CrystalChat 7B | 1.275T | 44.96 | 53.29 | 36.62 |
|
| 199 |
-
| Mistral-7B-Instruct-v0.1 | - | 44.34 | 54.86 | 30.62 |
|
| 200 |
-
| CodeLlama-7b-Instruct | 2.5T | 40.91 | 45.29 | 36.52 |
|
| 201 |
-
| Llama-2-7b-Chat | 2T | 34.11 | 52.86 | 15.35 |
|
| 202 |
-
| AmberChat 7B | 1.25T | - | 44.76 | - |
|
| 203 |
|
| 204 |
| Model | Trained Tokens | ARC | HellaSwag | MMLU (5-shot) | GSM8K | Winogrande(5-shot) | TruthfulQA | HumanEval (pass@1) | MBPP (pass@1) |
|
| 205 |
|:------------------------:|:--------------:|:------------:|:-------------:|:-----------:|:-----:|:---------:|:-------------:|:-----:|:------------------:|
|
|
|
|
| 193 |
|
| 194 |
# CrystalChat Performance
|
| 195 |
|
| 196 |
+
| Model | Trained Tokens | Avg. of Avg. | Language Avg. | Coding Avg.
|
| 197 |
+
|:------------------------:|:--------------:|:------------:|:-------------:|:-----------:|
|
| 198 |
+
| CrystalChat 7B | 1.275T | 44.96 | 53.29 | 36.62 |
|
| 199 |
+
| Mistral-7B-Instruct-v0.1 | - | 44.34 | 54.86 | 30.62 |
|
| 200 |
+
| CodeLlama-7b-Instruct | 2.5T | 40.91 | 45.29 | 36.52 |
|
| 201 |
+
| Llama-2-7b-Chat | 2T | 34.11 | 52.86 | 15.35 |
|
| 202 |
+
| AmberChat 7B | 1.25T | - | 44.76 | - |
|
| 203 |
|
| 204 |
| Model | Trained Tokens | ARC | HellaSwag | MMLU (5-shot) | GSM8K | Winogrande(5-shot) | TruthfulQA | HumanEval (pass@1) | MBPP (pass@1) |
|
| 205 |
|:------------------------:|:--------------:|:------------:|:-------------:|:-----------:|:-----:|:---------:|:-------------:|:-----:|:------------------:|
|