Update README.md
Browse files
README.md
CHANGED
|
@@ -130,18 +130,18 @@ This instruction model was built via parameter-efficient QLoRA finetuning of [ll
|
|
| 130 |
* Model license: Llama 2 Community License Agreement
|
| 131 |
* Basic usage: [notebook](assets/basic_inference_llama_2_dolphin.ipynb)
|
| 132 |
|
| 133 |
-
## Open LLM Leaderboard Evaluation Results
|
| 134 |
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_dfurman__llama-2-70b-dolphin-peft)
|
| 135 |
|
| 136 |
-
|
|
| 137 |
-
|
| 138 |
-
|
|
| 139 |
-
|
|
| 140 |
-
|
|
| 141 |
-
|
|
| 142 |
-
|
|
| 143 |
-
|
|
| 144 |
-
|
|
| 145 |
|
| 146 |
## Example prompts and responses
|
| 147 |
|
|
@@ -359,17 +359,3 @@ The license on this model does not constitute legal advice. We are not responsib
|
|
| 359 |
|
| 360 |
- PEFT 0.5.0.dev0
|
| 361 |
|
| 362 |
-
|
| 363 |
-
# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
|
| 364 |
-
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_dfurman__llama-2-70b-dolphin-peft)
|
| 365 |
-
|
| 366 |
-
| Metric |Value|
|
| 367 |
-
|---------------------------------|----:|
|
| 368 |
-
|Avg. |65.72|
|
| 369 |
-
|AI2 Reasoning Challenge (25-Shot)|69.62|
|
| 370 |
-
|HellaSwag (10-Shot) |86.82|
|
| 371 |
-
|MMLU (5-Shot) |69.18|
|
| 372 |
-
|TruthfulQA (0-shot) |57.43|
|
| 373 |
-
|Winogrande (5-shot) |83.90|
|
| 374 |
-
|GSM8k (5-shot) |27.37|
|
| 375 |
-
|
|
|
|
| 130 |
* Model license: Llama 2 Community License Agreement
|
| 131 |
* Basic usage: [notebook](assets/basic_inference_llama_2_dolphin.ipynb)
|
| 132 |
|
| 133 |
+
## [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
|
| 134 |
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_dfurman__llama-2-70b-dolphin-peft)
|
| 135 |
|
| 136 |
+
| Metric |Value|
|
| 137 |
+
|---------------------------------|----:|
|
| 138 |
+
|Avg. |65.72|
|
| 139 |
+
|AI2 Reasoning Challenge (25-Shot)|69.62|
|
| 140 |
+
|HellaSwag (10-Shot) |86.82|
|
| 141 |
+
|MMLU (5-Shot) |69.18|
|
| 142 |
+
|TruthfulQA (0-shot) |57.43|
|
| 143 |
+
|Winogrande (5-shot) |83.90|
|
| 144 |
+
|GSM8k (5-shot) |27.37|
|
| 145 |
|
| 146 |
## Example prompts and responses
|
| 147 |
|
|
|
|
| 359 |
|
| 360 |
- PEFT 0.5.0.dev0
|
| 361 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|