Update README.md
Browse files
README.md
CHANGED
|
@@ -14,6 +14,24 @@ tags:
|
|
| 14 |
|
| 15 |
experimental model to expose arco to some reasoning
|
| 16 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 17 |
# Uploaded model
|
| 18 |
|
| 19 |
- **Developed by:** appvoid
|
|
|
|
| 14 |
|
| 15 |
experimental model to expose arco to some reasoning
|
| 16 |
|
| 17 |
+
| Task | Score | Metric |
|
| 18 |
+
|--------------|-------|-----------|
|
| 19 |
+
| ARC Challenge| 0.3473| acc_norm |
|
| 20 |
+
| HellaSwag | 0.5986| acc_norm |
|
| 21 |
+
| MMLU | 0.2489| acc |
|
| 22 |
+
| PIQA | 0.7318| acc_norm |
|
| 23 |
+
| Winogrande | 0.6259| acc |
|
| 24 |
+
|
| 25 |
+
This table presents the extracted scores in a clear, tabular format. The "Task" column shows the name of each benchmark, the "Score" column displays the corresponding value, and the "Metric" column indicates whether the score is acc_norm or acc.
|
| 26 |
+
|
| 27 |
+
format is this:
|
| 28 |
+
|
| 29 |
+
```
|
| 30 |
+
Instruction: <your instruction>
|
| 31 |
+
Reasoning: // starting from here, the model will start to generate the resoning and output
|
| 32 |
+
Output:
|
| 33 |
+
```
|
| 34 |
+
|
| 35 |
# Uploaded model
|
| 36 |
|
| 37 |
- **Developed by:** appvoid
|