Create README.md
Browse files
README.md
ADDED
|
@@ -0,0 +1,32 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
language:
|
| 4 |
+
- en
|
| 5 |
+
pipeline_tag: text-generation
|
| 6 |
+
datasets:
|
| 7 |
+
- appvoid/no-prompt-50k
|
| 8 |
+
---
|
| 9 |
+

|
| 10 |
+
# palmer
|
| 11 |
+
### a better base model
|
| 12 |
+
This is a small improvement over palmer-002
|
| 13 |
+
|
| 14 |
+
### evaluation 🧪
|
| 15 |
+
note that this is a zero-shot setting as opposite to open llm leaderboard's few-shot evals
|
| 16 |
+
```
|
| 17 |
+
model ARC-C OBQA HellaSwag PIQA Winogrande Average
|
| 18 |
+
sheared-llama | 0.2910 | 0.3440 | 0.5809 | 0.7339 | 0.5809 | 0.5061 |
|
| 19 |
+
tinyllama | 0.3029 | 0.3600 | 0.5935 | 0.7329 | 0.5959 | 0.5170 |
|
| 20 |
+
palmer-002 | 0.3242 | 0.3700 | 0.5956 | 0.7345 | 0.5888 | 0.5226 |
|
| 21 |
+
palmer-002-2401 | 0.3294 | 0.3700 | 0.5950 | 0.7399 | 0.5896 | 0.5247 | (this)
|
| 22 |
+
babbage-002 | 0.3285 | 0.3620 | 0.6380 | 0.7606 | 0.6085 | 0.5395 |
|
| 23 |
+
```
|
| 24 |
+
|
| 25 |
+
### training 🦾
|
| 26 |
+
Training took 0.5 A100 gpu hours. It was trained on 50,000 gpt-4 shuffled samples. palmer was fine-tuned using lower learning rates ensuring it keeps as much general knowledge as possible.
|
| 27 |
+
|
| 28 |
+
### prompt 📝
|
| 29 |
+
```
|
| 30 |
+
no prompt 🚀
|
| 31 |
+
```
|
| 32 |
+
<a href="https://ko-fi.com/appvoid" target="_blank"><img src="https://cdn.buymeacoffee.com/buttons/v2/default-yellow.png" alt="Buy Me A Coffee" style="height: 48px !important;width: 180px !important; filter: invert(70%);" ></a>
|