Elaine
commited on
Update README.md
Browse files
README.md
CHANGED
|
@@ -1,3 +1,37 @@
|
|
| 1 |
-
---
|
| 2 |
-
license: apache-2.0
|
| 3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
language:
|
| 4 |
+
- en
|
| 5 |
+
- fi
|
| 6 |
+
base_model:
|
| 7 |
+
- LumiOpen/Poro-34B
|
| 8 |
+
datasets:
|
| 9 |
+
- sablo/oasst2_curated
|
| 10 |
+
- LumiOpen/instruction-collection-fin
|
| 11 |
+
---
|
| 12 |
+
|
| 13 |
+
This is an SFT-tuned model of [Poro-34B](https://huggingface.co/LumiOpen/Poro-34B) with English and Finnish data.
|
| 14 |
+
|
| 15 |
+
We use a curated subset of Open Assistant 2 and translated the dataset into Finnish using Poro-34B. We trained this model for experiments on the impact of multilingual instruction-tuning. For a better chat experience, we recommend using [Poro-34B-chat](https://huggingface.co/LumiOpen/Poro-34B-chat) instead.
|
| 16 |
+
|
| 17 |
+
**Datasets**
|
| 18 |
+
|
| 19 |
+
**Recipes**
|
| 20 |
+
|
| 21 |
+
|
| 22 |
+
|
| 23 |
+
**Evaluation**
|
| 24 |
+
TBA
|
| 25 |
+
|
| 26 |
+
|
| 27 |
+
**Citation**
|
| 28 |
+
```
|
| 29 |
+
@inproceedings{
|
| 30 |
+
zosa2024got,
|
| 31 |
+
title={Got Compute, but No Data: Lessons From Post-training a Finnish {LLM}},
|
| 32 |
+
author={Elaine Zosa and Ville Komulainen and Sampo Pyysalo},
|
| 33 |
+
booktitle={The Joint 25th Nordic Conference on Computational Linguistics and 11th Baltic Conference on Human Language Technologies},
|
| 34 |
+
year={2024},
|
| 35 |
+
url={https://openreview.net/forum?id=8wWlu1stNK}
|
| 36 |
+
}
|
| 37 |
+
```
|