Update README.md
Browse files
README.md
CHANGED
|
@@ -1,5 +1,10 @@
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 3 |
---
|
| 4 |
openllama_v2 3B second stage pre-trained on russian part of OSCAR with 4k sequence length. Model has seen about 5B tokens for now, weights will be updated as the training goes on.
|
| 5 |
Achieves 3.8 perplexity on the evaluation dataset. Will we further pre-trained on wiki dataset with 8K context length.
|
|
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
| 3 |
+
datasets:
|
| 4 |
+
- oscar-corpus/OSCAR-2301
|
| 5 |
+
- wikipedia
|
| 6 |
+
metrics:
|
| 7 |
+
- perplexity
|
| 8 |
---
|
| 9 |
openllama_v2 3B second stage pre-trained on russian part of OSCAR with 4k sequence length. Model has seen about 5B tokens for now, weights will be updated as the training goes on.
|
| 10 |
Achieves 3.8 perplexity on the evaluation dataset. Will we further pre-trained on wiki dataset with 8K context length.
|