added library to metadata
Browse files
README.md
CHANGED
|
@@ -5,12 +5,13 @@ license: llama2
|
|
| 5 |
base_model:
|
| 6 |
- NLPnorth/snakmodel-7b-base
|
| 7 |
pipeline_tag: text-generation
|
|
|
|
| 8 |
---
|
| 9 |

|
| 10 |
|
| 11 |
## Model Details
|
| 12 |
|
| 13 |
-
**SnakModel** is a 7B-parameter model specifically designed for the Danish language. This is the instruction-tuned variant: `SnakModel-7B (instruct)`. Our models
|
| 14 |
|
| 15 |
**Model Developers**
|
| 16 |
|
|
@@ -22,7 +23,7 @@ SnakModel comes as an instruction-tuned, and a base version. In addition, each m
|
|
| 22 |
|
| 23 |
**Input**
|
| 24 |
|
| 25 |
-
Text only.
|
| 26 |
|
| 27 |
**Output**
|
| 28 |
|
|
|
|
| 5 |
base_model:
|
| 6 |
- NLPnorth/snakmodel-7b-base
|
| 7 |
pipeline_tag: text-generation
|
| 8 |
+
library_name: transformers
|
| 9 |
---
|
| 10 |

|
| 11 |
|
| 12 |
## Model Details
|
| 13 |
|
| 14 |
+
**SnakModel** is a 7B-parameter model specifically designed for the Danish language. This is the instruction-tuned variant: `SnakModel-7B (instruct)`. Our models build upon [Llama 2](https://huggingface.co/meta-llama/Llama-2-7b-hf), which we continuously pre-train on a diverse collection of Danish corpora comprising 350M documents and 13.6B words, before tuning it on 3.7M Danish instruction-answer pairs.
|
| 15 |
|
| 16 |
**Model Developers**
|
| 17 |
|
|
|
|
| 23 |
|
| 24 |
**Input**
|
| 25 |
|
| 26 |
+
Text only, with instructions following the `[INST] {instruction} [/INST]` template.
|
| 27 |
|
| 28 |
**Output**
|
| 29 |
|