revana commited on
Commit
794d05e
·
verified ·
1 Parent(s): a5d3f2a

Upload folder using huggingface_hub

Browse files
Files changed (3) hide show
  1. README.md +1 -1
  2. model.safetensors +1 -1
  3. tokenizer_config.json +0 -1
README.md CHANGED
@@ -195,4 +195,4 @@ tokens = tokenizer("Hej varlden!", return_tensors="pt")
195
 
196
  ## License
197
 
198
- Apache 2.0 — see [LICENSE](LICENSE).
 
195
 
196
  ## License
197
 
198
+ Apache 2.0 — see [LICENSE](LICENSE).
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:175c08e1f6b45545532797be5294963d2bc66ee33af54a14a2c6600adbceff00
3
  size 1772319024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:91aa296097b7ec7968460e0ddbf0561d09da6620d5f52d47f536eee40f74f1a1
3
  size 1772319024
tokenizer_config.json CHANGED
@@ -5,7 +5,6 @@
5
  "eos_token": "[EOS]",
6
  "pad_token": "[PAD]",
7
  "unk_token": "[UNK]",
8
- "chat_template": "{% for message in messages %}{% if message['role'] == 'system' %}### System:\n{{ message['content'] }}\n\n{% elif message['role'] == 'user' %}### Instruction:\n{{ message['content'] }}\n\n{% elif message['role'] == 'assistant' %}### Response:\n{{ message['content'] }}\n\n{% endif %}{% endfor %}### Response:\n",
9
  "clean_up_tokenization_spaces": false,
10
  "added_tokens_decoder": {
11
  "0": {
 
5
  "eos_token": "[EOS]",
6
  "pad_token": "[PAD]",
7
  "unk_token": "[UNK]",
 
8
  "clean_up_tokenization_spaces": false,
9
  "added_tokens_decoder": {
10
  "0": {