ThomasTheMaker commited on
Commit
e88e4eb
·
verified ·
1 Parent(s): 77855e4

Upload folder using huggingface_hub

Browse files
chat_template.jinja CHANGED
@@ -1 +1,15 @@
1
- {{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '<|user|>\n' + message['content'] + '<|end|>\n' }}{% elif message['role'] == 'assistant' %}{{ '<|bot|>\n' + message['content'] + '<|end|>\n' }}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}{{ eos_token }}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {% for message in messages %}
2
+ {% if message['role'] == 'user' %}
3
+ {{ '<|user|>
4
+ ' + message['content'] + eos_token }}
5
+ {% elif message['role'] == 'system' %}
6
+ {{ '<|system|>
7
+ ' + message['content'] + eos_token }}
8
+ {% elif message['role'] == 'assistant' %}
9
+ {{ '<|assistant|>
10
+ ' + message['content'] + eos_token }}
11
+ {% endif %}
12
+ {% if loop.last and add_generation_prompt %}
13
+ {{ '<|assistant|>' }}
14
+ {% endif %}
15
+ {% endfor %}
config.json CHANGED
@@ -5,26 +5,26 @@
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
7
  "bos_token_id": 1,
8
- "dtype": "float32",
9
  "eos_token_id": 2,
10
- "head_dim": 128,
11
  "hidden_act": "silu",
12
- "hidden_size": 1536,
13
  "initializer_range": 0.02,
14
- "intermediate_size": 4096,
15
  "max_position_embeddings": 4096,
16
  "mlp_bias": false,
17
  "model_type": "llama",
18
  "num_attention_heads": 12,
19
- "num_hidden_layers": 22,
20
- "num_key_value_heads": 3,
21
- "pad_token_id": 35049,
22
  "pretraining_tp": 1,
23
  "rms_norm_eps": 1e-06,
24
  "rope_scaling": null,
25
  "rope_theta": 10000.0,
26
  "tie_word_embeddings": false,
27
  "transformers_version": "4.57.1",
28
- "use_cache": true,
29
- "vocab_size": 35054
30
  }
 
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
7
  "bos_token_id": 1,
8
+ "dtype": "bfloat16",
9
  "eos_token_id": 2,
10
+ "head_dim": 64,
11
  "hidden_act": "silu",
12
+ "hidden_size": 768,
13
  "initializer_range": 0.02,
14
+ "intermediate_size": 2048,
15
  "max_position_embeddings": 4096,
16
  "mlp_bias": false,
17
  "model_type": "llama",
18
  "num_attention_heads": 12,
19
+ "num_hidden_layers": 12,
20
+ "num_key_value_heads": 4,
21
+ "pad_token_id": 2,
22
  "pretraining_tp": 1,
23
  "rms_norm_eps": 1e-06,
24
  "rope_scaling": null,
25
  "rope_theta": 10000.0,
26
  "tie_word_embeddings": false,
27
  "transformers_version": "4.57.1",
28
+ "use_cache": false,
29
+ "vocab_size": 32000
30
  }
generation_config.json CHANGED
@@ -1,9 +1,8 @@
1
  {
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
- "eos_token_id": [
5
- 2
6
- ],
7
- "pad_token_id": 35049,
8
- "transformers_version": "4.57.1"
9
  }
 
1
  {
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "pad_token_id": 2,
6
+ "transformers_version": "4.57.1",
7
+ "use_cache": false
 
8
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9c707dceda61ef03e5388ff7134711c357b0d210baa4fd73e2414d7fa7478702
3
- size 2611032480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef6ca3e8ffbac339e2a537663a0211d4edcbfc5aa3f53ace6548ef56acf86431
3
+ size 249349768
special_tokens_map.json CHANGED
@@ -1,27 +1,4 @@
1
  {
2
- "additional_special_tokens": [
3
- {
4
- "content": "<|user|>",
5
- "lstrip": false,
6
- "normalized": false,
7
- "rstrip": false,
8
- "single_word": false
9
- },
10
- {
11
- "content": "<|bot|>",
12
- "lstrip": false,
13
- "normalized": false,
14
- "rstrip": false,
15
- "single_word": false
16
- },
17
- {
18
- "content": "<|end|>",
19
- "lstrip": false,
20
- "normalized": false,
21
- "rstrip": false,
22
- "single_word": false
23
- }
24
- ],
25
  "bos_token": {
26
  "content": "<s>",
27
  "lstrip": false,
@@ -36,15 +13,8 @@
36
  "rstrip": false,
37
  "single_word": false
38
  },
39
- "mask_token": {
40
- "content": "<mask>",
41
- "lstrip": false,
42
- "normalized": false,
43
- "rstrip": false,
44
- "single_word": false
45
- },
46
  "pad_token": {
47
- "content": "<pad>",
48
  "lstrip": false,
49
  "normalized": false,
50
  "rstrip": false,
 
1
  {
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  "bos_token": {
3
  "content": "<s>",
4
  "lstrip": false,
 
13
  "rstrip": false,
14
  "single_word": false
15
  },
 
 
 
 
 
 
 
16
  "pad_token": {
17
+ "content": "</s>",
18
  "lstrip": false,
19
  "normalized": false,
20
  "rstrip": false,
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -26,61 +26,15 @@
26
  "rstrip": false,
27
  "single_word": false,
28
  "special": true
29
- },
30
- "35049": {
31
- "content": "<pad>",
32
- "lstrip": false,
33
- "normalized": false,
34
- "rstrip": false,
35
- "single_word": false,
36
- "special": true
37
- },
38
- "35050": {
39
- "content": "<mask>",
40
- "lstrip": false,
41
- "normalized": false,
42
- "rstrip": false,
43
- "single_word": false,
44
- "special": true
45
- },
46
- "35051": {
47
- "content": "<|user|>",
48
- "lstrip": false,
49
- "normalized": false,
50
- "rstrip": false,
51
- "single_word": false,
52
- "special": true
53
- },
54
- "35052": {
55
- "content": "<|bot|>",
56
- "lstrip": false,
57
- "normalized": false,
58
- "rstrip": false,
59
- "single_word": false,
60
- "special": true
61
- },
62
- "35053": {
63
- "content": "<|end|>",
64
- "lstrip": false,
65
- "normalized": false,
66
- "rstrip": false,
67
- "single_word": false,
68
- "special": true
69
  }
70
  },
71
- "additional_special_tokens": [
72
- "<|user|>",
73
- "<|bot|>",
74
- "<|end|>"
75
- ],
76
  "bos_token": "<s>",
77
  "clean_up_tokenization_spaces": false,
78
  "eos_token": "</s>",
79
  "extra_special_tokens": {},
80
  "legacy": false,
81
- "mask_token": "<mask>",
82
  "model_max_length": 2048,
83
- "pad_token": "<pad>",
84
  "padding_side": "right",
85
  "sp_model_kwargs": {},
86
  "tokenizer_class": "LlamaTokenizer",
 
26
  "rstrip": false,
27
  "single_word": false,
28
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
29
  }
30
  },
 
 
 
 
 
31
  "bos_token": "<s>",
32
  "clean_up_tokenization_spaces": false,
33
  "eos_token": "</s>",
34
  "extra_special_tokens": {},
35
  "legacy": false,
 
36
  "model_max_length": 2048,
37
+ "pad_token": "</s>",
38
  "padding_side": "right",
39
  "sp_model_kwargs": {},
40
  "tokenizer_class": "LlamaTokenizer",
training_state.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22e983df10c34525cce3dc0cf17d479bdd2c05fab074c3bfdd1c30f3e928646f
3
+ size 498770858