Shannonjunior commited on
Commit
c3a5511
·
verified ·
1 Parent(s): c69e544

Upload checkpoint

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
added_tokens.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "<|endoftext|>": 151643,
3
+ "<|im_end|>": 151645,
4
+ "<|im_start|>": 151644
5
+ }
chat_template.jinja CHANGED
@@ -1,4 +1,6 @@
1
- {% for message in messages %}{{'<|im_start|>' + message['role'] + '
 
 
2
  ' + message['content'] + '<|im_end|>' + '
3
  '}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant
4
  ' }}{% endif %}
 
1
+ {% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system
2
+ You are a helpful assistant.<|im_end|>
3
+ ' }}{% endif %}{{'<|im_start|>' + message['role'] + '
4
  ' + message['content'] + '<|im_end|>' + '
5
  '}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant
6
  ' }}{% endif %}
config.json CHANGED
@@ -1,30 +1,53 @@
1
  {
2
  "architectures": [
3
- "LlamaForCausalLM"
4
  ],
5
- "attention_bias": false,
6
  "attention_dropout": 0.0,
7
- "bos_token_id": 1,
8
- "eos_token_id": 2,
9
- "head_dim": 64,
10
  "hidden_act": "silu",
11
- "hidden_size": 576,
12
  "initializer_range": 0.02,
13
- "intermediate_size": 1536,
14
- "max_position_embeddings": 4096,
15
- "mlp_bias": false,
16
- "model_type": "llama",
17
- "num_attention_heads": 9,
18
- "num_hidden_layers": 30,
19
- "num_key_value_heads": 3,
20
- "pad_token_id": 2,
21
- "pretraining_tp": 1,
22
- "rms_norm_eps": 1e-05,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
23
  "rope_scaling": null,
24
- "rope_theta": 10000.0,
 
25
  "tie_word_embeddings": true,
26
  "torch_dtype": "float32",
27
  "transformers_version": "4.53.0.dev0",
28
  "use_cache": false,
29
- "vocab_size": 49152
 
30
  }
 
1
  {
2
  "architectures": [
3
+ "Qwen2ForCausalLM"
4
  ],
 
5
  "attention_dropout": 0.0,
6
+ "eos_token_id": 151645,
 
 
7
  "hidden_act": "silu",
8
+ "hidden_size": 896,
9
  "initializer_range": 0.02,
10
+ "intermediate_size": 4864,
11
+ "layer_types": [
12
+ "full_attention",
13
+ "full_attention",
14
+ "full_attention",
15
+ "full_attention",
16
+ "full_attention",
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention"
36
+ ],
37
+ "max_position_embeddings": 32768,
38
+ "max_window_layers": 24,
39
+ "model_type": "qwen2",
40
+ "num_attention_heads": 14,
41
+ "num_hidden_layers": 24,
42
+ "num_key_value_heads": 2,
43
+ "rms_norm_eps": 1e-06,
44
  "rope_scaling": null,
45
+ "rope_theta": 1000000.0,
46
+ "sliding_window": null,
47
  "tie_word_embeddings": true,
48
  "torch_dtype": "float32",
49
  "transformers_version": "4.53.0.dev0",
50
  "use_cache": false,
51
+ "use_sliding_window": false,
52
+ "vocab_size": 151936
53
  }
generation_config.json CHANGED
@@ -1,9 +1,14 @@
1
  {
2
- "_from_model_config": true,
3
- "bos_token_id": 1,
4
  "do_sample": true,
5
- "eos_token_id": 2,
6
- "max_new_tokens": 40,
7
- "pad_token_id": 2,
 
 
 
 
 
 
8
  "transformers_version": "4.53.0.dev0"
9
  }
 
1
  {
2
+ "bos_token_id": 151643,
 
3
  "do_sample": true,
4
+ "eos_token_id": [
5
+ 151645,
6
+ 151643
7
+ ],
8
+ "pad_token_id": 151643,
9
+ "repetition_penalty": 1.1,
10
+ "temperature": 0.7,
11
+ "top_k": 20,
12
+ "top_p": 0.8,
13
  "transformers_version": "4.53.0.dev0"
14
  }
merges.txt CHANGED
The diff for this file is too large to render. See raw diff
 
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1bb544af3eb8a10044aab7260fc00b6be00a445ca67ec52f1e987bf42e46f993
3
- size 325753976
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:108254821a16f93440423d7e1f2dd21fe5d0c85063658ee06c6bc916541becf4
3
+ size 1260455072
special_tokens_map.json CHANGED
@@ -3,13 +3,6 @@
3
  "<|im_start|>",
4
  "<|im_end|>"
5
  ],
6
- "bos_token": {
7
- "content": "<|im_start|>",
8
- "lstrip": false,
9
- "normalized": false,
10
- "rstrip": false,
11
- "single_word": false
12
- },
13
  "eos_token": {
14
  "content": "<|im_end|>",
15
  "lstrip": false,
@@ -18,13 +11,6 @@
18
  "single_word": false
19
  },
20
  "pad_token": {
21
- "content": "<empty_output>",
22
- "lstrip": false,
23
- "normalized": false,
24
- "rstrip": false,
25
- "single_word": false
26
- },
27
- "unk_token": {
28
  "content": "<|endoftext|>",
29
  "lstrip": false,
30
  "normalized": false,
 
3
  "<|im_start|>",
4
  "<|im_end|>"
5
  ],
 
 
 
 
 
 
 
6
  "eos_token": {
7
  "content": "<|im_end|>",
8
  "lstrip": false,
 
11
  "single_word": false
12
  },
13
  "pad_token": {
 
 
 
 
 
 
 
14
  "content": "<|endoftext|>",
15
  "lstrip": false,
16
  "normalized": false,
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "add_prefix_space": false,
3
  "added_tokens_decoder": {
4
- "0": {
5
  "content": "<|endoftext|>",
6
  "lstrip": false,
7
  "normalized": false,
@@ -9,7 +9,7 @@
9
  "single_word": false,
10
  "special": true
11
  },
12
- "1": {
13
  "content": "<|im_start|>",
14
  "lstrip": false,
15
  "normalized": false,
@@ -17,139 +17,27 @@
17
  "single_word": false,
18
  "special": true
19
  },
20
- "2": {
21
  "content": "<|im_end|>",
22
  "lstrip": false,
23
  "normalized": false,
24
  "rstrip": false,
25
  "single_word": false,
26
  "special": true
27
- },
28
- "3": {
29
- "content": "<repo_name>",
30
- "lstrip": false,
31
- "normalized": false,
32
- "rstrip": false,
33
- "single_word": false,
34
- "special": true
35
- },
36
- "4": {
37
- "content": "<reponame>",
38
- "lstrip": false,
39
- "normalized": false,
40
- "rstrip": false,
41
- "single_word": false,
42
- "special": true
43
- },
44
- "5": {
45
- "content": "<file_sep>",
46
- "lstrip": false,
47
- "normalized": false,
48
- "rstrip": false,
49
- "single_word": false,
50
- "special": true
51
- },
52
- "6": {
53
- "content": "<filename>",
54
- "lstrip": false,
55
- "normalized": false,
56
- "rstrip": false,
57
- "single_word": false,
58
- "special": true
59
- },
60
- "7": {
61
- "content": "<gh_stars>",
62
- "lstrip": false,
63
- "normalized": false,
64
- "rstrip": false,
65
- "single_word": false,
66
- "special": true
67
- },
68
- "8": {
69
- "content": "<issue_start>",
70
- "lstrip": false,
71
- "normalized": false,
72
- "rstrip": false,
73
- "single_word": false,
74
- "special": true
75
- },
76
- "9": {
77
- "content": "<issue_comment>",
78
- "lstrip": false,
79
- "normalized": false,
80
- "rstrip": false,
81
- "single_word": false,
82
- "special": true
83
- },
84
- "10": {
85
- "content": "<issue_closed>",
86
- "lstrip": false,
87
- "normalized": false,
88
- "rstrip": false,
89
- "single_word": false,
90
- "special": true
91
- },
92
- "11": {
93
- "content": "<jupyter_start>",
94
- "lstrip": false,
95
- "normalized": false,
96
- "rstrip": false,
97
- "single_word": false,
98
- "special": true
99
- },
100
- "12": {
101
- "content": "<jupyter_text>",
102
- "lstrip": false,
103
- "normalized": false,
104
- "rstrip": false,
105
- "single_word": false,
106
- "special": true
107
- },
108
- "13": {
109
- "content": "<jupyter_code>",
110
- "lstrip": false,
111
- "normalized": false,
112
- "rstrip": false,
113
- "single_word": false,
114
- "special": true
115
- },
116
- "14": {
117
- "content": "<jupyter_output>",
118
- "lstrip": false,
119
- "normalized": false,
120
- "rstrip": false,
121
- "single_word": false,
122
- "special": true
123
- },
124
- "15": {
125
- "content": "<jupyter_script>",
126
- "lstrip": false,
127
- "normalized": false,
128
- "rstrip": false,
129
- "single_word": false,
130
- "special": true
131
- },
132
- "16": {
133
- "content": "<empty_output>",
134
- "lstrip": false,
135
- "normalized": false,
136
- "rstrip": false,
137
- "single_word": false,
138
- "special": true
139
  }
140
  },
141
  "additional_special_tokens": [
142
  "<|im_start|>",
143
  "<|im_end|>"
144
  ],
145
- "bos_token": "<|im_start|>",
146
  "clean_up_tokenization_spaces": false,
147
  "eos_token": "<|im_end|>",
 
148
  "extra_special_tokens": {},
149
- "model_max_length": 2048,
150
- "pad_token": "<empty_output>",
151
- "padding_side": "left",
152
- "tokenizer_class": "GPT2Tokenizer",
153
- "unk_token": "<|endoftext|>",
154
- "vocab_size": 49152
155
  }
 
1
  {
2
  "add_prefix_space": false,
3
  "added_tokens_decoder": {
4
+ "151643": {
5
  "content": "<|endoftext|>",
6
  "lstrip": false,
7
  "normalized": false,
 
9
  "single_word": false,
10
  "special": true
11
  },
12
+ "151644": {
13
  "content": "<|im_start|>",
14
  "lstrip": false,
15
  "normalized": false,
 
17
  "single_word": false,
18
  "special": true
19
  },
20
+ "151645": {
21
  "content": "<|im_end|>",
22
  "lstrip": false,
23
  "normalized": false,
24
  "rstrip": false,
25
  "single_word": false,
26
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
27
  }
28
  },
29
  "additional_special_tokens": [
30
  "<|im_start|>",
31
  "<|im_end|>"
32
  ],
33
+ "bos_token": null,
34
  "clean_up_tokenization_spaces": false,
35
  "eos_token": "<|im_end|>",
36
+ "errors": "replace",
37
  "extra_special_tokens": {},
38
+ "model_max_length": 32768,
39
+ "pad_token": "<|endoftext|>",
40
+ "split_special_tokens": false,
41
+ "tokenizer_class": "Qwen2Tokenizer",
42
+ "unk_token": null
 
43
  }
vocab.json CHANGED
The diff for this file is too large to render. See raw diff