rhgt1996 commited on
Commit
8bcf4e4
·
verified ·
1 Parent(s): 208039b

End of training

Browse files
.ipynb_checkpoints/chat_template-checkpoint.jinja ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {% for message in messages %}{{'<|im_start|>' + message['role'] + '
2
+ ' + message['content'] + '<|im_end|>' + '
3
+ '}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant
4
+ ' }}{% endif %}
README.md CHANGED
@@ -4,10 +4,10 @@ library_name: transformers
4
  model_name: math-SmolLM2-360M
5
  tags:
6
  - generated_from_trainer
7
- - camel-math
8
  - finetune
9
- - trl
10
  - sft
 
11
  licence: license
12
  ---
13
 
 
4
  model_name: math-SmolLM2-360M
5
  tags:
6
  - generated_from_trainer
 
7
  - finetune
8
+ - camel-math
9
  - sft
10
+ - trl
11
  licence: license
12
  ---
13
 
adapter_config.json CHANGED
@@ -25,12 +25,12 @@
25
  "rank_pattern": {},
26
  "revision": null,
27
  "target_modules": [
 
28
  "o_proj",
29
  "up_proj",
30
- "gate_proj",
31
  "q_proj",
32
  "v_proj",
33
- "down_proj",
34
  "k_proj"
35
  ],
36
  "target_parameters": null,
 
25
  "rank_pattern": {},
26
  "revision": null,
27
  "target_modules": [
28
+ "down_proj",
29
  "o_proj",
30
  "up_proj",
 
31
  "q_proj",
32
  "v_proj",
33
+ "gate_proj",
34
  "k_proj"
35
  ],
36
  "target_parameters": null,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:15514aefbe1f79fb4df924da0e43549a539dca4fabd3b4a3bebdb829ab640549
3
  size 8741392
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:556b015fcaf13e4682d92b6171209887c9d3088ced3a7612d9cf0911875386a8
3
  size 8741392
special_tokens_map.json CHANGED
@@ -1,23 +1,38 @@
1
  {
2
  "additional_special_tokens": [
3
- {
4
- "content": "<|im_start|>",
5
- "lstrip": false,
6
- "normalized": false,
7
- "rstrip": false,
8
- "single_word": false
9
- },
10
- {
11
- "content": "<|im_end|>",
12
- "lstrip": false,
13
- "normalized": false,
14
- "rstrip": false,
15
- "single_word": false
16
- }
 
 
 
17
  ],
18
- "bos_token": "<|im_start|>",
19
- "eos_token": "<|im_end|>",
20
- "pad_token": "<|im_end|>",
 
 
 
 
 
 
 
 
 
 
 
 
21
  "unk_token": {
22
  "content": "<|endoftext|>",
23
  "lstrip": false,
 
1
  {
2
  "additional_special_tokens": [
3
+ "<|endoftext|>",
4
+ "<|im_start|>",
5
+ "<|im_end|>",
6
+ "<repo_name>",
7
+ "<reponame>",
8
+ "<file_sep>",
9
+ "<filename>",
10
+ "<gh_stars>",
11
+ "<issue_start>",
12
+ "<issue_comment>",
13
+ "<issue_closed>",
14
+ "<jupyter_start>",
15
+ "<jupyter_text>",
16
+ "<jupyter_code>",
17
+ "<jupyter_output>",
18
+ "<jupyter_script>",
19
+ "<empty_output>"
20
  ],
21
+ "bos_token": {
22
+ "content": "<|endoftext|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false
27
+ },
28
+ "eos_token": {
29
+ "content": "<|endoftext|>",
30
+ "lstrip": false,
31
+ "normalized": false,
32
+ "rstrip": false,
33
+ "single_word": false
34
+ },
35
+ "pad_token": "<|endoftext|>",
36
  "unk_token": {
37
  "content": "<|endoftext|>",
38
  "lstrip": false,
tokenizer_config.json CHANGED
@@ -139,15 +139,30 @@
139
  }
140
  },
141
  "additional_special_tokens": [
 
142
  "<|im_start|>",
143
- "<|im_end|>"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
144
  ],
145
- "bos_token": "<|im_start|>",
146
  "clean_up_tokenization_spaces": false,
147
- "eos_token": "<|im_end|>",
148
  "extra_special_tokens": {},
149
  "model_max_length": 8192,
150
- "pad_token": "<|im_end|>",
151
  "tokenizer_class": "GPT2Tokenizer",
152
  "unk_token": "<|endoftext|>",
153
  "vocab_size": 49152
 
139
  }
140
  },
141
  "additional_special_tokens": [
142
+ "<|endoftext|>",
143
  "<|im_start|>",
144
+ "<|im_end|>",
145
+ "<repo_name>",
146
+ "<reponame>",
147
+ "<file_sep>",
148
+ "<filename>",
149
+ "<gh_stars>",
150
+ "<issue_start>",
151
+ "<issue_comment>",
152
+ "<issue_closed>",
153
+ "<jupyter_start>",
154
+ "<jupyter_text>",
155
+ "<jupyter_code>",
156
+ "<jupyter_output>",
157
+ "<jupyter_script>",
158
+ "<empty_output>"
159
  ],
160
+ "bos_token": "<|endoftext|>",
161
  "clean_up_tokenization_spaces": false,
162
+ "eos_token": "<|endoftext|>",
163
  "extra_special_tokens": {},
164
  "model_max_length": 8192,
165
+ "pad_token": "<|endoftext|>",
166
  "tokenizer_class": "GPT2Tokenizer",
167
  "unk_token": "<|endoftext|>",
168
  "vocab_size": 49152
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9792dd56196fc092e61bf4d4b40a15ba6cc93bcfd6285722f47c6f31460dad12
3
  size 6161
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:20db55b968a3c53f5dde9056413f58b86eb9ddf3bec0abe9827977712150c8ee
3
  size 6161