devJy commited on
Commit
0302d1f
·
verified ·
1 Parent(s): 7724bfb

RF Using GRPO

Browse files
config.json CHANGED
@@ -27,7 +27,7 @@
27
  "tie_word_embeddings": true,
28
  "torch_dtype": "bfloat16",
29
  "transformers_version": "4.51.3",
30
- "unsloth_version": "2025.4.1",
31
  "use_cache": true,
32
  "vocab_size": 110592
33
  }
 
27
  "tie_word_embeddings": true,
28
  "torch_dtype": "bfloat16",
29
  "transformers_version": "4.51.3",
30
+ "unsloth_version": "2025.4.3",
31
  "use_cache": true,
32
  "vocab_size": 110592
33
  }
generation_config.json CHANGED
@@ -3,6 +3,6 @@
3
  "bos_token_id": 100257,
4
  "eos_token_id": 100257,
5
  "max_length": 131072,
6
- "pad_token_id": 100257,
7
  "transformers_version": "4.51.3"
8
  }
 
3
  "bos_token_id": 100257,
4
  "eos_token_id": 100257,
5
  "max_length": 131072,
6
+ "pad_token_id": 0,
7
  "transformers_version": "4.51.3"
8
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c8387d63b6c1e210c4adf853759fe5fd0ee86d94198e6ed8e87552ca05dd3cf5
3
  size 3171119736
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9618b9b211c9b614698c288c6a30baa0cf6a4d78036dfd5e8a88bc593c32ef42
3
  size 3171119736
special_tokens_map.json CHANGED
@@ -1,6 +1,86 @@
1
  {
2
- "bos_token": "<|endoftext|>",
3
- "eos_token": "<|im_end|>",
4
- "pad_token": "<|endoftext|>",
5
- "unk_token": "<|endoftext|>"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6
  }
 
1
  {
2
+ "additional_special_tokens": [
3
+ "<|endoftext|>",
4
+ "<|fim_prefix|>",
5
+ "<|fim_middle|>",
6
+ "<|fim_suffix|>",
7
+ "<|endofprompt|>",
8
+ "<|_unuse_missing_100256|>",
9
+ "<|_unuse_missing_100261|>",
10
+ "<|_unuse_missing_100262|>",
11
+ "<|_unuse_missing_100263|>",
12
+ "<|_unuse_missing_100264|>",
13
+ "<|_unuse_missing_100265|>",
14
+ "<|_unuse_missing_100266|>",
15
+ "<|_unuse_missing_100267|>",
16
+ "<|_unuse_missing_100268|>",
17
+ "<|_unuse_missing_100269|>",
18
+ "<|_unuse_missing_100270|>",
19
+ "<|_unuse_missing_100271|>",
20
+ "<|im_start|>",
21
+ "<|im_end|>",
22
+ "<|stop|>",
23
+ "<|endofturn|>",
24
+ "<repo_name>",
25
+ "<file_sep>",
26
+ "<issue_start>",
27
+ "<issue_comment>",
28
+ "<issue_closed>",
29
+ "<jupyter_start>",
30
+ "<jupyter_text>",
31
+ "<jupyter_code>",
32
+ "<jupyter_output>",
33
+ "<jupyter_script>",
34
+ "<empty_output>",
35
+ "<code_to_intermediate>",
36
+ "<intermediate_to_code>",
37
+ "<pr>",
38
+ "<pr_status>",
39
+ "<pr_is_merged>",
40
+ "<pr_base>",
41
+ "<pr_file>",
42
+ "<pr_base_code>",
43
+ "<pr_diff>",
44
+ "<pr_diff_hunk>",
45
+ "<pr_comment>",
46
+ "<pr_event_id>",
47
+ "<pr_review>",
48
+ "<pr_review_state>",
49
+ "<pr_review_comment>",
50
+ "<pr_in_reply_to_review_id>",
51
+ "<pr_in_reply_to_comment_id>",
52
+ "<pr_diff_hunk_comment_line>",
53
+ "<NAME>",
54
+ "<EMAIL>",
55
+ "<KEY>",
56
+ "<PASSWORD>"
57
+ ],
58
+ "bos_token": {
59
+ "content": "<|endoftext|>",
60
+ "lstrip": false,
61
+ "normalized": false,
62
+ "rstrip": false,
63
+ "single_word": false
64
+ },
65
+ "eos_token": {
66
+ "content": "<|endofturn|>",
67
+ "lstrip": false,
68
+ "normalized": false,
69
+ "rstrip": false,
70
+ "single_word": false
71
+ },
72
+ "pad_token": {
73
+ "content": "<|endoftext|>",
74
+ "lstrip": false,
75
+ "normalized": false,
76
+ "rstrip": false,
77
+ "single_word": false
78
+ },
79
+ "unk_token": {
80
+ "content": "<|endoftext|>",
81
+ "lstrip": false,
82
+ "normalized": false,
83
+ "rstrip": false,
84
+ "single_word": false
85
+ }
86
  }
tokenizer.json CHANGED
@@ -158,7 +158,7 @@
158
  },
159
  {
160
  "id": 100273,
161
- "content": "<|endofturn|>",
162
  "single_word": false,
163
  "lstrip": false,
164
  "rstrip": false,
@@ -176,7 +176,7 @@
176
  },
177
  {
178
  "id": 100275,
179
- "content": "<|im_end|>",
180
  "single_word": false,
181
  "lstrip": false,
182
  "rstrip": false,
@@ -100807,9 +100807,9 @@
100807
  "<|_unuse_missing_100270|>": 100270,
100808
  "<|_unuse_missing_100271|>": 100271,
100809
  "<|im_start|>": 100272,
100810
- "<|endofturn|>": 100273,
100811
  "<|stop|>": 100274,
100812
- "<|im_end|>": 100275,
100813
  "<|endofprompt|>": 100276,
100814
  "Ġì§ĢìĽIJíķĺëĬĶ": 100277,
100815
  "ĠíķłìķĦë²Ħ": 100278,
 
158
  },
159
  {
160
  "id": 100273,
161
+ "content": "<|im_end|>",
162
  "single_word": false,
163
  "lstrip": false,
164
  "rstrip": false,
 
176
  },
177
  {
178
  "id": 100275,
179
+ "content": "<|endofturn|>",
180
  "single_word": false,
181
  "lstrip": false,
182
  "rstrip": false,
 
100807
  "<|_unuse_missing_100270|>": 100270,
100808
  "<|_unuse_missing_100271|>": 100271,
100809
  "<|im_start|>": 100272,
100810
+ "<|im_end|>": 100273,
100811
  "<|stop|>": 100274,
100812
+ "<|endofturn|>": 100275,
100813
  "<|endofprompt|>": 100276,
100814
  "Ġì§ĢìĽIJíķĺëĬĶ": 100277,
100815
  "ĠíķłìķĦë²Ħ": 100278,
tokenizer_config.json CHANGED
@@ -138,7 +138,7 @@
138
  "special": true
139
  },
140
  "100273": {
141
- "content": "<|endofturn|>",
142
  "lstrip": false,
143
  "normalized": false,
144
  "rstrip": false,
@@ -154,7 +154,7 @@
154
  "special": true
155
  },
156
  "100275": {
157
- "content": "<|im_end|>",
158
  "lstrip": false,
159
  "normalized": false,
160
  "rstrip": false,
@@ -434,13 +434,70 @@
434
  "special": true
435
  }
436
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
437
  "bos_token": "<|endoftext|>",
438
- "chat_template": "{% for message in messages %}{% if message['role'] == 'user' %}{{'<|im_start|>user\n' + message['content'] + '<|im_end|>\n'}}{% elif message['role'] == 'assistant' %}{{'<|im_start|>assistant\n' + message['content'] + '<|im_end|>\n' }}{% else %}{{ '<|im_start|>system\n' + message['content'] + '<|im_end|>\n' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
439
- "clean_up_tokenization_spaces": false,
440
- "eos_token": "<|im_end|>",
441
  "extra_special_tokens": {},
442
- "model_max_length": 1000000000000000019884624838656,
443
  "pad_token": "<|endoftext|>",
 
444
  "tokenizer_class": "GPT2Tokenizer",
445
  "unk_token": "<|endoftext|>"
446
  }
 
138
  "special": true
139
  },
140
  "100273": {
141
+ "content": "<|im_end|>",
142
  "lstrip": false,
143
  "normalized": false,
144
  "rstrip": false,
 
154
  "special": true
155
  },
156
  "100275": {
157
+ "content": "<|endofturn|>",
158
  "lstrip": false,
159
  "normalized": false,
160
  "rstrip": false,
 
434
  "special": true
435
  }
436
  },
437
+ "additional_special_tokens": [
438
+ "<|endoftext|>",
439
+ "<|fim_prefix|>",
440
+ "<|fim_middle|>",
441
+ "<|fim_suffix|>",
442
+ "<|endofprompt|>",
443
+ "<|_unuse_missing_100256|>",
444
+ "<|_unuse_missing_100261|>",
445
+ "<|_unuse_missing_100262|>",
446
+ "<|_unuse_missing_100263|>",
447
+ "<|_unuse_missing_100264|>",
448
+ "<|_unuse_missing_100265|>",
449
+ "<|_unuse_missing_100266|>",
450
+ "<|_unuse_missing_100267|>",
451
+ "<|_unuse_missing_100268|>",
452
+ "<|_unuse_missing_100269|>",
453
+ "<|_unuse_missing_100270|>",
454
+ "<|_unuse_missing_100271|>",
455
+ "<|im_start|>",
456
+ "<|im_end|>",
457
+ "<|stop|>",
458
+ "<|endofturn|>",
459
+ "<repo_name>",
460
+ "<file_sep>",
461
+ "<issue_start>",
462
+ "<issue_comment>",
463
+ "<issue_closed>",
464
+ "<jupyter_start>",
465
+ "<jupyter_text>",
466
+ "<jupyter_code>",
467
+ "<jupyter_output>",
468
+ "<jupyter_script>",
469
+ "<empty_output>",
470
+ "<code_to_intermediate>",
471
+ "<intermediate_to_code>",
472
+ "<pr>",
473
+ "<pr_status>",
474
+ "<pr_is_merged>",
475
+ "<pr_base>",
476
+ "<pr_file>",
477
+ "<pr_base_code>",
478
+ "<pr_diff>",
479
+ "<pr_diff_hunk>",
480
+ "<pr_comment>",
481
+ "<pr_event_id>",
482
+ "<pr_review>",
483
+ "<pr_review_state>",
484
+ "<pr_review_comment>",
485
+ "<pr_in_reply_to_review_id>",
486
+ "<pr_in_reply_to_comment_id>",
487
+ "<pr_diff_hunk_comment_line>",
488
+ "<NAME>",
489
+ "<EMAIL>",
490
+ "<KEY>",
491
+ "<PASSWORD>"
492
+ ],
493
  "bos_token": "<|endoftext|>",
494
+ "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
495
+ "clean_up_tokenization_spaces": true,
496
+ "eos_token": "<|endofturn|>",
497
  "extra_special_tokens": {},
498
+ "model_max_length": 131072,
499
  "pad_token": "<|endoftext|>",
500
+ "padding_side": "left",
501
  "tokenizer_class": "GPT2Tokenizer",
502
  "unk_token": "<|endoftext|>"
503
  }
vocab.json CHANGED
The diff for this file is too large to render. See raw diff