tocico28 commited on
Commit
2a09d72
·
verified ·
1 Parent(s): 26a44ff

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +4 -0
  2. checkpoint-1000/README.md +202 -0
  3. checkpoint-1000/adapter_config.json +39 -0
  4. checkpoint-1000/adapter_model.safetensors +3 -0
  5. checkpoint-1000/added_tokens.json +28 -0
  6. checkpoint-1000/merges.txt +0 -0
  7. checkpoint-1000/optimizer.pt +3 -0
  8. checkpoint-1000/rng_state.pth +3 -0
  9. checkpoint-1000/scaler.pt +3 -0
  10. checkpoint-1000/scheduler.pt +3 -0
  11. checkpoint-1000/special_tokens_map.json +31 -0
  12. checkpoint-1000/tokenizer.json +3 -0
  13. checkpoint-1000/tokenizer_config.json +240 -0
  14. checkpoint-1000/trainer_state.json +1534 -0
  15. checkpoint-1000/training_args.bin +3 -0
  16. checkpoint-1000/vocab.json +0 -0
  17. checkpoint-1500/README.md +202 -0
  18. checkpoint-1500/adapter_config.json +39 -0
  19. checkpoint-1500/adapter_model.safetensors +3 -0
  20. checkpoint-1500/added_tokens.json +28 -0
  21. checkpoint-1500/merges.txt +0 -0
  22. checkpoint-1500/optimizer.pt +3 -0
  23. checkpoint-1500/rng_state.pth +3 -0
  24. checkpoint-1500/scaler.pt +3 -0
  25. checkpoint-1500/scheduler.pt +3 -0
  26. checkpoint-1500/special_tokens_map.json +31 -0
  27. checkpoint-1500/tokenizer.json +3 -0
  28. checkpoint-1500/tokenizer_config.json +240 -0
  29. checkpoint-1500/trainer_state.json +2284 -0
  30. checkpoint-1500/training_args.bin +3 -0
  31. checkpoint-1500/vocab.json +0 -0
  32. checkpoint-1833/README.md +202 -0
  33. checkpoint-1833/adapter_config.json +39 -0
  34. checkpoint-1833/adapter_model.safetensors +3 -0
  35. checkpoint-1833/added_tokens.json +28 -0
  36. checkpoint-1833/merges.txt +0 -0
  37. checkpoint-1833/optimizer.pt +3 -0
  38. checkpoint-1833/rng_state.pth +3 -0
  39. checkpoint-1833/scaler.pt +3 -0
  40. checkpoint-1833/scheduler.pt +3 -0
  41. checkpoint-1833/special_tokens_map.json +31 -0
  42. checkpoint-1833/tokenizer.json +3 -0
  43. checkpoint-1833/tokenizer_config.json +240 -0
  44. checkpoint-1833/trainer_state.json +0 -0
  45. checkpoint-1833/training_args.bin +3 -0
  46. checkpoint-1833/vocab.json +0 -0
  47. checkpoint-500/README.md +202 -0
  48. checkpoint-500/adapter_config.json +39 -0
  49. checkpoint-500/adapter_model.safetensors +3 -0
  50. checkpoint-500/added_tokens.json +28 -0
.gitattributes CHANGED
@@ -33,3 +33,7 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ checkpoint-1000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
+ checkpoint-1500/tokenizer.json filter=lfs diff=lfs merge=lfs -text
38
+ checkpoint-1833/tokenizer.json filter=lfs diff=lfs merge=lfs -text
39
+ checkpoint-500/tokenizer.json filter=lfs diff=lfs merge=lfs -text
checkpoint-1000/README.md ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: unsloth/qwen3-0.6b-base-unsloth-bnb-4bit
3
+ library_name: peft
4
+ ---
5
+
6
+ # Model Card for Model ID
7
+
8
+ <!-- Provide a quick summary of what the model is/does. -->
9
+
10
+
11
+
12
+ ## Model Details
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+
19
+
20
+ - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
+ - **Shared by [optional]:** [More Information Needed]
23
+ - **Model type:** [More Information Needed]
24
+ - **Language(s) (NLP):** [More Information Needed]
25
+ - **License:** [More Information Needed]
26
+ - **Finetuned from model [optional]:** [More Information Needed]
27
+
28
+ ### Model Sources [optional]
29
+
30
+ <!-- Provide the basic links for the model. -->
31
+
32
+ - **Repository:** [More Information Needed]
33
+ - **Paper [optional]:** [More Information Needed]
34
+ - **Demo [optional]:** [More Information Needed]
35
+
36
+ ## Uses
37
+
38
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
+
40
+ ### Direct Use
41
+
42
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
+
44
+ [More Information Needed]
45
+
46
+ ### Downstream Use [optional]
47
+
48
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Out-of-Scope Use
53
+
54
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
+
56
+ [More Information Needed]
57
+
58
+ ## Bias, Risks, and Limitations
59
+
60
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ### Recommendations
65
+
66
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
+
68
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ Use the code below to get started with the model.
73
+
74
+ [More Information Needed]
75
+
76
+ ## Training Details
77
+
78
+ ### Training Data
79
+
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
+
82
+ [More Information Needed]
83
+
84
+ ### Training Procedure
85
+
86
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
+
88
+ #### Preprocessing [optional]
89
+
90
+ [More Information Needed]
91
+
92
+
93
+ #### Training Hyperparameters
94
+
95
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
+
97
+ #### Speeds, Sizes, Times [optional]
98
+
99
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
+
101
+ [More Information Needed]
102
+
103
+ ## Evaluation
104
+
105
+ <!-- This section describes the evaluation protocols and provides the results. -->
106
+
107
+ ### Testing Data, Factors & Metrics
108
+
109
+ #### Testing Data
110
+
111
+ <!-- This should link to a Dataset Card if possible. -->
112
+
113
+ [More Information Needed]
114
+
115
+ #### Factors
116
+
117
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Metrics
122
+
123
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
+
125
+ [More Information Needed]
126
+
127
+ ### Results
128
+
129
+ [More Information Needed]
130
+
131
+ #### Summary
132
+
133
+
134
+
135
+ ## Model Examination [optional]
136
+
137
+ <!-- Relevant interpretability work for the model goes here -->
138
+
139
+ [More Information Needed]
140
+
141
+ ## Environmental Impact
142
+
143
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
+
145
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
+
147
+ - **Hardware Type:** [More Information Needed]
148
+ - **Hours used:** [More Information Needed]
149
+ - **Cloud Provider:** [More Information Needed]
150
+ - **Compute Region:** [More Information Needed]
151
+ - **Carbon Emitted:** [More Information Needed]
152
+
153
+ ## Technical Specifications [optional]
154
+
155
+ ### Model Architecture and Objective
156
+
157
+ [More Information Needed]
158
+
159
+ ### Compute Infrastructure
160
+
161
+ [More Information Needed]
162
+
163
+ #### Hardware
164
+
165
+ [More Information Needed]
166
+
167
+ #### Software
168
+
169
+ [More Information Needed]
170
+
171
+ ## Citation [optional]
172
+
173
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
+
175
+ **BibTeX:**
176
+
177
+ [More Information Needed]
178
+
179
+ **APA:**
180
+
181
+ [More Information Needed]
182
+
183
+ ## Glossary [optional]
184
+
185
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
+
187
+ [More Information Needed]
188
+
189
+ ## More Information [optional]
190
+
191
+ [More Information Needed]
192
+
193
+ ## Model Card Authors [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Contact
198
+
199
+ [More Information Needed]
200
+ ### Framework versions
201
+
202
+ - PEFT 0.15.2
checkpoint-1000/adapter_config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "unsloth/qwen3-0.6b-base-unsloth-bnb-4bit",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 16,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "r": 16,
24
+ "rank_pattern": {},
25
+ "revision": null,
26
+ "target_modules": [
27
+ "q_proj",
28
+ "up_proj",
29
+ "down_proj",
30
+ "o_proj",
31
+ "gate_proj",
32
+ "k_proj",
33
+ "v_proj"
34
+ ],
35
+ "task_type": "CAUSAL_LM",
36
+ "trainable_token_indices": null,
37
+ "use_dora": false,
38
+ "use_rslora": false
39
+ }
checkpoint-1000/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f5baac6a5f0170a892f732eedad54637771d5e578f9643c8eecdb5b701314928
3
+ size 40422168
checkpoint-1000/added_tokens.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</think>": 151668,
3
+ "</tool_call>": 151658,
4
+ "</tool_response>": 151666,
5
+ "<think>": 151667,
6
+ "<tool_call>": 151657,
7
+ "<tool_response>": 151665,
8
+ "<|box_end|>": 151649,
9
+ "<|box_start|>": 151648,
10
+ "<|endoftext|>": 151643,
11
+ "<|file_sep|>": 151664,
12
+ "<|fim_middle|>": 151660,
13
+ "<|fim_pad|>": 151662,
14
+ "<|fim_prefix|>": 151659,
15
+ "<|fim_suffix|>": 151661,
16
+ "<|im_end|>": 151645,
17
+ "<|im_start|>": 151644,
18
+ "<|image_pad|>": 151655,
19
+ "<|object_ref_end|>": 151647,
20
+ "<|object_ref_start|>": 151646,
21
+ "<|quad_end|>": 151651,
22
+ "<|quad_start|>": 151650,
23
+ "<|repo_name|>": 151663,
24
+ "<|video_pad|>": 151656,
25
+ "<|vision_end|>": 151653,
26
+ "<|vision_pad|>": 151654,
27
+ "<|vision_start|>": 151652
28
+ }
checkpoint-1000/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-1000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1fa537cf6e90f70723512ae3157f8b7aafe0e6e6f1b82f71b77a62f224fd399
3
+ size 20959365
checkpoint-1000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c800b778fa7e115e4c34de8529902de8b61c9a1b4bab3eb8295d06dafff030e
3
+ size 14645
checkpoint-1000/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64ea44886479de845f62bfe26343b4b4fac05aabefaba1c4f934cff2d66d9d47
3
+ size 1383
checkpoint-1000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:341de82911758a81e7a119c105bbdd42c1129fec9388a31aef12a3974dc1cab7
3
+ size 1465
checkpoint-1000/special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|endoftext|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|vision_pad|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
checkpoint-1000/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aeb13307a71acd8fe81861d94ad54ab689df773318809eed3cbe794b4492dae4
3
+ size 11422654
checkpoint-1000/tokenizer_config.json ADDED
@@ -0,0 +1,240 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ },
181
+ "151665": {
182
+ "content": "<tool_response>",
183
+ "lstrip": false,
184
+ "normalized": false,
185
+ "rstrip": false,
186
+ "single_word": false,
187
+ "special": false
188
+ },
189
+ "151666": {
190
+ "content": "</tool_response>",
191
+ "lstrip": false,
192
+ "normalized": false,
193
+ "rstrip": false,
194
+ "single_word": false,
195
+ "special": false
196
+ },
197
+ "151667": {
198
+ "content": "<think>",
199
+ "lstrip": false,
200
+ "normalized": false,
201
+ "rstrip": false,
202
+ "single_word": false,
203
+ "special": false
204
+ },
205
+ "151668": {
206
+ "content": "</think>",
207
+ "lstrip": false,
208
+ "normalized": false,
209
+ "rstrip": false,
210
+ "single_word": false,
211
+ "special": false
212
+ }
213
+ },
214
+ "additional_special_tokens": [
215
+ "<|im_start|>",
216
+ "<|im_end|>",
217
+ "<|object_ref_start|>",
218
+ "<|object_ref_end|>",
219
+ "<|box_start|>",
220
+ "<|box_end|>",
221
+ "<|quad_start|>",
222
+ "<|quad_end|>",
223
+ "<|vision_start|>",
224
+ "<|vision_end|>",
225
+ "<|vision_pad|>",
226
+ "<|image_pad|>",
227
+ "<|video_pad|>"
228
+ ],
229
+ "bos_token": null,
230
+ "clean_up_tokenization_spaces": false,
231
+ "eos_token": "<|endoftext|>",
232
+ "errors": "replace",
233
+ "extra_special_tokens": {},
234
+ "model_max_length": 32768,
235
+ "pad_token": "<|vision_pad|>",
236
+ "padding_side": "right",
237
+ "split_special_tokens": false,
238
+ "tokenizer_class": "Qwen2Tokenizer",
239
+ "unk_token": null
240
+ }
checkpoint-1000/trainer_state.json ADDED
@@ -0,0 +1,1534 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 1.6366612111292962,
6
+ "eval_steps": 500,
7
+ "global_step": 1000,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.016366612111292964,
14
+ "grad_norm": NaN,
15
+ "learning_rate": 1.6304347826086957e-06,
16
+ "logits/chosen": -1.2515310049057007,
17
+ "logits/rejected": -0.8573959469795227,
18
+ "logps/chosen": -455.1468811035156,
19
+ "logps/rejected": -387.0008239746094,
20
+ "loss": 0.6923,
21
+ "rewards/accuracies": 0.44999998807907104,
22
+ "rewards/chosen": 0.0002114105736836791,
23
+ "rewards/margins": 0.00180077797267586,
24
+ "rewards/rejected": -0.0015893673989921808,
25
+ "step": 10
26
+ },
27
+ {
28
+ "epoch": 0.03273322422258593,
29
+ "grad_norm": 10.154256820678711,
30
+ "learning_rate": 3.2608695652173914e-06,
31
+ "logits/chosen": -1.145971417427063,
32
+ "logits/rejected": -1.082183837890625,
33
+ "logps/chosen": -446.249755859375,
34
+ "logps/rejected": -426.87237548828125,
35
+ "loss": 0.6942,
36
+ "rewards/accuracies": 0.375,
37
+ "rewards/chosen": -0.0005732345744036138,
38
+ "rewards/margins": -0.002082281280308962,
39
+ "rewards/rejected": 0.0015090465312823653,
40
+ "step": 20
41
+ },
42
+ {
43
+ "epoch": 0.049099836333878884,
44
+ "grad_norm": 8.393162727355957,
45
+ "learning_rate": 5.4347826086956525e-06,
46
+ "logits/chosen": -0.9614235162734985,
47
+ "logits/rejected": -0.8527517318725586,
48
+ "logps/chosen": -460.367431640625,
49
+ "logps/rejected": -433.61285400390625,
50
+ "loss": 0.6934,
51
+ "rewards/accuracies": 0.4625000059604645,
52
+ "rewards/chosen": 0.0062791816890239716,
53
+ "rewards/margins": -0.00045429245801642537,
54
+ "rewards/rejected": 0.006733474787324667,
55
+ "step": 30
56
+ },
57
+ {
58
+ "epoch": 0.06546644844517185,
59
+ "grad_norm": 9.398829460144043,
60
+ "learning_rate": 8.15217391304348e-06,
61
+ "logits/chosen": -1.171631097793579,
62
+ "logits/rejected": -1.0616101026535034,
63
+ "logps/chosen": -421.583984375,
64
+ "logps/rejected": -424.787109375,
65
+ "loss": 0.6902,
66
+ "rewards/accuracies": 0.48750001192092896,
67
+ "rewards/chosen": 0.020799484103918076,
68
+ "rewards/margins": 0.006221642717719078,
69
+ "rewards/rejected": 0.014577841386198997,
70
+ "step": 40
71
+ },
72
+ {
73
+ "epoch": 0.08183306055646482,
74
+ "grad_norm": 9.43005657196045,
75
+ "learning_rate": 1.0869565217391305e-05,
76
+ "logits/chosen": -1.2450587749481201,
77
+ "logits/rejected": -1.1417832374572754,
78
+ "logps/chosen": -476.83538818359375,
79
+ "logps/rejected": -444.6556091308594,
80
+ "loss": 0.6883,
81
+ "rewards/accuracies": 0.5,
82
+ "rewards/chosen": 0.042164258658885956,
83
+ "rewards/margins": 0.010725794360041618,
84
+ "rewards/rejected": 0.03143846243619919,
85
+ "step": 50
86
+ },
87
+ {
88
+ "epoch": 0.09819967266775777,
89
+ "grad_norm": 8.825221061706543,
90
+ "learning_rate": 1.331521739130435e-05,
91
+ "logits/chosen": -1.3223823308944702,
92
+ "logits/rejected": NaN,
93
+ "logps/chosen": -383.4811706542969,
94
+ "logps/rejected": -454.43780517578125,
95
+ "loss": 0.6965,
96
+ "rewards/accuracies": 0.48750001192092896,
97
+ "rewards/chosen": 0.03668512776494026,
98
+ "rewards/margins": -0.003995636478066444,
99
+ "rewards/rejected": 0.04068076238036156,
100
+ "step": 60
101
+ },
102
+ {
103
+ "epoch": 0.11456628477905073,
104
+ "grad_norm": 9.5093994140625,
105
+ "learning_rate": 1.6032608695652173e-05,
106
+ "logits/chosen": -1.2538468837738037,
107
+ "logits/rejected": -1.2771422863006592,
108
+ "logps/chosen": -468.79400634765625,
109
+ "logps/rejected": -402.565673828125,
110
+ "loss": 0.6864,
111
+ "rewards/accuracies": 0.5,
112
+ "rewards/chosen": 0.04952714219689369,
113
+ "rewards/margins": 0.01824185810983181,
114
+ "rewards/rejected": 0.03128528222441673,
115
+ "step": 70
116
+ },
117
+ {
118
+ "epoch": 0.1309328968903437,
119
+ "grad_norm": 7.945922374725342,
120
+ "learning_rate": 1.8750000000000002e-05,
121
+ "logits/chosen": -1.386731743812561,
122
+ "logits/rejected": -0.8966825604438782,
123
+ "logps/chosen": -446.54736328125,
124
+ "logps/rejected": -430.45025634765625,
125
+ "loss": 0.6719,
126
+ "rewards/accuracies": 0.612500011920929,
127
+ "rewards/chosen": 0.06767664849758148,
128
+ "rewards/margins": 0.048766594380140305,
129
+ "rewards/rejected": 0.01891004480421543,
130
+ "step": 80
131
+ },
132
+ {
133
+ "epoch": 0.14729950900163666,
134
+ "grad_norm": 8.496461868286133,
135
+ "learning_rate": 2.1467391304347828e-05,
136
+ "logits/chosen": -1.3730494976043701,
137
+ "logits/rejected": -1.1826181411743164,
138
+ "logps/chosen": -413.87835693359375,
139
+ "logps/rejected": -385.1138916015625,
140
+ "loss": 0.69,
141
+ "rewards/accuracies": 0.48750001192092896,
142
+ "rewards/chosen": 0.07122664153575897,
143
+ "rewards/margins": 0.014941570349037647,
144
+ "rewards/rejected": 0.0562850646674633,
145
+ "step": 90
146
+ },
147
+ {
148
+ "epoch": 0.16366612111292964,
149
+ "grad_norm": 10.744088172912598,
150
+ "learning_rate": 2.4184782608695653e-05,
151
+ "logits/chosen": -1.2254825830459595,
152
+ "logits/rejected": -1.1136434078216553,
153
+ "logps/chosen": -495.34295654296875,
154
+ "logps/rejected": -435.0003967285156,
155
+ "loss": 0.6618,
156
+ "rewards/accuracies": 0.612500011920929,
157
+ "rewards/chosen": 0.1682208627462387,
158
+ "rewards/margins": 0.08198796212673187,
159
+ "rewards/rejected": 0.08623288571834564,
160
+ "step": 100
161
+ },
162
+ {
163
+ "epoch": 0.18003273322422259,
164
+ "grad_norm": 11.016289710998535,
165
+ "learning_rate": 2.6902173913043476e-05,
166
+ "logits/chosen": -1.3138881921768188,
167
+ "logits/rejected": -1.0289314985275269,
168
+ "logps/chosen": -476.26153564453125,
169
+ "logps/rejected": -426.837646484375,
170
+ "loss": 0.6355,
171
+ "rewards/accuracies": 0.6499999761581421,
172
+ "rewards/chosen": 0.21744687855243683,
173
+ "rewards/margins": 0.15108919143676758,
174
+ "rewards/rejected": 0.06635768711566925,
175
+ "step": 110
176
+ },
177
+ {
178
+ "epoch": 0.19639934533551553,
179
+ "grad_norm": 11.619248390197754,
180
+ "learning_rate": 2.9619565217391305e-05,
181
+ "logits/chosen": -1.2751747369766235,
182
+ "logits/rejected": -1.1858090162277222,
183
+ "logps/chosen": -444.4823303222656,
184
+ "logps/rejected": -404.58123779296875,
185
+ "loss": 0.6716,
186
+ "rewards/accuracies": 0.550000011920929,
187
+ "rewards/chosen": 0.2203199863433838,
188
+ "rewards/margins": 0.06618332862854004,
189
+ "rewards/rejected": 0.15413665771484375,
190
+ "step": 120
191
+ },
192
+ {
193
+ "epoch": 0.2127659574468085,
194
+ "grad_norm": 8.900009155273438,
195
+ "learning_rate": 3.233695652173913e-05,
196
+ "logits/chosen": -1.2365785837173462,
197
+ "logits/rejected": -1.2031922340393066,
198
+ "logps/chosen": -477.8089294433594,
199
+ "logps/rejected": -456.80120849609375,
200
+ "loss": 0.6794,
201
+ "rewards/accuracies": 0.5625,
202
+ "rewards/chosen": 0.20705166459083557,
203
+ "rewards/margins": 0.060810189694166183,
204
+ "rewards/rejected": 0.1462414562702179,
205
+ "step": 130
206
+ },
207
+ {
208
+ "epoch": 0.22913256955810146,
209
+ "grad_norm": 8.928816795349121,
210
+ "learning_rate": 3.505434782608696e-05,
211
+ "logits/chosen": -1.2421447038650513,
212
+ "logits/rejected": -1.136252522468567,
213
+ "logps/chosen": -547.4909057617188,
214
+ "logps/rejected": -460.38275146484375,
215
+ "loss": 0.6047,
216
+ "rewards/accuracies": 0.6875,
217
+ "rewards/chosen": 0.21343636512756348,
218
+ "rewards/margins": 0.24689535796642303,
219
+ "rewards/rejected": -0.033459004014730453,
220
+ "step": 140
221
+ },
222
+ {
223
+ "epoch": 0.24549918166939444,
224
+ "grad_norm": 11.977150917053223,
225
+ "learning_rate": 3.777173913043478e-05,
226
+ "logits/chosen": -1.338332176208496,
227
+ "logits/rejected": -1.1015567779541016,
228
+ "logps/chosen": -432.2693786621094,
229
+ "logps/rejected": -423.0946350097656,
230
+ "loss": 0.6946,
231
+ "rewards/accuracies": 0.4749999940395355,
232
+ "rewards/chosen": 0.15119442343711853,
233
+ "rewards/margins": 0.07870660722255707,
234
+ "rewards/rejected": 0.07248781621456146,
235
+ "step": 150
236
+ },
237
+ {
238
+ "epoch": 0.2618657937806874,
239
+ "grad_norm": 13.040871620178223,
240
+ "learning_rate": 4.0489130434782615e-05,
241
+ "logits/chosen": -1.0111101865768433,
242
+ "logits/rejected": NaN,
243
+ "logps/chosen": -472.6917419433594,
244
+ "logps/rejected": -426.6333923339844,
245
+ "loss": 0.6598,
246
+ "rewards/accuracies": 0.5625,
247
+ "rewards/chosen": 0.24765905737876892,
248
+ "rewards/margins": 0.13647916913032532,
249
+ "rewards/rejected": 0.11117993295192719,
250
+ "step": 160
251
+ },
252
+ {
253
+ "epoch": 0.27823240589198034,
254
+ "grad_norm": 8.028325080871582,
255
+ "learning_rate": 4.3206521739130434e-05,
256
+ "logits/chosen": -0.9152072668075562,
257
+ "logits/rejected": -0.8340185284614563,
258
+ "logps/chosen": -438.38555908203125,
259
+ "logps/rejected": -392.93426513671875,
260
+ "loss": 0.6272,
261
+ "rewards/accuracies": 0.612500011920929,
262
+ "rewards/chosen": 0.3285876512527466,
263
+ "rewards/margins": 0.20840699970722198,
264
+ "rewards/rejected": 0.1201806515455246,
265
+ "step": 170
266
+ },
267
+ {
268
+ "epoch": 0.2945990180032733,
269
+ "grad_norm": 9.436712265014648,
270
+ "learning_rate": 4.592391304347826e-05,
271
+ "logits/chosen": -1.0703661441802979,
272
+ "logits/rejected": -1.0127884149551392,
273
+ "logps/chosen": -449.83905029296875,
274
+ "logps/rejected": -437.92822265625,
275
+ "loss": 0.6684,
276
+ "rewards/accuracies": 0.5375000238418579,
277
+ "rewards/chosen": 0.34186336398124695,
278
+ "rewards/margins": 0.14258861541748047,
279
+ "rewards/rejected": 0.19927474856376648,
280
+ "step": 180
281
+ },
282
+ {
283
+ "epoch": 0.3109656301145663,
284
+ "grad_norm": 8.914007186889648,
285
+ "learning_rate": 4.8641304347826086e-05,
286
+ "logits/chosen": -0.8762575387954712,
287
+ "logits/rejected": -0.5631248354911804,
288
+ "logps/chosen": -473.806396484375,
289
+ "logps/rejected": -443.6448669433594,
290
+ "loss": 0.6243,
291
+ "rewards/accuracies": 0.574999988079071,
292
+ "rewards/chosen": 0.3772050440311432,
293
+ "rewards/margins": 0.24287982285022736,
294
+ "rewards/rejected": 0.13432522118091583,
295
+ "step": 190
296
+ },
297
+ {
298
+ "epoch": 0.32733224222585927,
299
+ "grad_norm": 14.76502513885498,
300
+ "learning_rate": 4.98483929654336e-05,
301
+ "logits/chosen": NaN,
302
+ "logits/rejected": -0.9821416139602661,
303
+ "logps/chosen": -447.65386962890625,
304
+ "logps/rejected": -400.1768493652344,
305
+ "loss": 0.6805,
306
+ "rewards/accuracies": 0.5874999761581421,
307
+ "rewards/chosen": 0.2871859073638916,
308
+ "rewards/margins": 0.17125040292739868,
309
+ "rewards/rejected": 0.11593550443649292,
310
+ "step": 200
311
+ },
312
+ {
313
+ "epoch": 0.3436988543371522,
314
+ "grad_norm": 11.80213451385498,
315
+ "learning_rate": 4.954517889630079e-05,
316
+ "logits/chosen": -1.2535641193389893,
317
+ "logits/rejected": -1.0489479303359985,
318
+ "logps/chosen": -457.010986328125,
319
+ "logps/rejected": -432.8558044433594,
320
+ "loss": 0.6769,
321
+ "rewards/accuracies": 0.512499988079071,
322
+ "rewards/chosen": 0.24471895396709442,
323
+ "rewards/margins": 0.16205154359340668,
324
+ "rewards/rejected": 0.08266742527484894,
325
+ "step": 210
326
+ },
327
+ {
328
+ "epoch": 0.36006546644844517,
329
+ "grad_norm": 7.534430980682373,
330
+ "learning_rate": 4.924196482716799e-05,
331
+ "logits/chosen": -1.5050650835037231,
332
+ "logits/rejected": -1.5133517980575562,
333
+ "logps/chosen": -445.3924865722656,
334
+ "logps/rejected": -420.99114990234375,
335
+ "loss": 0.7003,
336
+ "rewards/accuracies": 0.5625,
337
+ "rewards/chosen": 0.18017885088920593,
338
+ "rewards/margins": 0.1387586146593094,
339
+ "rewards/rejected": 0.04142025485634804,
340
+ "step": 220
341
+ },
342
+ {
343
+ "epoch": 0.37643207855973815,
344
+ "grad_norm": 7.207632541656494,
345
+ "learning_rate": 4.893875075803517e-05,
346
+ "logits/chosen": -1.3496973514556885,
347
+ "logits/rejected": -1.2030962705612183,
348
+ "logps/chosen": -436.5039978027344,
349
+ "logps/rejected": -396.91534423828125,
350
+ "loss": 0.6192,
351
+ "rewards/accuracies": 0.574999988079071,
352
+ "rewards/chosen": 0.19479091465473175,
353
+ "rewards/margins": 0.353401243686676,
354
+ "rewards/rejected": -0.15861034393310547,
355
+ "step": 230
356
+ },
357
+ {
358
+ "epoch": 0.39279869067103107,
359
+ "grad_norm": 9.134820938110352,
360
+ "learning_rate": 4.8665858095815645e-05,
361
+ "logits/chosen": -1.2978178262710571,
362
+ "logits/rejected": -1.1864020824432373,
363
+ "logps/chosen": -486.7569885253906,
364
+ "logps/rejected": -457.94586181640625,
365
+ "loss": 0.6931,
366
+ "rewards/accuracies": 0.4749999940395355,
367
+ "rewards/chosen": 0.29108190536499023,
368
+ "rewards/margins": 0.15347807109355927,
369
+ "rewards/rejected": 0.13760384917259216,
370
+ "step": 240
371
+ },
372
+ {
373
+ "epoch": 0.40916530278232405,
374
+ "grad_norm": 10.825005531311035,
375
+ "learning_rate": 4.836264402668284e-05,
376
+ "logits/chosen": -1.3271738290786743,
377
+ "logits/rejected": -1.030890703201294,
378
+ "logps/chosen": -421.89349365234375,
379
+ "logps/rejected": -400.1674499511719,
380
+ "loss": 0.6536,
381
+ "rewards/accuracies": 0.5,
382
+ "rewards/chosen": 0.42532652616500854,
383
+ "rewards/margins": 0.34351587295532227,
384
+ "rewards/rejected": 0.08181065320968628,
385
+ "step": 250
386
+ },
387
+ {
388
+ "epoch": 0.425531914893617,
389
+ "grad_norm": 9.450189590454102,
390
+ "learning_rate": 4.8059429957550034e-05,
391
+ "logits/chosen": -1.0212231874465942,
392
+ "logits/rejected": -1.0058345794677734,
393
+ "logps/chosen": -422.1316833496094,
394
+ "logps/rejected": -390.0425109863281,
395
+ "loss": 0.6384,
396
+ "rewards/accuracies": 0.5874999761581421,
397
+ "rewards/chosen": 0.490867555141449,
398
+ "rewards/margins": 0.35965341329574585,
399
+ "rewards/rejected": 0.13121409714221954,
400
+ "step": 260
401
+ },
402
+ {
403
+ "epoch": 0.44189852700491,
404
+ "grad_norm": 8.231124877929688,
405
+ "learning_rate": 4.7756215888417225e-05,
406
+ "logits/chosen": -0.9735445976257324,
407
+ "logits/rejected": -0.937743067741394,
408
+ "logps/chosen": -552.4168701171875,
409
+ "logps/rejected": -501.0535583496094,
410
+ "loss": 0.6694,
411
+ "rewards/accuracies": 0.574999988079071,
412
+ "rewards/chosen": 0.4536367952823639,
413
+ "rewards/margins": 0.19916944205760956,
414
+ "rewards/rejected": 0.2544674277305603,
415
+ "step": 270
416
+ },
417
+ {
418
+ "epoch": 0.4582651391162029,
419
+ "grad_norm": 9.207930564880371,
420
+ "learning_rate": 4.7453001819284416e-05,
421
+ "logits/chosen": -0.9774928092956543,
422
+ "logits/rejected": -0.8485246896743774,
423
+ "logps/chosen": -444.54827880859375,
424
+ "logps/rejected": -419.5541076660156,
425
+ "loss": 0.6607,
426
+ "rewards/accuracies": 0.5874999761581421,
427
+ "rewards/chosen": 0.3690536618232727,
428
+ "rewards/margins": 0.22144463658332825,
429
+ "rewards/rejected": 0.14760902523994446,
430
+ "step": 280
431
+ },
432
+ {
433
+ "epoch": 0.4746317512274959,
434
+ "grad_norm": 6.876308441162109,
435
+ "learning_rate": 4.714978775015161e-05,
436
+ "logits/chosen": -1.2095867395401,
437
+ "logits/rejected": -0.8773800730705261,
438
+ "logps/chosen": -437.869384765625,
439
+ "logps/rejected": -419.9366149902344,
440
+ "loss": 0.5867,
441
+ "rewards/accuracies": 0.612500011920929,
442
+ "rewards/chosen": 0.5094749331474304,
443
+ "rewards/margins": 0.5250319242477417,
444
+ "rewards/rejected": -0.015556998550891876,
445
+ "step": 290
446
+ },
447
+ {
448
+ "epoch": 0.4909983633387889,
449
+ "grad_norm": 12.918950080871582,
450
+ "learning_rate": 4.68465736810188e-05,
451
+ "logits/chosen": -0.9076827764511108,
452
+ "logits/rejected": -0.7181829214096069,
453
+ "logps/chosen": -469.4864196777344,
454
+ "logps/rejected": -420.38885498046875,
455
+ "loss": 0.6833,
456
+ "rewards/accuracies": 0.5625,
457
+ "rewards/chosen": 0.5444287657737732,
458
+ "rewards/margins": 0.2743019461631775,
459
+ "rewards/rejected": 0.2701267600059509,
460
+ "step": 300
461
+ },
462
+ {
463
+ "epoch": 0.5073649754500819,
464
+ "grad_norm": 4.9246110916137695,
465
+ "learning_rate": 4.6543359611885996e-05,
466
+ "logits/chosen": -1.214247226715088,
467
+ "logits/rejected": -0.9910456538200378,
468
+ "logps/chosen": -385.4250183105469,
469
+ "logps/rejected": -390.26043701171875,
470
+ "loss": 0.6419,
471
+ "rewards/accuracies": 0.675000011920929,
472
+ "rewards/chosen": 0.4199371337890625,
473
+ "rewards/margins": 0.34462255239486694,
474
+ "rewards/rejected": 0.07531457394361496,
475
+ "step": 310
476
+ },
477
+ {
478
+ "epoch": 0.5237315875613748,
479
+ "grad_norm": 8.829249382019043,
480
+ "learning_rate": 4.624014554275319e-05,
481
+ "logits/chosen": -1.0396336317062378,
482
+ "logits/rejected": -0.9439867734909058,
483
+ "logps/chosen": -452.1678161621094,
484
+ "logps/rejected": -379.70745849609375,
485
+ "loss": 0.6396,
486
+ "rewards/accuracies": 0.5625,
487
+ "rewards/chosen": 0.472791850566864,
488
+ "rewards/margins": 0.23509594798088074,
489
+ "rewards/rejected": 0.23769588768482208,
490
+ "step": 320
491
+ },
492
+ {
493
+ "epoch": 0.5400981996726678,
494
+ "grad_norm": 13.249287605285645,
495
+ "learning_rate": 4.593693147362038e-05,
496
+ "logits/chosen": -0.927478015422821,
497
+ "logits/rejected": -0.7090943455696106,
498
+ "logps/chosen": -488.009765625,
499
+ "logps/rejected": -473.47833251953125,
500
+ "loss": 0.6833,
501
+ "rewards/accuracies": 0.550000011920929,
502
+ "rewards/chosen": 0.4911763668060303,
503
+ "rewards/margins": 0.25351986289024353,
504
+ "rewards/rejected": 0.23765654861927032,
505
+ "step": 330
506
+ },
507
+ {
508
+ "epoch": 0.5564648117839607,
509
+ "grad_norm": 9.580172538757324,
510
+ "learning_rate": 4.566403881140085e-05,
511
+ "logits/chosen": -0.9790302515029907,
512
+ "logits/rejected": -0.8282285928726196,
513
+ "logps/chosen": -481.6971740722656,
514
+ "logps/rejected": -441.3619689941406,
515
+ "loss": 0.6922,
516
+ "rewards/accuracies": 0.550000011920929,
517
+ "rewards/chosen": 0.4195849299430847,
518
+ "rewards/margins": 0.210377499461174,
519
+ "rewards/rejected": 0.20920750498771667,
520
+ "step": 340
521
+ },
522
+ {
523
+ "epoch": 0.5728314238952537,
524
+ "grad_norm": 6.872241973876953,
525
+ "learning_rate": 4.536082474226804e-05,
526
+ "logits/chosen": -0.9942235946655273,
527
+ "logits/rejected": -0.824662983417511,
528
+ "logps/chosen": -437.9869079589844,
529
+ "logps/rejected": -417.3473205566406,
530
+ "loss": 0.6071,
531
+ "rewards/accuracies": 0.6000000238418579,
532
+ "rewards/chosen": 0.49881643056869507,
533
+ "rewards/margins": 0.41778793931007385,
534
+ "rewards/rejected": 0.08102855831384659,
535
+ "step": 350
536
+ },
537
+ {
538
+ "epoch": 0.5891980360065466,
539
+ "grad_norm": 7.748345375061035,
540
+ "learning_rate": 4.5057610673135234e-05,
541
+ "logits/chosen": -1.0704625844955444,
542
+ "logits/rejected": -0.7879603505134583,
543
+ "logps/chosen": -428.53021240234375,
544
+ "logps/rejected": -435.9026794433594,
545
+ "loss": 0.6084,
546
+ "rewards/accuracies": 0.612500011920929,
547
+ "rewards/chosen": 0.4484317898750305,
548
+ "rewards/margins": 0.4125775396823883,
549
+ "rewards/rejected": 0.035854239016771317,
550
+ "step": 360
551
+ },
552
+ {
553
+ "epoch": 0.6055646481178396,
554
+ "grad_norm": 7.428864479064941,
555
+ "learning_rate": 4.475439660400243e-05,
556
+ "logits/chosen": -1.098026990890503,
557
+ "logits/rejected": -0.9232645034790039,
558
+ "logps/chosen": -441.7071228027344,
559
+ "logps/rejected": -402.95770263671875,
560
+ "loss": 0.6901,
561
+ "rewards/accuracies": 0.550000011920929,
562
+ "rewards/chosen": 0.37514036893844604,
563
+ "rewards/margins": 0.2525361478328705,
564
+ "rewards/rejected": 0.12260414659976959,
565
+ "step": 370
566
+ },
567
+ {
568
+ "epoch": 0.6219312602291326,
569
+ "grad_norm": 9.591557502746582,
570
+ "learning_rate": 4.445118253486962e-05,
571
+ "logits/chosen": -1.2409312725067139,
572
+ "logits/rejected": -0.8442489504814148,
573
+ "logps/chosen": -466.13311767578125,
574
+ "logps/rejected": -420.9102478027344,
575
+ "loss": 0.7684,
576
+ "rewards/accuracies": 0.4625000059604645,
577
+ "rewards/chosen": 0.2286566197872162,
578
+ "rewards/margins": 0.11457403004169464,
579
+ "rewards/rejected": 0.11408257484436035,
580
+ "step": 380
581
+ },
582
+ {
583
+ "epoch": 0.6382978723404256,
584
+ "grad_norm": 7.025293350219727,
585
+ "learning_rate": 4.414796846573681e-05,
586
+ "logits/chosen": -1.1844158172607422,
587
+ "logits/rejected": -1.1069843769073486,
588
+ "logps/chosen": -428.5450134277344,
589
+ "logps/rejected": -371.85614013671875,
590
+ "loss": 0.6921,
591
+ "rewards/accuracies": 0.550000011920929,
592
+ "rewards/chosen": 0.2637328803539276,
593
+ "rewards/margins": 0.10889965295791626,
594
+ "rewards/rejected": 0.15483322739601135,
595
+ "step": 390
596
+ },
597
+ {
598
+ "epoch": 0.6546644844517185,
599
+ "grad_norm": 19.2535343170166,
600
+ "learning_rate": 4.3844754396604006e-05,
601
+ "logits/chosen": -1.1141570806503296,
602
+ "logits/rejected": -0.8798454999923706,
603
+ "logps/chosen": -493.08197021484375,
604
+ "logps/rejected": -465.0943298339844,
605
+ "loss": 0.6545,
606
+ "rewards/accuracies": 0.5874999761581421,
607
+ "rewards/chosen": 0.4174181818962097,
608
+ "rewards/margins": 0.25259822607040405,
609
+ "rewards/rejected": 0.16481998562812805,
610
+ "step": 400
611
+ },
612
+ {
613
+ "epoch": 0.6710310965630114,
614
+ "grad_norm": 8.808697700500488,
615
+ "learning_rate": 4.35415403274712e-05,
616
+ "logits/chosen": -0.9878439903259277,
617
+ "logits/rejected": -0.7299278974533081,
618
+ "logps/chosen": -523.9139404296875,
619
+ "logps/rejected": -482.0597229003906,
620
+ "loss": 0.552,
621
+ "rewards/accuracies": 0.699999988079071,
622
+ "rewards/chosen": 0.4945390820503235,
623
+ "rewards/margins": 0.6175243854522705,
624
+ "rewards/rejected": -0.1229853481054306,
625
+ "step": 410
626
+ },
627
+ {
628
+ "epoch": 0.6873977086743044,
629
+ "grad_norm": 9.99456787109375,
630
+ "learning_rate": 4.323832625833839e-05,
631
+ "logits/chosen": -1.0902565717697144,
632
+ "logits/rejected": -0.9221500158309937,
633
+ "logps/chosen": -443.0076599121094,
634
+ "logps/rejected": -375.8013610839844,
635
+ "loss": 0.5684,
636
+ "rewards/accuracies": 0.5874999761581421,
637
+ "rewards/chosen": 0.46177297830581665,
638
+ "rewards/margins": 0.509793221950531,
639
+ "rewards/rejected": -0.048020295798778534,
640
+ "step": 420
641
+ },
642
+ {
643
+ "epoch": 0.7037643207855974,
644
+ "grad_norm": 5.767345428466797,
645
+ "learning_rate": 4.293511218920558e-05,
646
+ "logits/chosen": -1.0450439453125,
647
+ "logits/rejected": -0.7902520298957825,
648
+ "logps/chosen": -458.2330017089844,
649
+ "logps/rejected": -382.6787109375,
650
+ "loss": 0.6126,
651
+ "rewards/accuracies": 0.6499999761581421,
652
+ "rewards/chosen": 0.5793684124946594,
653
+ "rewards/margins": 0.35584205389022827,
654
+ "rewards/rejected": 0.22352632880210876,
655
+ "step": 430
656
+ },
657
+ {
658
+ "epoch": 0.7201309328968903,
659
+ "grad_norm": 11.51710033416748,
660
+ "learning_rate": 4.263189812007277e-05,
661
+ "logits/chosen": -0.8710979223251343,
662
+ "logits/rejected": -0.783151388168335,
663
+ "logps/chosen": -451.32122802734375,
664
+ "logps/rejected": -416.0005798339844,
665
+ "loss": 0.6751,
666
+ "rewards/accuracies": 0.5625,
667
+ "rewards/chosen": 0.5284687280654907,
668
+ "rewards/margins": 0.2338237315416336,
669
+ "rewards/rejected": 0.2946449816226959,
670
+ "step": 440
671
+ },
672
+ {
673
+ "epoch": 0.7364975450081833,
674
+ "grad_norm": 8.408599853515625,
675
+ "learning_rate": 4.232868405093997e-05,
676
+ "logits/chosen": -0.9611997604370117,
677
+ "logits/rejected": -0.8005222082138062,
678
+ "logps/chosen": -390.06817626953125,
679
+ "logps/rejected": -376.57183837890625,
680
+ "loss": 0.723,
681
+ "rewards/accuracies": 0.48750001192092896,
682
+ "rewards/chosen": 0.35626164078712463,
683
+ "rewards/margins": 0.07882154732942581,
684
+ "rewards/rejected": 0.27744007110595703,
685
+ "step": 450
686
+ },
687
+ {
688
+ "epoch": 0.7528641571194763,
689
+ "grad_norm": 7.085683822631836,
690
+ "learning_rate": 4.202546998180716e-05,
691
+ "logits/chosen": -1.181065559387207,
692
+ "logits/rejected": -0.9188621640205383,
693
+ "logps/chosen": -446.1695861816406,
694
+ "logps/rejected": -461.0537109375,
695
+ "loss": 0.6376,
696
+ "rewards/accuracies": 0.574999988079071,
697
+ "rewards/chosen": 0.6099992990493774,
698
+ "rewards/margins": 0.40477481484413147,
699
+ "rewards/rejected": 0.20522448420524597,
700
+ "step": 460
701
+ },
702
+ {
703
+ "epoch": 0.7692307692307693,
704
+ "grad_norm": 7.778029918670654,
705
+ "learning_rate": 4.172225591267435e-05,
706
+ "logits/chosen": -1.1612136363983154,
707
+ "logits/rejected": -0.7909379601478577,
708
+ "logps/chosen": -415.356201171875,
709
+ "logps/rejected": -405.0163879394531,
710
+ "loss": 0.639,
711
+ "rewards/accuracies": 0.512499988079071,
712
+ "rewards/chosen": 0.5554541349411011,
713
+ "rewards/margins": 0.49387890100479126,
714
+ "rewards/rejected": 0.06157524511218071,
715
+ "step": 470
716
+ },
717
+ {
718
+ "epoch": 0.7855973813420621,
719
+ "grad_norm": 9.144447326660156,
720
+ "learning_rate": 4.141904184354154e-05,
721
+ "logits/chosen": -1.1227318048477173,
722
+ "logits/rejected": -0.8980533480644226,
723
+ "logps/chosen": -470.1067810058594,
724
+ "logps/rejected": -385.7353210449219,
725
+ "loss": 0.6301,
726
+ "rewards/accuracies": 0.5874999761581421,
727
+ "rewards/chosen": 0.5011923909187317,
728
+ "rewards/margins": 0.36913150548934937,
729
+ "rewards/rejected": 0.1320609152317047,
730
+ "step": 480
731
+ },
732
+ {
733
+ "epoch": 0.8019639934533551,
734
+ "grad_norm": 9.528294563293457,
735
+ "learning_rate": 4.111582777440873e-05,
736
+ "logits/chosen": -1.4198505878448486,
737
+ "logits/rejected": -1.1528406143188477,
738
+ "logps/chosen": -430.28668212890625,
739
+ "logps/rejected": -385.2845764160156,
740
+ "loss": 0.634,
741
+ "rewards/accuracies": 0.5625,
742
+ "rewards/chosen": 0.6369161605834961,
743
+ "rewards/margins": 0.4485607147216797,
744
+ "rewards/rejected": 0.18835540115833282,
745
+ "step": 490
746
+ },
747
+ {
748
+ "epoch": 0.8183306055646481,
749
+ "grad_norm": 6.964690208435059,
750
+ "learning_rate": 4.0812613705275924e-05,
751
+ "logits/chosen": -1.1329933404922485,
752
+ "logits/rejected": -1.066542625427246,
753
+ "logps/chosen": -423.4967346191406,
754
+ "logps/rejected": -376.92803955078125,
755
+ "loss": 0.5904,
756
+ "rewards/accuracies": 0.550000011920929,
757
+ "rewards/chosen": 0.6795145273208618,
758
+ "rewards/margins": 0.3617789149284363,
759
+ "rewards/rejected": 0.3177356719970703,
760
+ "step": 500
761
+ },
762
+ {
763
+ "epoch": 0.8346972176759411,
764
+ "grad_norm": 11.78022289276123,
765
+ "learning_rate": 4.050939963614312e-05,
766
+ "logits/chosen": -1.1565020084381104,
767
+ "logits/rejected": -1.064300298690796,
768
+ "logps/chosen": -465.0497131347656,
769
+ "logps/rejected": -438.4883728027344,
770
+ "loss": 0.6384,
771
+ "rewards/accuracies": 0.5249999761581421,
772
+ "rewards/chosen": 0.6123272180557251,
773
+ "rewards/margins": 0.3706541657447815,
774
+ "rewards/rejected": 0.241673082113266,
775
+ "step": 510
776
+ },
777
+ {
778
+ "epoch": 0.851063829787234,
779
+ "grad_norm": 6.378419876098633,
780
+ "learning_rate": 4.020618556701031e-05,
781
+ "logits/chosen": -1.0936065912246704,
782
+ "logits/rejected": -1.157414436340332,
783
+ "logps/chosen": -431.05657958984375,
784
+ "logps/rejected": -380.1803283691406,
785
+ "loss": 0.6866,
786
+ "rewards/accuracies": 0.5,
787
+ "rewards/chosen": 0.514751672744751,
788
+ "rewards/margins": 0.2915990948677063,
789
+ "rewards/rejected": 0.22315259277820587,
790
+ "step": 520
791
+ },
792
+ {
793
+ "epoch": 0.867430441898527,
794
+ "grad_norm": 6.811694145202637,
795
+ "learning_rate": 3.9902971497877504e-05,
796
+ "logits/chosen": -1.0564637184143066,
797
+ "logits/rejected": -0.8537761569023132,
798
+ "logps/chosen": -400.8801574707031,
799
+ "logps/rejected": -400.9678039550781,
800
+ "loss": 0.6036,
801
+ "rewards/accuracies": 0.625,
802
+ "rewards/chosen": 0.5905786156654358,
803
+ "rewards/margins": 0.49240899085998535,
804
+ "rewards/rejected": 0.09816964715719223,
805
+ "step": 530
806
+ },
807
+ {
808
+ "epoch": 0.88379705400982,
809
+ "grad_norm": 8.516433715820312,
810
+ "learning_rate": 3.9599757428744695e-05,
811
+ "logits/chosen": -1.0425868034362793,
812
+ "logits/rejected": -1.0106391906738281,
813
+ "logps/chosen": -494.23681640625,
814
+ "logps/rejected": -451.3269958496094,
815
+ "loss": 0.7295,
816
+ "rewards/accuracies": 0.5,
817
+ "rewards/chosen": 0.480445921421051,
818
+ "rewards/margins": 0.12438355386257172,
819
+ "rewards/rejected": 0.3560623824596405,
820
+ "step": 540
821
+ },
822
+ {
823
+ "epoch": 0.900163666121113,
824
+ "grad_norm": 9.584715843200684,
825
+ "learning_rate": 3.9296543359611886e-05,
826
+ "logits/chosen": -1.1160447597503662,
827
+ "logits/rejected": -0.898239016532898,
828
+ "logps/chosen": -471.17718505859375,
829
+ "logps/rejected": -404.9838562011719,
830
+ "loss": 0.5973,
831
+ "rewards/accuracies": 0.625,
832
+ "rewards/chosen": 0.4906229078769684,
833
+ "rewards/margins": 0.3674588203430176,
834
+ "rewards/rejected": 0.1231641173362732,
835
+ "step": 550
836
+ },
837
+ {
838
+ "epoch": 0.9165302782324058,
839
+ "grad_norm": 9.050358772277832,
840
+ "learning_rate": 3.899332929047908e-05,
841
+ "logits/chosen": -1.0188591480255127,
842
+ "logits/rejected": NaN,
843
+ "logps/chosen": -505.81146240234375,
844
+ "logps/rejected": -455.074462890625,
845
+ "loss": 0.672,
846
+ "rewards/accuracies": 0.48750001192092896,
847
+ "rewards/chosen": 0.5001901388168335,
848
+ "rewards/margins": 0.3212779462337494,
849
+ "rewards/rejected": 0.1789121776819229,
850
+ "step": 560
851
+ },
852
+ {
853
+ "epoch": 0.9328968903436988,
854
+ "grad_norm": 10.636481285095215,
855
+ "learning_rate": 3.8690115221346276e-05,
856
+ "logits/chosen": -1.0234591960906982,
857
+ "logits/rejected": -0.9390643835067749,
858
+ "logps/chosen": -423.03594970703125,
859
+ "logps/rejected": -383.97369384765625,
860
+ "loss": 0.6515,
861
+ "rewards/accuracies": 0.6000000238418579,
862
+ "rewards/chosen": 0.40482932329177856,
863
+ "rewards/margins": 0.20538429915905,
864
+ "rewards/rejected": 0.19944503903388977,
865
+ "step": 570
866
+ },
867
+ {
868
+ "epoch": 0.9492635024549918,
869
+ "grad_norm": 12.001391410827637,
870
+ "learning_rate": 3.838690115221347e-05,
871
+ "logits/chosen": -1.254690170288086,
872
+ "logits/rejected": -1.098749041557312,
873
+ "logps/chosen": -400.72021484375,
874
+ "logps/rejected": -348.39678955078125,
875
+ "loss": 0.6332,
876
+ "rewards/accuracies": 0.5874999761581421,
877
+ "rewards/chosen": 0.5642321705818176,
878
+ "rewards/margins": 0.34682947397232056,
879
+ "rewards/rejected": 0.21740269660949707,
880
+ "step": 580
881
+ },
882
+ {
883
+ "epoch": 0.9656301145662848,
884
+ "grad_norm": 9.584880828857422,
885
+ "learning_rate": 3.808368708308066e-05,
886
+ "logits/chosen": -0.9567238092422485,
887
+ "logits/rejected": -0.8445194959640503,
888
+ "logps/chosen": -475.3368225097656,
889
+ "logps/rejected": -418.3975524902344,
890
+ "loss": 0.609,
891
+ "rewards/accuracies": 0.5874999761581421,
892
+ "rewards/chosen": 0.7818068861961365,
893
+ "rewards/margins": 0.4438362717628479,
894
+ "rewards/rejected": 0.3379705548286438,
895
+ "step": 590
896
+ },
897
+ {
898
+ "epoch": 0.9819967266775778,
899
+ "grad_norm": 8.225208282470703,
900
+ "learning_rate": 3.778047301394785e-05,
901
+ "logits/chosen": -0.8194648623466492,
902
+ "logits/rejected": -0.6516283750534058,
903
+ "logps/chosen": -464.95648193359375,
904
+ "logps/rejected": -454.74896240234375,
905
+ "loss": 0.6792,
906
+ "rewards/accuracies": 0.612500011920929,
907
+ "rewards/chosen": 0.673040509223938,
908
+ "rewards/margins": 0.3073888421058655,
909
+ "rewards/rejected": 0.3656516671180725,
910
+ "step": 600
911
+ },
912
+ {
913
+ "epoch": 0.9983633387888707,
914
+ "grad_norm": 7.938261032104492,
915
+ "learning_rate": 3.747725894481504e-05,
916
+ "logits/chosen": -0.9632278680801392,
917
+ "logits/rejected": -0.7192381620407104,
918
+ "logps/chosen": -428.2535095214844,
919
+ "logps/rejected": -426.05780029296875,
920
+ "loss": 0.5784,
921
+ "rewards/accuracies": 0.675000011920929,
922
+ "rewards/chosen": 0.778376042842865,
923
+ "rewards/margins": 0.44757041335105896,
924
+ "rewards/rejected": 0.33080559968948364,
925
+ "step": 610
926
+ },
927
+ {
928
+ "epoch": 1.0147299509001637,
929
+ "grad_norm": 6.6163482666015625,
930
+ "learning_rate": 3.717404487568223e-05,
931
+ "logits/chosen": -0.8798831105232239,
932
+ "logits/rejected": -0.716325581073761,
933
+ "logps/chosen": -401.0724182128906,
934
+ "logps/rejected": -393.04925537109375,
935
+ "loss": 0.5397,
936
+ "rewards/accuracies": 0.5874999761581421,
937
+ "rewards/chosen": 0.6812083721160889,
938
+ "rewards/margins": 0.6590946912765503,
939
+ "rewards/rejected": 0.022113658487796783,
940
+ "step": 620
941
+ },
942
+ {
943
+ "epoch": 1.0310965630114566,
944
+ "grad_norm": 4.5942792892456055,
945
+ "learning_rate": 3.687083080654943e-05,
946
+ "logits/chosen": -0.595669150352478,
947
+ "logits/rejected": -0.6483780741691589,
948
+ "logps/chosen": -519.7013549804688,
949
+ "logps/rejected": -459.6736755371094,
950
+ "loss": 0.3798,
951
+ "rewards/accuracies": 0.8125,
952
+ "rewards/chosen": 0.9690455198287964,
953
+ "rewards/margins": 1.2637821435928345,
954
+ "rewards/rejected": -0.2947366237640381,
955
+ "step": 630
956
+ },
957
+ {
958
+ "epoch": 1.0474631751227497,
959
+ "grad_norm": 5.822299003601074,
960
+ "learning_rate": 3.6567616737416614e-05,
961
+ "logits/chosen": -1.1080423593521118,
962
+ "logits/rejected": -0.8690654635429382,
963
+ "logps/chosen": -430.8399963378906,
964
+ "logps/rejected": -382.16131591796875,
965
+ "loss": 0.3725,
966
+ "rewards/accuracies": 0.8500000238418579,
967
+ "rewards/chosen": 0.926641583442688,
968
+ "rewards/margins": 1.1296237707138062,
969
+ "rewards/rejected": -0.20298215746879578,
970
+ "step": 640
971
+ },
972
+ {
973
+ "epoch": 1.0638297872340425,
974
+ "grad_norm": 5.2786078453063965,
975
+ "learning_rate": 3.626440266828381e-05,
976
+ "logits/chosen": -0.986005961894989,
977
+ "logits/rejected": -0.8763575553894043,
978
+ "logps/chosen": -407.85260009765625,
979
+ "logps/rejected": -399.2723693847656,
980
+ "loss": 0.4381,
981
+ "rewards/accuracies": 0.6875,
982
+ "rewards/chosen": 0.8758468627929688,
983
+ "rewards/margins": 0.9750221371650696,
984
+ "rewards/rejected": -0.09917531907558441,
985
+ "step": 650
986
+ },
987
+ {
988
+ "epoch": 1.0801963993453354,
989
+ "grad_norm": 11.273237228393555,
990
+ "learning_rate": 3.5961188599151e-05,
991
+ "logits/chosen": -1.3430869579315186,
992
+ "logits/rejected": -1.056379795074463,
993
+ "logps/chosen": -410.79791259765625,
994
+ "logps/rejected": -434.85845947265625,
995
+ "loss": 0.4005,
996
+ "rewards/accuracies": 0.8125,
997
+ "rewards/chosen": 0.9130380749702454,
998
+ "rewards/margins": 0.9950882196426392,
999
+ "rewards/rejected": -0.08205002546310425,
1000
+ "step": 660
1001
+ },
1002
+ {
1003
+ "epoch": 1.0965630114566285,
1004
+ "grad_norm": 7.220339775085449,
1005
+ "learning_rate": 3.5657974530018194e-05,
1006
+ "logits/chosen": -1.0506983995437622,
1007
+ "logits/rejected": -1.0361990928649902,
1008
+ "logps/chosen": -408.18035888671875,
1009
+ "logps/rejected": -391.70562744140625,
1010
+ "loss": 0.4471,
1011
+ "rewards/accuracies": 0.7749999761581421,
1012
+ "rewards/chosen": 0.7731181979179382,
1013
+ "rewards/margins": 0.9366775751113892,
1014
+ "rewards/rejected": -0.16355939209461212,
1015
+ "step": 670
1016
+ },
1017
+ {
1018
+ "epoch": 1.1129296235679214,
1019
+ "grad_norm": 5.74124813079834,
1020
+ "learning_rate": 3.535476046088539e-05,
1021
+ "logits/chosen": -1.206479787826538,
1022
+ "logits/rejected": -1.1451863050460815,
1023
+ "logps/chosen": -444.80438232421875,
1024
+ "logps/rejected": -439.0628356933594,
1025
+ "loss": 0.4043,
1026
+ "rewards/accuracies": 0.800000011920929,
1027
+ "rewards/chosen": 0.8405340313911438,
1028
+ "rewards/margins": 1.1449936628341675,
1029
+ "rewards/rejected": -0.30445969104766846,
1030
+ "step": 680
1031
+ },
1032
+ {
1033
+ "epoch": 1.1292962356792144,
1034
+ "grad_norm": 7.32386589050293,
1035
+ "learning_rate": 3.5051546391752576e-05,
1036
+ "logits/chosen": -1.0227863788604736,
1037
+ "logits/rejected": -0.9067742228507996,
1038
+ "logps/chosen": -476.1012268066406,
1039
+ "logps/rejected": -466.14971923828125,
1040
+ "loss": 0.4261,
1041
+ "rewards/accuracies": 0.800000011920929,
1042
+ "rewards/chosen": 0.8802071809768677,
1043
+ "rewards/margins": 1.1421091556549072,
1044
+ "rewards/rejected": -0.2619021236896515,
1045
+ "step": 690
1046
+ },
1047
+ {
1048
+ "epoch": 1.1456628477905073,
1049
+ "grad_norm": 4.43737268447876,
1050
+ "learning_rate": 3.474833232261977e-05,
1051
+ "logits/chosen": -1.135181188583374,
1052
+ "logits/rejected": -1.0297483205795288,
1053
+ "logps/chosen": -506.5615234375,
1054
+ "logps/rejected": -483.1039123535156,
1055
+ "loss": 0.3875,
1056
+ "rewards/accuracies": 0.8125,
1057
+ "rewards/chosen": 0.5547564029693604,
1058
+ "rewards/margins": 1.281551480293274,
1059
+ "rewards/rejected": -0.7267951369285583,
1060
+ "step": 700
1061
+ },
1062
+ {
1063
+ "epoch": 1.1620294599018004,
1064
+ "grad_norm": 2.298978328704834,
1065
+ "learning_rate": 3.4445118253486965e-05,
1066
+ "logits/chosen": -1.426138162612915,
1067
+ "logits/rejected": -1.091489553451538,
1068
+ "logps/chosen": -400.107177734375,
1069
+ "logps/rejected": -401.62109375,
1070
+ "loss": 0.4072,
1071
+ "rewards/accuracies": 0.75,
1072
+ "rewards/chosen": 0.6946335434913635,
1073
+ "rewards/margins": 1.441784143447876,
1074
+ "rewards/rejected": -0.7471505999565125,
1075
+ "step": 710
1076
+ },
1077
+ {
1078
+ "epoch": 1.1783960720130933,
1079
+ "grad_norm": 4.7108154296875,
1080
+ "learning_rate": 3.4141904184354156e-05,
1081
+ "logits/chosen": -1.1755582094192505,
1082
+ "logits/rejected": -0.901213526725769,
1083
+ "logps/chosen": -425.28759765625,
1084
+ "logps/rejected": -402.042236328125,
1085
+ "loss": 0.4615,
1086
+ "rewards/accuracies": 0.762499988079071,
1087
+ "rewards/chosen": 0.6500431299209595,
1088
+ "rewards/margins": 0.9347268342971802,
1089
+ "rewards/rejected": -0.2846837639808655,
1090
+ "step": 720
1091
+ },
1092
+ {
1093
+ "epoch": 1.1947626841243864,
1094
+ "grad_norm": 12.316296577453613,
1095
+ "learning_rate": 3.383869011522135e-05,
1096
+ "logits/chosen": -1.284589409828186,
1097
+ "logits/rejected": -1.0795433521270752,
1098
+ "logps/chosen": -413.8037109375,
1099
+ "logps/rejected": -491.56146240234375,
1100
+ "loss": 0.4385,
1101
+ "rewards/accuracies": 0.737500011920929,
1102
+ "rewards/chosen": 0.6515102386474609,
1103
+ "rewards/margins": 1.2281266450881958,
1104
+ "rewards/rejected": -0.5766164660453796,
1105
+ "step": 730
1106
+ },
1107
+ {
1108
+ "epoch": 1.2111292962356792,
1109
+ "grad_norm": 9.386224746704102,
1110
+ "learning_rate": 3.353547604608854e-05,
1111
+ "logits/chosen": -1.257037878036499,
1112
+ "logits/rejected": -1.0705680847167969,
1113
+ "logps/chosen": -505.6416931152344,
1114
+ "logps/rejected": -459.56805419921875,
1115
+ "loss": 0.3726,
1116
+ "rewards/accuracies": 0.7749999761581421,
1117
+ "rewards/chosen": 0.7937537431716919,
1118
+ "rewards/margins": 1.4233770370483398,
1119
+ "rewards/rejected": -0.629623293876648,
1120
+ "step": 740
1121
+ },
1122
+ {
1123
+ "epoch": 1.227495908346972,
1124
+ "grad_norm": 6.265888214111328,
1125
+ "learning_rate": 3.323226197695573e-05,
1126
+ "logits/chosen": -1.332099199295044,
1127
+ "logits/rejected": -1.0086009502410889,
1128
+ "logps/chosen": -457.0177307128906,
1129
+ "logps/rejected": -449.7435607910156,
1130
+ "loss": 0.4034,
1131
+ "rewards/accuracies": 0.7875000238418579,
1132
+ "rewards/chosen": 0.790605366230011,
1133
+ "rewards/margins": 1.2190719842910767,
1134
+ "rewards/rejected": -0.4284667372703552,
1135
+ "step": 750
1136
+ },
1137
+ {
1138
+ "epoch": 1.2438625204582652,
1139
+ "grad_norm": 6.070497035980225,
1140
+ "learning_rate": 3.292904790782292e-05,
1141
+ "logits/chosen": -1.2865736484527588,
1142
+ "logits/rejected": -1.072819471359253,
1143
+ "logps/chosen": -432.710693359375,
1144
+ "logps/rejected": -467.6375427246094,
1145
+ "loss": 0.3773,
1146
+ "rewards/accuracies": 0.800000011920929,
1147
+ "rewards/chosen": 0.7443059682846069,
1148
+ "rewards/margins": 1.3002400398254395,
1149
+ "rewards/rejected": -0.5559341907501221,
1150
+ "step": 760
1151
+ },
1152
+ {
1153
+ "epoch": 1.260229132569558,
1154
+ "grad_norm": 5.7988128662109375,
1155
+ "learning_rate": 3.262583383869012e-05,
1156
+ "logits/chosen": -1.0803556442260742,
1157
+ "logits/rejected": -0.8561310768127441,
1158
+ "logps/chosen": -514.7871704101562,
1159
+ "logps/rejected": -503.19500732421875,
1160
+ "loss": 0.3216,
1161
+ "rewards/accuracies": 0.8500000238418579,
1162
+ "rewards/chosen": 0.7713959217071533,
1163
+ "rewards/margins": 1.6244882345199585,
1164
+ "rewards/rejected": -0.8530920743942261,
1165
+ "step": 770
1166
+ },
1167
+ {
1168
+ "epoch": 1.2765957446808511,
1169
+ "grad_norm": 4.323042869567871,
1170
+ "learning_rate": 3.232261976955731e-05,
1171
+ "logits/chosen": -1.408276915550232,
1172
+ "logits/rejected": -0.9835014343261719,
1173
+ "logps/chosen": -456.21142578125,
1174
+ "logps/rejected": -429.10467529296875,
1175
+ "loss": 0.3832,
1176
+ "rewards/accuracies": 0.824999988079071,
1177
+ "rewards/chosen": 0.5816126465797424,
1178
+ "rewards/margins": 1.3987071514129639,
1179
+ "rewards/rejected": -0.8170945048332214,
1180
+ "step": 780
1181
+ },
1182
+ {
1183
+ "epoch": 1.292962356792144,
1184
+ "grad_norm": 7.249789237976074,
1185
+ "learning_rate": 3.20194057004245e-05,
1186
+ "logits/chosen": -1.1671850681304932,
1187
+ "logits/rejected": -0.653769314289093,
1188
+ "logps/chosen": -481.74859619140625,
1189
+ "logps/rejected": -441.5457458496094,
1190
+ "loss": 0.3293,
1191
+ "rewards/accuracies": 0.887499988079071,
1192
+ "rewards/chosen": 0.5319666862487793,
1193
+ "rewards/margins": 1.7422034740447998,
1194
+ "rewards/rejected": -1.21023690700531,
1195
+ "step": 790
1196
+ },
1197
+ {
1198
+ "epoch": 1.3093289689034369,
1199
+ "grad_norm": 9.989459037780762,
1200
+ "learning_rate": 3.171619163129169e-05,
1201
+ "logits/chosen": -1.2330873012542725,
1202
+ "logits/rejected": -1.1575677394866943,
1203
+ "logps/chosen": -462.0821228027344,
1204
+ "logps/rejected": -419.34149169921875,
1205
+ "loss": 0.4737,
1206
+ "rewards/accuracies": 0.75,
1207
+ "rewards/chosen": 0.7070258855819702,
1208
+ "rewards/margins": 1.2338263988494873,
1209
+ "rewards/rejected": -0.5268004536628723,
1210
+ "step": 800
1211
+ },
1212
+ {
1213
+ "epoch": 1.32569558101473,
1214
+ "grad_norm": 5.914211750030518,
1215
+ "learning_rate": 3.141297756215888e-05,
1216
+ "logits/chosen": -1.257925271987915,
1217
+ "logits/rejected": -1.1667311191558838,
1218
+ "logps/chosen": -492.0726623535156,
1219
+ "logps/rejected": -463.2273864746094,
1220
+ "loss": 0.3597,
1221
+ "rewards/accuracies": 0.8125,
1222
+ "rewards/chosen": 0.7776979207992554,
1223
+ "rewards/margins": 1.3328384160995483,
1224
+ "rewards/rejected": -0.555140495300293,
1225
+ "step": 810
1226
+ },
1227
+ {
1228
+ "epoch": 1.342062193126023,
1229
+ "grad_norm": 3.587632179260254,
1230
+ "learning_rate": 3.110976349302608e-05,
1231
+ "logits/chosen": -1.2870782613754272,
1232
+ "logits/rejected": -0.9442241787910461,
1233
+ "logps/chosen": -493.3888244628906,
1234
+ "logps/rejected": -464.4835510253906,
1235
+ "loss": 0.3151,
1236
+ "rewards/accuracies": 0.824999988079071,
1237
+ "rewards/chosen": 0.9334325790405273,
1238
+ "rewards/margins": 1.9022817611694336,
1239
+ "rewards/rejected": -0.9688493013381958,
1240
+ "step": 820
1241
+ },
1242
+ {
1243
+ "epoch": 1.358428805237316,
1244
+ "grad_norm": 6.462862968444824,
1245
+ "learning_rate": 3.080654942389327e-05,
1246
+ "logits/chosen": -1.4893635511398315,
1247
+ "logits/rejected": -1.0626758337020874,
1248
+ "logps/chosen": -373.52288818359375,
1249
+ "logps/rejected": -411.831298828125,
1250
+ "loss": 0.438,
1251
+ "rewards/accuracies": 0.7749999761581421,
1252
+ "rewards/chosen": 0.5565814971923828,
1253
+ "rewards/margins": 1.1154134273529053,
1254
+ "rewards/rejected": -0.5588318705558777,
1255
+ "step": 830
1256
+ },
1257
+ {
1258
+ "epoch": 1.3747954173486088,
1259
+ "grad_norm": 6.482163429260254,
1260
+ "learning_rate": 3.0503335354760464e-05,
1261
+ "logits/chosen": -1.2426273822784424,
1262
+ "logits/rejected": -1.0962947607040405,
1263
+ "logps/chosen": -418.48577880859375,
1264
+ "logps/rejected": -414.7897033691406,
1265
+ "loss": 0.4516,
1266
+ "rewards/accuracies": 0.75,
1267
+ "rewards/chosen": 0.7719891667366028,
1268
+ "rewards/margins": 1.0017683506011963,
1269
+ "rewards/rejected": -0.22977924346923828,
1270
+ "step": 840
1271
+ },
1272
+ {
1273
+ "epoch": 1.3911620294599019,
1274
+ "grad_norm": 8.918207168579102,
1275
+ "learning_rate": 3.020012128562765e-05,
1276
+ "logits/chosen": -1.3620984554290771,
1277
+ "logits/rejected": -1.1984447240829468,
1278
+ "logps/chosen": -465.46270751953125,
1279
+ "logps/rejected": -446.5343322753906,
1280
+ "loss": 0.4053,
1281
+ "rewards/accuracies": 0.8125,
1282
+ "rewards/chosen": 0.6932386159896851,
1283
+ "rewards/margins": 1.3631744384765625,
1284
+ "rewards/rejected": -0.6699355840682983,
1285
+ "step": 850
1286
+ },
1287
+ {
1288
+ "epoch": 1.4075286415711947,
1289
+ "grad_norm": 19.940839767456055,
1290
+ "learning_rate": 2.9896907216494846e-05,
1291
+ "logits/chosen": -1.2039759159088135,
1292
+ "logits/rejected": -0.9377444386482239,
1293
+ "logps/chosen": -574.588623046875,
1294
+ "logps/rejected": -502.51214599609375,
1295
+ "loss": 0.4461,
1296
+ "rewards/accuracies": 0.824999988079071,
1297
+ "rewards/chosen": 0.8642657399177551,
1298
+ "rewards/margins": 1.3982216119766235,
1299
+ "rewards/rejected": -0.5339558124542236,
1300
+ "step": 860
1301
+ },
1302
+ {
1303
+ "epoch": 1.4238952536824878,
1304
+ "grad_norm": 6.069481372833252,
1305
+ "learning_rate": 2.959369314736204e-05,
1306
+ "logits/chosen": -1.3913848400115967,
1307
+ "logits/rejected": -1.2518432140350342,
1308
+ "logps/chosen": -458.3941345214844,
1309
+ "logps/rejected": -452.67352294921875,
1310
+ "loss": 0.3604,
1311
+ "rewards/accuracies": 0.762499988079071,
1312
+ "rewards/chosen": 1.043166160583496,
1313
+ "rewards/margins": 1.7257499694824219,
1314
+ "rewards/rejected": -0.6825838685035706,
1315
+ "step": 870
1316
+ },
1317
+ {
1318
+ "epoch": 1.4402618657937807,
1319
+ "grad_norm": 3.334609031677246,
1320
+ "learning_rate": 2.929047907822923e-05,
1321
+ "logits/chosen": -1.3935604095458984,
1322
+ "logits/rejected": NaN,
1323
+ "logps/chosen": -490.9991149902344,
1324
+ "logps/rejected": -452.0387268066406,
1325
+ "loss": 0.3173,
1326
+ "rewards/accuracies": 0.8500000238418579,
1327
+ "rewards/chosen": 1.0780329704284668,
1328
+ "rewards/margins": 1.919669508934021,
1329
+ "rewards/rejected": -0.8416364789009094,
1330
+ "step": 880
1331
+ },
1332
+ {
1333
+ "epoch": 1.4566284779050735,
1334
+ "grad_norm": 5.730930805206299,
1335
+ "learning_rate": 2.8987265009096426e-05,
1336
+ "logits/chosen": -1.3444249629974365,
1337
+ "logits/rejected": -1.124793291091919,
1338
+ "logps/chosen": -409.4859619140625,
1339
+ "logps/rejected": -429.79150390625,
1340
+ "loss": 0.4307,
1341
+ "rewards/accuracies": 0.762499988079071,
1342
+ "rewards/chosen": 0.8416284322738647,
1343
+ "rewards/margins": 1.227608561515808,
1344
+ "rewards/rejected": -0.3859800696372986,
1345
+ "step": 890
1346
+ },
1347
+ {
1348
+ "epoch": 1.4729950900163666,
1349
+ "grad_norm": 4.7208333015441895,
1350
+ "learning_rate": 2.8684050939963614e-05,
1351
+ "logits/chosen": -1.2483845949172974,
1352
+ "logits/rejected": -1.162945032119751,
1353
+ "logps/chosen": -438.02978515625,
1354
+ "logps/rejected": -483.64776611328125,
1355
+ "loss": 0.4359,
1356
+ "rewards/accuracies": 0.8374999761581421,
1357
+ "rewards/chosen": 0.6855508089065552,
1358
+ "rewards/margins": 1.3711297512054443,
1359
+ "rewards/rejected": -0.6855790019035339,
1360
+ "step": 900
1361
+ },
1362
+ {
1363
+ "epoch": 1.4893617021276595,
1364
+ "grad_norm": 4.603919982910156,
1365
+ "learning_rate": 2.838083687083081e-05,
1366
+ "logits/chosen": -1.3359647989273071,
1367
+ "logits/rejected": -1.2829469442367554,
1368
+ "logps/chosen": -478.17022705078125,
1369
+ "logps/rejected": -408.0987243652344,
1370
+ "loss": 0.431,
1371
+ "rewards/accuracies": 0.6499999761581421,
1372
+ "rewards/chosen": 0.9056881666183472,
1373
+ "rewards/margins": 1.124969244003296,
1374
+ "rewards/rejected": -0.21928110718727112,
1375
+ "step": 910
1376
+ },
1377
+ {
1378
+ "epoch": 1.5057283142389526,
1379
+ "grad_norm": 7.60992431640625,
1380
+ "learning_rate": 2.8077622801698e-05,
1381
+ "logits/chosen": -1.4223133325576782,
1382
+ "logits/rejected": -1.2066584825515747,
1383
+ "logps/chosen": -428.1543884277344,
1384
+ "logps/rejected": -370.4458923339844,
1385
+ "loss": 0.4003,
1386
+ "rewards/accuracies": 0.675000011920929,
1387
+ "rewards/chosen": 0.8528293371200562,
1388
+ "rewards/margins": 1.363084077835083,
1389
+ "rewards/rejected": -0.5102548599243164,
1390
+ "step": 920
1391
+ },
1392
+ {
1393
+ "epoch": 1.5220949263502455,
1394
+ "grad_norm": 4.768836498260498,
1395
+ "learning_rate": 2.7774408732565194e-05,
1396
+ "logits/chosen": -1.2342239618301392,
1397
+ "logits/rejected": -1.1602811813354492,
1398
+ "logps/chosen": -440.5987854003906,
1399
+ "logps/rejected": -405.0908508300781,
1400
+ "loss": 0.3879,
1401
+ "rewards/accuracies": 0.8374999761581421,
1402
+ "rewards/chosen": 0.9851611256599426,
1403
+ "rewards/margins": 1.1822245121002197,
1404
+ "rewards/rejected": -0.19706343114376068,
1405
+ "step": 930
1406
+ },
1407
+ {
1408
+ "epoch": 1.5384615384615383,
1409
+ "grad_norm": 8.149114608764648,
1410
+ "learning_rate": 2.7471194663432385e-05,
1411
+ "logits/chosen": -1.1173126697540283,
1412
+ "logits/rejected": -1.014575481414795,
1413
+ "logps/chosen": -402.3595275878906,
1414
+ "logps/rejected": -410.60003662109375,
1415
+ "loss": 0.4529,
1416
+ "rewards/accuracies": 0.8125,
1417
+ "rewards/chosen": 0.7943775653839111,
1418
+ "rewards/margins": 1.2599968910217285,
1419
+ "rewards/rejected": -0.4656193256378174,
1420
+ "step": 940
1421
+ },
1422
+ {
1423
+ "epoch": 1.5548281505728314,
1424
+ "grad_norm": 7.4743852615356445,
1425
+ "learning_rate": 2.7167980594299573e-05,
1426
+ "logits/chosen": NaN,
1427
+ "logits/rejected": -1.143424391746521,
1428
+ "logps/chosen": -446.26629638671875,
1429
+ "logps/rejected": -469.7017517089844,
1430
+ "loss": 0.491,
1431
+ "rewards/accuracies": 0.675000011920929,
1432
+ "rewards/chosen": 0.797395646572113,
1433
+ "rewards/margins": 1.1872708797454834,
1434
+ "rewards/rejected": -0.38987529277801514,
1435
+ "step": 950
1436
+ },
1437
+ {
1438
+ "epoch": 1.5711947626841245,
1439
+ "grad_norm": 8.159062385559082,
1440
+ "learning_rate": 2.6864766525166768e-05,
1441
+ "logits/chosen": -1.14289128780365,
1442
+ "logits/rejected": NaN,
1443
+ "logps/chosen": -450.77117919921875,
1444
+ "logps/rejected": -438.8096618652344,
1445
+ "loss": 0.4282,
1446
+ "rewards/accuracies": 0.737500011920929,
1447
+ "rewards/chosen": 0.7298113107681274,
1448
+ "rewards/margins": 1.092790126800537,
1449
+ "rewards/rejected": -0.3629787862300873,
1450
+ "step": 960
1451
+ },
1452
+ {
1453
+ "epoch": 1.5875613747954174,
1454
+ "grad_norm": 9.274547576904297,
1455
+ "learning_rate": 2.6561552456033962e-05,
1456
+ "logits/chosen": -0.9688614010810852,
1457
+ "logits/rejected": -0.897729754447937,
1458
+ "logps/chosen": -515.4688720703125,
1459
+ "logps/rejected": -436.51275634765625,
1460
+ "loss": 0.4169,
1461
+ "rewards/accuracies": 0.7749999761581421,
1462
+ "rewards/chosen": 0.7370742559432983,
1463
+ "rewards/margins": 1.1543539762496948,
1464
+ "rewards/rejected": -0.4172796607017517,
1465
+ "step": 970
1466
+ },
1467
+ {
1468
+ "epoch": 1.6039279869067102,
1469
+ "grad_norm": 5.051360130310059,
1470
+ "learning_rate": 2.6258338386901153e-05,
1471
+ "logits/chosen": -1.0737905502319336,
1472
+ "logits/rejected": -0.9251714944839478,
1473
+ "logps/chosen": -471.31146240234375,
1474
+ "logps/rejected": -432.572509765625,
1475
+ "loss": 0.3841,
1476
+ "rewards/accuracies": 0.762499988079071,
1477
+ "rewards/chosen": 0.8669958114624023,
1478
+ "rewards/margins": 1.4004991054534912,
1479
+ "rewards/rejected": -0.5335032939910889,
1480
+ "step": 980
1481
+ },
1482
+ {
1483
+ "epoch": 1.6202945990180033,
1484
+ "grad_norm": 6.768540859222412,
1485
+ "learning_rate": 2.5955124317768348e-05,
1486
+ "logits/chosen": -1.1147942543029785,
1487
+ "logits/rejected": -1.1377754211425781,
1488
+ "logps/chosen": -462.5465393066406,
1489
+ "logps/rejected": -390.2607727050781,
1490
+ "loss": 0.3919,
1491
+ "rewards/accuracies": 0.8125,
1492
+ "rewards/chosen": 0.8462246060371399,
1493
+ "rewards/margins": 1.2740905284881592,
1494
+ "rewards/rejected": -0.4278659224510193,
1495
+ "step": 990
1496
+ },
1497
+ {
1498
+ "epoch": 1.6366612111292962,
1499
+ "grad_norm": 8.313865661621094,
1500
+ "learning_rate": 2.5651910248635535e-05,
1501
+ "logits/chosen": -1.0041131973266602,
1502
+ "logits/rejected": -0.8046473264694214,
1503
+ "logps/chosen": -461.7525939941406,
1504
+ "logps/rejected": -443.0941467285156,
1505
+ "loss": 0.4222,
1506
+ "rewards/accuracies": 0.824999988079071,
1507
+ "rewards/chosen": 0.8126258850097656,
1508
+ "rewards/margins": 1.1657203435897827,
1509
+ "rewards/rejected": -0.3530944585800171,
1510
+ "step": 1000
1511
+ }
1512
+ ],
1513
+ "logging_steps": 10,
1514
+ "max_steps": 1833,
1515
+ "num_input_tokens_seen": 0,
1516
+ "num_train_epochs": 3,
1517
+ "save_steps": 500,
1518
+ "stateful_callbacks": {
1519
+ "TrainerControl": {
1520
+ "args": {
1521
+ "should_epoch_stop": false,
1522
+ "should_evaluate": false,
1523
+ "should_log": false,
1524
+ "should_save": true,
1525
+ "should_training_stop": false
1526
+ },
1527
+ "attributes": {}
1528
+ }
1529
+ },
1530
+ "total_flos": 0.0,
1531
+ "train_batch_size": 4,
1532
+ "trial_name": null,
1533
+ "trial_params": null
1534
+ }
checkpoint-1000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:adb07c9cf31b39264fa1c39cdcedb4f8f2cd0a21811a03b1f63968e1dd128142
3
+ size 6673
checkpoint-1000/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-1500/README.md ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: unsloth/qwen3-0.6b-base-unsloth-bnb-4bit
3
+ library_name: peft
4
+ ---
5
+
6
+ # Model Card for Model ID
7
+
8
+ <!-- Provide a quick summary of what the model is/does. -->
9
+
10
+
11
+
12
+ ## Model Details
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+
19
+
20
+ - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
+ - **Shared by [optional]:** [More Information Needed]
23
+ - **Model type:** [More Information Needed]
24
+ - **Language(s) (NLP):** [More Information Needed]
25
+ - **License:** [More Information Needed]
26
+ - **Finetuned from model [optional]:** [More Information Needed]
27
+
28
+ ### Model Sources [optional]
29
+
30
+ <!-- Provide the basic links for the model. -->
31
+
32
+ - **Repository:** [More Information Needed]
33
+ - **Paper [optional]:** [More Information Needed]
34
+ - **Demo [optional]:** [More Information Needed]
35
+
36
+ ## Uses
37
+
38
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
+
40
+ ### Direct Use
41
+
42
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
+
44
+ [More Information Needed]
45
+
46
+ ### Downstream Use [optional]
47
+
48
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Out-of-Scope Use
53
+
54
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
+
56
+ [More Information Needed]
57
+
58
+ ## Bias, Risks, and Limitations
59
+
60
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ### Recommendations
65
+
66
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
+
68
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ Use the code below to get started with the model.
73
+
74
+ [More Information Needed]
75
+
76
+ ## Training Details
77
+
78
+ ### Training Data
79
+
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
+
82
+ [More Information Needed]
83
+
84
+ ### Training Procedure
85
+
86
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
+
88
+ #### Preprocessing [optional]
89
+
90
+ [More Information Needed]
91
+
92
+
93
+ #### Training Hyperparameters
94
+
95
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
+
97
+ #### Speeds, Sizes, Times [optional]
98
+
99
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
+
101
+ [More Information Needed]
102
+
103
+ ## Evaluation
104
+
105
+ <!-- This section describes the evaluation protocols and provides the results. -->
106
+
107
+ ### Testing Data, Factors & Metrics
108
+
109
+ #### Testing Data
110
+
111
+ <!-- This should link to a Dataset Card if possible. -->
112
+
113
+ [More Information Needed]
114
+
115
+ #### Factors
116
+
117
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Metrics
122
+
123
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
+
125
+ [More Information Needed]
126
+
127
+ ### Results
128
+
129
+ [More Information Needed]
130
+
131
+ #### Summary
132
+
133
+
134
+
135
+ ## Model Examination [optional]
136
+
137
+ <!-- Relevant interpretability work for the model goes here -->
138
+
139
+ [More Information Needed]
140
+
141
+ ## Environmental Impact
142
+
143
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
+
145
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
+
147
+ - **Hardware Type:** [More Information Needed]
148
+ - **Hours used:** [More Information Needed]
149
+ - **Cloud Provider:** [More Information Needed]
150
+ - **Compute Region:** [More Information Needed]
151
+ - **Carbon Emitted:** [More Information Needed]
152
+
153
+ ## Technical Specifications [optional]
154
+
155
+ ### Model Architecture and Objective
156
+
157
+ [More Information Needed]
158
+
159
+ ### Compute Infrastructure
160
+
161
+ [More Information Needed]
162
+
163
+ #### Hardware
164
+
165
+ [More Information Needed]
166
+
167
+ #### Software
168
+
169
+ [More Information Needed]
170
+
171
+ ## Citation [optional]
172
+
173
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
+
175
+ **BibTeX:**
176
+
177
+ [More Information Needed]
178
+
179
+ **APA:**
180
+
181
+ [More Information Needed]
182
+
183
+ ## Glossary [optional]
184
+
185
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
+
187
+ [More Information Needed]
188
+
189
+ ## More Information [optional]
190
+
191
+ [More Information Needed]
192
+
193
+ ## Model Card Authors [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Contact
198
+
199
+ [More Information Needed]
200
+ ### Framework versions
201
+
202
+ - PEFT 0.15.2
checkpoint-1500/adapter_config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "unsloth/qwen3-0.6b-base-unsloth-bnb-4bit",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 16,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "r": 16,
24
+ "rank_pattern": {},
25
+ "revision": null,
26
+ "target_modules": [
27
+ "q_proj",
28
+ "up_proj",
29
+ "down_proj",
30
+ "o_proj",
31
+ "gate_proj",
32
+ "k_proj",
33
+ "v_proj"
34
+ ],
35
+ "task_type": "CAUSAL_LM",
36
+ "trainable_token_indices": null,
37
+ "use_dora": false,
38
+ "use_rslora": false
39
+ }
checkpoint-1500/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e77d054037f29662b91216f6c5f9406697d2701032795d47399c49bafacd7d0
3
+ size 40422168
checkpoint-1500/added_tokens.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</think>": 151668,
3
+ "</tool_call>": 151658,
4
+ "</tool_response>": 151666,
5
+ "<think>": 151667,
6
+ "<tool_call>": 151657,
7
+ "<tool_response>": 151665,
8
+ "<|box_end|>": 151649,
9
+ "<|box_start|>": 151648,
10
+ "<|endoftext|>": 151643,
11
+ "<|file_sep|>": 151664,
12
+ "<|fim_middle|>": 151660,
13
+ "<|fim_pad|>": 151662,
14
+ "<|fim_prefix|>": 151659,
15
+ "<|fim_suffix|>": 151661,
16
+ "<|im_end|>": 151645,
17
+ "<|im_start|>": 151644,
18
+ "<|image_pad|>": 151655,
19
+ "<|object_ref_end|>": 151647,
20
+ "<|object_ref_start|>": 151646,
21
+ "<|quad_end|>": 151651,
22
+ "<|quad_start|>": 151650,
23
+ "<|repo_name|>": 151663,
24
+ "<|video_pad|>": 151656,
25
+ "<|vision_end|>": 151653,
26
+ "<|vision_pad|>": 151654,
27
+ "<|vision_start|>": 151652
28
+ }
checkpoint-1500/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-1500/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfe899f20a6bb1b399697aef05305c1279775eeb4165480022288a33f17e0d35
3
+ size 20959365
checkpoint-1500/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:181c5f0270cf39930062ddfa3767a2481d0c360f120b11f8e25dbf533a1cdaba
3
+ size 14645
checkpoint-1500/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:11e8ed66ea322c1704fd9c995b59f3c398a1dadf5b26079d1a0030603756c87c
3
+ size 1383
checkpoint-1500/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08a7d8687def3cfcf5ccbbd742c18a147aa6b07b0e64f6e36353773e8567f6a9
3
+ size 1465
checkpoint-1500/special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|endoftext|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|vision_pad|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
checkpoint-1500/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aeb13307a71acd8fe81861d94ad54ab689df773318809eed3cbe794b4492dae4
3
+ size 11422654
checkpoint-1500/tokenizer_config.json ADDED
@@ -0,0 +1,240 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ },
181
+ "151665": {
182
+ "content": "<tool_response>",
183
+ "lstrip": false,
184
+ "normalized": false,
185
+ "rstrip": false,
186
+ "single_word": false,
187
+ "special": false
188
+ },
189
+ "151666": {
190
+ "content": "</tool_response>",
191
+ "lstrip": false,
192
+ "normalized": false,
193
+ "rstrip": false,
194
+ "single_word": false,
195
+ "special": false
196
+ },
197
+ "151667": {
198
+ "content": "<think>",
199
+ "lstrip": false,
200
+ "normalized": false,
201
+ "rstrip": false,
202
+ "single_word": false,
203
+ "special": false
204
+ },
205
+ "151668": {
206
+ "content": "</think>",
207
+ "lstrip": false,
208
+ "normalized": false,
209
+ "rstrip": false,
210
+ "single_word": false,
211
+ "special": false
212
+ }
213
+ },
214
+ "additional_special_tokens": [
215
+ "<|im_start|>",
216
+ "<|im_end|>",
217
+ "<|object_ref_start|>",
218
+ "<|object_ref_end|>",
219
+ "<|box_start|>",
220
+ "<|box_end|>",
221
+ "<|quad_start|>",
222
+ "<|quad_end|>",
223
+ "<|vision_start|>",
224
+ "<|vision_end|>",
225
+ "<|vision_pad|>",
226
+ "<|image_pad|>",
227
+ "<|video_pad|>"
228
+ ],
229
+ "bos_token": null,
230
+ "clean_up_tokenization_spaces": false,
231
+ "eos_token": "<|endoftext|>",
232
+ "errors": "replace",
233
+ "extra_special_tokens": {},
234
+ "model_max_length": 32768,
235
+ "pad_token": "<|vision_pad|>",
236
+ "padding_side": "right",
237
+ "split_special_tokens": false,
238
+ "tokenizer_class": "Qwen2Tokenizer",
239
+ "unk_token": null
240
+ }
checkpoint-1500/trainer_state.json ADDED
@@ -0,0 +1,2284 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 2.454991816693944,
6
+ "eval_steps": 500,
7
+ "global_step": 1500,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.016366612111292964,
14
+ "grad_norm": NaN,
15
+ "learning_rate": 1.6304347826086957e-06,
16
+ "logits/chosen": -1.2515310049057007,
17
+ "logits/rejected": -0.8573959469795227,
18
+ "logps/chosen": -455.1468811035156,
19
+ "logps/rejected": -387.0008239746094,
20
+ "loss": 0.6923,
21
+ "rewards/accuracies": 0.44999998807907104,
22
+ "rewards/chosen": 0.0002114105736836791,
23
+ "rewards/margins": 0.00180077797267586,
24
+ "rewards/rejected": -0.0015893673989921808,
25
+ "step": 10
26
+ },
27
+ {
28
+ "epoch": 0.03273322422258593,
29
+ "grad_norm": 10.154256820678711,
30
+ "learning_rate": 3.2608695652173914e-06,
31
+ "logits/chosen": -1.145971417427063,
32
+ "logits/rejected": -1.082183837890625,
33
+ "logps/chosen": -446.249755859375,
34
+ "logps/rejected": -426.87237548828125,
35
+ "loss": 0.6942,
36
+ "rewards/accuracies": 0.375,
37
+ "rewards/chosen": -0.0005732345744036138,
38
+ "rewards/margins": -0.002082281280308962,
39
+ "rewards/rejected": 0.0015090465312823653,
40
+ "step": 20
41
+ },
42
+ {
43
+ "epoch": 0.049099836333878884,
44
+ "grad_norm": 8.393162727355957,
45
+ "learning_rate": 5.4347826086956525e-06,
46
+ "logits/chosen": -0.9614235162734985,
47
+ "logits/rejected": -0.8527517318725586,
48
+ "logps/chosen": -460.367431640625,
49
+ "logps/rejected": -433.61285400390625,
50
+ "loss": 0.6934,
51
+ "rewards/accuracies": 0.4625000059604645,
52
+ "rewards/chosen": 0.0062791816890239716,
53
+ "rewards/margins": -0.00045429245801642537,
54
+ "rewards/rejected": 0.006733474787324667,
55
+ "step": 30
56
+ },
57
+ {
58
+ "epoch": 0.06546644844517185,
59
+ "grad_norm": 9.398829460144043,
60
+ "learning_rate": 8.15217391304348e-06,
61
+ "logits/chosen": -1.171631097793579,
62
+ "logits/rejected": -1.0616101026535034,
63
+ "logps/chosen": -421.583984375,
64
+ "logps/rejected": -424.787109375,
65
+ "loss": 0.6902,
66
+ "rewards/accuracies": 0.48750001192092896,
67
+ "rewards/chosen": 0.020799484103918076,
68
+ "rewards/margins": 0.006221642717719078,
69
+ "rewards/rejected": 0.014577841386198997,
70
+ "step": 40
71
+ },
72
+ {
73
+ "epoch": 0.08183306055646482,
74
+ "grad_norm": 9.43005657196045,
75
+ "learning_rate": 1.0869565217391305e-05,
76
+ "logits/chosen": -1.2450587749481201,
77
+ "logits/rejected": -1.1417832374572754,
78
+ "logps/chosen": -476.83538818359375,
79
+ "logps/rejected": -444.6556091308594,
80
+ "loss": 0.6883,
81
+ "rewards/accuracies": 0.5,
82
+ "rewards/chosen": 0.042164258658885956,
83
+ "rewards/margins": 0.010725794360041618,
84
+ "rewards/rejected": 0.03143846243619919,
85
+ "step": 50
86
+ },
87
+ {
88
+ "epoch": 0.09819967266775777,
89
+ "grad_norm": 8.825221061706543,
90
+ "learning_rate": 1.331521739130435e-05,
91
+ "logits/chosen": -1.3223823308944702,
92
+ "logits/rejected": NaN,
93
+ "logps/chosen": -383.4811706542969,
94
+ "logps/rejected": -454.43780517578125,
95
+ "loss": 0.6965,
96
+ "rewards/accuracies": 0.48750001192092896,
97
+ "rewards/chosen": 0.03668512776494026,
98
+ "rewards/margins": -0.003995636478066444,
99
+ "rewards/rejected": 0.04068076238036156,
100
+ "step": 60
101
+ },
102
+ {
103
+ "epoch": 0.11456628477905073,
104
+ "grad_norm": 9.5093994140625,
105
+ "learning_rate": 1.6032608695652173e-05,
106
+ "logits/chosen": -1.2538468837738037,
107
+ "logits/rejected": -1.2771422863006592,
108
+ "logps/chosen": -468.79400634765625,
109
+ "logps/rejected": -402.565673828125,
110
+ "loss": 0.6864,
111
+ "rewards/accuracies": 0.5,
112
+ "rewards/chosen": 0.04952714219689369,
113
+ "rewards/margins": 0.01824185810983181,
114
+ "rewards/rejected": 0.03128528222441673,
115
+ "step": 70
116
+ },
117
+ {
118
+ "epoch": 0.1309328968903437,
119
+ "grad_norm": 7.945922374725342,
120
+ "learning_rate": 1.8750000000000002e-05,
121
+ "logits/chosen": -1.386731743812561,
122
+ "logits/rejected": -0.8966825604438782,
123
+ "logps/chosen": -446.54736328125,
124
+ "logps/rejected": -430.45025634765625,
125
+ "loss": 0.6719,
126
+ "rewards/accuracies": 0.612500011920929,
127
+ "rewards/chosen": 0.06767664849758148,
128
+ "rewards/margins": 0.048766594380140305,
129
+ "rewards/rejected": 0.01891004480421543,
130
+ "step": 80
131
+ },
132
+ {
133
+ "epoch": 0.14729950900163666,
134
+ "grad_norm": 8.496461868286133,
135
+ "learning_rate": 2.1467391304347828e-05,
136
+ "logits/chosen": -1.3730494976043701,
137
+ "logits/rejected": -1.1826181411743164,
138
+ "logps/chosen": -413.87835693359375,
139
+ "logps/rejected": -385.1138916015625,
140
+ "loss": 0.69,
141
+ "rewards/accuracies": 0.48750001192092896,
142
+ "rewards/chosen": 0.07122664153575897,
143
+ "rewards/margins": 0.014941570349037647,
144
+ "rewards/rejected": 0.0562850646674633,
145
+ "step": 90
146
+ },
147
+ {
148
+ "epoch": 0.16366612111292964,
149
+ "grad_norm": 10.744088172912598,
150
+ "learning_rate": 2.4184782608695653e-05,
151
+ "logits/chosen": -1.2254825830459595,
152
+ "logits/rejected": -1.1136434078216553,
153
+ "logps/chosen": -495.34295654296875,
154
+ "logps/rejected": -435.0003967285156,
155
+ "loss": 0.6618,
156
+ "rewards/accuracies": 0.612500011920929,
157
+ "rewards/chosen": 0.1682208627462387,
158
+ "rewards/margins": 0.08198796212673187,
159
+ "rewards/rejected": 0.08623288571834564,
160
+ "step": 100
161
+ },
162
+ {
163
+ "epoch": 0.18003273322422259,
164
+ "grad_norm": 11.016289710998535,
165
+ "learning_rate": 2.6902173913043476e-05,
166
+ "logits/chosen": -1.3138881921768188,
167
+ "logits/rejected": -1.0289314985275269,
168
+ "logps/chosen": -476.26153564453125,
169
+ "logps/rejected": -426.837646484375,
170
+ "loss": 0.6355,
171
+ "rewards/accuracies": 0.6499999761581421,
172
+ "rewards/chosen": 0.21744687855243683,
173
+ "rewards/margins": 0.15108919143676758,
174
+ "rewards/rejected": 0.06635768711566925,
175
+ "step": 110
176
+ },
177
+ {
178
+ "epoch": 0.19639934533551553,
179
+ "grad_norm": 11.619248390197754,
180
+ "learning_rate": 2.9619565217391305e-05,
181
+ "logits/chosen": -1.2751747369766235,
182
+ "logits/rejected": -1.1858090162277222,
183
+ "logps/chosen": -444.4823303222656,
184
+ "logps/rejected": -404.58123779296875,
185
+ "loss": 0.6716,
186
+ "rewards/accuracies": 0.550000011920929,
187
+ "rewards/chosen": 0.2203199863433838,
188
+ "rewards/margins": 0.06618332862854004,
189
+ "rewards/rejected": 0.15413665771484375,
190
+ "step": 120
191
+ },
192
+ {
193
+ "epoch": 0.2127659574468085,
194
+ "grad_norm": 8.900009155273438,
195
+ "learning_rate": 3.233695652173913e-05,
196
+ "logits/chosen": -1.2365785837173462,
197
+ "logits/rejected": -1.2031922340393066,
198
+ "logps/chosen": -477.8089294433594,
199
+ "logps/rejected": -456.80120849609375,
200
+ "loss": 0.6794,
201
+ "rewards/accuracies": 0.5625,
202
+ "rewards/chosen": 0.20705166459083557,
203
+ "rewards/margins": 0.060810189694166183,
204
+ "rewards/rejected": 0.1462414562702179,
205
+ "step": 130
206
+ },
207
+ {
208
+ "epoch": 0.22913256955810146,
209
+ "grad_norm": 8.928816795349121,
210
+ "learning_rate": 3.505434782608696e-05,
211
+ "logits/chosen": -1.2421447038650513,
212
+ "logits/rejected": -1.136252522468567,
213
+ "logps/chosen": -547.4909057617188,
214
+ "logps/rejected": -460.38275146484375,
215
+ "loss": 0.6047,
216
+ "rewards/accuracies": 0.6875,
217
+ "rewards/chosen": 0.21343636512756348,
218
+ "rewards/margins": 0.24689535796642303,
219
+ "rewards/rejected": -0.033459004014730453,
220
+ "step": 140
221
+ },
222
+ {
223
+ "epoch": 0.24549918166939444,
224
+ "grad_norm": 11.977150917053223,
225
+ "learning_rate": 3.777173913043478e-05,
226
+ "logits/chosen": -1.338332176208496,
227
+ "logits/rejected": -1.1015567779541016,
228
+ "logps/chosen": -432.2693786621094,
229
+ "logps/rejected": -423.0946350097656,
230
+ "loss": 0.6946,
231
+ "rewards/accuracies": 0.4749999940395355,
232
+ "rewards/chosen": 0.15119442343711853,
233
+ "rewards/margins": 0.07870660722255707,
234
+ "rewards/rejected": 0.07248781621456146,
235
+ "step": 150
236
+ },
237
+ {
238
+ "epoch": 0.2618657937806874,
239
+ "grad_norm": 13.040871620178223,
240
+ "learning_rate": 4.0489130434782615e-05,
241
+ "logits/chosen": -1.0111101865768433,
242
+ "logits/rejected": NaN,
243
+ "logps/chosen": -472.6917419433594,
244
+ "logps/rejected": -426.6333923339844,
245
+ "loss": 0.6598,
246
+ "rewards/accuracies": 0.5625,
247
+ "rewards/chosen": 0.24765905737876892,
248
+ "rewards/margins": 0.13647916913032532,
249
+ "rewards/rejected": 0.11117993295192719,
250
+ "step": 160
251
+ },
252
+ {
253
+ "epoch": 0.27823240589198034,
254
+ "grad_norm": 8.028325080871582,
255
+ "learning_rate": 4.3206521739130434e-05,
256
+ "logits/chosen": -0.9152072668075562,
257
+ "logits/rejected": -0.8340185284614563,
258
+ "logps/chosen": -438.38555908203125,
259
+ "logps/rejected": -392.93426513671875,
260
+ "loss": 0.6272,
261
+ "rewards/accuracies": 0.612500011920929,
262
+ "rewards/chosen": 0.3285876512527466,
263
+ "rewards/margins": 0.20840699970722198,
264
+ "rewards/rejected": 0.1201806515455246,
265
+ "step": 170
266
+ },
267
+ {
268
+ "epoch": 0.2945990180032733,
269
+ "grad_norm": 9.436712265014648,
270
+ "learning_rate": 4.592391304347826e-05,
271
+ "logits/chosen": -1.0703661441802979,
272
+ "logits/rejected": -1.0127884149551392,
273
+ "logps/chosen": -449.83905029296875,
274
+ "logps/rejected": -437.92822265625,
275
+ "loss": 0.6684,
276
+ "rewards/accuracies": 0.5375000238418579,
277
+ "rewards/chosen": 0.34186336398124695,
278
+ "rewards/margins": 0.14258861541748047,
279
+ "rewards/rejected": 0.19927474856376648,
280
+ "step": 180
281
+ },
282
+ {
283
+ "epoch": 0.3109656301145663,
284
+ "grad_norm": 8.914007186889648,
285
+ "learning_rate": 4.8641304347826086e-05,
286
+ "logits/chosen": -0.8762575387954712,
287
+ "logits/rejected": -0.5631248354911804,
288
+ "logps/chosen": -473.806396484375,
289
+ "logps/rejected": -443.6448669433594,
290
+ "loss": 0.6243,
291
+ "rewards/accuracies": 0.574999988079071,
292
+ "rewards/chosen": 0.3772050440311432,
293
+ "rewards/margins": 0.24287982285022736,
294
+ "rewards/rejected": 0.13432522118091583,
295
+ "step": 190
296
+ },
297
+ {
298
+ "epoch": 0.32733224222585927,
299
+ "grad_norm": 14.76502513885498,
300
+ "learning_rate": 4.98483929654336e-05,
301
+ "logits/chosen": NaN,
302
+ "logits/rejected": -0.9821416139602661,
303
+ "logps/chosen": -447.65386962890625,
304
+ "logps/rejected": -400.1768493652344,
305
+ "loss": 0.6805,
306
+ "rewards/accuracies": 0.5874999761581421,
307
+ "rewards/chosen": 0.2871859073638916,
308
+ "rewards/margins": 0.17125040292739868,
309
+ "rewards/rejected": 0.11593550443649292,
310
+ "step": 200
311
+ },
312
+ {
313
+ "epoch": 0.3436988543371522,
314
+ "grad_norm": 11.80213451385498,
315
+ "learning_rate": 4.954517889630079e-05,
316
+ "logits/chosen": -1.2535641193389893,
317
+ "logits/rejected": -1.0489479303359985,
318
+ "logps/chosen": -457.010986328125,
319
+ "logps/rejected": -432.8558044433594,
320
+ "loss": 0.6769,
321
+ "rewards/accuracies": 0.512499988079071,
322
+ "rewards/chosen": 0.24471895396709442,
323
+ "rewards/margins": 0.16205154359340668,
324
+ "rewards/rejected": 0.08266742527484894,
325
+ "step": 210
326
+ },
327
+ {
328
+ "epoch": 0.36006546644844517,
329
+ "grad_norm": 7.534430980682373,
330
+ "learning_rate": 4.924196482716799e-05,
331
+ "logits/chosen": -1.5050650835037231,
332
+ "logits/rejected": -1.5133517980575562,
333
+ "logps/chosen": -445.3924865722656,
334
+ "logps/rejected": -420.99114990234375,
335
+ "loss": 0.7003,
336
+ "rewards/accuracies": 0.5625,
337
+ "rewards/chosen": 0.18017885088920593,
338
+ "rewards/margins": 0.1387586146593094,
339
+ "rewards/rejected": 0.04142025485634804,
340
+ "step": 220
341
+ },
342
+ {
343
+ "epoch": 0.37643207855973815,
344
+ "grad_norm": 7.207632541656494,
345
+ "learning_rate": 4.893875075803517e-05,
346
+ "logits/chosen": -1.3496973514556885,
347
+ "logits/rejected": -1.2030962705612183,
348
+ "logps/chosen": -436.5039978027344,
349
+ "logps/rejected": -396.91534423828125,
350
+ "loss": 0.6192,
351
+ "rewards/accuracies": 0.574999988079071,
352
+ "rewards/chosen": 0.19479091465473175,
353
+ "rewards/margins": 0.353401243686676,
354
+ "rewards/rejected": -0.15861034393310547,
355
+ "step": 230
356
+ },
357
+ {
358
+ "epoch": 0.39279869067103107,
359
+ "grad_norm": 9.134820938110352,
360
+ "learning_rate": 4.8665858095815645e-05,
361
+ "logits/chosen": -1.2978178262710571,
362
+ "logits/rejected": -1.1864020824432373,
363
+ "logps/chosen": -486.7569885253906,
364
+ "logps/rejected": -457.94586181640625,
365
+ "loss": 0.6931,
366
+ "rewards/accuracies": 0.4749999940395355,
367
+ "rewards/chosen": 0.29108190536499023,
368
+ "rewards/margins": 0.15347807109355927,
369
+ "rewards/rejected": 0.13760384917259216,
370
+ "step": 240
371
+ },
372
+ {
373
+ "epoch": 0.40916530278232405,
374
+ "grad_norm": 10.825005531311035,
375
+ "learning_rate": 4.836264402668284e-05,
376
+ "logits/chosen": -1.3271738290786743,
377
+ "logits/rejected": -1.030890703201294,
378
+ "logps/chosen": -421.89349365234375,
379
+ "logps/rejected": -400.1674499511719,
380
+ "loss": 0.6536,
381
+ "rewards/accuracies": 0.5,
382
+ "rewards/chosen": 0.42532652616500854,
383
+ "rewards/margins": 0.34351587295532227,
384
+ "rewards/rejected": 0.08181065320968628,
385
+ "step": 250
386
+ },
387
+ {
388
+ "epoch": 0.425531914893617,
389
+ "grad_norm": 9.450189590454102,
390
+ "learning_rate": 4.8059429957550034e-05,
391
+ "logits/chosen": -1.0212231874465942,
392
+ "logits/rejected": -1.0058345794677734,
393
+ "logps/chosen": -422.1316833496094,
394
+ "logps/rejected": -390.0425109863281,
395
+ "loss": 0.6384,
396
+ "rewards/accuracies": 0.5874999761581421,
397
+ "rewards/chosen": 0.490867555141449,
398
+ "rewards/margins": 0.35965341329574585,
399
+ "rewards/rejected": 0.13121409714221954,
400
+ "step": 260
401
+ },
402
+ {
403
+ "epoch": 0.44189852700491,
404
+ "grad_norm": 8.231124877929688,
405
+ "learning_rate": 4.7756215888417225e-05,
406
+ "logits/chosen": -0.9735445976257324,
407
+ "logits/rejected": -0.937743067741394,
408
+ "logps/chosen": -552.4168701171875,
409
+ "logps/rejected": -501.0535583496094,
410
+ "loss": 0.6694,
411
+ "rewards/accuracies": 0.574999988079071,
412
+ "rewards/chosen": 0.4536367952823639,
413
+ "rewards/margins": 0.19916944205760956,
414
+ "rewards/rejected": 0.2544674277305603,
415
+ "step": 270
416
+ },
417
+ {
418
+ "epoch": 0.4582651391162029,
419
+ "grad_norm": 9.207930564880371,
420
+ "learning_rate": 4.7453001819284416e-05,
421
+ "logits/chosen": -0.9774928092956543,
422
+ "logits/rejected": -0.8485246896743774,
423
+ "logps/chosen": -444.54827880859375,
424
+ "logps/rejected": -419.5541076660156,
425
+ "loss": 0.6607,
426
+ "rewards/accuracies": 0.5874999761581421,
427
+ "rewards/chosen": 0.3690536618232727,
428
+ "rewards/margins": 0.22144463658332825,
429
+ "rewards/rejected": 0.14760902523994446,
430
+ "step": 280
431
+ },
432
+ {
433
+ "epoch": 0.4746317512274959,
434
+ "grad_norm": 6.876308441162109,
435
+ "learning_rate": 4.714978775015161e-05,
436
+ "logits/chosen": -1.2095867395401,
437
+ "logits/rejected": -0.8773800730705261,
438
+ "logps/chosen": -437.869384765625,
439
+ "logps/rejected": -419.9366149902344,
440
+ "loss": 0.5867,
441
+ "rewards/accuracies": 0.612500011920929,
442
+ "rewards/chosen": 0.5094749331474304,
443
+ "rewards/margins": 0.5250319242477417,
444
+ "rewards/rejected": -0.015556998550891876,
445
+ "step": 290
446
+ },
447
+ {
448
+ "epoch": 0.4909983633387889,
449
+ "grad_norm": 12.918950080871582,
450
+ "learning_rate": 4.68465736810188e-05,
451
+ "logits/chosen": -0.9076827764511108,
452
+ "logits/rejected": -0.7181829214096069,
453
+ "logps/chosen": -469.4864196777344,
454
+ "logps/rejected": -420.38885498046875,
455
+ "loss": 0.6833,
456
+ "rewards/accuracies": 0.5625,
457
+ "rewards/chosen": 0.5444287657737732,
458
+ "rewards/margins": 0.2743019461631775,
459
+ "rewards/rejected": 0.2701267600059509,
460
+ "step": 300
461
+ },
462
+ {
463
+ "epoch": 0.5073649754500819,
464
+ "grad_norm": 4.9246110916137695,
465
+ "learning_rate": 4.6543359611885996e-05,
466
+ "logits/chosen": -1.214247226715088,
467
+ "logits/rejected": -0.9910456538200378,
468
+ "logps/chosen": -385.4250183105469,
469
+ "logps/rejected": -390.26043701171875,
470
+ "loss": 0.6419,
471
+ "rewards/accuracies": 0.675000011920929,
472
+ "rewards/chosen": 0.4199371337890625,
473
+ "rewards/margins": 0.34462255239486694,
474
+ "rewards/rejected": 0.07531457394361496,
475
+ "step": 310
476
+ },
477
+ {
478
+ "epoch": 0.5237315875613748,
479
+ "grad_norm": 8.829249382019043,
480
+ "learning_rate": 4.624014554275319e-05,
481
+ "logits/chosen": -1.0396336317062378,
482
+ "logits/rejected": -0.9439867734909058,
483
+ "logps/chosen": -452.1678161621094,
484
+ "logps/rejected": -379.70745849609375,
485
+ "loss": 0.6396,
486
+ "rewards/accuracies": 0.5625,
487
+ "rewards/chosen": 0.472791850566864,
488
+ "rewards/margins": 0.23509594798088074,
489
+ "rewards/rejected": 0.23769588768482208,
490
+ "step": 320
491
+ },
492
+ {
493
+ "epoch": 0.5400981996726678,
494
+ "grad_norm": 13.249287605285645,
495
+ "learning_rate": 4.593693147362038e-05,
496
+ "logits/chosen": -0.927478015422821,
497
+ "logits/rejected": -0.7090943455696106,
498
+ "logps/chosen": -488.009765625,
499
+ "logps/rejected": -473.47833251953125,
500
+ "loss": 0.6833,
501
+ "rewards/accuracies": 0.550000011920929,
502
+ "rewards/chosen": 0.4911763668060303,
503
+ "rewards/margins": 0.25351986289024353,
504
+ "rewards/rejected": 0.23765654861927032,
505
+ "step": 330
506
+ },
507
+ {
508
+ "epoch": 0.5564648117839607,
509
+ "grad_norm": 9.580172538757324,
510
+ "learning_rate": 4.566403881140085e-05,
511
+ "logits/chosen": -0.9790302515029907,
512
+ "logits/rejected": -0.8282285928726196,
513
+ "logps/chosen": -481.6971740722656,
514
+ "logps/rejected": -441.3619689941406,
515
+ "loss": 0.6922,
516
+ "rewards/accuracies": 0.550000011920929,
517
+ "rewards/chosen": 0.4195849299430847,
518
+ "rewards/margins": 0.210377499461174,
519
+ "rewards/rejected": 0.20920750498771667,
520
+ "step": 340
521
+ },
522
+ {
523
+ "epoch": 0.5728314238952537,
524
+ "grad_norm": 6.872241973876953,
525
+ "learning_rate": 4.536082474226804e-05,
526
+ "logits/chosen": -0.9942235946655273,
527
+ "logits/rejected": -0.824662983417511,
528
+ "logps/chosen": -437.9869079589844,
529
+ "logps/rejected": -417.3473205566406,
530
+ "loss": 0.6071,
531
+ "rewards/accuracies": 0.6000000238418579,
532
+ "rewards/chosen": 0.49881643056869507,
533
+ "rewards/margins": 0.41778793931007385,
534
+ "rewards/rejected": 0.08102855831384659,
535
+ "step": 350
536
+ },
537
+ {
538
+ "epoch": 0.5891980360065466,
539
+ "grad_norm": 7.748345375061035,
540
+ "learning_rate": 4.5057610673135234e-05,
541
+ "logits/chosen": -1.0704625844955444,
542
+ "logits/rejected": -0.7879603505134583,
543
+ "logps/chosen": -428.53021240234375,
544
+ "logps/rejected": -435.9026794433594,
545
+ "loss": 0.6084,
546
+ "rewards/accuracies": 0.612500011920929,
547
+ "rewards/chosen": 0.4484317898750305,
548
+ "rewards/margins": 0.4125775396823883,
549
+ "rewards/rejected": 0.035854239016771317,
550
+ "step": 360
551
+ },
552
+ {
553
+ "epoch": 0.6055646481178396,
554
+ "grad_norm": 7.428864479064941,
555
+ "learning_rate": 4.475439660400243e-05,
556
+ "logits/chosen": -1.098026990890503,
557
+ "logits/rejected": -0.9232645034790039,
558
+ "logps/chosen": -441.7071228027344,
559
+ "logps/rejected": -402.95770263671875,
560
+ "loss": 0.6901,
561
+ "rewards/accuracies": 0.550000011920929,
562
+ "rewards/chosen": 0.37514036893844604,
563
+ "rewards/margins": 0.2525361478328705,
564
+ "rewards/rejected": 0.12260414659976959,
565
+ "step": 370
566
+ },
567
+ {
568
+ "epoch": 0.6219312602291326,
569
+ "grad_norm": 9.591557502746582,
570
+ "learning_rate": 4.445118253486962e-05,
571
+ "logits/chosen": -1.2409312725067139,
572
+ "logits/rejected": -0.8442489504814148,
573
+ "logps/chosen": -466.13311767578125,
574
+ "logps/rejected": -420.9102478027344,
575
+ "loss": 0.7684,
576
+ "rewards/accuracies": 0.4625000059604645,
577
+ "rewards/chosen": 0.2286566197872162,
578
+ "rewards/margins": 0.11457403004169464,
579
+ "rewards/rejected": 0.11408257484436035,
580
+ "step": 380
581
+ },
582
+ {
583
+ "epoch": 0.6382978723404256,
584
+ "grad_norm": 7.025293350219727,
585
+ "learning_rate": 4.414796846573681e-05,
586
+ "logits/chosen": -1.1844158172607422,
587
+ "logits/rejected": -1.1069843769073486,
588
+ "logps/chosen": -428.5450134277344,
589
+ "logps/rejected": -371.85614013671875,
590
+ "loss": 0.6921,
591
+ "rewards/accuracies": 0.550000011920929,
592
+ "rewards/chosen": 0.2637328803539276,
593
+ "rewards/margins": 0.10889965295791626,
594
+ "rewards/rejected": 0.15483322739601135,
595
+ "step": 390
596
+ },
597
+ {
598
+ "epoch": 0.6546644844517185,
599
+ "grad_norm": 19.2535343170166,
600
+ "learning_rate": 4.3844754396604006e-05,
601
+ "logits/chosen": -1.1141570806503296,
602
+ "logits/rejected": -0.8798454999923706,
603
+ "logps/chosen": -493.08197021484375,
604
+ "logps/rejected": -465.0943298339844,
605
+ "loss": 0.6545,
606
+ "rewards/accuracies": 0.5874999761581421,
607
+ "rewards/chosen": 0.4174181818962097,
608
+ "rewards/margins": 0.25259822607040405,
609
+ "rewards/rejected": 0.16481998562812805,
610
+ "step": 400
611
+ },
612
+ {
613
+ "epoch": 0.6710310965630114,
614
+ "grad_norm": 8.808697700500488,
615
+ "learning_rate": 4.35415403274712e-05,
616
+ "logits/chosen": -0.9878439903259277,
617
+ "logits/rejected": -0.7299278974533081,
618
+ "logps/chosen": -523.9139404296875,
619
+ "logps/rejected": -482.0597229003906,
620
+ "loss": 0.552,
621
+ "rewards/accuracies": 0.699999988079071,
622
+ "rewards/chosen": 0.4945390820503235,
623
+ "rewards/margins": 0.6175243854522705,
624
+ "rewards/rejected": -0.1229853481054306,
625
+ "step": 410
626
+ },
627
+ {
628
+ "epoch": 0.6873977086743044,
629
+ "grad_norm": 9.99456787109375,
630
+ "learning_rate": 4.323832625833839e-05,
631
+ "logits/chosen": -1.0902565717697144,
632
+ "logits/rejected": -0.9221500158309937,
633
+ "logps/chosen": -443.0076599121094,
634
+ "logps/rejected": -375.8013610839844,
635
+ "loss": 0.5684,
636
+ "rewards/accuracies": 0.5874999761581421,
637
+ "rewards/chosen": 0.46177297830581665,
638
+ "rewards/margins": 0.509793221950531,
639
+ "rewards/rejected": -0.048020295798778534,
640
+ "step": 420
641
+ },
642
+ {
643
+ "epoch": 0.7037643207855974,
644
+ "grad_norm": 5.767345428466797,
645
+ "learning_rate": 4.293511218920558e-05,
646
+ "logits/chosen": -1.0450439453125,
647
+ "logits/rejected": -0.7902520298957825,
648
+ "logps/chosen": -458.2330017089844,
649
+ "logps/rejected": -382.6787109375,
650
+ "loss": 0.6126,
651
+ "rewards/accuracies": 0.6499999761581421,
652
+ "rewards/chosen": 0.5793684124946594,
653
+ "rewards/margins": 0.35584205389022827,
654
+ "rewards/rejected": 0.22352632880210876,
655
+ "step": 430
656
+ },
657
+ {
658
+ "epoch": 0.7201309328968903,
659
+ "grad_norm": 11.51710033416748,
660
+ "learning_rate": 4.263189812007277e-05,
661
+ "logits/chosen": -0.8710979223251343,
662
+ "logits/rejected": -0.783151388168335,
663
+ "logps/chosen": -451.32122802734375,
664
+ "logps/rejected": -416.0005798339844,
665
+ "loss": 0.6751,
666
+ "rewards/accuracies": 0.5625,
667
+ "rewards/chosen": 0.5284687280654907,
668
+ "rewards/margins": 0.2338237315416336,
669
+ "rewards/rejected": 0.2946449816226959,
670
+ "step": 440
671
+ },
672
+ {
673
+ "epoch": 0.7364975450081833,
674
+ "grad_norm": 8.408599853515625,
675
+ "learning_rate": 4.232868405093997e-05,
676
+ "logits/chosen": -0.9611997604370117,
677
+ "logits/rejected": -0.8005222082138062,
678
+ "logps/chosen": -390.06817626953125,
679
+ "logps/rejected": -376.57183837890625,
680
+ "loss": 0.723,
681
+ "rewards/accuracies": 0.48750001192092896,
682
+ "rewards/chosen": 0.35626164078712463,
683
+ "rewards/margins": 0.07882154732942581,
684
+ "rewards/rejected": 0.27744007110595703,
685
+ "step": 450
686
+ },
687
+ {
688
+ "epoch": 0.7528641571194763,
689
+ "grad_norm": 7.085683822631836,
690
+ "learning_rate": 4.202546998180716e-05,
691
+ "logits/chosen": -1.181065559387207,
692
+ "logits/rejected": -0.9188621640205383,
693
+ "logps/chosen": -446.1695861816406,
694
+ "logps/rejected": -461.0537109375,
695
+ "loss": 0.6376,
696
+ "rewards/accuracies": 0.574999988079071,
697
+ "rewards/chosen": 0.6099992990493774,
698
+ "rewards/margins": 0.40477481484413147,
699
+ "rewards/rejected": 0.20522448420524597,
700
+ "step": 460
701
+ },
702
+ {
703
+ "epoch": 0.7692307692307693,
704
+ "grad_norm": 7.778029918670654,
705
+ "learning_rate": 4.172225591267435e-05,
706
+ "logits/chosen": -1.1612136363983154,
707
+ "logits/rejected": -0.7909379601478577,
708
+ "logps/chosen": -415.356201171875,
709
+ "logps/rejected": -405.0163879394531,
710
+ "loss": 0.639,
711
+ "rewards/accuracies": 0.512499988079071,
712
+ "rewards/chosen": 0.5554541349411011,
713
+ "rewards/margins": 0.49387890100479126,
714
+ "rewards/rejected": 0.06157524511218071,
715
+ "step": 470
716
+ },
717
+ {
718
+ "epoch": 0.7855973813420621,
719
+ "grad_norm": 9.144447326660156,
720
+ "learning_rate": 4.141904184354154e-05,
721
+ "logits/chosen": -1.1227318048477173,
722
+ "logits/rejected": -0.8980533480644226,
723
+ "logps/chosen": -470.1067810058594,
724
+ "logps/rejected": -385.7353210449219,
725
+ "loss": 0.6301,
726
+ "rewards/accuracies": 0.5874999761581421,
727
+ "rewards/chosen": 0.5011923909187317,
728
+ "rewards/margins": 0.36913150548934937,
729
+ "rewards/rejected": 0.1320609152317047,
730
+ "step": 480
731
+ },
732
+ {
733
+ "epoch": 0.8019639934533551,
734
+ "grad_norm": 9.528294563293457,
735
+ "learning_rate": 4.111582777440873e-05,
736
+ "logits/chosen": -1.4198505878448486,
737
+ "logits/rejected": -1.1528406143188477,
738
+ "logps/chosen": -430.28668212890625,
739
+ "logps/rejected": -385.2845764160156,
740
+ "loss": 0.634,
741
+ "rewards/accuracies": 0.5625,
742
+ "rewards/chosen": 0.6369161605834961,
743
+ "rewards/margins": 0.4485607147216797,
744
+ "rewards/rejected": 0.18835540115833282,
745
+ "step": 490
746
+ },
747
+ {
748
+ "epoch": 0.8183306055646481,
749
+ "grad_norm": 6.964690208435059,
750
+ "learning_rate": 4.0812613705275924e-05,
751
+ "logits/chosen": -1.1329933404922485,
752
+ "logits/rejected": -1.066542625427246,
753
+ "logps/chosen": -423.4967346191406,
754
+ "logps/rejected": -376.92803955078125,
755
+ "loss": 0.5904,
756
+ "rewards/accuracies": 0.550000011920929,
757
+ "rewards/chosen": 0.6795145273208618,
758
+ "rewards/margins": 0.3617789149284363,
759
+ "rewards/rejected": 0.3177356719970703,
760
+ "step": 500
761
+ },
762
+ {
763
+ "epoch": 0.8346972176759411,
764
+ "grad_norm": 11.78022289276123,
765
+ "learning_rate": 4.050939963614312e-05,
766
+ "logits/chosen": -1.1565020084381104,
767
+ "logits/rejected": -1.064300298690796,
768
+ "logps/chosen": -465.0497131347656,
769
+ "logps/rejected": -438.4883728027344,
770
+ "loss": 0.6384,
771
+ "rewards/accuracies": 0.5249999761581421,
772
+ "rewards/chosen": 0.6123272180557251,
773
+ "rewards/margins": 0.3706541657447815,
774
+ "rewards/rejected": 0.241673082113266,
775
+ "step": 510
776
+ },
777
+ {
778
+ "epoch": 0.851063829787234,
779
+ "grad_norm": 6.378419876098633,
780
+ "learning_rate": 4.020618556701031e-05,
781
+ "logits/chosen": -1.0936065912246704,
782
+ "logits/rejected": -1.157414436340332,
783
+ "logps/chosen": -431.05657958984375,
784
+ "logps/rejected": -380.1803283691406,
785
+ "loss": 0.6866,
786
+ "rewards/accuracies": 0.5,
787
+ "rewards/chosen": 0.514751672744751,
788
+ "rewards/margins": 0.2915990948677063,
789
+ "rewards/rejected": 0.22315259277820587,
790
+ "step": 520
791
+ },
792
+ {
793
+ "epoch": 0.867430441898527,
794
+ "grad_norm": 6.811694145202637,
795
+ "learning_rate": 3.9902971497877504e-05,
796
+ "logits/chosen": -1.0564637184143066,
797
+ "logits/rejected": -0.8537761569023132,
798
+ "logps/chosen": -400.8801574707031,
799
+ "logps/rejected": -400.9678039550781,
800
+ "loss": 0.6036,
801
+ "rewards/accuracies": 0.625,
802
+ "rewards/chosen": 0.5905786156654358,
803
+ "rewards/margins": 0.49240899085998535,
804
+ "rewards/rejected": 0.09816964715719223,
805
+ "step": 530
806
+ },
807
+ {
808
+ "epoch": 0.88379705400982,
809
+ "grad_norm": 8.516433715820312,
810
+ "learning_rate": 3.9599757428744695e-05,
811
+ "logits/chosen": -1.0425868034362793,
812
+ "logits/rejected": -1.0106391906738281,
813
+ "logps/chosen": -494.23681640625,
814
+ "logps/rejected": -451.3269958496094,
815
+ "loss": 0.7295,
816
+ "rewards/accuracies": 0.5,
817
+ "rewards/chosen": 0.480445921421051,
818
+ "rewards/margins": 0.12438355386257172,
819
+ "rewards/rejected": 0.3560623824596405,
820
+ "step": 540
821
+ },
822
+ {
823
+ "epoch": 0.900163666121113,
824
+ "grad_norm": 9.584715843200684,
825
+ "learning_rate": 3.9296543359611886e-05,
826
+ "logits/chosen": -1.1160447597503662,
827
+ "logits/rejected": -0.898239016532898,
828
+ "logps/chosen": -471.17718505859375,
829
+ "logps/rejected": -404.9838562011719,
830
+ "loss": 0.5973,
831
+ "rewards/accuracies": 0.625,
832
+ "rewards/chosen": 0.4906229078769684,
833
+ "rewards/margins": 0.3674588203430176,
834
+ "rewards/rejected": 0.1231641173362732,
835
+ "step": 550
836
+ },
837
+ {
838
+ "epoch": 0.9165302782324058,
839
+ "grad_norm": 9.050358772277832,
840
+ "learning_rate": 3.899332929047908e-05,
841
+ "logits/chosen": -1.0188591480255127,
842
+ "logits/rejected": NaN,
843
+ "logps/chosen": -505.81146240234375,
844
+ "logps/rejected": -455.074462890625,
845
+ "loss": 0.672,
846
+ "rewards/accuracies": 0.48750001192092896,
847
+ "rewards/chosen": 0.5001901388168335,
848
+ "rewards/margins": 0.3212779462337494,
849
+ "rewards/rejected": 0.1789121776819229,
850
+ "step": 560
851
+ },
852
+ {
853
+ "epoch": 0.9328968903436988,
854
+ "grad_norm": 10.636481285095215,
855
+ "learning_rate": 3.8690115221346276e-05,
856
+ "logits/chosen": -1.0234591960906982,
857
+ "logits/rejected": -0.9390643835067749,
858
+ "logps/chosen": -423.03594970703125,
859
+ "logps/rejected": -383.97369384765625,
860
+ "loss": 0.6515,
861
+ "rewards/accuracies": 0.6000000238418579,
862
+ "rewards/chosen": 0.40482932329177856,
863
+ "rewards/margins": 0.20538429915905,
864
+ "rewards/rejected": 0.19944503903388977,
865
+ "step": 570
866
+ },
867
+ {
868
+ "epoch": 0.9492635024549918,
869
+ "grad_norm": 12.001391410827637,
870
+ "learning_rate": 3.838690115221347e-05,
871
+ "logits/chosen": -1.254690170288086,
872
+ "logits/rejected": -1.098749041557312,
873
+ "logps/chosen": -400.72021484375,
874
+ "logps/rejected": -348.39678955078125,
875
+ "loss": 0.6332,
876
+ "rewards/accuracies": 0.5874999761581421,
877
+ "rewards/chosen": 0.5642321705818176,
878
+ "rewards/margins": 0.34682947397232056,
879
+ "rewards/rejected": 0.21740269660949707,
880
+ "step": 580
881
+ },
882
+ {
883
+ "epoch": 0.9656301145662848,
884
+ "grad_norm": 9.584880828857422,
885
+ "learning_rate": 3.808368708308066e-05,
886
+ "logits/chosen": -0.9567238092422485,
887
+ "logits/rejected": -0.8445194959640503,
888
+ "logps/chosen": -475.3368225097656,
889
+ "logps/rejected": -418.3975524902344,
890
+ "loss": 0.609,
891
+ "rewards/accuracies": 0.5874999761581421,
892
+ "rewards/chosen": 0.7818068861961365,
893
+ "rewards/margins": 0.4438362717628479,
894
+ "rewards/rejected": 0.3379705548286438,
895
+ "step": 590
896
+ },
897
+ {
898
+ "epoch": 0.9819967266775778,
899
+ "grad_norm": 8.225208282470703,
900
+ "learning_rate": 3.778047301394785e-05,
901
+ "logits/chosen": -0.8194648623466492,
902
+ "logits/rejected": -0.6516283750534058,
903
+ "logps/chosen": -464.95648193359375,
904
+ "logps/rejected": -454.74896240234375,
905
+ "loss": 0.6792,
906
+ "rewards/accuracies": 0.612500011920929,
907
+ "rewards/chosen": 0.673040509223938,
908
+ "rewards/margins": 0.3073888421058655,
909
+ "rewards/rejected": 0.3656516671180725,
910
+ "step": 600
911
+ },
912
+ {
913
+ "epoch": 0.9983633387888707,
914
+ "grad_norm": 7.938261032104492,
915
+ "learning_rate": 3.747725894481504e-05,
916
+ "logits/chosen": -0.9632278680801392,
917
+ "logits/rejected": -0.7192381620407104,
918
+ "logps/chosen": -428.2535095214844,
919
+ "logps/rejected": -426.05780029296875,
920
+ "loss": 0.5784,
921
+ "rewards/accuracies": 0.675000011920929,
922
+ "rewards/chosen": 0.778376042842865,
923
+ "rewards/margins": 0.44757041335105896,
924
+ "rewards/rejected": 0.33080559968948364,
925
+ "step": 610
926
+ },
927
+ {
928
+ "epoch": 1.0147299509001637,
929
+ "grad_norm": 6.6163482666015625,
930
+ "learning_rate": 3.717404487568223e-05,
931
+ "logits/chosen": -0.8798831105232239,
932
+ "logits/rejected": -0.716325581073761,
933
+ "logps/chosen": -401.0724182128906,
934
+ "logps/rejected": -393.04925537109375,
935
+ "loss": 0.5397,
936
+ "rewards/accuracies": 0.5874999761581421,
937
+ "rewards/chosen": 0.6812083721160889,
938
+ "rewards/margins": 0.6590946912765503,
939
+ "rewards/rejected": 0.022113658487796783,
940
+ "step": 620
941
+ },
942
+ {
943
+ "epoch": 1.0310965630114566,
944
+ "grad_norm": 4.5942792892456055,
945
+ "learning_rate": 3.687083080654943e-05,
946
+ "logits/chosen": -0.595669150352478,
947
+ "logits/rejected": -0.6483780741691589,
948
+ "logps/chosen": -519.7013549804688,
949
+ "logps/rejected": -459.6736755371094,
950
+ "loss": 0.3798,
951
+ "rewards/accuracies": 0.8125,
952
+ "rewards/chosen": 0.9690455198287964,
953
+ "rewards/margins": 1.2637821435928345,
954
+ "rewards/rejected": -0.2947366237640381,
955
+ "step": 630
956
+ },
957
+ {
958
+ "epoch": 1.0474631751227497,
959
+ "grad_norm": 5.822299003601074,
960
+ "learning_rate": 3.6567616737416614e-05,
961
+ "logits/chosen": -1.1080423593521118,
962
+ "logits/rejected": -0.8690654635429382,
963
+ "logps/chosen": -430.8399963378906,
964
+ "logps/rejected": -382.16131591796875,
965
+ "loss": 0.3725,
966
+ "rewards/accuracies": 0.8500000238418579,
967
+ "rewards/chosen": 0.926641583442688,
968
+ "rewards/margins": 1.1296237707138062,
969
+ "rewards/rejected": -0.20298215746879578,
970
+ "step": 640
971
+ },
972
+ {
973
+ "epoch": 1.0638297872340425,
974
+ "grad_norm": 5.2786078453063965,
975
+ "learning_rate": 3.626440266828381e-05,
976
+ "logits/chosen": -0.986005961894989,
977
+ "logits/rejected": -0.8763575553894043,
978
+ "logps/chosen": -407.85260009765625,
979
+ "logps/rejected": -399.2723693847656,
980
+ "loss": 0.4381,
981
+ "rewards/accuracies": 0.6875,
982
+ "rewards/chosen": 0.8758468627929688,
983
+ "rewards/margins": 0.9750221371650696,
984
+ "rewards/rejected": -0.09917531907558441,
985
+ "step": 650
986
+ },
987
+ {
988
+ "epoch": 1.0801963993453354,
989
+ "grad_norm": 11.273237228393555,
990
+ "learning_rate": 3.5961188599151e-05,
991
+ "logits/chosen": -1.3430869579315186,
992
+ "logits/rejected": -1.056379795074463,
993
+ "logps/chosen": -410.79791259765625,
994
+ "logps/rejected": -434.85845947265625,
995
+ "loss": 0.4005,
996
+ "rewards/accuracies": 0.8125,
997
+ "rewards/chosen": 0.9130380749702454,
998
+ "rewards/margins": 0.9950882196426392,
999
+ "rewards/rejected": -0.08205002546310425,
1000
+ "step": 660
1001
+ },
1002
+ {
1003
+ "epoch": 1.0965630114566285,
1004
+ "grad_norm": 7.220339775085449,
1005
+ "learning_rate": 3.5657974530018194e-05,
1006
+ "logits/chosen": -1.0506983995437622,
1007
+ "logits/rejected": -1.0361990928649902,
1008
+ "logps/chosen": -408.18035888671875,
1009
+ "logps/rejected": -391.70562744140625,
1010
+ "loss": 0.4471,
1011
+ "rewards/accuracies": 0.7749999761581421,
1012
+ "rewards/chosen": 0.7731181979179382,
1013
+ "rewards/margins": 0.9366775751113892,
1014
+ "rewards/rejected": -0.16355939209461212,
1015
+ "step": 670
1016
+ },
1017
+ {
1018
+ "epoch": 1.1129296235679214,
1019
+ "grad_norm": 5.74124813079834,
1020
+ "learning_rate": 3.535476046088539e-05,
1021
+ "logits/chosen": -1.206479787826538,
1022
+ "logits/rejected": -1.1451863050460815,
1023
+ "logps/chosen": -444.80438232421875,
1024
+ "logps/rejected": -439.0628356933594,
1025
+ "loss": 0.4043,
1026
+ "rewards/accuracies": 0.800000011920929,
1027
+ "rewards/chosen": 0.8405340313911438,
1028
+ "rewards/margins": 1.1449936628341675,
1029
+ "rewards/rejected": -0.30445969104766846,
1030
+ "step": 680
1031
+ },
1032
+ {
1033
+ "epoch": 1.1292962356792144,
1034
+ "grad_norm": 7.32386589050293,
1035
+ "learning_rate": 3.5051546391752576e-05,
1036
+ "logits/chosen": -1.0227863788604736,
1037
+ "logits/rejected": -0.9067742228507996,
1038
+ "logps/chosen": -476.1012268066406,
1039
+ "logps/rejected": -466.14971923828125,
1040
+ "loss": 0.4261,
1041
+ "rewards/accuracies": 0.800000011920929,
1042
+ "rewards/chosen": 0.8802071809768677,
1043
+ "rewards/margins": 1.1421091556549072,
1044
+ "rewards/rejected": -0.2619021236896515,
1045
+ "step": 690
1046
+ },
1047
+ {
1048
+ "epoch": 1.1456628477905073,
1049
+ "grad_norm": 4.43737268447876,
1050
+ "learning_rate": 3.474833232261977e-05,
1051
+ "logits/chosen": -1.135181188583374,
1052
+ "logits/rejected": -1.0297483205795288,
1053
+ "logps/chosen": -506.5615234375,
1054
+ "logps/rejected": -483.1039123535156,
1055
+ "loss": 0.3875,
1056
+ "rewards/accuracies": 0.8125,
1057
+ "rewards/chosen": 0.5547564029693604,
1058
+ "rewards/margins": 1.281551480293274,
1059
+ "rewards/rejected": -0.7267951369285583,
1060
+ "step": 700
1061
+ },
1062
+ {
1063
+ "epoch": 1.1620294599018004,
1064
+ "grad_norm": 2.298978328704834,
1065
+ "learning_rate": 3.4445118253486965e-05,
1066
+ "logits/chosen": -1.426138162612915,
1067
+ "logits/rejected": -1.091489553451538,
1068
+ "logps/chosen": -400.107177734375,
1069
+ "logps/rejected": -401.62109375,
1070
+ "loss": 0.4072,
1071
+ "rewards/accuracies": 0.75,
1072
+ "rewards/chosen": 0.6946335434913635,
1073
+ "rewards/margins": 1.441784143447876,
1074
+ "rewards/rejected": -0.7471505999565125,
1075
+ "step": 710
1076
+ },
1077
+ {
1078
+ "epoch": 1.1783960720130933,
1079
+ "grad_norm": 4.7108154296875,
1080
+ "learning_rate": 3.4141904184354156e-05,
1081
+ "logits/chosen": -1.1755582094192505,
1082
+ "logits/rejected": -0.901213526725769,
1083
+ "logps/chosen": -425.28759765625,
1084
+ "logps/rejected": -402.042236328125,
1085
+ "loss": 0.4615,
1086
+ "rewards/accuracies": 0.762499988079071,
1087
+ "rewards/chosen": 0.6500431299209595,
1088
+ "rewards/margins": 0.9347268342971802,
1089
+ "rewards/rejected": -0.2846837639808655,
1090
+ "step": 720
1091
+ },
1092
+ {
1093
+ "epoch": 1.1947626841243864,
1094
+ "grad_norm": 12.316296577453613,
1095
+ "learning_rate": 3.383869011522135e-05,
1096
+ "logits/chosen": -1.284589409828186,
1097
+ "logits/rejected": -1.0795433521270752,
1098
+ "logps/chosen": -413.8037109375,
1099
+ "logps/rejected": -491.56146240234375,
1100
+ "loss": 0.4385,
1101
+ "rewards/accuracies": 0.737500011920929,
1102
+ "rewards/chosen": 0.6515102386474609,
1103
+ "rewards/margins": 1.2281266450881958,
1104
+ "rewards/rejected": -0.5766164660453796,
1105
+ "step": 730
1106
+ },
1107
+ {
1108
+ "epoch": 1.2111292962356792,
1109
+ "grad_norm": 9.386224746704102,
1110
+ "learning_rate": 3.353547604608854e-05,
1111
+ "logits/chosen": -1.257037878036499,
1112
+ "logits/rejected": -1.0705680847167969,
1113
+ "logps/chosen": -505.6416931152344,
1114
+ "logps/rejected": -459.56805419921875,
1115
+ "loss": 0.3726,
1116
+ "rewards/accuracies": 0.7749999761581421,
1117
+ "rewards/chosen": 0.7937537431716919,
1118
+ "rewards/margins": 1.4233770370483398,
1119
+ "rewards/rejected": -0.629623293876648,
1120
+ "step": 740
1121
+ },
1122
+ {
1123
+ "epoch": 1.227495908346972,
1124
+ "grad_norm": 6.265888214111328,
1125
+ "learning_rate": 3.323226197695573e-05,
1126
+ "logits/chosen": -1.332099199295044,
1127
+ "logits/rejected": -1.0086009502410889,
1128
+ "logps/chosen": -457.0177307128906,
1129
+ "logps/rejected": -449.7435607910156,
1130
+ "loss": 0.4034,
1131
+ "rewards/accuracies": 0.7875000238418579,
1132
+ "rewards/chosen": 0.790605366230011,
1133
+ "rewards/margins": 1.2190719842910767,
1134
+ "rewards/rejected": -0.4284667372703552,
1135
+ "step": 750
1136
+ },
1137
+ {
1138
+ "epoch": 1.2438625204582652,
1139
+ "grad_norm": 6.070497035980225,
1140
+ "learning_rate": 3.292904790782292e-05,
1141
+ "logits/chosen": -1.2865736484527588,
1142
+ "logits/rejected": -1.072819471359253,
1143
+ "logps/chosen": -432.710693359375,
1144
+ "logps/rejected": -467.6375427246094,
1145
+ "loss": 0.3773,
1146
+ "rewards/accuracies": 0.800000011920929,
1147
+ "rewards/chosen": 0.7443059682846069,
1148
+ "rewards/margins": 1.3002400398254395,
1149
+ "rewards/rejected": -0.5559341907501221,
1150
+ "step": 760
1151
+ },
1152
+ {
1153
+ "epoch": 1.260229132569558,
1154
+ "grad_norm": 5.7988128662109375,
1155
+ "learning_rate": 3.262583383869012e-05,
1156
+ "logits/chosen": -1.0803556442260742,
1157
+ "logits/rejected": -0.8561310768127441,
1158
+ "logps/chosen": -514.7871704101562,
1159
+ "logps/rejected": -503.19500732421875,
1160
+ "loss": 0.3216,
1161
+ "rewards/accuracies": 0.8500000238418579,
1162
+ "rewards/chosen": 0.7713959217071533,
1163
+ "rewards/margins": 1.6244882345199585,
1164
+ "rewards/rejected": -0.8530920743942261,
1165
+ "step": 770
1166
+ },
1167
+ {
1168
+ "epoch": 1.2765957446808511,
1169
+ "grad_norm": 4.323042869567871,
1170
+ "learning_rate": 3.232261976955731e-05,
1171
+ "logits/chosen": -1.408276915550232,
1172
+ "logits/rejected": -0.9835014343261719,
1173
+ "logps/chosen": -456.21142578125,
1174
+ "logps/rejected": -429.10467529296875,
1175
+ "loss": 0.3832,
1176
+ "rewards/accuracies": 0.824999988079071,
1177
+ "rewards/chosen": 0.5816126465797424,
1178
+ "rewards/margins": 1.3987071514129639,
1179
+ "rewards/rejected": -0.8170945048332214,
1180
+ "step": 780
1181
+ },
1182
+ {
1183
+ "epoch": 1.292962356792144,
1184
+ "grad_norm": 7.249789237976074,
1185
+ "learning_rate": 3.20194057004245e-05,
1186
+ "logits/chosen": -1.1671850681304932,
1187
+ "logits/rejected": -0.653769314289093,
1188
+ "logps/chosen": -481.74859619140625,
1189
+ "logps/rejected": -441.5457458496094,
1190
+ "loss": 0.3293,
1191
+ "rewards/accuracies": 0.887499988079071,
1192
+ "rewards/chosen": 0.5319666862487793,
1193
+ "rewards/margins": 1.7422034740447998,
1194
+ "rewards/rejected": -1.21023690700531,
1195
+ "step": 790
1196
+ },
1197
+ {
1198
+ "epoch": 1.3093289689034369,
1199
+ "grad_norm": 9.989459037780762,
1200
+ "learning_rate": 3.171619163129169e-05,
1201
+ "logits/chosen": -1.2330873012542725,
1202
+ "logits/rejected": -1.1575677394866943,
1203
+ "logps/chosen": -462.0821228027344,
1204
+ "logps/rejected": -419.34149169921875,
1205
+ "loss": 0.4737,
1206
+ "rewards/accuracies": 0.75,
1207
+ "rewards/chosen": 0.7070258855819702,
1208
+ "rewards/margins": 1.2338263988494873,
1209
+ "rewards/rejected": -0.5268004536628723,
1210
+ "step": 800
1211
+ },
1212
+ {
1213
+ "epoch": 1.32569558101473,
1214
+ "grad_norm": 5.914211750030518,
1215
+ "learning_rate": 3.141297756215888e-05,
1216
+ "logits/chosen": -1.257925271987915,
1217
+ "logits/rejected": -1.1667311191558838,
1218
+ "logps/chosen": -492.0726623535156,
1219
+ "logps/rejected": -463.2273864746094,
1220
+ "loss": 0.3597,
1221
+ "rewards/accuracies": 0.8125,
1222
+ "rewards/chosen": 0.7776979207992554,
1223
+ "rewards/margins": 1.3328384160995483,
1224
+ "rewards/rejected": -0.555140495300293,
1225
+ "step": 810
1226
+ },
1227
+ {
1228
+ "epoch": 1.342062193126023,
1229
+ "grad_norm": 3.587632179260254,
1230
+ "learning_rate": 3.110976349302608e-05,
1231
+ "logits/chosen": -1.2870782613754272,
1232
+ "logits/rejected": -0.9442241787910461,
1233
+ "logps/chosen": -493.3888244628906,
1234
+ "logps/rejected": -464.4835510253906,
1235
+ "loss": 0.3151,
1236
+ "rewards/accuracies": 0.824999988079071,
1237
+ "rewards/chosen": 0.9334325790405273,
1238
+ "rewards/margins": 1.9022817611694336,
1239
+ "rewards/rejected": -0.9688493013381958,
1240
+ "step": 820
1241
+ },
1242
+ {
1243
+ "epoch": 1.358428805237316,
1244
+ "grad_norm": 6.462862968444824,
1245
+ "learning_rate": 3.080654942389327e-05,
1246
+ "logits/chosen": -1.4893635511398315,
1247
+ "logits/rejected": -1.0626758337020874,
1248
+ "logps/chosen": -373.52288818359375,
1249
+ "logps/rejected": -411.831298828125,
1250
+ "loss": 0.438,
1251
+ "rewards/accuracies": 0.7749999761581421,
1252
+ "rewards/chosen": 0.5565814971923828,
1253
+ "rewards/margins": 1.1154134273529053,
1254
+ "rewards/rejected": -0.5588318705558777,
1255
+ "step": 830
1256
+ },
1257
+ {
1258
+ "epoch": 1.3747954173486088,
1259
+ "grad_norm": 6.482163429260254,
1260
+ "learning_rate": 3.0503335354760464e-05,
1261
+ "logits/chosen": -1.2426273822784424,
1262
+ "logits/rejected": -1.0962947607040405,
1263
+ "logps/chosen": -418.48577880859375,
1264
+ "logps/rejected": -414.7897033691406,
1265
+ "loss": 0.4516,
1266
+ "rewards/accuracies": 0.75,
1267
+ "rewards/chosen": 0.7719891667366028,
1268
+ "rewards/margins": 1.0017683506011963,
1269
+ "rewards/rejected": -0.22977924346923828,
1270
+ "step": 840
1271
+ },
1272
+ {
1273
+ "epoch": 1.3911620294599019,
1274
+ "grad_norm": 8.918207168579102,
1275
+ "learning_rate": 3.020012128562765e-05,
1276
+ "logits/chosen": -1.3620984554290771,
1277
+ "logits/rejected": -1.1984447240829468,
1278
+ "logps/chosen": -465.46270751953125,
1279
+ "logps/rejected": -446.5343322753906,
1280
+ "loss": 0.4053,
1281
+ "rewards/accuracies": 0.8125,
1282
+ "rewards/chosen": 0.6932386159896851,
1283
+ "rewards/margins": 1.3631744384765625,
1284
+ "rewards/rejected": -0.6699355840682983,
1285
+ "step": 850
1286
+ },
1287
+ {
1288
+ "epoch": 1.4075286415711947,
1289
+ "grad_norm": 19.940839767456055,
1290
+ "learning_rate": 2.9896907216494846e-05,
1291
+ "logits/chosen": -1.2039759159088135,
1292
+ "logits/rejected": -0.9377444386482239,
1293
+ "logps/chosen": -574.588623046875,
1294
+ "logps/rejected": -502.51214599609375,
1295
+ "loss": 0.4461,
1296
+ "rewards/accuracies": 0.824999988079071,
1297
+ "rewards/chosen": 0.8642657399177551,
1298
+ "rewards/margins": 1.3982216119766235,
1299
+ "rewards/rejected": -0.5339558124542236,
1300
+ "step": 860
1301
+ },
1302
+ {
1303
+ "epoch": 1.4238952536824878,
1304
+ "grad_norm": 6.069481372833252,
1305
+ "learning_rate": 2.959369314736204e-05,
1306
+ "logits/chosen": -1.3913848400115967,
1307
+ "logits/rejected": -1.2518432140350342,
1308
+ "logps/chosen": -458.3941345214844,
1309
+ "logps/rejected": -452.67352294921875,
1310
+ "loss": 0.3604,
1311
+ "rewards/accuracies": 0.762499988079071,
1312
+ "rewards/chosen": 1.043166160583496,
1313
+ "rewards/margins": 1.7257499694824219,
1314
+ "rewards/rejected": -0.6825838685035706,
1315
+ "step": 870
1316
+ },
1317
+ {
1318
+ "epoch": 1.4402618657937807,
1319
+ "grad_norm": 3.334609031677246,
1320
+ "learning_rate": 2.929047907822923e-05,
1321
+ "logits/chosen": -1.3935604095458984,
1322
+ "logits/rejected": NaN,
1323
+ "logps/chosen": -490.9991149902344,
1324
+ "logps/rejected": -452.0387268066406,
1325
+ "loss": 0.3173,
1326
+ "rewards/accuracies": 0.8500000238418579,
1327
+ "rewards/chosen": 1.0780329704284668,
1328
+ "rewards/margins": 1.919669508934021,
1329
+ "rewards/rejected": -0.8416364789009094,
1330
+ "step": 880
1331
+ },
1332
+ {
1333
+ "epoch": 1.4566284779050735,
1334
+ "grad_norm": 5.730930805206299,
1335
+ "learning_rate": 2.8987265009096426e-05,
1336
+ "logits/chosen": -1.3444249629974365,
1337
+ "logits/rejected": -1.124793291091919,
1338
+ "logps/chosen": -409.4859619140625,
1339
+ "logps/rejected": -429.79150390625,
1340
+ "loss": 0.4307,
1341
+ "rewards/accuracies": 0.762499988079071,
1342
+ "rewards/chosen": 0.8416284322738647,
1343
+ "rewards/margins": 1.227608561515808,
1344
+ "rewards/rejected": -0.3859800696372986,
1345
+ "step": 890
1346
+ },
1347
+ {
1348
+ "epoch": 1.4729950900163666,
1349
+ "grad_norm": 4.7208333015441895,
1350
+ "learning_rate": 2.8684050939963614e-05,
1351
+ "logits/chosen": -1.2483845949172974,
1352
+ "logits/rejected": -1.162945032119751,
1353
+ "logps/chosen": -438.02978515625,
1354
+ "logps/rejected": -483.64776611328125,
1355
+ "loss": 0.4359,
1356
+ "rewards/accuracies": 0.8374999761581421,
1357
+ "rewards/chosen": 0.6855508089065552,
1358
+ "rewards/margins": 1.3711297512054443,
1359
+ "rewards/rejected": -0.6855790019035339,
1360
+ "step": 900
1361
+ },
1362
+ {
1363
+ "epoch": 1.4893617021276595,
1364
+ "grad_norm": 4.603919982910156,
1365
+ "learning_rate": 2.838083687083081e-05,
1366
+ "logits/chosen": -1.3359647989273071,
1367
+ "logits/rejected": -1.2829469442367554,
1368
+ "logps/chosen": -478.17022705078125,
1369
+ "logps/rejected": -408.0987243652344,
1370
+ "loss": 0.431,
1371
+ "rewards/accuracies": 0.6499999761581421,
1372
+ "rewards/chosen": 0.9056881666183472,
1373
+ "rewards/margins": 1.124969244003296,
1374
+ "rewards/rejected": -0.21928110718727112,
1375
+ "step": 910
1376
+ },
1377
+ {
1378
+ "epoch": 1.5057283142389526,
1379
+ "grad_norm": 7.60992431640625,
1380
+ "learning_rate": 2.8077622801698e-05,
1381
+ "logits/chosen": -1.4223133325576782,
1382
+ "logits/rejected": -1.2066584825515747,
1383
+ "logps/chosen": -428.1543884277344,
1384
+ "logps/rejected": -370.4458923339844,
1385
+ "loss": 0.4003,
1386
+ "rewards/accuracies": 0.675000011920929,
1387
+ "rewards/chosen": 0.8528293371200562,
1388
+ "rewards/margins": 1.363084077835083,
1389
+ "rewards/rejected": -0.5102548599243164,
1390
+ "step": 920
1391
+ },
1392
+ {
1393
+ "epoch": 1.5220949263502455,
1394
+ "grad_norm": 4.768836498260498,
1395
+ "learning_rate": 2.7774408732565194e-05,
1396
+ "logits/chosen": -1.2342239618301392,
1397
+ "logits/rejected": -1.1602811813354492,
1398
+ "logps/chosen": -440.5987854003906,
1399
+ "logps/rejected": -405.0908508300781,
1400
+ "loss": 0.3879,
1401
+ "rewards/accuracies": 0.8374999761581421,
1402
+ "rewards/chosen": 0.9851611256599426,
1403
+ "rewards/margins": 1.1822245121002197,
1404
+ "rewards/rejected": -0.19706343114376068,
1405
+ "step": 930
1406
+ },
1407
+ {
1408
+ "epoch": 1.5384615384615383,
1409
+ "grad_norm": 8.149114608764648,
1410
+ "learning_rate": 2.7471194663432385e-05,
1411
+ "logits/chosen": -1.1173126697540283,
1412
+ "logits/rejected": -1.014575481414795,
1413
+ "logps/chosen": -402.3595275878906,
1414
+ "logps/rejected": -410.60003662109375,
1415
+ "loss": 0.4529,
1416
+ "rewards/accuracies": 0.8125,
1417
+ "rewards/chosen": 0.7943775653839111,
1418
+ "rewards/margins": 1.2599968910217285,
1419
+ "rewards/rejected": -0.4656193256378174,
1420
+ "step": 940
1421
+ },
1422
+ {
1423
+ "epoch": 1.5548281505728314,
1424
+ "grad_norm": 7.4743852615356445,
1425
+ "learning_rate": 2.7167980594299573e-05,
1426
+ "logits/chosen": NaN,
1427
+ "logits/rejected": -1.143424391746521,
1428
+ "logps/chosen": -446.26629638671875,
1429
+ "logps/rejected": -469.7017517089844,
1430
+ "loss": 0.491,
1431
+ "rewards/accuracies": 0.675000011920929,
1432
+ "rewards/chosen": 0.797395646572113,
1433
+ "rewards/margins": 1.1872708797454834,
1434
+ "rewards/rejected": -0.38987529277801514,
1435
+ "step": 950
1436
+ },
1437
+ {
1438
+ "epoch": 1.5711947626841245,
1439
+ "grad_norm": 8.159062385559082,
1440
+ "learning_rate": 2.6864766525166768e-05,
1441
+ "logits/chosen": -1.14289128780365,
1442
+ "logits/rejected": NaN,
1443
+ "logps/chosen": -450.77117919921875,
1444
+ "logps/rejected": -438.8096618652344,
1445
+ "loss": 0.4282,
1446
+ "rewards/accuracies": 0.737500011920929,
1447
+ "rewards/chosen": 0.7298113107681274,
1448
+ "rewards/margins": 1.092790126800537,
1449
+ "rewards/rejected": -0.3629787862300873,
1450
+ "step": 960
1451
+ },
1452
+ {
1453
+ "epoch": 1.5875613747954174,
1454
+ "grad_norm": 9.274547576904297,
1455
+ "learning_rate": 2.6561552456033962e-05,
1456
+ "logits/chosen": -0.9688614010810852,
1457
+ "logits/rejected": -0.897729754447937,
1458
+ "logps/chosen": -515.4688720703125,
1459
+ "logps/rejected": -436.51275634765625,
1460
+ "loss": 0.4169,
1461
+ "rewards/accuracies": 0.7749999761581421,
1462
+ "rewards/chosen": 0.7370742559432983,
1463
+ "rewards/margins": 1.1543539762496948,
1464
+ "rewards/rejected": -0.4172796607017517,
1465
+ "step": 970
1466
+ },
1467
+ {
1468
+ "epoch": 1.6039279869067102,
1469
+ "grad_norm": 5.051360130310059,
1470
+ "learning_rate": 2.6258338386901153e-05,
1471
+ "logits/chosen": -1.0737905502319336,
1472
+ "logits/rejected": -0.9251714944839478,
1473
+ "logps/chosen": -471.31146240234375,
1474
+ "logps/rejected": -432.572509765625,
1475
+ "loss": 0.3841,
1476
+ "rewards/accuracies": 0.762499988079071,
1477
+ "rewards/chosen": 0.8669958114624023,
1478
+ "rewards/margins": 1.4004991054534912,
1479
+ "rewards/rejected": -0.5335032939910889,
1480
+ "step": 980
1481
+ },
1482
+ {
1483
+ "epoch": 1.6202945990180033,
1484
+ "grad_norm": 6.768540859222412,
1485
+ "learning_rate": 2.5955124317768348e-05,
1486
+ "logits/chosen": -1.1147942543029785,
1487
+ "logits/rejected": -1.1377754211425781,
1488
+ "logps/chosen": -462.5465393066406,
1489
+ "logps/rejected": -390.2607727050781,
1490
+ "loss": 0.3919,
1491
+ "rewards/accuracies": 0.8125,
1492
+ "rewards/chosen": 0.8462246060371399,
1493
+ "rewards/margins": 1.2740905284881592,
1494
+ "rewards/rejected": -0.4278659224510193,
1495
+ "step": 990
1496
+ },
1497
+ {
1498
+ "epoch": 1.6366612111292962,
1499
+ "grad_norm": 8.313865661621094,
1500
+ "learning_rate": 2.5651910248635535e-05,
1501
+ "logits/chosen": -1.0041131973266602,
1502
+ "logits/rejected": -0.8046473264694214,
1503
+ "logps/chosen": -461.7525939941406,
1504
+ "logps/rejected": -443.0941467285156,
1505
+ "loss": 0.4222,
1506
+ "rewards/accuracies": 0.824999988079071,
1507
+ "rewards/chosen": 0.8126258850097656,
1508
+ "rewards/margins": 1.1657203435897827,
1509
+ "rewards/rejected": -0.3530944585800171,
1510
+ "step": 1000
1511
+ },
1512
+ {
1513
+ "epoch": 1.6530278232405893,
1514
+ "grad_norm": 11.539511680603027,
1515
+ "learning_rate": 2.534869617950273e-05,
1516
+ "logits/chosen": -1.2518326044082642,
1517
+ "logits/rejected": -0.9322371482849121,
1518
+ "logps/chosen": -477.0799255371094,
1519
+ "logps/rejected": -450.06976318359375,
1520
+ "loss": 0.3355,
1521
+ "rewards/accuracies": 0.7875000238418579,
1522
+ "rewards/chosen": 0.9717556834220886,
1523
+ "rewards/margins": 1.7565772533416748,
1524
+ "rewards/rejected": -0.784821629524231,
1525
+ "step": 1010
1526
+ },
1527
+ {
1528
+ "epoch": 1.6693944353518821,
1529
+ "grad_norm": 8.115518569946289,
1530
+ "learning_rate": 2.504548211036992e-05,
1531
+ "logits/chosen": -1.4984604120254517,
1532
+ "logits/rejected": -1.33262038230896,
1533
+ "logps/chosen": -469.9319763183594,
1534
+ "logps/rejected": -408.9674072265625,
1535
+ "loss": 0.3737,
1536
+ "rewards/accuracies": 0.824999988079071,
1537
+ "rewards/chosen": 0.8488012552261353,
1538
+ "rewards/margins": 1.4102153778076172,
1539
+ "rewards/rejected": -0.5614141225814819,
1540
+ "step": 1020
1541
+ },
1542
+ {
1543
+ "epoch": 1.685761047463175,
1544
+ "grad_norm": 7.375257968902588,
1545
+ "learning_rate": 2.4742268041237116e-05,
1546
+ "logits/chosen": -1.078918695449829,
1547
+ "logits/rejected": -0.8934460878372192,
1548
+ "logps/chosen": -492.44769287109375,
1549
+ "logps/rejected": -451.58795166015625,
1550
+ "loss": 0.3332,
1551
+ "rewards/accuracies": 0.8374999761581421,
1552
+ "rewards/chosen": 1.0642842054367065,
1553
+ "rewards/margins": 1.744214415550232,
1554
+ "rewards/rejected": -0.6799300909042358,
1555
+ "step": 1030
1556
+ },
1557
+ {
1558
+ "epoch": 1.702127659574468,
1559
+ "grad_norm": 3.4032139778137207,
1560
+ "learning_rate": 2.4439053972104307e-05,
1561
+ "logits/chosen": -1.1205917596817017,
1562
+ "logits/rejected": -0.8298323750495911,
1563
+ "logps/chosen": -420.125,
1564
+ "logps/rejected": -380.03729248046875,
1565
+ "loss": 0.4017,
1566
+ "rewards/accuracies": 0.7875000238418579,
1567
+ "rewards/chosen": 0.7596099376678467,
1568
+ "rewards/margins": 1.2399921417236328,
1569
+ "rewards/rejected": -0.48038214445114136,
1570
+ "step": 1040
1571
+ },
1572
+ {
1573
+ "epoch": 1.7184942716857612,
1574
+ "grad_norm": 3.1656646728515625,
1575
+ "learning_rate": 2.4135839902971498e-05,
1576
+ "logits/chosen": -1.0980288982391357,
1577
+ "logits/rejected": -0.9042747616767883,
1578
+ "logps/chosen": -448.61029052734375,
1579
+ "logps/rejected": -430.3683166503906,
1580
+ "loss": 0.3857,
1581
+ "rewards/accuracies": 0.7875000238418579,
1582
+ "rewards/chosen": 0.9228833317756653,
1583
+ "rewards/margins": 1.3001716136932373,
1584
+ "rewards/rejected": -0.37728843092918396,
1585
+ "step": 1050
1586
+ },
1587
+ {
1588
+ "epoch": 1.734860883797054,
1589
+ "grad_norm": 6.291409015655518,
1590
+ "learning_rate": 2.3832625833838693e-05,
1591
+ "logits/chosen": -1.2553077936172485,
1592
+ "logits/rejected": -0.9304355382919312,
1593
+ "logps/chosen": -443.744140625,
1594
+ "logps/rejected": -375.67840576171875,
1595
+ "loss": 0.4043,
1596
+ "rewards/accuracies": 0.8125,
1597
+ "rewards/chosen": 0.7896615862846375,
1598
+ "rewards/margins": 1.1822974681854248,
1599
+ "rewards/rejected": -0.39263591170310974,
1600
+ "step": 1060
1601
+ },
1602
+ {
1603
+ "epoch": 1.751227495908347,
1604
+ "grad_norm": 4.611414909362793,
1605
+ "learning_rate": 2.3529411764705884e-05,
1606
+ "logits/chosen": -1.3921210765838623,
1607
+ "logits/rejected": -1.2024405002593994,
1608
+ "logps/chosen": -438.8999938964844,
1609
+ "logps/rejected": -379.8556213378906,
1610
+ "loss": 0.3666,
1611
+ "rewards/accuracies": 0.8125,
1612
+ "rewards/chosen": 0.7746474742889404,
1613
+ "rewards/margins": 1.3105146884918213,
1614
+ "rewards/rejected": -0.5358672142028809,
1615
+ "step": 1070
1616
+ },
1617
+ {
1618
+ "epoch": 1.7675941080196398,
1619
+ "grad_norm": 2.6192657947540283,
1620
+ "learning_rate": 2.3226197695573075e-05,
1621
+ "logits/chosen": -1.2743886709213257,
1622
+ "logits/rejected": -1.193303108215332,
1623
+ "logps/chosen": -469.51220703125,
1624
+ "logps/rejected": -446.6626892089844,
1625
+ "loss": 0.4615,
1626
+ "rewards/accuracies": 0.7124999761581421,
1627
+ "rewards/chosen": 0.9081875085830688,
1628
+ "rewards/margins": 1.3380632400512695,
1629
+ "rewards/rejected": -0.4298757016658783,
1630
+ "step": 1080
1631
+ },
1632
+ {
1633
+ "epoch": 1.7839607201309329,
1634
+ "grad_norm": 5.901863098144531,
1635
+ "learning_rate": 2.2922983626440266e-05,
1636
+ "logits/chosen": -1.249284267425537,
1637
+ "logits/rejected": -1.0896720886230469,
1638
+ "logps/chosen": -388.6412048339844,
1639
+ "logps/rejected": -381.84747314453125,
1640
+ "loss": 0.4718,
1641
+ "rewards/accuracies": 0.625,
1642
+ "rewards/chosen": 0.62062668800354,
1643
+ "rewards/margins": 1.1487479209899902,
1644
+ "rewards/rejected": -0.5281211733818054,
1645
+ "step": 1090
1646
+ },
1647
+ {
1648
+ "epoch": 1.800327332242226,
1649
+ "grad_norm": 4.81393575668335,
1650
+ "learning_rate": 2.261976955730746e-05,
1651
+ "logits/chosen": -1.2179973125457764,
1652
+ "logits/rejected": -1.1146996021270752,
1653
+ "logps/chosen": -432.94970703125,
1654
+ "logps/rejected": -404.9814453125,
1655
+ "loss": 0.4061,
1656
+ "rewards/accuracies": 0.7875000238418579,
1657
+ "rewards/chosen": 0.49649685621261597,
1658
+ "rewards/margins": 1.0791733264923096,
1659
+ "rewards/rejected": -0.5826765298843384,
1660
+ "step": 1100
1661
+ },
1662
+ {
1663
+ "epoch": 1.8166939443535188,
1664
+ "grad_norm": 7.300076484680176,
1665
+ "learning_rate": 2.2316555488174655e-05,
1666
+ "logits/chosen": -1.4200313091278076,
1667
+ "logits/rejected": -1.3520318269729614,
1668
+ "logps/chosen": -398.1583557128906,
1669
+ "logps/rejected": -411.41864013671875,
1670
+ "loss": 0.4397,
1671
+ "rewards/accuracies": 0.6875,
1672
+ "rewards/chosen": 0.8094931840896606,
1673
+ "rewards/margins": 1.2237316370010376,
1674
+ "rewards/rejected": -0.41423845291137695,
1675
+ "step": 1110
1676
+ },
1677
+ {
1678
+ "epoch": 1.8330605564648117,
1679
+ "grad_norm": 5.185066223144531,
1680
+ "learning_rate": 2.2013341419041843e-05,
1681
+ "logits/chosen": -1.282606840133667,
1682
+ "logits/rejected": -1.3600891828536987,
1683
+ "logps/chosen": -397.7345275878906,
1684
+ "logps/rejected": -368.4806823730469,
1685
+ "loss": 0.4605,
1686
+ "rewards/accuracies": 0.800000011920929,
1687
+ "rewards/chosen": 0.7671560049057007,
1688
+ "rewards/margins": 0.9713778495788574,
1689
+ "rewards/rejected": -0.20422179996967316,
1690
+ "step": 1120
1691
+ },
1692
+ {
1693
+ "epoch": 1.8494271685761048,
1694
+ "grad_norm": 6.64644718170166,
1695
+ "learning_rate": 2.1710127349909037e-05,
1696
+ "logits/chosen": -1.309072732925415,
1697
+ "logits/rejected": -1.237240195274353,
1698
+ "logps/chosen": -508.67864990234375,
1699
+ "logps/rejected": -486.6451110839844,
1700
+ "loss": 0.3327,
1701
+ "rewards/accuracies": 0.8374999761581421,
1702
+ "rewards/chosen": 1.003493070602417,
1703
+ "rewards/margins": 1.6196180582046509,
1704
+ "rewards/rejected": -0.6161248683929443,
1705
+ "step": 1130
1706
+ },
1707
+ {
1708
+ "epoch": 1.8657937806873979,
1709
+ "grad_norm": 8.295884132385254,
1710
+ "learning_rate": 2.140691328077623e-05,
1711
+ "logits/chosen": -1.2569806575775146,
1712
+ "logits/rejected": -1.0191644430160522,
1713
+ "logps/chosen": -450.56884765625,
1714
+ "logps/rejected": -396.7924499511719,
1715
+ "loss": 0.4093,
1716
+ "rewards/accuracies": 0.7875000238418579,
1717
+ "rewards/chosen": 0.6329465508460999,
1718
+ "rewards/margins": 1.3668501377105713,
1719
+ "rewards/rejected": -0.7339035868644714,
1720
+ "step": 1140
1721
+ },
1722
+ {
1723
+ "epoch": 1.8821603927986907,
1724
+ "grad_norm": 5.191346168518066,
1725
+ "learning_rate": 2.110369921164342e-05,
1726
+ "logits/chosen": -1.2032971382141113,
1727
+ "logits/rejected": -0.9198576211929321,
1728
+ "logps/chosen": -401.7949523925781,
1729
+ "logps/rejected": -387.2808532714844,
1730
+ "loss": 0.3817,
1731
+ "rewards/accuracies": 0.875,
1732
+ "rewards/chosen": 0.5356651544570923,
1733
+ "rewards/margins": 1.356689214706421,
1734
+ "rewards/rejected": -0.8210241198539734,
1735
+ "step": 1150
1736
+ },
1737
+ {
1738
+ "epoch": 1.8985270049099836,
1739
+ "grad_norm": 4.5822367668151855,
1740
+ "learning_rate": 2.0800485142510614e-05,
1741
+ "logits/chosen": -1.16325843334198,
1742
+ "logits/rejected": -0.8339241147041321,
1743
+ "logps/chosen": -385.2162170410156,
1744
+ "logps/rejected": -399.42401123046875,
1745
+ "loss": 0.4329,
1746
+ "rewards/accuracies": 0.7124999761581421,
1747
+ "rewards/chosen": 0.5546852946281433,
1748
+ "rewards/margins": 1.410249948501587,
1749
+ "rewards/rejected": -0.8555648922920227,
1750
+ "step": 1160
1751
+ },
1752
+ {
1753
+ "epoch": 1.9148936170212765,
1754
+ "grad_norm": 7.598694324493408,
1755
+ "learning_rate": 2.0497271073377805e-05,
1756
+ "logits/chosen": -1.2841596603393555,
1757
+ "logits/rejected": -1.0715458393096924,
1758
+ "logps/chosen": -426.3335876464844,
1759
+ "logps/rejected": -392.77947998046875,
1760
+ "loss": 0.3999,
1761
+ "rewards/accuracies": 0.824999988079071,
1762
+ "rewards/chosen": 0.6999267339706421,
1763
+ "rewards/margins": 1.2335584163665771,
1764
+ "rewards/rejected": -0.5336317420005798,
1765
+ "step": 1170
1766
+ },
1767
+ {
1768
+ "epoch": 1.9312602291325696,
1769
+ "grad_norm": 12.754505157470703,
1770
+ "learning_rate": 2.0194057004245e-05,
1771
+ "logits/chosen": -1.2506614923477173,
1772
+ "logits/rejected": -1.1044342517852783,
1773
+ "logps/chosen": -464.2216796875,
1774
+ "logps/rejected": -420.13336181640625,
1775
+ "loss": 0.4258,
1776
+ "rewards/accuracies": 0.824999988079071,
1777
+ "rewards/chosen": 0.7730525135993958,
1778
+ "rewards/margins": 1.1193318367004395,
1779
+ "rewards/rejected": -0.3462793827056885,
1780
+ "step": 1180
1781
+ },
1782
+ {
1783
+ "epoch": 1.9476268412438626,
1784
+ "grad_norm": 5.040411949157715,
1785
+ "learning_rate": 1.9890842935112188e-05,
1786
+ "logits/chosen": -1.5200366973876953,
1787
+ "logits/rejected": -1.453028917312622,
1788
+ "logps/chosen": -471.79180908203125,
1789
+ "logps/rejected": -421.40899658203125,
1790
+ "loss": 0.3997,
1791
+ "rewards/accuracies": 0.762499988079071,
1792
+ "rewards/chosen": 0.7299527525901794,
1793
+ "rewards/margins": 1.2998378276824951,
1794
+ "rewards/rejected": -0.56988525390625,
1795
+ "step": 1190
1796
+ },
1797
+ {
1798
+ "epoch": 1.9639934533551555,
1799
+ "grad_norm": 12.064972877502441,
1800
+ "learning_rate": 1.9587628865979382e-05,
1801
+ "logits/chosen": -1.2272264957427979,
1802
+ "logits/rejected": -0.9471932649612427,
1803
+ "logps/chosen": -501.19842529296875,
1804
+ "logps/rejected": -451.31658935546875,
1805
+ "loss": 0.4069,
1806
+ "rewards/accuracies": 0.800000011920929,
1807
+ "rewards/chosen": 0.8413068056106567,
1808
+ "rewards/margins": 1.3548190593719482,
1809
+ "rewards/rejected": -0.5135123133659363,
1810
+ "step": 1200
1811
+ },
1812
+ {
1813
+ "epoch": 1.9803600654664484,
1814
+ "grad_norm": 7.827969551086426,
1815
+ "learning_rate": 1.9284414796846577e-05,
1816
+ "logits/chosen": -1.1897993087768555,
1817
+ "logits/rejected": -1.085766077041626,
1818
+ "logps/chosen": -467.1263122558594,
1819
+ "logps/rejected": -420.46417236328125,
1820
+ "loss": 0.4029,
1821
+ "rewards/accuracies": 0.824999988079071,
1822
+ "rewards/chosen": 0.8640726208686829,
1823
+ "rewards/margins": 1.2383184432983398,
1824
+ "rewards/rejected": -0.3742457330226898,
1825
+ "step": 1210
1826
+ },
1827
+ {
1828
+ "epoch": 1.9967266775777412,
1829
+ "grad_norm": 4.033712863922119,
1830
+ "learning_rate": 1.8981200727713764e-05,
1831
+ "logits/chosen": -1.1938722133636475,
1832
+ "logits/rejected": -0.7925646305084229,
1833
+ "logps/chosen": -434.82916259765625,
1834
+ "logps/rejected": -415.64178466796875,
1835
+ "loss": 0.4062,
1836
+ "rewards/accuracies": 0.8500000238418579,
1837
+ "rewards/chosen": 0.6897398233413696,
1838
+ "rewards/margins": 1.3555059432983398,
1839
+ "rewards/rejected": -0.6657661199569702,
1840
+ "step": 1220
1841
+ },
1842
+ {
1843
+ "epoch": 2.0130932896890346,
1844
+ "grad_norm": 2.060443162918091,
1845
+ "learning_rate": 1.867798665858096e-05,
1846
+ "logits/chosen": NaN,
1847
+ "logits/rejected": -1.3225266933441162,
1848
+ "logps/chosen": -402.45654296875,
1849
+ "logps/rejected": -432.16448974609375,
1850
+ "loss": 0.3632,
1851
+ "rewards/accuracies": 0.7749999761581421,
1852
+ "rewards/chosen": 0.8448646664619446,
1853
+ "rewards/margins": 1.4718005657196045,
1854
+ "rewards/rejected": -0.6269359588623047,
1855
+ "step": 1230
1856
+ },
1857
+ {
1858
+ "epoch": 2.0294599018003274,
1859
+ "grad_norm": 2.7568705081939697,
1860
+ "learning_rate": 1.8374772589448153e-05,
1861
+ "logits/chosen": -1.589988112449646,
1862
+ "logits/rejected": -1.4496564865112305,
1863
+ "logps/chosen": -415.6414489746094,
1864
+ "logps/rejected": -382.269287109375,
1865
+ "loss": 0.3286,
1866
+ "rewards/accuracies": 0.7875000238418579,
1867
+ "rewards/chosen": 1.015729308128357,
1868
+ "rewards/margins": 1.6888082027435303,
1869
+ "rewards/rejected": -0.6730789542198181,
1870
+ "step": 1240
1871
+ },
1872
+ {
1873
+ "epoch": 2.0458265139116203,
1874
+ "grad_norm": 7.693350315093994,
1875
+ "learning_rate": 1.8071558520315345e-05,
1876
+ "logits/chosen": -1.2396166324615479,
1877
+ "logits/rejected": -1.0914833545684814,
1878
+ "logps/chosen": -428.75146484375,
1879
+ "logps/rejected": -439.94537353515625,
1880
+ "loss": 0.3436,
1881
+ "rewards/accuracies": 0.824999988079071,
1882
+ "rewards/chosen": 0.9149911999702454,
1883
+ "rewards/margins": 1.669695496559143,
1884
+ "rewards/rejected": -0.7547044157981873,
1885
+ "step": 1250
1886
+ },
1887
+ {
1888
+ "epoch": 2.062193126022913,
1889
+ "grad_norm": 6.800998210906982,
1890
+ "learning_rate": 1.7768344451182536e-05,
1891
+ "logits/chosen": -1.3096355199813843,
1892
+ "logits/rejected": -1.2225430011749268,
1893
+ "logps/chosen": -544.235595703125,
1894
+ "logps/rejected": -471.3172302246094,
1895
+ "loss": 0.2387,
1896
+ "rewards/accuracies": 0.9125000238418579,
1897
+ "rewards/chosen": 1.1924231052398682,
1898
+ "rewards/margins": 1.9907211065292358,
1899
+ "rewards/rejected": -0.7982980608940125,
1900
+ "step": 1260
1901
+ },
1902
+ {
1903
+ "epoch": 2.078559738134206,
1904
+ "grad_norm": 2.9473204612731934,
1905
+ "learning_rate": 1.7465130382049727e-05,
1906
+ "logits/chosen": -1.5475558042526245,
1907
+ "logits/rejected": -1.2119872570037842,
1908
+ "logps/chosen": -425.4368591308594,
1909
+ "logps/rejected": -431.3377990722656,
1910
+ "loss": 0.1952,
1911
+ "rewards/accuracies": 0.9125000238418579,
1912
+ "rewards/chosen": 1.1395524740219116,
1913
+ "rewards/margins": 2.476022243499756,
1914
+ "rewards/rejected": -1.3364696502685547,
1915
+ "step": 1270
1916
+ },
1917
+ {
1918
+ "epoch": 2.0949263502454993,
1919
+ "grad_norm": 5.591909885406494,
1920
+ "learning_rate": 1.716191631291692e-05,
1921
+ "logits/chosen": -1.5607810020446777,
1922
+ "logits/rejected": -1.4012185335159302,
1923
+ "logps/chosen": -413.8076171875,
1924
+ "logps/rejected": -411.9990234375,
1925
+ "loss": 0.2864,
1926
+ "rewards/accuracies": 0.824999988079071,
1927
+ "rewards/chosen": 1.065088152885437,
1928
+ "rewards/margins": 1.9613841772079468,
1929
+ "rewards/rejected": -0.8962962031364441,
1930
+ "step": 1280
1931
+ },
1932
+ {
1933
+ "epoch": 2.111292962356792,
1934
+ "grad_norm": 4.462930679321289,
1935
+ "learning_rate": 1.6858702243784113e-05,
1936
+ "logits/chosen": -1.4985883235931396,
1937
+ "logits/rejected": -1.2055768966674805,
1938
+ "logps/chosen": -402.1160583496094,
1939
+ "logps/rejected": -397.81402587890625,
1940
+ "loss": 0.2957,
1941
+ "rewards/accuracies": 0.8125,
1942
+ "rewards/chosen": 0.8168094754219055,
1943
+ "rewards/margins": 1.9579219818115234,
1944
+ "rewards/rejected": -1.1411124467849731,
1945
+ "step": 1290
1946
+ },
1947
+ {
1948
+ "epoch": 2.127659574468085,
1949
+ "grad_norm": 3.3461086750030518,
1950
+ "learning_rate": 1.6555488174651304e-05,
1951
+ "logits/chosen": -1.5269520282745361,
1952
+ "logits/rejected": -1.0539462566375732,
1953
+ "logps/chosen": -507.36834716796875,
1954
+ "logps/rejected": -461.356689453125,
1955
+ "loss": 0.2565,
1956
+ "rewards/accuracies": 0.9125000238418579,
1957
+ "rewards/chosen": 0.6355259418487549,
1958
+ "rewards/margins": 2.068328380584717,
1959
+ "rewards/rejected": -1.432802438735962,
1960
+ "step": 1300
1961
+ },
1962
+ {
1963
+ "epoch": 2.144026186579378,
1964
+ "grad_norm": 4.208719253540039,
1965
+ "learning_rate": 1.6252274105518498e-05,
1966
+ "logits/chosen": -1.403276801109314,
1967
+ "logits/rejected": -1.199331283569336,
1968
+ "logps/chosen": -427.9095153808594,
1969
+ "logps/rejected": -373.31756591796875,
1970
+ "loss": 0.2796,
1971
+ "rewards/accuracies": 0.8500000238418579,
1972
+ "rewards/chosen": 0.8879494667053223,
1973
+ "rewards/margins": 1.912683129310608,
1974
+ "rewards/rejected": -1.0247336626052856,
1975
+ "step": 1310
1976
+ },
1977
+ {
1978
+ "epoch": 2.160392798690671,
1979
+ "grad_norm": 3.667640447616577,
1980
+ "learning_rate": 1.594906003638569e-05,
1981
+ "logits/chosen": -1.4091194868087769,
1982
+ "logits/rejected": -1.123011827468872,
1983
+ "logps/chosen": -457.79461669921875,
1984
+ "logps/rejected": -444.686279296875,
1985
+ "loss": 0.2564,
1986
+ "rewards/accuracies": 0.8500000238418579,
1987
+ "rewards/chosen": 0.8721135258674622,
1988
+ "rewards/margins": 2.0980000495910645,
1989
+ "rewards/rejected": -1.2258864641189575,
1990
+ "step": 1320
1991
+ },
1992
+ {
1993
+ "epoch": 2.176759410801964,
1994
+ "grad_norm": 2.9427664279937744,
1995
+ "learning_rate": 1.564584596725288e-05,
1996
+ "logits/chosen": -1.4273817539215088,
1997
+ "logits/rejected": -1.2326085567474365,
1998
+ "logps/chosen": -431.9471130371094,
1999
+ "logps/rejected": -421.22747802734375,
2000
+ "loss": 0.3017,
2001
+ "rewards/accuracies": 0.862500011920929,
2002
+ "rewards/chosen": 0.7032436728477478,
2003
+ "rewards/margins": 1.7488292455673218,
2004
+ "rewards/rejected": -1.0455853939056396,
2005
+ "step": 1330
2006
+ },
2007
+ {
2008
+ "epoch": 2.193126022913257,
2009
+ "grad_norm": 3.592177629470825,
2010
+ "learning_rate": 1.5342631898120075e-05,
2011
+ "logits/chosen": -1.3208458423614502,
2012
+ "logits/rejected": -1.1550036668777466,
2013
+ "logps/chosen": -442.4117736816406,
2014
+ "logps/rejected": -370.07708740234375,
2015
+ "loss": 0.3644,
2016
+ "rewards/accuracies": 0.7749999761581421,
2017
+ "rewards/chosen": 0.9467415809631348,
2018
+ "rewards/margins": 1.7952324151992798,
2019
+ "rewards/rejected": -0.8484910726547241,
2020
+ "step": 1340
2021
+ },
2022
+ {
2023
+ "epoch": 2.20949263502455,
2024
+ "grad_norm": 2.6388635635375977,
2025
+ "learning_rate": 1.5039417828987265e-05,
2026
+ "logits/chosen": -1.305037021636963,
2027
+ "logits/rejected": -1.2544772624969482,
2028
+ "logps/chosen": -438.3041076660156,
2029
+ "logps/rejected": -412.14923095703125,
2030
+ "loss": 0.2553,
2031
+ "rewards/accuracies": 0.875,
2032
+ "rewards/chosen": 1.2883623838424683,
2033
+ "rewards/margins": 1.9844329357147217,
2034
+ "rewards/rejected": -0.6960704326629639,
2035
+ "step": 1350
2036
+ },
2037
+ {
2038
+ "epoch": 2.2258592471358427,
2039
+ "grad_norm": 5.4221954345703125,
2040
+ "learning_rate": 1.4736203759854459e-05,
2041
+ "logits/chosen": -1.3623788356781006,
2042
+ "logits/rejected": -1.030826210975647,
2043
+ "logps/chosen": -437.98150634765625,
2044
+ "logps/rejected": -413.3927307128906,
2045
+ "loss": 0.3294,
2046
+ "rewards/accuracies": 0.8125,
2047
+ "rewards/chosen": 0.8824920654296875,
2048
+ "rewards/margins": 1.7985508441925049,
2049
+ "rewards/rejected": -0.9160588383674622,
2050
+ "step": 1360
2051
+ },
2052
+ {
2053
+ "epoch": 2.242225859247136,
2054
+ "grad_norm": 1.4540634155273438,
2055
+ "learning_rate": 1.4432989690721649e-05,
2056
+ "logits/chosen": -1.4040526151657104,
2057
+ "logits/rejected": -1.2125266790390015,
2058
+ "logps/chosen": -444.2418518066406,
2059
+ "logps/rejected": -445.39154052734375,
2060
+ "loss": 0.2202,
2061
+ "rewards/accuracies": 0.887499988079071,
2062
+ "rewards/chosen": 1.2369893789291382,
2063
+ "rewards/margins": 2.244361162185669,
2064
+ "rewards/rejected": -1.0073716640472412,
2065
+ "step": 1370
2066
+ },
2067
+ {
2068
+ "epoch": 2.258592471358429,
2069
+ "grad_norm": 5.328054904937744,
2070
+ "learning_rate": 1.4129775621588841e-05,
2071
+ "logits/chosen": -1.1481372117996216,
2072
+ "logits/rejected": -1.1810128688812256,
2073
+ "logps/chosen": -483.92901611328125,
2074
+ "logps/rejected": -435.12408447265625,
2075
+ "loss": 0.2465,
2076
+ "rewards/accuracies": 0.887499988079071,
2077
+ "rewards/chosen": 1.3297772407531738,
2078
+ "rewards/margins": 2.0740952491760254,
2079
+ "rewards/rejected": -0.7443181872367859,
2080
+ "step": 1380
2081
+ },
2082
+ {
2083
+ "epoch": 2.2749590834697218,
2084
+ "grad_norm": 5.371053695678711,
2085
+ "learning_rate": 1.3826561552456036e-05,
2086
+ "logits/chosen": -1.341086506843567,
2087
+ "logits/rejected": -1.2301025390625,
2088
+ "logps/chosen": -441.9823303222656,
2089
+ "logps/rejected": -371.228271484375,
2090
+ "loss": 0.2796,
2091
+ "rewards/accuracies": 0.8374999761581421,
2092
+ "rewards/chosen": 1.0860683917999268,
2093
+ "rewards/margins": 1.9399855136871338,
2094
+ "rewards/rejected": -0.853917121887207,
2095
+ "step": 1390
2096
+ },
2097
+ {
2098
+ "epoch": 2.2913256955810146,
2099
+ "grad_norm": 4.199862480163574,
2100
+ "learning_rate": 1.3523347483323225e-05,
2101
+ "logits/chosen": -1.2742938995361328,
2102
+ "logits/rejected": -1.0367982387542725,
2103
+ "logps/chosen": -441.23193359375,
2104
+ "logps/rejected": -432.3246154785156,
2105
+ "loss": 0.2161,
2106
+ "rewards/accuracies": 0.9125000238418579,
2107
+ "rewards/chosen": 1.0744221210479736,
2108
+ "rewards/margins": 2.1862921714782715,
2109
+ "rewards/rejected": -1.1118695735931396,
2110
+ "step": 1400
2111
+ },
2112
+ {
2113
+ "epoch": 2.3076923076923075,
2114
+ "grad_norm": 4.229262351989746,
2115
+ "learning_rate": 1.322013341419042e-05,
2116
+ "logits/chosen": -1.4364268779754639,
2117
+ "logits/rejected": -1.0992625951766968,
2118
+ "logps/chosen": -425.8427734375,
2119
+ "logps/rejected": -459.83642578125,
2120
+ "loss": 0.2952,
2121
+ "rewards/accuracies": 0.8125,
2122
+ "rewards/chosen": 0.9090694189071655,
2123
+ "rewards/margins": 1.8446950912475586,
2124
+ "rewards/rejected": -0.9356255531311035,
2125
+ "step": 1410
2126
+ },
2127
+ {
2128
+ "epoch": 2.324058919803601,
2129
+ "grad_norm": 2.794767379760742,
2130
+ "learning_rate": 1.2916919345057613e-05,
2131
+ "logits/chosen": -1.354282021522522,
2132
+ "logits/rejected": -1.2035588026046753,
2133
+ "logps/chosen": -420.01800537109375,
2134
+ "logps/rejected": -375.9340515136719,
2135
+ "loss": 0.2299,
2136
+ "rewards/accuracies": 0.9125000238418579,
2137
+ "rewards/chosen": 1.1438004970550537,
2138
+ "rewards/margins": 2.1549389362335205,
2139
+ "rewards/rejected": -1.0111383199691772,
2140
+ "step": 1420
2141
+ },
2142
+ {
2143
+ "epoch": 2.3404255319148937,
2144
+ "grad_norm": 1.970015525817871,
2145
+ "learning_rate": 1.2613705275924804e-05,
2146
+ "logits/chosen": -1.3893083333969116,
2147
+ "logits/rejected": -1.3427255153656006,
2148
+ "logps/chosen": -428.79400634765625,
2149
+ "logps/rejected": -409.81500244140625,
2150
+ "loss": 0.2522,
2151
+ "rewards/accuracies": 0.875,
2152
+ "rewards/chosen": 1.1792007684707642,
2153
+ "rewards/margins": 2.0457987785339355,
2154
+ "rewards/rejected": -0.8665979504585266,
2155
+ "step": 1430
2156
+ },
2157
+ {
2158
+ "epoch": 2.3567921440261865,
2159
+ "grad_norm": 3.0225942134857178,
2160
+ "learning_rate": 1.2310491206791997e-05,
2161
+ "logits/chosen": -1.5218456983566284,
2162
+ "logits/rejected": -1.236580491065979,
2163
+ "logps/chosen": -474.56829833984375,
2164
+ "logps/rejected": -513.43408203125,
2165
+ "loss": 0.2505,
2166
+ "rewards/accuracies": 0.9125000238418579,
2167
+ "rewards/chosen": 1.0795310735702515,
2168
+ "rewards/margins": 2.1252081394195557,
2169
+ "rewards/rejected": -1.0456770658493042,
2170
+ "step": 1440
2171
+ },
2172
+ {
2173
+ "epoch": 2.3731587561374794,
2174
+ "grad_norm": 4.234305381774902,
2175
+ "learning_rate": 1.2007277137659188e-05,
2176
+ "logits/chosen": -1.3646624088287354,
2177
+ "logits/rejected": -1.2389724254608154,
2178
+ "logps/chosen": -431.1107482910156,
2179
+ "logps/rejected": -442.53997802734375,
2180
+ "loss": 0.3115,
2181
+ "rewards/accuracies": 0.824999988079071,
2182
+ "rewards/chosen": 0.8404763340950012,
2183
+ "rewards/margins": 2.168391466140747,
2184
+ "rewards/rejected": -1.3279149532318115,
2185
+ "step": 1450
2186
+ },
2187
+ {
2188
+ "epoch": 2.3895253682487727,
2189
+ "grad_norm": 9.905987739562988,
2190
+ "learning_rate": 1.170406306852638e-05,
2191
+ "logits/chosen": -1.4754550457000732,
2192
+ "logits/rejected": -1.4108117818832397,
2193
+ "logps/chosen": -453.54364013671875,
2194
+ "logps/rejected": -432.04693603515625,
2195
+ "loss": 0.3368,
2196
+ "rewards/accuracies": 0.7749999761581421,
2197
+ "rewards/chosen": 0.6598278284072876,
2198
+ "rewards/margins": 1.6955798864364624,
2199
+ "rewards/rejected": -1.0357519388198853,
2200
+ "step": 1460
2201
+ },
2202
+ {
2203
+ "epoch": 2.4058919803600656,
2204
+ "grad_norm": 4.8183746337890625,
2205
+ "learning_rate": 1.1400848999393572e-05,
2206
+ "logits/chosen": -1.3298962116241455,
2207
+ "logits/rejected": -1.2613223791122437,
2208
+ "logps/chosen": -480.92205810546875,
2209
+ "logps/rejected": -446.716796875,
2210
+ "loss": 0.3334,
2211
+ "rewards/accuracies": 0.8374999761581421,
2212
+ "rewards/chosen": 0.8991876840591431,
2213
+ "rewards/margins": 1.7357298135757446,
2214
+ "rewards/rejected": -0.8365424275398254,
2215
+ "step": 1470
2216
+ },
2217
+ {
2218
+ "epoch": 2.4222585924713584,
2219
+ "grad_norm": 6.064842700958252,
2220
+ "learning_rate": 1.1097634930260765e-05,
2221
+ "logits/chosen": -1.4571373462677002,
2222
+ "logits/rejected": -1.1927803754806519,
2223
+ "logps/chosen": -447.56787109375,
2224
+ "logps/rejected": -440.72064208984375,
2225
+ "loss": 0.2964,
2226
+ "rewards/accuracies": 0.8500000238418579,
2227
+ "rewards/chosen": 1.0991294384002686,
2228
+ "rewards/margins": 2.0829381942749023,
2229
+ "rewards/rejected": -0.983808696269989,
2230
+ "step": 1480
2231
+ },
2232
+ {
2233
+ "epoch": 2.4386252045826513,
2234
+ "grad_norm": 2.8285984992980957,
2235
+ "learning_rate": 1.0794420861127958e-05,
2236
+ "logits/chosen": -1.4223463535308838,
2237
+ "logits/rejected": -1.2678343057632446,
2238
+ "logps/chosen": -438.255126953125,
2239
+ "logps/rejected": -454.03948974609375,
2240
+ "loss": 0.2041,
2241
+ "rewards/accuracies": 0.9125000238418579,
2242
+ "rewards/chosen": 1.2381608486175537,
2243
+ "rewards/margins": 2.366389036178589,
2244
+ "rewards/rejected": -1.1282285451889038,
2245
+ "step": 1490
2246
+ },
2247
+ {
2248
+ "epoch": 2.454991816693944,
2249
+ "grad_norm": 3.3027286529541016,
2250
+ "learning_rate": 1.0491206791995149e-05,
2251
+ "logits/chosen": -1.3270039558410645,
2252
+ "logits/rejected": -1.2212953567504883,
2253
+ "logps/chosen": -447.6775817871094,
2254
+ "logps/rejected": -454.34918212890625,
2255
+ "loss": 0.246,
2256
+ "rewards/accuracies": 0.8500000238418579,
2257
+ "rewards/chosen": 0.9780976176261902,
2258
+ "rewards/margins": 1.9729305505752563,
2259
+ "rewards/rejected": -0.9948328137397766,
2260
+ "step": 1500
2261
+ }
2262
+ ],
2263
+ "logging_steps": 10,
2264
+ "max_steps": 1833,
2265
+ "num_input_tokens_seen": 0,
2266
+ "num_train_epochs": 3,
2267
+ "save_steps": 500,
2268
+ "stateful_callbacks": {
2269
+ "TrainerControl": {
2270
+ "args": {
2271
+ "should_epoch_stop": false,
2272
+ "should_evaluate": false,
2273
+ "should_log": false,
2274
+ "should_save": true,
2275
+ "should_training_stop": false
2276
+ },
2277
+ "attributes": {}
2278
+ }
2279
+ },
2280
+ "total_flos": 0.0,
2281
+ "train_batch_size": 4,
2282
+ "trial_name": null,
2283
+ "trial_params": null
2284
+ }
checkpoint-1500/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:adb07c9cf31b39264fa1c39cdcedb4f8f2cd0a21811a03b1f63968e1dd128142
3
+ size 6673
checkpoint-1500/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-1833/README.md ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: unsloth/qwen3-0.6b-base-unsloth-bnb-4bit
3
+ library_name: peft
4
+ ---
5
+
6
+ # Model Card for Model ID
7
+
8
+ <!-- Provide a quick summary of what the model is/does. -->
9
+
10
+
11
+
12
+ ## Model Details
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+
19
+
20
+ - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
+ - **Shared by [optional]:** [More Information Needed]
23
+ - **Model type:** [More Information Needed]
24
+ - **Language(s) (NLP):** [More Information Needed]
25
+ - **License:** [More Information Needed]
26
+ - **Finetuned from model [optional]:** [More Information Needed]
27
+
28
+ ### Model Sources [optional]
29
+
30
+ <!-- Provide the basic links for the model. -->
31
+
32
+ - **Repository:** [More Information Needed]
33
+ - **Paper [optional]:** [More Information Needed]
34
+ - **Demo [optional]:** [More Information Needed]
35
+
36
+ ## Uses
37
+
38
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
+
40
+ ### Direct Use
41
+
42
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
+
44
+ [More Information Needed]
45
+
46
+ ### Downstream Use [optional]
47
+
48
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Out-of-Scope Use
53
+
54
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
+
56
+ [More Information Needed]
57
+
58
+ ## Bias, Risks, and Limitations
59
+
60
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ### Recommendations
65
+
66
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
+
68
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ Use the code below to get started with the model.
73
+
74
+ [More Information Needed]
75
+
76
+ ## Training Details
77
+
78
+ ### Training Data
79
+
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
+
82
+ [More Information Needed]
83
+
84
+ ### Training Procedure
85
+
86
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
+
88
+ #### Preprocessing [optional]
89
+
90
+ [More Information Needed]
91
+
92
+
93
+ #### Training Hyperparameters
94
+
95
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
+
97
+ #### Speeds, Sizes, Times [optional]
98
+
99
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
+
101
+ [More Information Needed]
102
+
103
+ ## Evaluation
104
+
105
+ <!-- This section describes the evaluation protocols and provides the results. -->
106
+
107
+ ### Testing Data, Factors & Metrics
108
+
109
+ #### Testing Data
110
+
111
+ <!-- This should link to a Dataset Card if possible. -->
112
+
113
+ [More Information Needed]
114
+
115
+ #### Factors
116
+
117
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Metrics
122
+
123
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
+
125
+ [More Information Needed]
126
+
127
+ ### Results
128
+
129
+ [More Information Needed]
130
+
131
+ #### Summary
132
+
133
+
134
+
135
+ ## Model Examination [optional]
136
+
137
+ <!-- Relevant interpretability work for the model goes here -->
138
+
139
+ [More Information Needed]
140
+
141
+ ## Environmental Impact
142
+
143
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
+
145
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
+
147
+ - **Hardware Type:** [More Information Needed]
148
+ - **Hours used:** [More Information Needed]
149
+ - **Cloud Provider:** [More Information Needed]
150
+ - **Compute Region:** [More Information Needed]
151
+ - **Carbon Emitted:** [More Information Needed]
152
+
153
+ ## Technical Specifications [optional]
154
+
155
+ ### Model Architecture and Objective
156
+
157
+ [More Information Needed]
158
+
159
+ ### Compute Infrastructure
160
+
161
+ [More Information Needed]
162
+
163
+ #### Hardware
164
+
165
+ [More Information Needed]
166
+
167
+ #### Software
168
+
169
+ [More Information Needed]
170
+
171
+ ## Citation [optional]
172
+
173
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
+
175
+ **BibTeX:**
176
+
177
+ [More Information Needed]
178
+
179
+ **APA:**
180
+
181
+ [More Information Needed]
182
+
183
+ ## Glossary [optional]
184
+
185
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
+
187
+ [More Information Needed]
188
+
189
+ ## More Information [optional]
190
+
191
+ [More Information Needed]
192
+
193
+ ## Model Card Authors [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Contact
198
+
199
+ [More Information Needed]
200
+ ### Framework versions
201
+
202
+ - PEFT 0.15.2
checkpoint-1833/adapter_config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "unsloth/qwen3-0.6b-base-unsloth-bnb-4bit",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 16,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "r": 16,
24
+ "rank_pattern": {},
25
+ "revision": null,
26
+ "target_modules": [
27
+ "q_proj",
28
+ "up_proj",
29
+ "down_proj",
30
+ "o_proj",
31
+ "gate_proj",
32
+ "k_proj",
33
+ "v_proj"
34
+ ],
35
+ "task_type": "CAUSAL_LM",
36
+ "trainable_token_indices": null,
37
+ "use_dora": false,
38
+ "use_rslora": false
39
+ }
checkpoint-1833/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:485fa41bedc9335878e9c7b4f0c4f52385a522583f401de1c5a2956133f230f8
3
+ size 40422168
checkpoint-1833/added_tokens.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</think>": 151668,
3
+ "</tool_call>": 151658,
4
+ "</tool_response>": 151666,
5
+ "<think>": 151667,
6
+ "<tool_call>": 151657,
7
+ "<tool_response>": 151665,
8
+ "<|box_end|>": 151649,
9
+ "<|box_start|>": 151648,
10
+ "<|endoftext|>": 151643,
11
+ "<|file_sep|>": 151664,
12
+ "<|fim_middle|>": 151660,
13
+ "<|fim_pad|>": 151662,
14
+ "<|fim_prefix|>": 151659,
15
+ "<|fim_suffix|>": 151661,
16
+ "<|im_end|>": 151645,
17
+ "<|im_start|>": 151644,
18
+ "<|image_pad|>": 151655,
19
+ "<|object_ref_end|>": 151647,
20
+ "<|object_ref_start|>": 151646,
21
+ "<|quad_end|>": 151651,
22
+ "<|quad_start|>": 151650,
23
+ "<|repo_name|>": 151663,
24
+ "<|video_pad|>": 151656,
25
+ "<|vision_end|>": 151653,
26
+ "<|vision_pad|>": 151654,
27
+ "<|vision_start|>": 151652
28
+ }
checkpoint-1833/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-1833/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af38f592ee12af3adb138359a44a2e5b21a789cec34f8c8573868b748aaf1ce0
3
+ size 20959365
checkpoint-1833/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:181c5f0270cf39930062ddfa3767a2481d0c360f120b11f8e25dbf533a1cdaba
3
+ size 14645
checkpoint-1833/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0558c543294a429c780f4ca9db6afecbc8f9865e5c4bc93fd3a8b8c2532011b2
3
+ size 1383
checkpoint-1833/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b28aa5bac6afb5a97da9997a0dd243453fcbfa53975cc9fb9d08a0733692ce8
3
+ size 1465
checkpoint-1833/special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|endoftext|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|vision_pad|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
checkpoint-1833/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aeb13307a71acd8fe81861d94ad54ab689df773318809eed3cbe794b4492dae4
3
+ size 11422654
checkpoint-1833/tokenizer_config.json ADDED
@@ -0,0 +1,240 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ },
181
+ "151665": {
182
+ "content": "<tool_response>",
183
+ "lstrip": false,
184
+ "normalized": false,
185
+ "rstrip": false,
186
+ "single_word": false,
187
+ "special": false
188
+ },
189
+ "151666": {
190
+ "content": "</tool_response>",
191
+ "lstrip": false,
192
+ "normalized": false,
193
+ "rstrip": false,
194
+ "single_word": false,
195
+ "special": false
196
+ },
197
+ "151667": {
198
+ "content": "<think>",
199
+ "lstrip": false,
200
+ "normalized": false,
201
+ "rstrip": false,
202
+ "single_word": false,
203
+ "special": false
204
+ },
205
+ "151668": {
206
+ "content": "</think>",
207
+ "lstrip": false,
208
+ "normalized": false,
209
+ "rstrip": false,
210
+ "single_word": false,
211
+ "special": false
212
+ }
213
+ },
214
+ "additional_special_tokens": [
215
+ "<|im_start|>",
216
+ "<|im_end|>",
217
+ "<|object_ref_start|>",
218
+ "<|object_ref_end|>",
219
+ "<|box_start|>",
220
+ "<|box_end|>",
221
+ "<|quad_start|>",
222
+ "<|quad_end|>",
223
+ "<|vision_start|>",
224
+ "<|vision_end|>",
225
+ "<|vision_pad|>",
226
+ "<|image_pad|>",
227
+ "<|video_pad|>"
228
+ ],
229
+ "bos_token": null,
230
+ "clean_up_tokenization_spaces": false,
231
+ "eos_token": "<|endoftext|>",
232
+ "errors": "replace",
233
+ "extra_special_tokens": {},
234
+ "model_max_length": 32768,
235
+ "pad_token": "<|vision_pad|>",
236
+ "padding_side": "right",
237
+ "split_special_tokens": false,
238
+ "tokenizer_class": "Qwen2Tokenizer",
239
+ "unk_token": null
240
+ }
checkpoint-1833/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-1833/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:adb07c9cf31b39264fa1c39cdcedb4f8f2cd0a21811a03b1f63968e1dd128142
3
+ size 6673
checkpoint-1833/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-500/README.md ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: unsloth/qwen3-0.6b-base-unsloth-bnb-4bit
3
+ library_name: peft
4
+ ---
5
+
6
+ # Model Card for Model ID
7
+
8
+ <!-- Provide a quick summary of what the model is/does. -->
9
+
10
+
11
+
12
+ ## Model Details
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+
19
+
20
+ - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
+ - **Shared by [optional]:** [More Information Needed]
23
+ - **Model type:** [More Information Needed]
24
+ - **Language(s) (NLP):** [More Information Needed]
25
+ - **License:** [More Information Needed]
26
+ - **Finetuned from model [optional]:** [More Information Needed]
27
+
28
+ ### Model Sources [optional]
29
+
30
+ <!-- Provide the basic links for the model. -->
31
+
32
+ - **Repository:** [More Information Needed]
33
+ - **Paper [optional]:** [More Information Needed]
34
+ - **Demo [optional]:** [More Information Needed]
35
+
36
+ ## Uses
37
+
38
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
+
40
+ ### Direct Use
41
+
42
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
+
44
+ [More Information Needed]
45
+
46
+ ### Downstream Use [optional]
47
+
48
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Out-of-Scope Use
53
+
54
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
+
56
+ [More Information Needed]
57
+
58
+ ## Bias, Risks, and Limitations
59
+
60
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ### Recommendations
65
+
66
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
+
68
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ Use the code below to get started with the model.
73
+
74
+ [More Information Needed]
75
+
76
+ ## Training Details
77
+
78
+ ### Training Data
79
+
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
+
82
+ [More Information Needed]
83
+
84
+ ### Training Procedure
85
+
86
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
+
88
+ #### Preprocessing [optional]
89
+
90
+ [More Information Needed]
91
+
92
+
93
+ #### Training Hyperparameters
94
+
95
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
+
97
+ #### Speeds, Sizes, Times [optional]
98
+
99
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
+
101
+ [More Information Needed]
102
+
103
+ ## Evaluation
104
+
105
+ <!-- This section describes the evaluation protocols and provides the results. -->
106
+
107
+ ### Testing Data, Factors & Metrics
108
+
109
+ #### Testing Data
110
+
111
+ <!-- This should link to a Dataset Card if possible. -->
112
+
113
+ [More Information Needed]
114
+
115
+ #### Factors
116
+
117
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Metrics
122
+
123
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
+
125
+ [More Information Needed]
126
+
127
+ ### Results
128
+
129
+ [More Information Needed]
130
+
131
+ #### Summary
132
+
133
+
134
+
135
+ ## Model Examination [optional]
136
+
137
+ <!-- Relevant interpretability work for the model goes here -->
138
+
139
+ [More Information Needed]
140
+
141
+ ## Environmental Impact
142
+
143
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
+
145
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
+
147
+ - **Hardware Type:** [More Information Needed]
148
+ - **Hours used:** [More Information Needed]
149
+ - **Cloud Provider:** [More Information Needed]
150
+ - **Compute Region:** [More Information Needed]
151
+ - **Carbon Emitted:** [More Information Needed]
152
+
153
+ ## Technical Specifications [optional]
154
+
155
+ ### Model Architecture and Objective
156
+
157
+ [More Information Needed]
158
+
159
+ ### Compute Infrastructure
160
+
161
+ [More Information Needed]
162
+
163
+ #### Hardware
164
+
165
+ [More Information Needed]
166
+
167
+ #### Software
168
+
169
+ [More Information Needed]
170
+
171
+ ## Citation [optional]
172
+
173
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
+
175
+ **BibTeX:**
176
+
177
+ [More Information Needed]
178
+
179
+ **APA:**
180
+
181
+ [More Information Needed]
182
+
183
+ ## Glossary [optional]
184
+
185
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
+
187
+ [More Information Needed]
188
+
189
+ ## More Information [optional]
190
+
191
+ [More Information Needed]
192
+
193
+ ## Model Card Authors [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Contact
198
+
199
+ [More Information Needed]
200
+ ### Framework versions
201
+
202
+ - PEFT 0.15.2
checkpoint-500/adapter_config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "unsloth/qwen3-0.6b-base-unsloth-bnb-4bit",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 16,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "r": 16,
24
+ "rank_pattern": {},
25
+ "revision": null,
26
+ "target_modules": [
27
+ "q_proj",
28
+ "up_proj",
29
+ "down_proj",
30
+ "o_proj",
31
+ "gate_proj",
32
+ "k_proj",
33
+ "v_proj"
34
+ ],
35
+ "task_type": "CAUSAL_LM",
36
+ "trainable_token_indices": null,
37
+ "use_dora": false,
38
+ "use_rslora": false
39
+ }
checkpoint-500/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa8535923a33bd7ce1d03778028cceb99333b8e9f594005a6ef6e9d6e9e070df
3
+ size 40422168
checkpoint-500/added_tokens.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</think>": 151668,
3
+ "</tool_call>": 151658,
4
+ "</tool_response>": 151666,
5
+ "<think>": 151667,
6
+ "<tool_call>": 151657,
7
+ "<tool_response>": 151665,
8
+ "<|box_end|>": 151649,
9
+ "<|box_start|>": 151648,
10
+ "<|endoftext|>": 151643,
11
+ "<|file_sep|>": 151664,
12
+ "<|fim_middle|>": 151660,
13
+ "<|fim_pad|>": 151662,
14
+ "<|fim_prefix|>": 151659,
15
+ "<|fim_suffix|>": 151661,
16
+ "<|im_end|>": 151645,
17
+ "<|im_start|>": 151644,
18
+ "<|image_pad|>": 151655,
19
+ "<|object_ref_end|>": 151647,
20
+ "<|object_ref_start|>": 151646,
21
+ "<|quad_end|>": 151651,
22
+ "<|quad_start|>": 151650,
23
+ "<|repo_name|>": 151663,
24
+ "<|video_pad|>": 151656,
25
+ "<|vision_end|>": 151653,
26
+ "<|vision_pad|>": 151654,
27
+ "<|vision_start|>": 151652
28
+ }