hai2131 commited on
Commit
a174eeb
·
verified ·
1 Parent(s): 778e4c4

Upload checkpoint with token

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: hai2131/sailor2_1b_mt_sft
3
+ library_name: peft
4
+ ---
5
+
6
+ # Model Card for Model ID
7
+
8
+ <!-- Provide a quick summary of what the model is/does. -->
9
+
10
+
11
+
12
+ ## Model Details
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+
19
+
20
+ - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
+ - **Shared by [optional]:** [More Information Needed]
23
+ - **Model type:** [More Information Needed]
24
+ - **Language(s) (NLP):** [More Information Needed]
25
+ - **License:** [More Information Needed]
26
+ - **Finetuned from model [optional]:** [More Information Needed]
27
+
28
+ ### Model Sources [optional]
29
+
30
+ <!-- Provide the basic links for the model. -->
31
+
32
+ - **Repository:** [More Information Needed]
33
+ - **Paper [optional]:** [More Information Needed]
34
+ - **Demo [optional]:** [More Information Needed]
35
+
36
+ ## Uses
37
+
38
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
+
40
+ ### Direct Use
41
+
42
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
+
44
+ [More Information Needed]
45
+
46
+ ### Downstream Use [optional]
47
+
48
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Out-of-Scope Use
53
+
54
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
+
56
+ [More Information Needed]
57
+
58
+ ## Bias, Risks, and Limitations
59
+
60
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ### Recommendations
65
+
66
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
+
68
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ Use the code below to get started with the model.
73
+
74
+ [More Information Needed]
75
+
76
+ ## Training Details
77
+
78
+ ### Training Data
79
+
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
+
82
+ [More Information Needed]
83
+
84
+ ### Training Procedure
85
+
86
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
+
88
+ #### Preprocessing [optional]
89
+
90
+ [More Information Needed]
91
+
92
+
93
+ #### Training Hyperparameters
94
+
95
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
+
97
+ #### Speeds, Sizes, Times [optional]
98
+
99
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
+
101
+ [More Information Needed]
102
+
103
+ ## Evaluation
104
+
105
+ <!-- This section describes the evaluation protocols and provides the results. -->
106
+
107
+ ### Testing Data, Factors & Metrics
108
+
109
+ #### Testing Data
110
+
111
+ <!-- This should link to a Dataset Card if possible. -->
112
+
113
+ [More Information Needed]
114
+
115
+ #### Factors
116
+
117
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Metrics
122
+
123
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
+
125
+ [More Information Needed]
126
+
127
+ ### Results
128
+
129
+ [More Information Needed]
130
+
131
+ #### Summary
132
+
133
+
134
+
135
+ ## Model Examination [optional]
136
+
137
+ <!-- Relevant interpretability work for the model goes here -->
138
+
139
+ [More Information Needed]
140
+
141
+ ## Environmental Impact
142
+
143
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
+
145
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
+
147
+ - **Hardware Type:** [More Information Needed]
148
+ - **Hours used:** [More Information Needed]
149
+ - **Cloud Provider:** [More Information Needed]
150
+ - **Compute Region:** [More Information Needed]
151
+ - **Carbon Emitted:** [More Information Needed]
152
+
153
+ ## Technical Specifications [optional]
154
+
155
+ ### Model Architecture and Objective
156
+
157
+ [More Information Needed]
158
+
159
+ ### Compute Infrastructure
160
+
161
+ [More Information Needed]
162
+
163
+ #### Hardware
164
+
165
+ [More Information Needed]
166
+
167
+ #### Software
168
+
169
+ [More Information Needed]
170
+
171
+ ## Citation [optional]
172
+
173
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
+
175
+ **BibTeX:**
176
+
177
+ [More Information Needed]
178
+
179
+ **APA:**
180
+
181
+ [More Information Needed]
182
+
183
+ ## Glossary [optional]
184
+
185
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
+
187
+ [More Information Needed]
188
+
189
+ ## More Information [optional]
190
+
191
+ [More Information Needed]
192
+
193
+ ## Model Card Authors [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Contact
198
+
199
+ [More Information Needed]
200
+ ### Framework versions
201
+
202
+ - PEFT 0.15.2
adapter_config.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": {
4
+ "base_model_class": "Qwen2ForCausalLM",
5
+ "parent_library": "transformers.models.qwen2.modeling_qwen2",
6
+ "unsloth_fixed": true
7
+ },
8
+ "base_model_name_or_path": "hai2131/sailor2_1b_mt_sft",
9
+ "bias": "none",
10
+ "corda_config": null,
11
+ "eva_config": null,
12
+ "exclude_modules": null,
13
+ "fan_in_fan_out": false,
14
+ "inference_mode": true,
15
+ "init_lora_weights": true,
16
+ "layer_replication": null,
17
+ "layers_pattern": null,
18
+ "layers_to_transform": null,
19
+ "loftq_config": {},
20
+ "lora_alpha": 64,
21
+ "lora_bias": false,
22
+ "lora_dropout": 0.0,
23
+ "megatron_config": null,
24
+ "megatron_core": "megatron.core",
25
+ "modules_to_save": null,
26
+ "peft_type": "LORA",
27
+ "r": 32,
28
+ "rank_pattern": {},
29
+ "revision": null,
30
+ "target_modules": [
31
+ "k_proj",
32
+ "up_proj",
33
+ "o_proj",
34
+ "v_proj",
35
+ "q_proj",
36
+ "down_proj",
37
+ "gate_proj"
38
+ ],
39
+ "task_type": "CAUSAL_LM",
40
+ "trainable_token_indices": null,
41
+ "use_dora": false,
42
+ "use_rslora": false
43
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b85280e0ed74c9882d21a5b538b98c48cbe403809e9fc9ed3d1bd7aa913a49f1
3
+ size 140860656
added_tokens.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</tool_call>": 151658,
3
+ "<tool_call>": 151657,
4
+ "<|box_end|>": 151649,
5
+ "<|box_start|>": 151648,
6
+ "<|endoftext|>": 151643,
7
+ "<|file_sep|>": 151664,
8
+ "<|fim_middle|>": 151660,
9
+ "<|fim_pad|>": 151662,
10
+ "<|fim_prefix|>": 151659,
11
+ "<|fim_suffix|>": 151661,
12
+ "<|im_end|>": 151645,
13
+ "<|im_start|>": 151644,
14
+ "<|image_pad|>": 151655,
15
+ "<|object_ref_end|>": 151647,
16
+ "<|object_ref_start|>": 151646,
17
+ "<|quad_end|>": 151651,
18
+ "<|quad_start|>": 151650,
19
+ "<|repo_name|>": 151663,
20
+ "<|video_pad|>": 151656,
21
+ "<|vision_end|>": 151653,
22
+ "<|vision_pad|>": 151654,
23
+ "<|vision_start|>": 151652
24
+ }
chat_template.jinja ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system
2
+ You are an AI assistant named Sailor2, created by Sea AI Lab. As an AI assistant, you can answer questions in English, Chinese, and Southeast Asian languages such as Burmese, Cebuano, Ilocano, Indonesian, Javanese, Khmer, Lao, Malay, Sundanese, Tagalog, Thai, Vietnamese, and Waray. Your responses should be friendly, unbiased, informative, detailed, and faithful.<|im_end|>
3
+ ' }}{% endif %}{{'<|im_start|>' + message['role'] + '
4
+ ' + message['content'] + '<|im_end|>' + '
5
+ '}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant
6
+ ' }}{% endif %}
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:965e74ce900db6f9d199af61739816aff53fe39e1768d57639ceaaaff2cd9af7
3
+ size 72272293
rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7cc3307602102ef5b4699c5d0d0877f4df4e9a8c6ed16159e0eb5f7c205b7a82
3
+ size 15365
rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a564fd29fdaea985856b48f282cabacb0dffa1dd1a79d98c9e6177e1de774cf
3
+ size 15365
rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c1fbc115b0e5dd3df1a482bef8a39fc8b8fc0f501ef419324e60d082c1133bc
3
+ size 15365
rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:476ddfa1fd95c624fc018bdacac3fd79e42329b309ca4076c70c58da0694461a
3
+ size 15365
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc3bfdba43e140aa8e5cd5fc947cfb23559f9daad7f487c0e1a19d5a1110e00e
3
+ size 1465
special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|im_end|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|endoftext|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c5ae00e602b8860cbd784ba82a8aa14e8feecec692e7076590d014d7b7fdafa
3
+ size 11421896
tokenizer_config.json ADDED
@@ -0,0 +1,212 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ }
181
+ },
182
+ "additional_special_tokens": [
183
+ "<|im_start|>",
184
+ "<|im_end|>",
185
+ "<|object_ref_start|>",
186
+ "<|object_ref_end|>",
187
+ "<|box_start|>",
188
+ "<|box_end|>",
189
+ "<|quad_start|>",
190
+ "<|quad_end|>",
191
+ "<|vision_start|>",
192
+ "<|vision_end|>",
193
+ "<|vision_pad|>",
194
+ "<|image_pad|>",
195
+ "<|video_pad|>"
196
+ ],
197
+ "bos_token": null,
198
+ "clean_up_tokenization_spaces": false,
199
+ "eos_token": "<|im_end|>",
200
+ "errors": "replace",
201
+ "extra_special_tokens": {},
202
+ "max_length": 3848,
203
+ "model_max_length": 4096,
204
+ "pad_token": "<|endoftext|>",
205
+ "padding_side": "right",
206
+ "split_special_tokens": false,
207
+ "stride": 0,
208
+ "tokenizer_class": "Qwen2Tokenizer",
209
+ "truncation_side": "right",
210
+ "truncation_strategy": "longest_first",
211
+ "unk_token": null
212
+ }
trainer_state.json ADDED
@@ -0,0 +1,1234 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 3.0,
6
+ "eval_steps": 500,
7
+ "global_step": 3750,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.04,
14
+ "grad_norm": 24.803234100341797,
15
+ "learning_rate": 1.3066666666666667e-06,
16
+ "logits/chosen": 2.3097052574157715,
17
+ "logits/rejected": 2.3757848739624023,
18
+ "logps/chosen": -67.16219329833984,
19
+ "logps/rejected": -80.8543930053711,
20
+ "loss": 1.0777,
21
+ "nll_loss": 0.41877487301826477,
22
+ "rewards/accuracies": 0.7237499952316284,
23
+ "rewards/chosen": -6.716219425201416,
24
+ "rewards/margins": 1.3692210912704468,
25
+ "rewards/rejected": -8.085439682006836,
26
+ "step": 50
27
+ },
28
+ {
29
+ "epoch": 0.08,
30
+ "grad_norm": 12.049943923950195,
31
+ "learning_rate": 2.64e-06,
32
+ "logits/chosen": 2.2844557762145996,
33
+ "logits/rejected": 2.364481210708618,
34
+ "logps/chosen": -61.06942367553711,
35
+ "logps/rejected": -75.95172119140625,
36
+ "loss": 1.0254,
37
+ "nll_loss": 0.38137227296829224,
38
+ "rewards/accuracies": 0.7174999713897705,
39
+ "rewards/chosen": -6.106942176818848,
40
+ "rewards/margins": 1.4882298707962036,
41
+ "rewards/rejected": -7.595171928405762,
42
+ "step": 100
43
+ },
44
+ {
45
+ "epoch": 0.12,
46
+ "grad_norm": 19.429025650024414,
47
+ "learning_rate": 3.973333333333333e-06,
48
+ "logits/chosen": 2.2412264347076416,
49
+ "logits/rejected": 2.2957522869110107,
50
+ "logps/chosen": -56.071502685546875,
51
+ "logps/rejected": -67.70003509521484,
52
+ "loss": 1.0618,
53
+ "nll_loss": 0.3372814655303955,
54
+ "rewards/accuracies": 0.6650000214576721,
55
+ "rewards/chosen": -5.607150077819824,
56
+ "rewards/margins": 1.1628532409667969,
57
+ "rewards/rejected": -6.770003795623779,
58
+ "step": 150
59
+ },
60
+ {
61
+ "epoch": 0.16,
62
+ "grad_norm": 14.493091583251953,
63
+ "learning_rate": 5.306666666666667e-06,
64
+ "logits/chosen": 2.1775848865509033,
65
+ "logits/rejected": 2.25844144821167,
66
+ "logps/chosen": -50.50484848022461,
67
+ "logps/rejected": -64.37080383300781,
68
+ "loss": 0.9298,
69
+ "nll_loss": 0.3056701719760895,
70
+ "rewards/accuracies": 0.7237499952316284,
71
+ "rewards/chosen": -5.050484657287598,
72
+ "rewards/margins": 1.3865959644317627,
73
+ "rewards/rejected": -6.437079906463623,
74
+ "step": 200
75
+ },
76
+ {
77
+ "epoch": 0.2,
78
+ "grad_norm": 17.336822509765625,
79
+ "learning_rate": 6.640000000000001e-06,
80
+ "logits/chosen": 2.203662872314453,
81
+ "logits/rejected": 2.28377103805542,
82
+ "logps/chosen": -51.80202102661133,
83
+ "logps/rejected": -65.30438232421875,
84
+ "loss": 0.9291,
85
+ "nll_loss": 0.2998887896537781,
86
+ "rewards/accuracies": 0.7149999737739563,
87
+ "rewards/chosen": -5.180202007293701,
88
+ "rewards/margins": 1.3502373695373535,
89
+ "rewards/rejected": -6.530439376831055,
90
+ "step": 250
91
+ },
92
+ {
93
+ "epoch": 0.24,
94
+ "grad_norm": 25.542375564575195,
95
+ "learning_rate": 7.973333333333334e-06,
96
+ "logits/chosen": 2.2555084228515625,
97
+ "logits/rejected": 2.3327550888061523,
98
+ "logps/chosen": -49.52461242675781,
99
+ "logps/rejected": -64.92198181152344,
100
+ "loss": 0.8288,
101
+ "nll_loss": 0.29204535484313965,
102
+ "rewards/accuracies": 0.7674999833106995,
103
+ "rewards/chosen": -4.9524617195129395,
104
+ "rewards/margins": 1.5397369861602783,
105
+ "rewards/rejected": -6.492198467254639,
106
+ "step": 300
107
+ },
108
+ {
109
+ "epoch": 0.28,
110
+ "grad_norm": 10.780914306640625,
111
+ "learning_rate": 9.306666666666667e-06,
112
+ "logits/chosen": 2.2995264530181885,
113
+ "logits/rejected": 2.3702523708343506,
114
+ "logps/chosen": -49.0943489074707,
115
+ "logps/rejected": -62.10791015625,
116
+ "loss": 0.9434,
117
+ "nll_loss": 0.29649874567985535,
118
+ "rewards/accuracies": 0.7250000238418579,
119
+ "rewards/chosen": -4.909435272216797,
120
+ "rewards/margins": 1.3013556003570557,
121
+ "rewards/rejected": -6.210791110992432,
122
+ "step": 350
123
+ },
124
+ {
125
+ "epoch": 0.32,
126
+ "grad_norm": 11.368802070617676,
127
+ "learning_rate": 9.998752338940612e-06,
128
+ "logits/chosen": 2.4777705669403076,
129
+ "logits/rejected": 2.5543289184570312,
130
+ "logps/chosen": -46.3427848815918,
131
+ "logps/rejected": -58.78908157348633,
132
+ "loss": 0.8661,
133
+ "nll_loss": 0.27519530057907104,
134
+ "rewards/accuracies": 0.7137500047683716,
135
+ "rewards/chosen": -4.634278297424316,
136
+ "rewards/margins": 1.2446298599243164,
137
+ "rewards/rejected": -5.878908634185791,
138
+ "step": 400
139
+ },
140
+ {
141
+ "epoch": 0.36,
142
+ "grad_norm": 13.878002166748047,
143
+ "learning_rate": 9.988142751731797e-06,
144
+ "logits/chosen": 2.4681615829467773,
145
+ "logits/rejected": 2.540001153945923,
146
+ "logps/chosen": -46.92790985107422,
147
+ "logps/rejected": -63.343467712402344,
148
+ "loss": 0.8081,
149
+ "nll_loss": 0.2872486710548401,
150
+ "rewards/accuracies": 0.7412499785423279,
151
+ "rewards/chosen": -4.692790985107422,
152
+ "rewards/margins": 1.6415550708770752,
153
+ "rewards/rejected": -6.334346294403076,
154
+ "step": 450
155
+ },
156
+ {
157
+ "epoch": 0.4,
158
+ "grad_norm": 14.774677276611328,
159
+ "learning_rate": 9.966729958067638e-06,
160
+ "logits/chosen": 2.509334087371826,
161
+ "logits/rejected": 2.5635790824890137,
162
+ "logps/chosen": -48.37749481201172,
163
+ "logps/rejected": -64.00433349609375,
164
+ "loss": 0.873,
165
+ "nll_loss": 0.2907007336616516,
166
+ "rewards/accuracies": 0.7200000286102295,
167
+ "rewards/chosen": -4.837749481201172,
168
+ "rewards/margins": 1.5626837015151978,
169
+ "rewards/rejected": -6.400432586669922,
170
+ "step": 500
171
+ },
172
+ {
173
+ "epoch": 0.44,
174
+ "grad_norm": 18.610092163085938,
175
+ "learning_rate": 9.934560333291077e-06,
176
+ "logits/chosen": 2.5441677570343018,
177
+ "logits/rejected": 2.613149404525757,
178
+ "logps/chosen": -48.265201568603516,
179
+ "logps/rejected": -63.49806213378906,
180
+ "loss": 0.8581,
181
+ "nll_loss": 0.2854183316230774,
182
+ "rewards/accuracies": 0.7287499904632568,
183
+ "rewards/chosen": -4.826520919799805,
184
+ "rewards/margins": 1.5232864618301392,
185
+ "rewards/rejected": -6.349807262420654,
186
+ "step": 550
187
+ },
188
+ {
189
+ "epoch": 0.48,
190
+ "grad_norm": 13.354905128479004,
191
+ "learning_rate": 9.89170354964594e-06,
192
+ "logits/chosen": 2.646564245223999,
193
+ "logits/rejected": 2.700756549835205,
194
+ "logps/chosen": -45.61009216308594,
195
+ "logps/rejected": -60.2086067199707,
196
+ "loss": 0.8363,
197
+ "nll_loss": 0.27716541290283203,
198
+ "rewards/accuracies": 0.7300000190734863,
199
+ "rewards/chosen": -4.561009883880615,
200
+ "rewards/margins": 1.4598517417907715,
201
+ "rewards/rejected": -6.0208611488342285,
202
+ "step": 600
203
+ },
204
+ {
205
+ "epoch": 0.52,
206
+ "grad_norm": 5.81559419631958,
207
+ "learning_rate": 9.83825242538238e-06,
208
+ "logits/chosen": 2.519411087036133,
209
+ "logits/rejected": 2.5865354537963867,
210
+ "logps/chosen": -46.41543960571289,
211
+ "logps/rejected": -63.667991638183594,
212
+ "loss": 0.7762,
213
+ "nll_loss": 0.27636775374412537,
214
+ "rewards/accuracies": 0.7587500214576721,
215
+ "rewards/chosen": -4.641544342041016,
216
+ "rewards/margins": 1.7252554893493652,
217
+ "rewards/rejected": -6.366799831390381,
218
+ "step": 650
219
+ },
220
+ {
221
+ "epoch": 0.56,
222
+ "grad_norm": 9.440147399902344,
223
+ "learning_rate": 9.774322723733216e-06,
224
+ "logits/chosen": 2.4503791332244873,
225
+ "logits/rejected": 2.5173845291137695,
226
+ "logps/chosen": -44.43014144897461,
227
+ "logps/rejected": -62.06157302856445,
228
+ "loss": 0.7447,
229
+ "nll_loss": 0.2715470492839813,
230
+ "rewards/accuracies": 0.7724999785423279,
231
+ "rewards/chosen": -4.443014621734619,
232
+ "rewards/margins": 1.7631423473358154,
233
+ "rewards/rejected": -6.206157207489014,
234
+ "step": 700
235
+ },
236
+ {
237
+ "epoch": 0.6,
238
+ "grad_norm": 9.054909706115723,
239
+ "learning_rate": 9.700052902196541e-06,
240
+ "logits/chosen": 2.3488733768463135,
241
+ "logits/rejected": 2.4224016666412354,
242
+ "logps/chosen": -45.02783203125,
243
+ "logps/rejected": -63.32554626464844,
244
+ "loss": 0.7256,
245
+ "nll_loss": 0.27446305751800537,
246
+ "rewards/accuracies": 0.8075000047683716,
247
+ "rewards/chosen": -4.502783298492432,
248
+ "rewards/margins": 1.8297713994979858,
249
+ "rewards/rejected": -6.332554817199707,
250
+ "step": 750
251
+ },
252
+ {
253
+ "epoch": 0.64,
254
+ "grad_norm": 16.773876190185547,
255
+ "learning_rate": 9.615603812667618e-06,
256
+ "logits/chosen": 2.4701356887817383,
257
+ "logits/rejected": 2.5356545448303223,
258
+ "logps/chosen": -44.635948181152344,
259
+ "logps/rejected": -62.55714416503906,
260
+ "loss": 0.7836,
261
+ "nll_loss": 0.2690128684043884,
262
+ "rewards/accuracies": 0.7524999976158142,
263
+ "rewards/chosen": -4.463594913482666,
264
+ "rewards/margins": 1.7921198606491089,
265
+ "rewards/rejected": -6.2557148933410645,
266
+ "step": 800
267
+ },
268
+ {
269
+ "epoch": 0.68,
270
+ "grad_norm": 12.325571060180664,
271
+ "learning_rate": 9.521158353069494e-06,
272
+ "logits/chosen": 2.450385332107544,
273
+ "logits/rejected": 2.5140254497528076,
274
+ "logps/chosen": -47.131587982177734,
275
+ "logps/rejected": -65.35118865966797,
276
+ "loss": 0.7507,
277
+ "nll_loss": 0.27970045804977417,
278
+ "rewards/accuracies": 0.7762500047683716,
279
+ "rewards/chosen": -4.713159561157227,
280
+ "rewards/margins": 1.8219600915908813,
281
+ "rewards/rejected": -6.535119533538818,
282
+ "step": 850
283
+ },
284
+ {
285
+ "epoch": 0.72,
286
+ "grad_norm": 18.231292724609375,
287
+ "learning_rate": 9.416921071236821e-06,
288
+ "logits/chosen": 2.4923791885375977,
289
+ "logits/rejected": 2.5595548152923584,
290
+ "logps/chosen": -46.485965728759766,
291
+ "logps/rejected": -63.454063415527344,
292
+ "loss": 0.8048,
293
+ "nll_loss": 0.27803030610084534,
294
+ "rewards/accuracies": 0.7437499761581421,
295
+ "rewards/chosen": -4.648596286773682,
296
+ "rewards/margins": 1.6968092918395996,
297
+ "rewards/rejected": -6.345406532287598,
298
+ "step": 900
299
+ },
300
+ {
301
+ "epoch": 0.76,
302
+ "grad_norm": 18.762359619140625,
303
+ "learning_rate": 9.303117721910801e-06,
304
+ "logits/chosen": 2.571411371231079,
305
+ "logits/rejected": 2.6245839595794678,
306
+ "logps/chosen": -44.32810974121094,
307
+ "logps/rejected": -61.370662689208984,
308
+ "loss": 0.7592,
309
+ "nll_loss": 0.2649703919887543,
310
+ "rewards/accuracies": 0.7662500143051147,
311
+ "rewards/chosen": -4.4328107833862305,
312
+ "rewards/margins": 1.7042557001113892,
313
+ "rewards/rejected": -6.137066841125488,
314
+ "step": 950
315
+ },
316
+ {
317
+ "epoch": 0.8,
318
+ "grad_norm": 13.24950122833252,
319
+ "learning_rate": 9.179994777804677e-06,
320
+ "logits/chosen": 2.5039403438568115,
321
+ "logits/rejected": 2.5774261951446533,
322
+ "logps/chosen": -45.20160675048828,
323
+ "logps/rejected": -62.881683349609375,
324
+ "loss": 0.7717,
325
+ "nll_loss": 0.2698582708835602,
326
+ "rewards/accuracies": 0.7787500023841858,
327
+ "rewards/chosen": -4.520160675048828,
328
+ "rewards/margins": 1.7680076360702515,
329
+ "rewards/rejected": -6.288168430328369,
330
+ "step": 1000
331
+ },
332
+ {
333
+ "epoch": 0.84,
334
+ "grad_norm": 12.429143905639648,
335
+ "learning_rate": 9.04781889579873e-06,
336
+ "logits/chosen": 2.5070714950561523,
337
+ "logits/rejected": 2.5675339698791504,
338
+ "logps/chosen": -45.09945297241211,
339
+ "logps/rejected": -64.37616729736328,
340
+ "loss": 0.7095,
341
+ "nll_loss": 0.2686724364757538,
342
+ "rewards/accuracies": 0.7912499904632568,
343
+ "rewards/chosen": -4.509944915771484,
344
+ "rewards/margins": 1.9276715517044067,
345
+ "rewards/rejected": -6.437616348266602,
346
+ "step": 1050
347
+ },
348
+ {
349
+ "epoch": 0.88,
350
+ "grad_norm": 13.30483341217041,
351
+ "learning_rate": 8.906876339420851e-06,
352
+ "logits/chosen": 2.388920783996582,
353
+ "logits/rejected": 2.4621076583862305,
354
+ "logps/chosen": -42.8404541015625,
355
+ "logps/rejected": -62.490779876708984,
356
+ "loss": 0.6913,
357
+ "nll_loss": 0.25781476497650146,
358
+ "rewards/accuracies": 0.8112499713897705,
359
+ "rewards/chosen": -4.284045219421387,
360
+ "rewards/margins": 1.9650330543518066,
361
+ "rewards/rejected": -6.249078750610352,
362
+ "step": 1100
363
+ },
364
+ {
365
+ "epoch": 0.92,
366
+ "grad_norm": 13.460766792297363,
367
+ "learning_rate": 8.757472358863481e-06,
368
+ "logits/chosen": 2.4224061965942383,
369
+ "logits/rejected": 2.5014472007751465,
370
+ "logps/chosen": -45.438575744628906,
371
+ "logps/rejected": -63.852115631103516,
372
+ "loss": 0.7312,
373
+ "nll_loss": 0.26547786593437195,
374
+ "rewards/accuracies": 0.7749999761581421,
375
+ "rewards/chosen": -4.543857574462891,
376
+ "rewards/margins": 1.841354250907898,
377
+ "rewards/rejected": -6.385211944580078,
378
+ "step": 1150
379
+ },
380
+ {
381
+ "epoch": 0.96,
382
+ "grad_norm": 9.995412826538086,
383
+ "learning_rate": 8.599930529879669e-06,
384
+ "logits/chosen": 2.289586067199707,
385
+ "logits/rejected": 2.3660268783569336,
386
+ "logps/chosen": -43.745357513427734,
387
+ "logps/rejected": -62.561038970947266,
388
+ "loss": 0.6825,
389
+ "nll_loss": 0.2639794647693634,
390
+ "rewards/accuracies": 0.8100000023841858,
391
+ "rewards/chosen": -4.37453556060791,
392
+ "rewards/margins": 1.8815685510635376,
393
+ "rewards/rejected": -6.256104946136475,
394
+ "step": 1200
395
+ },
396
+ {
397
+ "epoch": 1.0,
398
+ "grad_norm": 9.7256498336792,
399
+ "learning_rate": 8.434592052990044e-06,
400
+ "logits/chosen": 2.1411895751953125,
401
+ "logits/rejected": 2.216475248336792,
402
+ "logps/chosen": -41.174503326416016,
403
+ "logps/rejected": -59.75578308105469,
404
+ "loss": 0.687,
405
+ "nll_loss": 0.25248458981513977,
406
+ "rewards/accuracies": 0.7912499904632568,
407
+ "rewards/chosen": -4.11745023727417,
408
+ "rewards/margins": 1.8581281900405884,
409
+ "rewards/rejected": -5.9755778312683105,
410
+ "step": 1250
411
+ },
412
+ {
413
+ "epoch": 1.04,
414
+ "grad_norm": 5.8602166175842285,
415
+ "learning_rate": 8.261815014518465e-06,
416
+ "logits/chosen": 2.094317674636841,
417
+ "logits/rejected": 2.1707794666290283,
418
+ "logps/chosen": -42.47966003417969,
419
+ "logps/rejected": -67.64102172851562,
420
+ "loss": 0.5271,
421
+ "nll_loss": 0.25495222210884094,
422
+ "rewards/accuracies": 0.8837500214576721,
423
+ "rewards/chosen": -4.247966289520264,
424
+ "rewards/margins": 2.516136884689331,
425
+ "rewards/rejected": -6.764101982116699,
426
+ "step": 1300
427
+ },
428
+ {
429
+ "epoch": 1.08,
430
+ "grad_norm": 7.209846019744873,
431
+ "learning_rate": 8.081973611056784e-06,
432
+ "logits/chosen": 1.899792194366455,
433
+ "logits/rejected": 1.9656882286071777,
434
+ "logps/chosen": -41.67923355102539,
435
+ "logps/rejected": -67.45558166503906,
436
+ "loss": 0.5335,
437
+ "nll_loss": 0.24939106404781342,
438
+ "rewards/accuracies": 0.8774999976158142,
439
+ "rewards/chosen": -4.167923450469971,
440
+ "rewards/margins": 2.577634811401367,
441
+ "rewards/rejected": -6.745559215545654,
442
+ "step": 1350
443
+ },
444
+ {
445
+ "epoch": 1.12,
446
+ "grad_norm": 12.841861724853516,
447
+ "learning_rate": 7.89545733903834e-06,
448
+ "logits/chosen": 1.8675593137741089,
449
+ "logits/rejected": 1.9421709775924683,
450
+ "logps/chosen": -43.389644622802734,
451
+ "logps/rejected": -68.71646881103516,
452
+ "loss": 0.5398,
453
+ "nll_loss": 0.2613371014595032,
454
+ "rewards/accuracies": 0.8974999785423279,
455
+ "rewards/chosen": -4.338963985443115,
456
+ "rewards/margins": 2.5326828956604004,
457
+ "rewards/rejected": -6.871647834777832,
458
+ "step": 1400
459
+ },
460
+ {
461
+ "epoch": 1.16,
462
+ "grad_norm": 13.226041793823242,
463
+ "learning_rate": 7.702670151175435e-06,
464
+ "logits/chosen": 1.8003557920455933,
465
+ "logits/rejected": 1.877779245376587,
466
+ "logps/chosen": -42.74818801879883,
467
+ "logps/rejected": -67.07624053955078,
468
+ "loss": 0.5569,
469
+ "nll_loss": 0.25505152344703674,
470
+ "rewards/accuracies": 0.8899999856948853,
471
+ "rewards/chosen": -4.274819374084473,
472
+ "rewards/margins": 2.4328057765960693,
473
+ "rewards/rejected": -6.707624435424805,
474
+ "step": 1450
475
+ },
476
+ {
477
+ "epoch": 1.2,
478
+ "grad_norm": 6.528555393218994,
479
+ "learning_rate": 7.5040295815877e-06,
480
+ "logits/chosen": 1.8589718341827393,
481
+ "logits/rejected": 1.9315310716629028,
482
+ "logps/chosen": -42.12541198730469,
483
+ "logps/rejected": -68.34070587158203,
484
+ "loss": 0.5316,
485
+ "nll_loss": 0.25116249918937683,
486
+ "rewards/accuracies": 0.8849999904632568,
487
+ "rewards/chosen": -4.212540626525879,
488
+ "rewards/margins": 2.621530055999756,
489
+ "rewards/rejected": -6.834071636199951,
490
+ "step": 1500
491
+ },
492
+ {
493
+ "epoch": 1.24,
494
+ "grad_norm": 8.806782722473145,
495
+ "learning_rate": 7.299965841516164e-06,
496
+ "logits/chosen": 1.896395206451416,
497
+ "logits/rejected": 1.9738116264343262,
498
+ "logps/chosen": -42.76396560668945,
499
+ "logps/rejected": -68.06053161621094,
500
+ "loss": 0.5404,
501
+ "nll_loss": 0.2575341463088989,
502
+ "rewards/accuracies": 0.8799999952316284,
503
+ "rewards/chosen": -4.27639627456665,
504
+ "rewards/margins": 2.5296568870544434,
505
+ "rewards/rejected": -6.8060526847839355,
506
+ "step": 1550
507
+ },
508
+ {
509
+ "epoch": 1.28,
510
+ "grad_norm": 13.923650741577148,
511
+ "learning_rate": 7.090920887581507e-06,
512
+ "logits/chosen": 1.881974220275879,
513
+ "logits/rejected": 1.9411964416503906,
514
+ "logps/chosen": -41.719696044921875,
515
+ "logps/rejected": -66.24463653564453,
516
+ "loss": 0.5483,
517
+ "nll_loss": 0.24970975518226624,
518
+ "rewards/accuracies": 0.8762500286102295,
519
+ "rewards/chosen": -4.171969413757324,
520
+ "rewards/margins": 2.452493667602539,
521
+ "rewards/rejected": -6.62446403503418,
522
+ "step": 1600
523
+ },
524
+ {
525
+ "epoch": 1.32,
526
+ "grad_norm": 11.483983993530273,
527
+ "learning_rate": 6.877347464604446e-06,
528
+ "logits/chosen": 1.8503919839859009,
529
+ "logits/rejected": 1.9176833629608154,
530
+ "logps/chosen": -42.25181579589844,
531
+ "logps/rejected": -67.97564697265625,
532
+ "loss": 0.541,
533
+ "nll_loss": 0.25359034538269043,
534
+ "rewards/accuracies": 0.8812500238418579,
535
+ "rewards/chosen": -4.225182056427002,
536
+ "rewards/margins": 2.572382926940918,
537
+ "rewards/rejected": -6.797564506530762,
538
+ "step": 1650
539
+ },
540
+ {
541
+ "epoch": 1.3599999999999999,
542
+ "grad_norm": 9.980610847473145,
543
+ "learning_rate": 6.659708125061242e-06,
544
+ "logits/chosen": 1.7879056930541992,
545
+ "logits/rejected": 1.8512917757034302,
546
+ "logps/chosen": -41.724449157714844,
547
+ "logps/rejected": -68.18392181396484,
548
+ "loss": 0.5127,
549
+ "nll_loss": 0.2496194839477539,
550
+ "rewards/accuracies": 0.9075000286102295,
551
+ "rewards/chosen": -4.172445297241211,
552
+ "rewards/margins": 2.645946741104126,
553
+ "rewards/rejected": -6.818392276763916,
554
+ "step": 1700
555
+ },
556
+ {
557
+ "epoch": 1.4,
558
+ "grad_norm": 8.7877779006958,
559
+ "learning_rate": 6.438474227298065e-06,
560
+ "logits/chosen": 1.813609004020691,
561
+ "logits/rejected": 1.8792080879211426,
562
+ "logps/chosen": -42.19902420043945,
563
+ "logps/rejected": -67.10396575927734,
564
+ "loss": 0.5475,
565
+ "nll_loss": 0.2563365399837494,
566
+ "rewards/accuracies": 0.8887500166893005,
567
+ "rewards/chosen": -4.219902038574219,
568
+ "rewards/margins": 2.4904944896698,
569
+ "rewards/rejected": -6.710396766662598,
570
+ "step": 1750
571
+ },
572
+ {
573
+ "epoch": 1.44,
574
+ "grad_norm": 9.004395484924316,
575
+ "learning_rate": 6.2141249146737545e-06,
576
+ "logits/chosen": 1.8060498237609863,
577
+ "logits/rejected": 1.8766037225723267,
578
+ "logps/chosen": -40.13544845581055,
579
+ "logps/rejected": -64.01081085205078,
580
+ "loss": 0.5417,
581
+ "nll_loss": 0.23975929617881775,
582
+ "rewards/accuracies": 0.8712499737739563,
583
+ "rewards/chosen": -4.013545036315918,
584
+ "rewards/margins": 2.3875370025634766,
585
+ "rewards/rejected": -6.401081562042236,
586
+ "step": 1800
587
+ },
588
+ {
589
+ "epoch": 1.48,
590
+ "grad_norm": 14.659512519836426,
591
+ "learning_rate": 5.987146077842015e-06,
592
+ "logits/chosen": 1.7913141250610352,
593
+ "logits/rejected": 1.8706274032592773,
594
+ "logps/chosen": -41.8923454284668,
595
+ "logps/rejected": -68.07263946533203,
596
+ "loss": 0.5291,
597
+ "nll_loss": 0.24965395033359528,
598
+ "rewards/accuracies": 0.8812500238418579,
599
+ "rewards/chosen": -4.189234733581543,
600
+ "rewards/margins": 2.6180291175842285,
601
+ "rewards/rejected": -6.8072638511657715,
602
+ "step": 1850
603
+ },
604
+ {
605
+ "epoch": 1.52,
606
+ "grad_norm": 6.758883953094482,
607
+ "learning_rate": 5.7580293024204455e-06,
608
+ "logits/chosen": 1.7456320524215698,
609
+ "logits/rejected": 1.8188166618347168,
610
+ "logps/chosen": -40.59235763549805,
611
+ "logps/rejected": -67.43293762207031,
612
+ "loss": 0.4978,
613
+ "nll_loss": 0.2416204810142517,
614
+ "rewards/accuracies": 0.8924999833106995,
615
+ "rewards/chosen": -4.0592360496521,
616
+ "rewards/margins": 2.6840572357177734,
617
+ "rewards/rejected": -6.743292808532715,
618
+ "step": 1900
619
+ },
620
+ {
621
+ "epoch": 1.56,
622
+ "grad_norm": 11.839475631713867,
623
+ "learning_rate": 5.5272708043255605e-06,
624
+ "logits/chosen": 1.771327257156372,
625
+ "logits/rejected": 1.8334521055221558,
626
+ "logps/chosen": -40.4869499206543,
627
+ "logps/rejected": -64.32723999023438,
628
+ "loss": 0.5586,
629
+ "nll_loss": 0.24806198477745056,
630
+ "rewards/accuracies": 0.8675000071525574,
631
+ "rewards/chosen": -4.048694610595703,
632
+ "rewards/margins": 2.3840293884277344,
633
+ "rewards/rejected": -6.4327239990234375,
634
+ "step": 1950
635
+ },
636
+ {
637
+ "epoch": 1.6,
638
+ "grad_norm": 12.160818099975586,
639
+ "learning_rate": 5.295370355079615e-06,
640
+ "logits/chosen": 1.7191228866577148,
641
+ "logits/rejected": 1.7949659824371338,
642
+ "logps/chosen": -40.17674255371094,
643
+ "logps/rejected": -67.11283874511719,
644
+ "loss": 0.5123,
645
+ "nll_loss": 0.24404479563236237,
646
+ "rewards/accuracies": 0.8774999976158142,
647
+ "rewards/chosen": -4.017674446105957,
648
+ "rewards/margins": 2.6936092376708984,
649
+ "rewards/rejected": -6.711284160614014,
650
+ "step": 2000
651
+ },
652
+ {
653
+ "epoch": 1.6400000000000001,
654
+ "grad_norm": 12.399029731750488,
655
+ "learning_rate": 5.062830199416764e-06,
656
+ "logits/chosen": 1.7645864486694336,
657
+ "logits/rejected": 1.8290934562683105,
658
+ "logps/chosen": -41.4688720703125,
659
+ "logps/rejected": -67.27618408203125,
660
+ "loss": 0.5257,
661
+ "nll_loss": 0.24901245534420013,
662
+ "rewards/accuracies": 0.8949999809265137,
663
+ "rewards/chosen": -4.146886825561523,
664
+ "rewards/margins": 2.580731153488159,
665
+ "rewards/rejected": -6.727618217468262,
666
+ "step": 2050
667
+ },
668
+ {
669
+ "epoch": 1.6800000000000002,
670
+ "grad_norm": 13.637982368469238,
671
+ "learning_rate": 4.8301539675328205e-06,
672
+ "logits/chosen": 1.6826503276824951,
673
+ "logits/rejected": 1.7500274181365967,
674
+ "logps/chosen": -40.96113586425781,
675
+ "logps/rejected": -67.78974914550781,
676
+ "loss": 0.513,
677
+ "nll_loss": 0.2408064305782318,
678
+ "rewards/accuracies": 0.8987500071525574,
679
+ "rewards/chosen": -4.096114158630371,
680
+ "rewards/margins": 2.6828606128692627,
681
+ "rewards/rejected": -6.778974533081055,
682
+ "step": 2100
683
+ },
684
+ {
685
+ "epoch": 1.72,
686
+ "grad_norm": 5.442174911499023,
687
+ "learning_rate": 4.597845584334387e-06,
688
+ "logits/chosen": 1.6728637218475342,
689
+ "logits/rejected": 1.7252342700958252,
690
+ "logps/chosen": -42.251564025878906,
691
+ "logps/rejected": -69.24467468261719,
692
+ "loss": 0.5304,
693
+ "nll_loss": 0.25177592039108276,
694
+ "rewards/accuracies": 0.8899999856948853,
695
+ "rewards/chosen": -4.225156307220459,
696
+ "rewards/margins": 2.699312210083008,
697
+ "rewards/rejected": -6.92446756362915,
698
+ "step": 2150
699
+ },
700
+ {
701
+ "epoch": 1.76,
702
+ "grad_norm": 10.488734245300293,
703
+ "learning_rate": 4.366408178049728e-06,
704
+ "logits/chosen": 1.6670646667480469,
705
+ "logits/rejected": 1.7333457469940186,
706
+ "logps/chosen": -39.98529052734375,
707
+ "logps/rejected": -65.87380981445312,
708
+ "loss": 0.5198,
709
+ "nll_loss": 0.2439391314983368,
710
+ "rewards/accuracies": 0.8762500286102295,
711
+ "rewards/chosen": -3.9985289573669434,
712
+ "rewards/margins": 2.588852643966675,
713
+ "rewards/rejected": -6.587381362915039,
714
+ "step": 2200
715
+ },
716
+ {
717
+ "epoch": 1.8,
718
+ "grad_norm": 11.035846710205078,
719
+ "learning_rate": 4.136342990565055e-06,
720
+ "logits/chosen": 1.6644293069839478,
721
+ "logits/rejected": 1.7320383787155151,
722
+ "logps/chosen": -40.47782897949219,
723
+ "logps/rejected": -69.36231994628906,
724
+ "loss": 0.5059,
725
+ "nll_loss": 0.2478472888469696,
726
+ "rewards/accuracies": 0.8949999809265137,
727
+ "rewards/chosen": -4.047782897949219,
728
+ "rewards/margins": 2.888448476791382,
729
+ "rewards/rejected": -6.93623161315918,
730
+ "step": 2250
731
+ },
732
+ {
733
+ "epoch": 1.8399999999999999,
734
+ "grad_norm": 14.5468168258667,
735
+ "learning_rate": 3.908148291846225e-06,
736
+ "logits/chosen": 1.6440070867538452,
737
+ "logits/rejected": 1.6980115175247192,
738
+ "logps/chosen": -37.658931732177734,
739
+ "logps/rejected": -64.18938446044922,
740
+ "loss": 0.4975,
741
+ "nll_loss": 0.23423399031162262,
742
+ "rewards/accuracies": 0.8899999856948853,
743
+ "rewards/chosen": -3.7658934593200684,
744
+ "rewards/margins": 2.6530449390411377,
745
+ "rewards/rejected": -6.418938636779785,
746
+ "step": 2300
747
+ },
748
+ {
749
+ "epoch": 1.88,
750
+ "grad_norm": 15.06531810760498,
751
+ "learning_rate": 3.6823183007969375e-06,
752
+ "logits/chosen": 1.633602261543274,
753
+ "logits/rejected": 1.69327974319458,
754
+ "logps/chosen": -38.722007751464844,
755
+ "logps/rejected": -65.98956298828125,
756
+ "loss": 0.512,
757
+ "nll_loss": 0.23453904688358307,
758
+ "rewards/accuracies": 0.8837500214576721,
759
+ "rewards/chosen": -3.8722009658813477,
760
+ "rewards/margins": 2.726755380630493,
761
+ "rewards/rejected": -6.598956108093262,
762
+ "step": 2350
763
+ },
764
+ {
765
+ "epoch": 1.92,
766
+ "grad_norm": 9.41374397277832,
767
+ "learning_rate": 3.4593421148906523e-06,
768
+ "logits/chosen": 1.6280794143676758,
769
+ "logits/rejected": 1.694667935371399,
770
+ "logps/chosen": -41.9329833984375,
771
+ "logps/rejected": -68.24996948242188,
772
+ "loss": 0.5314,
773
+ "nll_loss": 0.24245284497737885,
774
+ "rewards/accuracies": 0.8887500166893005,
775
+ "rewards/chosen": -4.19329833984375,
776
+ "rewards/margins": 2.6316983699798584,
777
+ "rewards/rejected": -6.8249969482421875,
778
+ "step": 2400
779
+ },
780
+ {
781
+ "epoch": 1.96,
782
+ "grad_norm": 10.686447143554688,
783
+ "learning_rate": 3.239702650894364e-06,
784
+ "logits/chosen": 1.648103952407837,
785
+ "logits/rejected": 1.7120572328567505,
786
+ "logps/chosen": -39.89656066894531,
787
+ "logps/rejected": -67.84294128417969,
788
+ "loss": 0.5132,
789
+ "nll_loss": 0.23985882103443146,
790
+ "rewards/accuracies": 0.8799999952316284,
791
+ "rewards/chosen": -3.989656448364258,
792
+ "rewards/margins": 2.794638156890869,
793
+ "rewards/rejected": -6.784294605255127,
794
+ "step": 2450
795
+ },
796
+ {
797
+ "epoch": 2.0,
798
+ "grad_norm": 6.338383197784424,
799
+ "learning_rate": 3.023875598978419e-06,
800
+ "logits/chosen": 1.6715092658996582,
801
+ "logits/rejected": 1.7344988584518433,
802
+ "logps/chosen": -39.51519012451172,
803
+ "logps/rejected": -68.46276092529297,
804
+ "loss": 0.4834,
805
+ "nll_loss": 0.24019798636436462,
806
+ "rewards/accuracies": 0.9112499952316284,
807
+ "rewards/chosen": -3.951519250869751,
808
+ "rewards/margins": 2.8947577476501465,
809
+ "rewards/rejected": -6.846276760101318,
810
+ "step": 2500
811
+ },
812
+ {
813
+ "epoch": 2.04,
814
+ "grad_norm": 5.144768238067627,
815
+ "learning_rate": 2.812328392477536e-06,
816
+ "logits/chosen": 1.5945950746536255,
817
+ "logits/rejected": 1.6544932126998901,
818
+ "logps/chosen": -37.963443756103516,
819
+ "logps/rejected": -69.92556762695312,
820
+ "loss": 0.3913,
821
+ "nll_loss": 0.22887782752513885,
822
+ "rewards/accuracies": 0.9712499976158142,
823
+ "rewards/chosen": -3.796344041824341,
824
+ "rewards/margins": 3.196213722229004,
825
+ "rewards/rejected": -6.992558002471924,
826
+ "step": 2550
827
+ },
828
+ {
829
+ "epoch": 2.08,
830
+ "grad_norm": 5.904083251953125,
831
+ "learning_rate": 2.6055191955342886e-06,
832
+ "logits/chosen": 1.5594576597213745,
833
+ "logits/rejected": 1.6052813529968262,
834
+ "logps/chosen": -39.600093841552734,
835
+ "logps/rejected": -73.07125854492188,
836
+ "loss": 0.398,
837
+ "nll_loss": 0.23794788122177124,
838
+ "rewards/accuracies": 0.9662500023841858,
839
+ "rewards/chosen": -3.960009813308716,
840
+ "rewards/margins": 3.3471157550811768,
841
+ "rewards/rejected": -7.307126045227051,
842
+ "step": 2600
843
+ },
844
+ {
845
+ "epoch": 2.12,
846
+ "grad_norm": 5.723136901855469,
847
+ "learning_rate": 2.403895910817593e-06,
848
+ "logits/chosen": 1.532933235168457,
849
+ "logits/rejected": 1.5803064107894897,
850
+ "logps/chosen": -38.91565704345703,
851
+ "logps/rejected": -70.781982421875,
852
+ "loss": 0.4101,
853
+ "nll_loss": 0.23784016072750092,
854
+ "rewards/accuracies": 0.9574999809265137,
855
+ "rewards/chosen": -3.8915653228759766,
856
+ "rewards/margins": 3.186633348464966,
857
+ "rewards/rejected": -7.07819938659668,
858
+ "step": 2650
859
+ },
860
+ {
861
+ "epoch": 2.16,
862
+ "grad_norm": 7.6983561515808105,
863
+ "learning_rate": 2.2078952094652705e-06,
864
+ "logits/chosen": 1.4740697145462036,
865
+ "logits/rejected": 1.5456255674362183,
866
+ "logps/chosen": -39.75840377807617,
867
+ "logps/rejected": -75.84871673583984,
868
+ "loss": 0.3908,
869
+ "nll_loss": 0.23151804506778717,
870
+ "rewards/accuracies": 0.9612500071525574,
871
+ "rewards/chosen": -3.9758400917053223,
872
+ "rewards/margins": 3.6090316772460938,
873
+ "rewards/rejected": -7.584871292114258,
874
+ "step": 2700
875
+ },
876
+ {
877
+ "epoch": 2.2,
878
+ "grad_norm": 7.529278755187988,
879
+ "learning_rate": 2.017941585351591e-06,
880
+ "logits/chosen": 1.495976209640503,
881
+ "logits/rejected": 1.5389071702957153,
882
+ "logps/chosen": -39.2768440246582,
883
+ "logps/rejected": -71.7361831665039,
884
+ "loss": 0.4138,
885
+ "nll_loss": 0.24059143662452698,
886
+ "rewards/accuracies": 0.9574999809265137,
887
+ "rewards/chosen": -3.9276845455169678,
888
+ "rewards/margins": 3.245933771133423,
889
+ "rewards/rejected": -7.173618793487549,
890
+ "step": 2750
891
+ },
892
+ {
893
+ "epoch": 2.24,
894
+ "grad_norm": 6.076731204986572,
895
+ "learning_rate": 1.8344464357280722e-06,
896
+ "logits/chosen": 1.4751191139221191,
897
+ "logits/rejected": 1.5268608331680298,
898
+ "logps/chosen": -38.4973258972168,
899
+ "logps/rejected": -71.8733901977539,
900
+ "loss": 0.3943,
901
+ "nll_loss": 0.22699041664600372,
902
+ "rewards/accuracies": 0.9700000286102295,
903
+ "rewards/chosen": -3.8497328758239746,
904
+ "rewards/margins": 3.3376071453094482,
905
+ "rewards/rejected": -7.187338829040527,
906
+ "step": 2800
907
+ },
908
+ {
909
+ "epoch": 2.2800000000000002,
910
+ "grad_norm": 5.921169757843018,
911
+ "learning_rate": 1.6578071702286396e-06,
912
+ "logits/chosen": 1.4693129062652588,
913
+ "logits/rejected": 1.5106064081192017,
914
+ "logps/chosen": -38.8640022277832,
915
+ "logps/rejected": -72.8025131225586,
916
+ "loss": 0.4103,
917
+ "nll_loss": 0.2377665936946869,
918
+ "rewards/accuracies": 0.949999988079071,
919
+ "rewards/chosen": -3.8864006996154785,
920
+ "rewards/margins": 3.393850088119507,
921
+ "rewards/rejected": -7.280250072479248,
922
+ "step": 2850
923
+ },
924
+ {
925
+ "epoch": 2.32,
926
+ "grad_norm": 5.996314525604248,
927
+ "learning_rate": 1.4884063501688539e-06,
928
+ "logits/chosen": 1.4711755514144897,
929
+ "logits/rejected": 1.537732720375061,
930
+ "logps/chosen": -38.81364440917969,
931
+ "logps/rejected": -73.80380249023438,
932
+ "loss": 0.3908,
933
+ "nll_loss": 0.22893303632736206,
934
+ "rewards/accuracies": 0.9737499952316284,
935
+ "rewards/chosen": -3.881364345550537,
936
+ "rewards/margins": 3.499016523361206,
937
+ "rewards/rejected": -7.380380153656006,
938
+ "step": 2900
939
+ },
940
+ {
941
+ "epoch": 2.36,
942
+ "grad_norm": 10.235599517822266,
943
+ "learning_rate": 1.3266108600032928e-06,
944
+ "logits/chosen": 1.469245195388794,
945
+ "logits/rejected": 1.5233672857284546,
946
+ "logps/chosen": -39.08128356933594,
947
+ "logps/rejected": -70.8739242553711,
948
+ "loss": 0.4124,
949
+ "nll_loss": 0.2312445044517517,
950
+ "rewards/accuracies": 0.9512500166893005,
951
+ "rewards/chosen": -3.9081289768218994,
952
+ "rewards/margins": 3.1792635917663574,
953
+ "rewards/rejected": -7.087392807006836,
954
+ "step": 2950
955
+ },
956
+ {
957
+ "epoch": 2.4,
958
+ "grad_norm": 8.700451850891113,
959
+ "learning_rate": 1.1727711127355118e-06,
960
+ "logits/chosen": 1.445876955986023,
961
+ "logits/rejected": 1.494391679763794,
962
+ "logps/chosen": -38.116580963134766,
963
+ "logps/rejected": -70.0914535522461,
964
+ "loss": 0.3982,
965
+ "nll_loss": 0.22633756697177887,
966
+ "rewards/accuracies": 0.9574999809265137,
967
+ "rewards/chosen": -3.8116586208343506,
968
+ "rewards/margins": 3.1974875926971436,
969
+ "rewards/rejected": -7.009146213531494,
970
+ "step": 3000
971
+ },
972
+ {
973
+ "epoch": 2.44,
974
+ "grad_norm": 9.24150562286377,
975
+ "learning_rate": 1.0272202910015083e-06,
976
+ "logits/chosen": 1.4509037733078003,
977
+ "logits/rejected": 1.5117911100387573,
978
+ "logps/chosen": -39.70808410644531,
979
+ "logps/rejected": -74.1922836303711,
980
+ "loss": 0.4038,
981
+ "nll_loss": 0.23606754839420319,
982
+ "rewards/accuracies": 0.9662500023841858,
983
+ "rewards/chosen": -3.970808744430542,
984
+ "rewards/margins": 3.4484190940856934,
985
+ "rewards/rejected": -7.419227123260498,
986
+ "step": 3050
987
+ },
988
+ {
989
+ "epoch": 2.48,
990
+ "grad_norm": 10.763381958007812,
991
+ "learning_rate": 8.902736254703347e-07,
992
+ "logits/chosen": 1.4675711393356323,
993
+ "logits/rejected": 1.5065741539001465,
994
+ "logps/chosen": -36.786865234375,
995
+ "logps/rejected": -67.48006439208984,
996
+ "loss": 0.405,
997
+ "nll_loss": 0.22343981266021729,
998
+ "rewards/accuracies": 0.9624999761581421,
999
+ "rewards/chosen": -3.6786866188049316,
1000
+ "rewards/margins": 3.0693204402923584,
1001
+ "rewards/rejected": -6.748007297515869,
1002
+ "step": 3100
1003
+ },
1004
+ {
1005
+ "epoch": 2.52,
1006
+ "grad_norm": 5.758388996124268,
1007
+ "learning_rate": 7.622277121246513e-07,
1008
+ "logits/chosen": 1.483445405960083,
1009
+ "logits/rejected": 1.5389055013656616,
1010
+ "logps/chosen": -38.62077331542969,
1011
+ "logps/rejected": -71.79267883300781,
1012
+ "loss": 0.4121,
1013
+ "nll_loss": 0.23168590664863586,
1014
+ "rewards/accuracies": 0.9587500095367432,
1015
+ "rewards/chosen": -3.8620777130126953,
1016
+ "rewards/margins": 3.3171918392181396,
1017
+ "rewards/rejected": -7.179268836975098,
1018
+ "step": 3150
1019
+ },
1020
+ {
1021
+ "epoch": 2.56,
1022
+ "grad_norm": 7.307446002960205,
1023
+ "learning_rate": 6.433598698998766e-07,
1024
+ "logits/chosen": 1.4688174724578857,
1025
+ "logits/rejected": 1.5306994915008545,
1026
+ "logps/chosen": -39.013912200927734,
1027
+ "logps/rejected": -71.06527709960938,
1028
+ "loss": 0.416,
1029
+ "nll_loss": 0.23500658571720123,
1030
+ "rewards/accuracies": 0.9424999952316284,
1031
+ "rewards/chosen": -3.9013912677764893,
1032
+ "rewards/margins": 3.2051360607147217,
1033
+ "rewards/rejected": -7.106527328491211,
1034
+ "step": 3200
1035
+ },
1036
+ {
1037
+ "epoch": 2.6,
1038
+ "grad_norm": 6.810940742492676,
1039
+ "learning_rate": 5.339275400731331e-07,
1040
+ "logits/chosen": 1.4731531143188477,
1041
+ "logits/rejected": 1.5175625085830688,
1042
+ "logps/chosen": -39.05414581298828,
1043
+ "logps/rejected": -69.38569641113281,
1044
+ "loss": 0.4207,
1045
+ "nll_loss": 0.2375660389661789,
1046
+ "rewards/accuracies": 0.949999988079071,
1047
+ "rewards/chosen": -3.9054150581359863,
1048
+ "rewards/margins": 3.0331552028656006,
1049
+ "rewards/rejected": -6.93856954574585,
1050
+ "step": 3250
1051
+ },
1052
+ {
1053
+ "epoch": 2.64,
1054
+ "grad_norm": 5.935739517211914,
1055
+ "learning_rate": 4.3416772870275295e-07,
1056
+ "logits/chosen": 1.447837233543396,
1057
+ "logits/rejected": 1.5085737705230713,
1058
+ "logps/chosen": -38.072574615478516,
1059
+ "logps/rejected": -71.93220520019531,
1060
+ "loss": 0.3981,
1061
+ "nll_loss": 0.22540073096752167,
1062
+ "rewards/accuracies": 0.9537500143051147,
1063
+ "rewards/chosen": -3.807257652282715,
1064
+ "rewards/margins": 3.385963201522827,
1065
+ "rewards/rejected": -7.193220138549805,
1066
+ "step": 3300
1067
+ },
1068
+ {
1069
+ "epoch": 2.68,
1070
+ "grad_norm": 6.746799468994141,
1071
+ "learning_rate": 3.442964933259474e-07,
1072
+ "logits/chosen": 1.4588629007339478,
1073
+ "logits/rejected": 1.517850637435913,
1074
+ "logps/chosen": -38.37602615356445,
1075
+ "logps/rejected": -72.57811737060547,
1076
+ "loss": 0.3998,
1077
+ "nll_loss": 0.23044270277023315,
1078
+ "rewards/accuracies": 0.9587500095367432,
1079
+ "rewards/chosen": -3.8376026153564453,
1080
+ "rewards/margins": 3.4202094078063965,
1081
+ "rewards/rejected": -7.257812023162842,
1082
+ "step": 3350
1083
+ },
1084
+ {
1085
+ "epoch": 2.7199999999999998,
1086
+ "grad_norm": 7.958431243896484,
1087
+ "learning_rate": 2.6450847502627883e-07,
1088
+ "logits/chosen": 1.4416861534118652,
1089
+ "logits/rejected": 1.5023283958435059,
1090
+ "logps/chosen": -39.2740364074707,
1091
+ "logps/rejected": -73.73572540283203,
1092
+ "loss": 0.4036,
1093
+ "nll_loss": 0.22917690873146057,
1094
+ "rewards/accuracies": 0.9662500023841858,
1095
+ "rewards/chosen": -3.927403450012207,
1096
+ "rewards/margins": 3.446169376373291,
1097
+ "rewards/rejected": -7.373574256896973,
1098
+ "step": 3400
1099
+ },
1100
+ {
1101
+ "epoch": 2.76,
1102
+ "grad_norm": 8.157621383666992,
1103
+ "learning_rate": 1.9497647688442478e-07,
1104
+ "logits/chosen": 1.4702584743499756,
1105
+ "logits/rejected": 1.5254310369491577,
1106
+ "logps/chosen": -38.0818977355957,
1107
+ "logps/rejected": -70.6051254272461,
1108
+ "loss": 0.4055,
1109
+ "nll_loss": 0.23146916925907135,
1110
+ "rewards/accuracies": 0.9549999833106995,
1111
+ "rewards/chosen": -3.808190107345581,
1112
+ "rewards/margins": 3.252322196960449,
1113
+ "rewards/rejected": -7.060512065887451,
1114
+ "step": 3450
1115
+ },
1116
+ {
1117
+ "epoch": 2.8,
1118
+ "grad_norm": 6.671606540679932,
1119
+ "learning_rate": 1.358510897251808e-07,
1120
+ "logits/chosen": 1.4628223180770874,
1121
+ "logits/rejected": 1.525000810623169,
1122
+ "logps/chosen": -37.097007751464844,
1123
+ "logps/rejected": -72.08160400390625,
1124
+ "loss": 0.3906,
1125
+ "nll_loss": 0.22892680764198303,
1126
+ "rewards/accuracies": 0.9700000286102295,
1127
+ "rewards/chosen": -3.7097012996673584,
1128
+ "rewards/margins": 3.498459577560425,
1129
+ "rewards/rejected": -7.208160400390625,
1130
+ "step": 3500
1131
+ },
1132
+ {
1133
+ "epoch": 2.84,
1134
+ "grad_norm": 8.076554298400879,
1135
+ "learning_rate": 8.726036597126619e-08,
1136
+ "logits/chosen": 1.470629096031189,
1137
+ "logits/rejected": 1.535922646522522,
1138
+ "logps/chosen": -36.950157165527344,
1139
+ "logps/rejected": -68.89533996582031,
1140
+ "loss": 0.395,
1141
+ "nll_loss": 0.22603529691696167,
1142
+ "rewards/accuracies": 0.9549999833106995,
1143
+ "rewards/chosen": -3.6950161457061768,
1144
+ "rewards/margins": 3.1945183277130127,
1145
+ "rewards/rejected": -6.8895344734191895,
1146
+ "step": 3550
1147
+ },
1148
+ {
1149
+ "epoch": 2.88,
1150
+ "grad_norm": 6.550802707672119,
1151
+ "learning_rate": 4.93095423102935e-08,
1152
+ "logits/chosen": 1.4719560146331787,
1153
+ "logits/rejected": 1.5248959064483643,
1154
+ "logps/chosen": -40.710941314697266,
1155
+ "logps/rejected": -74.42230987548828,
1156
+ "loss": 0.4132,
1157
+ "nll_loss": 0.24109821021556854,
1158
+ "rewards/accuracies": 0.9524999856948853,
1159
+ "rewards/chosen": -4.071094512939453,
1160
+ "rewards/margins": 3.371136426925659,
1161
+ "rewards/rejected": -7.442230701446533,
1162
+ "step": 3600
1163
+ },
1164
+ {
1165
+ "epoch": 2.92,
1166
+ "grad_norm": 6.743546485900879,
1167
+ "learning_rate": 2.2080811775535006e-08,
1168
+ "logits/chosen": 1.4744410514831543,
1169
+ "logits/rejected": 1.5339540243148804,
1170
+ "logps/chosen": -39.543033599853516,
1171
+ "logps/rejected": -74.6586685180664,
1172
+ "loss": 0.3964,
1173
+ "nll_loss": 0.23712339997291565,
1174
+ "rewards/accuracies": 0.9674999713897705,
1175
+ "rewards/chosen": -3.954303026199341,
1176
+ "rewards/margins": 3.511564254760742,
1177
+ "rewards/rejected": -7.46586799621582,
1178
+ "step": 3650
1179
+ },
1180
+ {
1181
+ "epoch": 2.96,
1182
+ "grad_norm": 11.845396041870117,
1183
+ "learning_rate": 5.633145734114665e-09,
1184
+ "logits/chosen": 1.4899775981903076,
1185
+ "logits/rejected": 1.533370018005371,
1186
+ "logps/chosen": -38.53838348388672,
1187
+ "logps/rejected": -71.89134979248047,
1188
+ "loss": 0.4058,
1189
+ "nll_loss": 0.23659667372703552,
1190
+ "rewards/accuracies": 0.9674999713897705,
1191
+ "rewards/chosen": -3.853839159011841,
1192
+ "rewards/margins": 3.3352959156036377,
1193
+ "rewards/rejected": -7.18913459777832,
1194
+ "step": 3700
1195
+ },
1196
+ {
1197
+ "epoch": 3.0,
1198
+ "grad_norm": 9.659139633178711,
1199
+ "learning_rate": 2.1661681620654963e-12,
1200
+ "logits/chosen": 1.4696215391159058,
1201
+ "logits/rejected": 1.526814579963684,
1202
+ "logps/chosen": -38.22508239746094,
1203
+ "logps/rejected": -72.54460144042969,
1204
+ "loss": 0.3938,
1205
+ "nll_loss": 0.22828102111816406,
1206
+ "rewards/accuracies": 0.9637500047683716,
1207
+ "rewards/chosen": -3.8225083351135254,
1208
+ "rewards/margins": 3.4319519996643066,
1209
+ "rewards/rejected": -7.254461288452148,
1210
+ "step": 3750
1211
+ }
1212
+ ],
1213
+ "logging_steps": 50,
1214
+ "max_steps": 3750,
1215
+ "num_input_tokens_seen": 0,
1216
+ "num_train_epochs": 3,
1217
+ "save_steps": 200,
1218
+ "stateful_callbacks": {
1219
+ "TrainerControl": {
1220
+ "args": {
1221
+ "should_epoch_stop": false,
1222
+ "should_evaluate": false,
1223
+ "should_log": false,
1224
+ "should_save": true,
1225
+ "should_training_stop": true
1226
+ },
1227
+ "attributes": {}
1228
+ }
1229
+ },
1230
+ "total_flos": 0.0,
1231
+ "train_batch_size": 2,
1232
+ "trial_name": null,
1233
+ "trial_params": null
1234
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fdc46d7f4e510fac8857457512e9b21ca1bcc57fcd0eb82dff37e7aa86a9dc68
3
+ size 6289
vocab.json ADDED
The diff for this file is too large to render. See raw diff