3N3G commited on
Commit
193adc7
·
verified ·
1 Parent(s): da084b9

End of training

Browse files
README.md CHANGED
@@ -4,6 +4,7 @@ license: apache-2.0
4
  base_model: Qwen/Qwen3-4B
5
  tags:
6
  - llama-factory
 
7
  - generated_from_trainer
8
  model-index:
9
  - name: qwen3-4b-sft
@@ -15,9 +16,9 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  # qwen3-4b-sft
17
 
18
- This model is a fine-tuned version of [Qwen/Qwen3-4B](https://huggingface.co/Qwen/Qwen3-4B) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 0.6358
21
 
22
  ## Model description
23
 
 
4
  base_model: Qwen/Qwen3-4B
5
  tags:
6
  - llama-factory
7
+ - full
8
  - generated_from_trainer
9
  model-index:
10
  - name: qwen3-4b-sft
 
16
 
17
  # qwen3-4b-sft
18
 
19
+ This model is a fine-tuned version of [Qwen/Qwen3-4B](https://huggingface.co/Qwen/Qwen3-4B) on the qwen3-eft-data-new_train dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.6345
22
 
23
  ## Model description
24
 
all_results.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 4.0,
3
+ "eval_loss": 0.6345427632331848,
4
+ "eval_runtime": 3.9427,
5
+ "eval_samples_per_second": 12.682,
6
+ "eval_steps_per_second": 3.297,
7
+ "total_flos": 5574381731840.0,
8
+ "train_loss": 0.6609239322798592,
9
+ "train_runtime": 622.3863,
10
+ "train_samples_per_second": 2.86,
11
+ "train_steps_per_second": 0.045
12
+ }
eval_results.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 4.0,
3
+ "eval_loss": 0.6345427632331848,
4
+ "eval_runtime": 3.9427,
5
+ "eval_samples_per_second": 12.682,
6
+ "eval_steps_per_second": 3.297
7
+ }
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 4.0,
3
+ "total_flos": 5574381731840.0,
4
+ "train_loss": 0.6609239322798592,
5
+ "train_runtime": 622.3863,
6
+ "train_samples_per_second": 2.86,
7
+ "train_steps_per_second": 0.045
8
+ }
trainer_state.json ADDED
@@ -0,0 +1,247 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 4.0,
6
+ "eval_steps": 16,
7
+ "global_step": 28,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.14285714285714285,
14
+ "grad_norm": 5.135755695562593,
15
+ "learning_rate": 0.0,
16
+ "loss": 0.6626,
17
+ "step": 1
18
+ },
19
+ {
20
+ "epoch": 0.2857142857142857,
21
+ "grad_norm": 4.976422377741924,
22
+ "learning_rate": 6.666666666666665e-08,
23
+ "loss": 0.6659,
24
+ "step": 2
25
+ },
26
+ {
27
+ "epoch": 0.42857142857142855,
28
+ "grad_norm": 5.149566448288666,
29
+ "learning_rate": 1.333333333333333e-07,
30
+ "loss": 0.6934,
31
+ "step": 3
32
+ },
33
+ {
34
+ "epoch": 0.5714285714285714,
35
+ "grad_norm": 5.168205312692533,
36
+ "learning_rate": 2e-07,
37
+ "loss": 0.642,
38
+ "step": 4
39
+ },
40
+ {
41
+ "epoch": 0.7142857142857143,
42
+ "grad_norm": 5.219987568955151,
43
+ "learning_rate": 1.9929032311830301e-07,
44
+ "loss": 0.6757,
45
+ "step": 5
46
+ },
47
+ {
48
+ "epoch": 0.8571428571428571,
49
+ "grad_norm": 5.316918732370043,
50
+ "learning_rate": 1.971724845015768e-07,
51
+ "loss": 0.6625,
52
+ "step": 6
53
+ },
54
+ {
55
+ "epoch": 1.0,
56
+ "grad_norm": 4.9551247999375665,
57
+ "learning_rate": 1.9367988372994263e-07,
58
+ "loss": 0.6482,
59
+ "step": 7
60
+ },
61
+ {
62
+ "epoch": 1.1428571428571428,
63
+ "grad_norm": 5.346578649855734,
64
+ "learning_rate": 1.8886760120394772e-07,
65
+ "loss": 0.6805,
66
+ "step": 8
67
+ },
68
+ {
69
+ "epoch": 1.2857142857142856,
70
+ "grad_norm": 4.598120192548456,
71
+ "learning_rate": 1.8281152949374524e-07,
72
+ "loss": 0.635,
73
+ "step": 9
74
+ },
75
+ {
76
+ "epoch": 1.4285714285714286,
77
+ "grad_norm": 5.1151113017774374,
78
+ "learning_rate": 1.7560717646792702e-07,
79
+ "loss": 0.6743,
80
+ "step": 10
81
+ },
82
+ {
83
+ "epoch": 1.5714285714285714,
84
+ "grad_norm": 5.35558546306821,
85
+ "learning_rate": 1.6736815907738206e-07,
86
+ "loss": 0.6828,
87
+ "step": 11
88
+ },
89
+ {
90
+ "epoch": 1.7142857142857144,
91
+ "grad_norm": 5.2437631052769005,
92
+ "learning_rate": 1.5822441154810968e-07,
93
+ "loss": 0.6758,
94
+ "step": 12
95
+ },
96
+ {
97
+ "epoch": 1.8571428571428572,
98
+ "grad_norm": 4.7959856858897005,
99
+ "learning_rate": 1.4832013624085653e-07,
100
+ "loss": 0.643,
101
+ "step": 13
102
+ },
103
+ {
104
+ "epoch": 2.0,
105
+ "grad_norm": 5.146931489368684,
106
+ "learning_rate": 1.3781152949374525e-07,
107
+ "loss": 0.6467,
108
+ "step": 14
109
+ },
110
+ {
111
+ "epoch": 2.142857142857143,
112
+ "grad_norm": 5.565037540363902,
113
+ "learning_rate": 1.2686431831271523e-07,
114
+ "loss": 0.7021,
115
+ "step": 15
116
+ },
117
+ {
118
+ "epoch": 2.2857142857142856,
119
+ "grad_norm": 4.926934066030776,
120
+ "learning_rate": 1.1565114675763821e-07,
121
+ "loss": 0.6558,
122
+ "step": 16
123
+ },
124
+ {
125
+ "epoch": 2.2857142857142856,
126
+ "eval_loss": 0.6358002424240112,
127
+ "eval_runtime": 4.7554,
128
+ "eval_samples_per_second": 10.514,
129
+ "eval_steps_per_second": 2.734,
130
+ "step": 16
131
+ },
132
+ {
133
+ "epoch": 2.4285714285714284,
134
+ "grad_norm": 4.957804203750453,
135
+ "learning_rate": 1.043488532423618e-07,
136
+ "loss": 0.6317,
137
+ "step": 17
138
+ },
139
+ {
140
+ "epoch": 2.571428571428571,
141
+ "grad_norm": 4.89212984754668,
142
+ "learning_rate": 9.313568168728476e-08,
143
+ "loss": 0.669,
144
+ "step": 18
145
+ },
146
+ {
147
+ "epoch": 2.7142857142857144,
148
+ "grad_norm": 4.63661595623532,
149
+ "learning_rate": 8.218847050625475e-08,
150
+ "loss": 0.6435,
151
+ "step": 19
152
+ },
153
+ {
154
+ "epoch": 2.857142857142857,
155
+ "grad_norm": 5.237528783827469,
156
+ "learning_rate": 7.167986375914346e-08,
157
+ "loss": 0.6754,
158
+ "step": 20
159
+ },
160
+ {
161
+ "epoch": 3.0,
162
+ "grad_norm": 4.835537673366031,
163
+ "learning_rate": 6.177558845189028e-08,
164
+ "loss": 0.634,
165
+ "step": 21
166
+ },
167
+ {
168
+ "epoch": 3.142857142857143,
169
+ "grad_norm": 4.9632121977212025,
170
+ "learning_rate": 5.263184092261792e-08,
171
+ "loss": 0.6763,
172
+ "step": 22
173
+ },
174
+ {
175
+ "epoch": 3.2857142857142856,
176
+ "grad_norm": 5.154090826338096,
177
+ "learning_rate": 4.4392823532072976e-08,
178
+ "loss": 0.6697,
179
+ "step": 23
180
+ },
181
+ {
182
+ "epoch": 3.4285714285714284,
183
+ "grad_norm": 4.626173048013453,
184
+ "learning_rate": 3.7188470506254744e-08,
185
+ "loss": 0.6234,
186
+ "step": 24
187
+ },
188
+ {
189
+ "epoch": 3.571428571428571,
190
+ "grad_norm": 5.147858633287859,
191
+ "learning_rate": 3.1132398796052296e-08,
192
+ "loss": 0.6695,
193
+ "step": 25
194
+ },
195
+ {
196
+ "epoch": 3.7142857142857144,
197
+ "grad_norm": 5.058641331266572,
198
+ "learning_rate": 2.632011627005738e-08,
199
+ "loss": 0.6684,
200
+ "step": 26
201
+ },
202
+ {
203
+ "epoch": 3.857142857142857,
204
+ "grad_norm": 4.852574620795075,
205
+ "learning_rate": 2.2827515498423202e-08,
206
+ "loss": 0.6617,
207
+ "step": 27
208
+ },
209
+ {
210
+ "epoch": 4.0,
211
+ "grad_norm": 5.00072237542112,
212
+ "learning_rate": 2.0709676881697004e-08,
213
+ "loss": 0.6367,
214
+ "step": 28
215
+ },
216
+ {
217
+ "epoch": 4.0,
218
+ "step": 28,
219
+ "total_flos": 5574381731840.0,
220
+ "train_loss": 0.6609239322798592,
221
+ "train_runtime": 622.3863,
222
+ "train_samples_per_second": 2.86,
223
+ "train_steps_per_second": 0.045
224
+ }
225
+ ],
226
+ "logging_steps": 1,
227
+ "max_steps": 28,
228
+ "num_input_tokens_seen": 0,
229
+ "num_train_epochs": 4,
230
+ "save_steps": 14,
231
+ "stateful_callbacks": {
232
+ "TrainerControl": {
233
+ "args": {
234
+ "should_epoch_stop": false,
235
+ "should_evaluate": false,
236
+ "should_log": false,
237
+ "should_save": true,
238
+ "should_training_stop": true
239
+ },
240
+ "attributes": {}
241
+ }
242
+ },
243
+ "total_flos": 5574381731840.0,
244
+ "train_batch_size": 1,
245
+ "trial_name": null,
246
+ "trial_params": null
247
+ }
training_eval_loss.png ADDED
training_loss.png ADDED