gulaschnascher4000 commited on
Commit
fd37cb7
·
verified ·
1 Parent(s): 9ec8bec

End of training

Browse files
README.md CHANGED
@@ -4,6 +4,7 @@ license: llama3.2
4
  base_model: meta-llama/Llama-3.2-3B
5
  tags:
6
  - llama-factory
 
7
  - generated_from_trainer
8
  model-index:
9
  - name: lora_0-2_3B
@@ -15,7 +16,9 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  # lora_0-2_3B
17
 
18
- This model is a fine-tuned version of [meta-llama/Llama-3.2-3B](https://huggingface.co/meta-llama/Llama-3.2-3B) on the None dataset.
 
 
19
 
20
  ## Model description
21
 
 
4
  base_model: meta-llama/Llama-3.2-3B
5
  tags:
6
  - llama-factory
7
+ - lora
8
  - generated_from_trainer
9
  model-index:
10
  - name: lora_0-2_3B
 
16
 
17
  # lora_0-2_3B
18
 
19
+ This model is a fine-tuned version of [meta-llama/Llama-3.2-3B](https://huggingface.co/meta-llama/Llama-3.2-3B) on the gulaschnascher4000/stream-dataset-0-2, the identity-chatgulaschpt, the dolly_15k_de and the alpaca-gpt4_de datasets.
20
+ It achieves the following results on the evaluation set:
21
+ - Loss: 1.5461
22
 
23
  ## Model description
24
 
all_results.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 0.10006807351940095,
3
+ "eval_loss": 1.5460742712020874,
4
+ "eval_runtime": 135.5079,
5
+ "eval_samples_per_second": 77.066,
6
+ "eval_steps_per_second": 19.268,
7
+ "total_flos": 8.37378777242665e+16,
8
+ "train_loss": 1.3456703529876917,
9
+ "train_runtime": 360.87,
10
+ "train_samples_per_second": 26.045,
11
+ "train_steps_per_second": 0.815
12
+ }
eval_results.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 0.10006807351940095,
3
+ "eval_loss": 1.5460742712020874,
4
+ "eval_runtime": 135.5079,
5
+ "eval_samples_per_second": 77.066,
6
+ "eval_steps_per_second": 19.268
7
+ }
runs/Jan14_01-43-47_pss5xry2f0aq/events.out.tfevents.1736819545.pss5xry2f0aq.48508.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7955d9b0bde15152a3d7eafd961a5b568ae441898cd860c4cc32fdc475e847e9
3
+ size 359
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 0.10006807351940095,
3
+ "total_flos": 8.37378777242665e+16,
4
+ "train_loss": 1.3456703529876917,
5
+ "train_runtime": 360.87,
6
+ "train_samples_per_second": 26.045,
7
+ "train_steps_per_second": 0.815
8
+ }
trainer_state.json ADDED
@@ -0,0 +1,245 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 0.10006807351940095,
5
+ "eval_steps": 500,
6
+ "global_step": 294,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.0034036759700476512,
13
+ "grad_norm": 0.3913409113883972,
14
+ "learning_rate": 3.3333333333333335e-05,
15
+ "loss": 1.517,
16
+ "step": 10
17
+ },
18
+ {
19
+ "epoch": 0.0068073519400953025,
20
+ "grad_norm": 0.42031803727149963,
21
+ "learning_rate": 6.666666666666667e-05,
22
+ "loss": 1.4607,
23
+ "step": 20
24
+ },
25
+ {
26
+ "epoch": 0.010211027910142955,
27
+ "grad_norm": 0.33206892013549805,
28
+ "learning_rate": 0.0001,
29
+ "loss": 1.3907,
30
+ "step": 30
31
+ },
32
+ {
33
+ "epoch": 0.013614703880190605,
34
+ "grad_norm": 0.38453030586242676,
35
+ "learning_rate": 9.964639423366442e-05,
36
+ "loss": 1.3338,
37
+ "step": 40
38
+ },
39
+ {
40
+ "epoch": 0.01701837985023826,
41
+ "grad_norm": 0.5517820715904236,
42
+ "learning_rate": 9.859057841617709e-05,
43
+ "loss": 1.3894,
44
+ "step": 50
45
+ },
46
+ {
47
+ "epoch": 0.02042205582028591,
48
+ "grad_norm": 0.3676927983760834,
49
+ "learning_rate": 9.68474862499881e-05,
50
+ "loss": 1.3543,
51
+ "step": 60
52
+ },
53
+ {
54
+ "epoch": 0.023825731790333562,
55
+ "grad_norm": 1.1669039726257324,
56
+ "learning_rate": 9.444177243274618e-05,
57
+ "loss": 1.3556,
58
+ "step": 70
59
+ },
60
+ {
61
+ "epoch": 0.02722940776038121,
62
+ "grad_norm": 0.4291529953479767,
63
+ "learning_rate": 9.140746393556854e-05,
64
+ "loss": 1.3505,
65
+ "step": 80
66
+ },
67
+ {
68
+ "epoch": 0.03063308373042886,
69
+ "grad_norm": 0.6486052870750427,
70
+ "learning_rate": 8.778747871771292e-05,
71
+ "loss": 1.3473,
72
+ "step": 90
73
+ },
74
+ {
75
+ "epoch": 0.03403675970047652,
76
+ "grad_norm": 0.42231065034866333,
77
+ "learning_rate": 8.363301868506264e-05,
78
+ "loss": 1.3506,
79
+ "step": 100
80
+ },
81
+ {
82
+ "epoch": 0.037440435670524165,
83
+ "grad_norm": 0.5026513934135437,
84
+ "learning_rate": 7.900284547855991e-05,
85
+ "loss": 1.3029,
86
+ "step": 110
87
+ },
88
+ {
89
+ "epoch": 0.04084411164057182,
90
+ "grad_norm": 0.6771671175956726,
91
+ "learning_rate": 7.396244933600285e-05,
92
+ "loss": 1.3336,
93
+ "step": 120
94
+ },
95
+ {
96
+ "epoch": 0.04424778761061947,
97
+ "grad_norm": 0.5867034792900085,
98
+ "learning_rate": 6.858312278301637e-05,
99
+ "loss": 1.3177,
100
+ "step": 130
101
+ },
102
+ {
103
+ "epoch": 0.047651463580667124,
104
+ "grad_norm": 0.5451478958129883,
105
+ "learning_rate": 6.294095225512603e-05,
106
+ "loss": 1.2924,
107
+ "step": 140
108
+ },
109
+ {
110
+ "epoch": 0.05105513955071477,
111
+ "grad_norm": 0.7218224406242371,
112
+ "learning_rate": 5.7115741913664264e-05,
113
+ "loss": 1.3566,
114
+ "step": 150
115
+ },
116
+ {
117
+ "epoch": 0.05445881552076242,
118
+ "grad_norm": 0.5916715264320374,
119
+ "learning_rate": 5.1189884877305375e-05,
120
+ "loss": 1.3093,
121
+ "step": 160
122
+ },
123
+ {
124
+ "epoch": 0.057862491490810075,
125
+ "grad_norm": 0.45692551136016846,
126
+ "learning_rate": 4.5247197834790876e-05,
127
+ "loss": 1.3207,
128
+ "step": 170
129
+ },
130
+ {
131
+ "epoch": 0.06126616746085772,
132
+ "grad_norm": 0.840484082698822,
133
+ "learning_rate": 3.937173552235117e-05,
134
+ "loss": 1.3437,
135
+ "step": 180
136
+ },
137
+ {
138
+ "epoch": 0.06466984343090537,
139
+ "grad_norm": 0.6113607883453369,
140
+ "learning_rate": 3.364660183412892e-05,
141
+ "loss": 1.3109,
142
+ "step": 190
143
+ },
144
+ {
145
+ "epoch": 0.06807351940095303,
146
+ "grad_norm": 0.5302855968475342,
147
+ "learning_rate": 2.8152774381532033e-05,
148
+ "loss": 1.308,
149
+ "step": 200
150
+ },
151
+ {
152
+ "epoch": 0.07147719537100068,
153
+ "grad_norm": 0.5346827507019043,
154
+ "learning_rate": 2.296795912722014e-05,
155
+ "loss": 1.331,
156
+ "step": 210
157
+ },
158
+ {
159
+ "epoch": 0.07488087134104833,
160
+ "grad_norm": 0.6379411816596985,
161
+ "learning_rate": 1.8165491294045593e-05,
162
+ "loss": 1.33,
163
+ "step": 220
164
+ },
165
+ {
166
+ "epoch": 0.07828454731109598,
167
+ "grad_norm": 0.4734562635421753,
168
+ "learning_rate": 1.3813298094746491e-05,
169
+ "loss": 1.3551,
170
+ "step": 230
171
+ },
172
+ {
173
+ "epoch": 0.08168822328114364,
174
+ "grad_norm": 0.680587887763977,
175
+ "learning_rate": 9.972937953781986e-06,
176
+ "loss": 1.3258,
177
+ "step": 240
178
+ },
179
+ {
180
+ "epoch": 0.08509189925119129,
181
+ "grad_norm": 0.5624794960021973,
182
+ "learning_rate": 6.698729810778065e-06,
183
+ "loss": 1.3268,
184
+ "step": 250
185
+ },
186
+ {
187
+ "epoch": 0.08849557522123894,
188
+ "grad_norm": 0.5939269065856934,
189
+ "learning_rate": 4.036984820916723e-06,
190
+ "loss": 1.3556,
191
+ "step": 260
192
+ },
193
+ {
194
+ "epoch": 0.09189925119128659,
195
+ "grad_norm": 0.6130989193916321,
196
+ "learning_rate": 2.0253513192751373e-06,
197
+ "loss": 1.2661,
198
+ "step": 270
199
+ },
200
+ {
201
+ "epoch": 0.09530292716133425,
202
+ "grad_norm": 0.8308888673782349,
203
+ "learning_rate": 6.922823140906753e-07,
204
+ "loss": 1.3755,
205
+ "step": 280
206
+ },
207
+ {
208
+ "epoch": 0.0987066031313819,
209
+ "grad_norm": 0.8799161911010742,
210
+ "learning_rate": 5.663304084960186e-08,
211
+ "loss": 1.3005,
212
+ "step": 290
213
+ },
214
+ {
215
+ "epoch": 0.10006807351940095,
216
+ "step": 294,
217
+ "total_flos": 8.37378777242665e+16,
218
+ "train_loss": 1.3456703529876917,
219
+ "train_runtime": 360.87,
220
+ "train_samples_per_second": 26.045,
221
+ "train_steps_per_second": 0.815
222
+ }
223
+ ],
224
+ "logging_steps": 10,
225
+ "max_steps": 294,
226
+ "num_input_tokens_seen": 0,
227
+ "num_train_epochs": 1,
228
+ "save_steps": 50000,
229
+ "stateful_callbacks": {
230
+ "TrainerControl": {
231
+ "args": {
232
+ "should_epoch_stop": false,
233
+ "should_evaluate": false,
234
+ "should_log": false,
235
+ "should_save": true,
236
+ "should_training_stop": true
237
+ },
238
+ "attributes": {}
239
+ }
240
+ },
241
+ "total_flos": 8.37378777242665e+16,
242
+ "train_batch_size": 8,
243
+ "trial_name": null,
244
+ "trial_params": null
245
+ }
training_loss.png ADDED