Kei5uke commited on
Commit
dd9e05b
·
verified ·
1 Parent(s): 3ebccd9

Training in progress, step 18, checkpoint

Browse files
checkpoint-18/adapter_config.json CHANGED
@@ -23,13 +23,13 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
- "k_proj",
27
- "o_proj",
28
  "v_proj",
29
- "gate_proj",
 
30
  "up_proj",
31
- "down_proj",
32
- "q_proj"
33
  ],
34
  "task_type": "CAUSAL_LM",
35
  "use_dora": false,
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
+ "down_proj",
 
27
  "v_proj",
28
+ "o_proj",
29
+ "q_proj",
30
  "up_proj",
31
+ "gate_proj",
32
+ "k_proj"
33
  ],
34
  "task_type": "CAUSAL_LM",
35
  "use_dora": false,
checkpoint-18/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0c972cd6fab27aea164e71f9a48c6f1af2cc2e0e32c88be7ef2ab94466515133
3
  size 137714904
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4524e9b2c899dc015bf07d93f59e3f5d00aededff20ba91876cf61f8355de1cc
3
  size 137714904
checkpoint-18/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ca92a88267f9552bdc278d73150f4314ff1631a886afe72b3eab9f96318b161b
3
  size 275817586
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b09bd0f42a76a2127e001170084d5e977cf1383c5845d802590a36964934a13c
3
  size 275817586
checkpoint-18/trainer_state.json CHANGED
@@ -10,176 +10,176 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.0006200589055960316,
13
- "grad_norm": 0.599152684211731,
14
  "learning_rate": 4.000000000000001e-06,
15
- "loss": 1.4569,
16
  "step": 1
17
  },
18
  {
19
  "epoch": 0.0012401178111920632,
20
- "grad_norm": 0.41296300292015076,
21
  "learning_rate": 8.000000000000001e-06,
22
- "loss": 1.1609,
23
  "step": 2
24
  },
25
  {
26
  "epoch": 0.0018601767167880949,
27
- "grad_norm": 0.3926275372505188,
28
  "learning_rate": 1.2e-05,
29
- "loss": 0.9938,
30
  "step": 3
31
  },
32
  {
33
  "epoch": 0.0018601767167880949,
34
- "eval_loss": 1.2600003480911255,
35
- "eval_runtime": 49.3684,
36
- "eval_samples_per_second": 2.026,
37
- "eval_steps_per_second": 2.026,
38
  "step": 3
39
  },
40
  {
41
  "epoch": 0.0024802356223841263,
42
- "grad_norm": 0.4559924006462097,
43
  "learning_rate": 1.6000000000000003e-05,
44
- "loss": 1.1897,
45
  "step": 4
46
  },
47
  {
48
  "epoch": 0.0031002945279801583,
49
- "grad_norm": 0.40210390090942383,
50
  "learning_rate": 2e-05,
51
- "loss": 0.9835,
52
  "step": 5
53
  },
54
  {
55
  "epoch": 0.0037203534335761897,
56
- "grad_norm": 0.45071661472320557,
57
  "learning_rate": 1.9200000000000003e-05,
58
- "loss": 1.112,
59
  "step": 6
60
  },
61
  {
62
  "epoch": 0.0037203534335761897,
63
- "eval_loss": 1.2538120746612549,
64
- "eval_runtime": 49.623,
65
- "eval_samples_per_second": 2.015,
66
- "eval_steps_per_second": 2.015,
67
  "step": 6
68
  },
69
  {
70
  "epoch": 0.004340412339172222,
71
- "grad_norm": 0.44490283727645874,
72
  "learning_rate": 1.8400000000000003e-05,
73
- "loss": 1.0809,
74
  "step": 7
75
  },
76
  {
77
  "epoch": 0.004960471244768253,
78
- "grad_norm": 0.6036495566368103,
79
  "learning_rate": 1.76e-05,
80
- "loss": 1.2205,
81
  "step": 8
82
  },
83
  {
84
  "epoch": 0.005580530150364285,
85
- "grad_norm": 0.4536096751689911,
86
  "learning_rate": 1.6800000000000002e-05,
87
- "loss": 1.0715,
88
  "step": 9
89
  },
90
  {
91
  "epoch": 0.005580530150364285,
92
- "eval_loss": 1.2402353286743164,
93
- "eval_runtime": 49.777,
94
- "eval_samples_per_second": 2.009,
95
- "eval_steps_per_second": 2.009,
96
  "step": 9
97
  },
98
  {
99
  "epoch": 0.0062005890559603165,
100
- "grad_norm": 0.5870596170425415,
101
  "learning_rate": 1.6000000000000003e-05,
102
- "loss": 1.3001,
103
  "step": 10
104
  },
105
  {
106
  "epoch": 0.0068206479615563476,
107
- "grad_norm": 0.5543638467788696,
108
  "learning_rate": 1.5200000000000002e-05,
109
- "loss": 1.0459,
110
  "step": 11
111
  },
112
  {
113
  "epoch": 0.0074407068671523795,
114
- "grad_norm": 0.6657857894897461,
115
  "learning_rate": 1.4400000000000001e-05,
116
- "loss": 1.4109,
117
  "step": 12
118
  },
119
  {
120
  "epoch": 0.0074407068671523795,
121
- "eval_loss": 1.2233326435089111,
122
- "eval_runtime": 49.0096,
123
- "eval_samples_per_second": 2.04,
124
- "eval_steps_per_second": 2.04,
125
  "step": 12
126
  },
127
  {
128
  "epoch": 0.008060765772748411,
129
- "grad_norm": 0.5861024856567383,
130
  "learning_rate": 1.3600000000000002e-05,
131
- "loss": 1.1955,
132
  "step": 13
133
  },
134
  {
135
  "epoch": 0.008680824678344443,
136
- "grad_norm": 0.436432808637619,
137
  "learning_rate": 1.2800000000000001e-05,
138
- "loss": 0.9885,
139
  "step": 14
140
  },
141
  {
142
  "epoch": 0.009300883583940475,
143
- "grad_norm": 0.5418331623077393,
144
  "learning_rate": 1.2e-05,
145
- "loss": 1.0504,
146
  "step": 15
147
  },
148
  {
149
  "epoch": 0.009300883583940475,
150
- "eval_loss": 1.2044346332550049,
151
- "eval_runtime": 49.0412,
152
- "eval_samples_per_second": 2.039,
153
- "eval_steps_per_second": 2.039,
154
  "step": 15
155
  },
156
  {
157
  "epoch": 0.009920942489536505,
158
- "grad_norm": 0.4635182023048401,
159
  "learning_rate": 1.1200000000000001e-05,
160
- "loss": 0.9267,
161
  "step": 16
162
  },
163
  {
164
  "epoch": 0.010541001395132537,
165
- "grad_norm": 0.4401405155658722,
166
  "learning_rate": 1.04e-05,
167
- "loss": 1.0594,
168
  "step": 17
169
  },
170
  {
171
  "epoch": 0.01116106030072857,
172
- "grad_norm": 0.49569207429885864,
173
  "learning_rate": 9.600000000000001e-06,
174
- "loss": 0.9993,
175
  "step": 18
176
  },
177
  {
178
  "epoch": 0.01116106030072857,
179
- "eval_loss": 1.1875802278518677,
180
- "eval_runtime": 49.6265,
181
- "eval_samples_per_second": 2.015,
182
- "eval_steps_per_second": 2.015,
183
  "step": 18
184
  }
185
  ],
@@ -200,7 +200,7 @@
200
  "attributes": {}
201
  }
202
  },
203
- "total_flos": 3.856581922449408e+16,
204
  "train_batch_size": 2,
205
  "trial_name": null,
206
  "trial_params": null
 
10
  "log_history": [
11
  {
12
  "epoch": 0.0006200589055960316,
13
+ "grad_norm": 0.6006524562835693,
14
  "learning_rate": 4.000000000000001e-06,
15
+ "loss": 1.4363,
16
  "step": 1
17
  },
18
  {
19
  "epoch": 0.0012401178111920632,
20
+ "grad_norm": 0.4452102482318878,
21
  "learning_rate": 8.000000000000001e-06,
22
+ "loss": 1.1641,
23
  "step": 2
24
  },
25
  {
26
  "epoch": 0.0018601767167880949,
27
+ "grad_norm": 0.418393075466156,
28
  "learning_rate": 1.2e-05,
29
+ "loss": 1.0005,
30
  "step": 3
31
  },
32
  {
33
  "epoch": 0.0018601767167880949,
34
+ "eval_loss": 1.2559138536453247,
35
+ "eval_runtime": 50.2632,
36
+ "eval_samples_per_second": 1.99,
37
+ "eval_steps_per_second": 1.99,
38
  "step": 3
39
  },
40
  {
41
  "epoch": 0.0024802356223841263,
42
+ "grad_norm": 0.4599858820438385,
43
  "learning_rate": 1.6000000000000003e-05,
44
+ "loss": 1.1938,
45
  "step": 4
46
  },
47
  {
48
  "epoch": 0.0031002945279801583,
49
+ "grad_norm": 0.4313387870788574,
50
  "learning_rate": 2e-05,
51
+ "loss": 0.9897,
52
  "step": 5
53
  },
54
  {
55
  "epoch": 0.0037203534335761897,
56
+ "grad_norm": 0.4715091586112976,
57
  "learning_rate": 1.9200000000000003e-05,
58
+ "loss": 1.1105,
59
  "step": 6
60
  },
61
  {
62
  "epoch": 0.0037203534335761897,
63
+ "eval_loss": 1.2492942810058594,
64
+ "eval_runtime": 49.6589,
65
+ "eval_samples_per_second": 2.014,
66
+ "eval_steps_per_second": 2.014,
67
  "step": 6
68
  },
69
  {
70
  "epoch": 0.004340412339172222,
71
+ "grad_norm": 0.4590495824813843,
72
  "learning_rate": 1.8400000000000003e-05,
73
+ "loss": 1.0824,
74
  "step": 7
75
  },
76
  {
77
  "epoch": 0.004960471244768253,
78
+ "grad_norm": 0.6232957243919373,
79
  "learning_rate": 1.76e-05,
80
+ "loss": 1.2176,
81
  "step": 8
82
  },
83
  {
84
  "epoch": 0.005580530150364285,
85
+ "grad_norm": 0.4820755422115326,
86
  "learning_rate": 1.6800000000000002e-05,
87
+ "loss": 1.0774,
88
  "step": 9
89
  },
90
  {
91
  "epoch": 0.005580530150364285,
92
+ "eval_loss": 1.2356064319610596,
93
+ "eval_runtime": 49.9029,
94
+ "eval_samples_per_second": 2.004,
95
+ "eval_steps_per_second": 2.004,
96
  "step": 9
97
  },
98
  {
99
  "epoch": 0.0062005890559603165,
100
+ "grad_norm": 0.614038348197937,
101
  "learning_rate": 1.6000000000000003e-05,
102
+ "loss": 1.2989,
103
  "step": 10
104
  },
105
  {
106
  "epoch": 0.0068206479615563476,
107
+ "grad_norm": 0.5838705897331238,
108
  "learning_rate": 1.5200000000000002e-05,
109
+ "loss": 1.05,
110
  "step": 11
111
  },
112
  {
113
  "epoch": 0.0074407068671523795,
114
+ "grad_norm": 0.6922520995140076,
115
  "learning_rate": 1.4400000000000001e-05,
116
+ "loss": 1.3979,
117
  "step": 12
118
  },
119
  {
120
  "epoch": 0.0074407068671523795,
121
+ "eval_loss": 1.2183419466018677,
122
+ "eval_runtime": 50.5422,
123
+ "eval_samples_per_second": 1.979,
124
+ "eval_steps_per_second": 1.979,
125
  "step": 12
126
  },
127
  {
128
  "epoch": 0.008060765772748411,
129
+ "grad_norm": 0.6026104688644409,
130
  "learning_rate": 1.3600000000000002e-05,
131
+ "loss": 1.1974,
132
  "step": 13
133
  },
134
  {
135
  "epoch": 0.008680824678344443,
136
+ "grad_norm": 0.46571722626686096,
137
  "learning_rate": 1.2800000000000001e-05,
138
+ "loss": 0.9946,
139
  "step": 14
140
  },
141
  {
142
  "epoch": 0.009300883583940475,
143
+ "grad_norm": 0.5651541948318481,
144
  "learning_rate": 1.2e-05,
145
+ "loss": 1.0568,
146
  "step": 15
147
  },
148
  {
149
  "epoch": 0.009300883583940475,
150
+ "eval_loss": 1.1998839378356934,
151
+ "eval_runtime": 50.192,
152
+ "eval_samples_per_second": 1.992,
153
+ "eval_steps_per_second": 1.992,
154
  "step": 15
155
  },
156
  {
157
  "epoch": 0.009920942489536505,
158
+ "grad_norm": 0.4988487958908081,
159
  "learning_rate": 1.1200000000000001e-05,
160
+ "loss": 0.9318,
161
  "step": 16
162
  },
163
  {
164
  "epoch": 0.010541001395132537,
165
+ "grad_norm": 0.4823167324066162,
166
  "learning_rate": 1.04e-05,
167
+ "loss": 1.0622,
168
  "step": 17
169
  },
170
  {
171
  "epoch": 0.01116106030072857,
172
+ "grad_norm": 0.5371574759483337,
173
  "learning_rate": 9.600000000000001e-06,
174
+ "loss": 1.0033,
175
  "step": 18
176
  },
177
  {
178
  "epoch": 0.01116106030072857,
179
+ "eval_loss": 1.1833301782608032,
180
+ "eval_runtime": 50.0653,
181
+ "eval_samples_per_second": 1.997,
182
+ "eval_steps_per_second": 1.997,
183
  "step": 18
184
  }
185
  ],
 
200
  "attributes": {}
201
  }
202
  },
203
+ "total_flos": 3.911602614637363e+16,
204
  "train_batch_size": 2,
205
  "trial_name": null,
206
  "trial_params": null
checkpoint-18/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bf98514c81a51106969007e868d7ab8fb036a6bedc2987ddadf4a06b2a43f04f
3
  size 5752
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:522b553cd97c4d175d86850f43dd8230595c851fb1c776adff1402d853af53e9
3
  size 5752