Kei5uke commited on
Commit
b33e52d
·
verified ·
1 Parent(s): 06cc5e0

Training in progress, step 21, checkpoint

Browse files
checkpoint-21/adapter_config.json CHANGED
@@ -23,13 +23,13 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
- "down_proj",
27
- "q_proj",
28
- "up_proj",
29
  "k_proj",
30
- "gate_proj",
31
  "o_proj",
32
- "v_proj"
 
 
 
 
33
  ],
34
  "task_type": "CAUSAL_LM",
35
  "use_dora": false,
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
 
 
 
26
  "k_proj",
 
27
  "o_proj",
28
+ "v_proj",
29
+ "gate_proj",
30
+ "up_proj",
31
+ "down_proj",
32
+ "q_proj"
33
  ],
34
  "task_type": "CAUSAL_LM",
35
  "use_dora": false,
checkpoint-21/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:44e71a3c293f33d3d00ba5d07bf5e221f22eef99efabe5ba42d900420cbda86a
3
  size 83945296
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:675eddc86571193eb0870fe0e5c88ece07e4fa00ca342f5222e3d9c04722d273
3
  size 83945296
checkpoint-21/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0fd303d88d6316b401f13059794be833aabda5daa320f1db186bbc64962351e7
3
  size 168149074
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b7b1563f73cd76301fbeba65c91e5e2c22bad6a25919a99713b66b8089d4442a
3
  size 168149074
checkpoint-21/trainer_state.json CHANGED
@@ -10,205 +10,205 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.0006200589055960316,
13
- "grad_norm": 0.7733756303787231,
14
  "learning_rate": 4.000000000000001e-06,
15
  "loss": 1.166,
16
  "step": 1
17
  },
18
  {
19
  "epoch": 0.0012401178111920632,
20
- "grad_norm": 0.34648507833480835,
21
  "learning_rate": 8.000000000000001e-06,
22
  "loss": 0.9594,
23
  "step": 2
24
  },
25
  {
26
  "epoch": 0.0018601767167880949,
27
- "grad_norm": 0.3973380923271179,
28
  "learning_rate": 1.2e-05,
29
- "loss": 0.8281,
30
  "step": 3
31
  },
32
  {
33
  "epoch": 0.0018601767167880949,
34
- "eval_loss": 1.0490500926971436,
35
- "eval_runtime": 28.8129,
36
- "eval_samples_per_second": 3.471,
37
- "eval_steps_per_second": 3.471,
38
  "step": 3
39
  },
40
  {
41
  "epoch": 0.0024802356223841263,
42
- "grad_norm": 0.342887282371521,
43
  "learning_rate": 1.6000000000000003e-05,
44
- "loss": 0.9881,
45
  "step": 4
46
  },
47
  {
48
  "epoch": 0.0031002945279801583,
49
- "grad_norm": 0.46208909153938293,
50
  "learning_rate": 2e-05,
51
  "loss": 0.8319,
52
  "step": 5
53
  },
54
  {
55
  "epoch": 0.0037203534335761897,
56
- "grad_norm": 0.4722931981086731,
57
  "learning_rate": 1.9200000000000003e-05,
58
- "loss": 0.9364,
59
  "step": 6
60
  },
61
  {
62
  "epoch": 0.0037203534335761897,
63
- "eval_loss": 1.0428661108016968,
64
- "eval_runtime": 26.3402,
65
- "eval_samples_per_second": 3.796,
66
- "eval_steps_per_second": 3.796,
67
  "step": 6
68
  },
69
  {
70
  "epoch": 0.004340412339172222,
71
- "grad_norm": 0.4379231333732605,
72
  "learning_rate": 1.8400000000000003e-05,
73
- "loss": 0.9073,
74
  "step": 7
75
  },
76
  {
77
  "epoch": 0.004960471244768253,
78
- "grad_norm": 0.5576643943786621,
79
  "learning_rate": 1.76e-05,
80
- "loss": 0.9823,
81
  "step": 8
82
  },
83
  {
84
  "epoch": 0.005580530150364285,
85
- "grad_norm": 0.4029313325881958,
86
  "learning_rate": 1.6800000000000002e-05,
87
  "loss": 0.9045,
88
  "step": 9
89
  },
90
  {
91
  "epoch": 0.005580530150364285,
92
- "eval_loss": 1.0290312767028809,
93
- "eval_runtime": 26.3678,
94
- "eval_samples_per_second": 3.793,
95
- "eval_steps_per_second": 3.793,
96
  "step": 9
97
  },
98
  {
99
  "epoch": 0.0062005890559603165,
100
- "grad_norm": 0.4460880160331726,
101
  "learning_rate": 1.6000000000000003e-05,
102
- "loss": 1.0763,
103
  "step": 10
104
  },
105
  {
106
  "epoch": 0.0068206479615563476,
107
- "grad_norm": 0.7060808539390564,
108
  "learning_rate": 1.5200000000000002e-05,
109
- "loss": 0.8713,
110
  "step": 11
111
  },
112
  {
113
  "epoch": 0.0074407068671523795,
114
- "grad_norm": 0.6190696954727173,
115
  "learning_rate": 1.4400000000000001e-05,
116
- "loss": 1.1753,
117
  "step": 12
118
  },
119
  {
120
  "epoch": 0.0074407068671523795,
121
- "eval_loss": 1.0143623352050781,
122
- "eval_runtime": 26.7829,
123
- "eval_samples_per_second": 3.734,
124
- "eval_steps_per_second": 3.734,
125
  "step": 12
126
  },
127
  {
128
  "epoch": 0.008060765772748411,
129
- "grad_norm": 0.5669320225715637,
130
  "learning_rate": 1.3600000000000002e-05,
131
- "loss": 0.991,
132
  "step": 13
133
  },
134
  {
135
  "epoch": 0.008680824678344443,
136
- "grad_norm": 0.49686795473098755,
137
  "learning_rate": 1.2800000000000001e-05,
138
- "loss": 0.8497,
139
  "step": 14
140
  },
141
  {
142
  "epoch": 0.009300883583940475,
143
- "grad_norm": 0.5733391642570496,
144
  "learning_rate": 1.2e-05,
145
- "loss": 0.8679,
146
  "step": 15
147
  },
148
  {
149
  "epoch": 0.009300883583940475,
150
- "eval_loss": 0.9997775554656982,
151
- "eval_runtime": 26.7666,
152
- "eval_samples_per_second": 3.736,
153
- "eval_steps_per_second": 3.736,
154
  "step": 15
155
  },
156
  {
157
  "epoch": 0.009920942489536505,
158
- "grad_norm": 0.5733532309532166,
159
  "learning_rate": 1.1200000000000001e-05,
160
- "loss": 0.8173,
161
  "step": 16
162
  },
163
  {
164
  "epoch": 0.010541001395132537,
165
- "grad_norm": 0.375442773103714,
166
  "learning_rate": 1.04e-05,
167
- "loss": 0.9054,
168
  "step": 17
169
  },
170
  {
171
  "epoch": 0.01116106030072857,
172
- "grad_norm": 0.4431408643722534,
173
  "learning_rate": 9.600000000000001e-06,
174
- "loss": 0.8451,
175
  "step": 18
176
  },
177
  {
178
  "epoch": 0.01116106030072857,
179
- "eval_loss": 0.986146092414856,
180
- "eval_runtime": 26.7104,
181
- "eval_samples_per_second": 3.744,
182
- "eval_steps_per_second": 3.744,
183
  "step": 18
184
  },
185
  {
186
  "epoch": 0.011781119206324601,
187
- "grad_norm": 0.5397400259971619,
188
  "learning_rate": 8.8e-06,
189
- "loss": 0.8274,
190
  "step": 19
191
  },
192
  {
193
  "epoch": 0.012401178111920633,
194
- "grad_norm": 0.3542424440383911,
195
  "learning_rate": 8.000000000000001e-06,
196
- "loss": 0.6586,
197
  "step": 20
198
  },
199
  {
200
  "epoch": 0.013021237017516665,
201
- "grad_norm": 0.45993754267692566,
202
  "learning_rate": 7.2000000000000005e-06,
203
- "loss": 0.937,
204
  "step": 21
205
  },
206
  {
207
  "epoch": 0.013021237017516665,
208
- "eval_loss": 0.9746847748756409,
209
- "eval_runtime": 26.6243,
210
- "eval_samples_per_second": 3.756,
211
- "eval_steps_per_second": 3.756,
212
  "step": 21
213
  }
214
  ],
 
10
  "log_history": [
11
  {
12
  "epoch": 0.0006200589055960316,
13
+ "grad_norm": 0.7731499075889587,
14
  "learning_rate": 4.000000000000001e-06,
15
  "loss": 1.166,
16
  "step": 1
17
  },
18
  {
19
  "epoch": 0.0012401178111920632,
20
+ "grad_norm": 0.346652090549469,
21
  "learning_rate": 8.000000000000001e-06,
22
  "loss": 0.9594,
23
  "step": 2
24
  },
25
  {
26
  "epoch": 0.0018601767167880949,
27
+ "grad_norm": 0.39937612414360046,
28
  "learning_rate": 1.2e-05,
29
+ "loss": 0.828,
30
  "step": 3
31
  },
32
  {
33
  "epoch": 0.0018601767167880949,
34
+ "eval_loss": 1.0491173267364502,
35
+ "eval_runtime": 26.0459,
36
+ "eval_samples_per_second": 3.839,
37
+ "eval_steps_per_second": 3.839,
38
  "step": 3
39
  },
40
  {
41
  "epoch": 0.0024802356223841263,
42
+ "grad_norm": 0.3457990288734436,
43
  "learning_rate": 1.6000000000000003e-05,
44
+ "loss": 0.988,
45
  "step": 4
46
  },
47
  {
48
  "epoch": 0.0031002945279801583,
49
+ "grad_norm": 0.513918399810791,
50
  "learning_rate": 2e-05,
51
  "loss": 0.8319,
52
  "step": 5
53
  },
54
  {
55
  "epoch": 0.0037203534335761897,
56
+ "grad_norm": 0.47683727741241455,
57
  "learning_rate": 1.9200000000000003e-05,
58
+ "loss": 0.9366,
59
  "step": 6
60
  },
61
  {
62
  "epoch": 0.0037203534335761897,
63
+ "eval_loss": 1.0431987047195435,
64
+ "eval_runtime": 26.2352,
65
+ "eval_samples_per_second": 3.812,
66
+ "eval_steps_per_second": 3.812,
67
  "step": 6
68
  },
69
  {
70
  "epoch": 0.004340412339172222,
71
+ "grad_norm": 0.4349649250507355,
72
  "learning_rate": 1.8400000000000003e-05,
73
+ "loss": 0.9076,
74
  "step": 7
75
  },
76
  {
77
  "epoch": 0.004960471244768253,
78
+ "grad_norm": 0.5443339943885803,
79
  "learning_rate": 1.76e-05,
80
+ "loss": 0.9829,
81
  "step": 8
82
  },
83
  {
84
  "epoch": 0.005580530150364285,
85
+ "grad_norm": 0.3990451395511627,
86
  "learning_rate": 1.6800000000000002e-05,
87
  "loss": 0.9045,
88
  "step": 9
89
  },
90
  {
91
  "epoch": 0.005580530150364285,
92
+ "eval_loss": 1.0292819738388062,
93
+ "eval_runtime": 26.3477,
94
+ "eval_samples_per_second": 3.795,
95
+ "eval_steps_per_second": 3.795,
96
  "step": 9
97
  },
98
  {
99
  "epoch": 0.0062005890559603165,
100
+ "grad_norm": 0.44750896096229553,
101
  "learning_rate": 1.6000000000000003e-05,
102
+ "loss": 1.0768,
103
  "step": 10
104
  },
105
  {
106
  "epoch": 0.0068206479615563476,
107
+ "grad_norm": 0.6865904331207275,
108
  "learning_rate": 1.5200000000000002e-05,
109
+ "loss": 0.8717,
110
  "step": 11
111
  },
112
  {
113
  "epoch": 0.0074407068671523795,
114
+ "grad_norm": 0.548491895198822,
115
  "learning_rate": 1.4400000000000001e-05,
116
+ "loss": 1.1758,
117
  "step": 12
118
  },
119
  {
120
  "epoch": 0.0074407068671523795,
121
+ "eval_loss": 1.0153621435165405,
122
+ "eval_runtime": 26.5025,
123
+ "eval_samples_per_second": 3.773,
124
+ "eval_steps_per_second": 3.773,
125
  "step": 12
126
  },
127
  {
128
  "epoch": 0.008060765772748411,
129
+ "grad_norm": 0.5520397424697876,
130
  "learning_rate": 1.3600000000000002e-05,
131
+ "loss": 0.992,
132
  "step": 13
133
  },
134
  {
135
  "epoch": 0.008680824678344443,
136
+ "grad_norm": 0.381074458360672,
137
  "learning_rate": 1.2800000000000001e-05,
138
+ "loss": 0.8504,
139
  "step": 14
140
  },
141
  {
142
  "epoch": 0.009300883583940475,
143
+ "grad_norm": 0.41114744544029236,
144
  "learning_rate": 1.2e-05,
145
+ "loss": 0.8685,
146
  "step": 15
147
  },
148
  {
149
  "epoch": 0.009300883583940475,
150
+ "eval_loss": 1.0006165504455566,
151
+ "eval_runtime": 26.4086,
152
+ "eval_samples_per_second": 3.787,
153
+ "eval_steps_per_second": 3.787,
154
  "step": 15
155
  },
156
  {
157
  "epoch": 0.009920942489536505,
158
+ "grad_norm": 0.5197786688804626,
159
  "learning_rate": 1.1200000000000001e-05,
160
+ "loss": 0.8176,
161
  "step": 16
162
  },
163
  {
164
  "epoch": 0.010541001395132537,
165
+ "grad_norm": 0.3905779719352722,
166
  "learning_rate": 1.04e-05,
167
+ "loss": 0.9055,
168
  "step": 17
169
  },
170
  {
171
  "epoch": 0.01116106030072857,
172
+ "grad_norm": 0.43444159626960754,
173
  "learning_rate": 9.600000000000001e-06,
174
+ "loss": 0.8452,
175
  "step": 18
176
  },
177
  {
178
  "epoch": 0.01116106030072857,
179
+ "eval_loss": 0.9866559505462646,
180
+ "eval_runtime": 26.3639,
181
+ "eval_samples_per_second": 3.793,
182
+ "eval_steps_per_second": 3.793,
183
  "step": 18
184
  },
185
  {
186
  "epoch": 0.011781119206324601,
187
+ "grad_norm": 0.5130135416984558,
188
  "learning_rate": 8.8e-06,
189
+ "loss": 0.8276,
190
  "step": 19
191
  },
192
  {
193
  "epoch": 0.012401178111920633,
194
+ "grad_norm": 0.35828444361686707,
195
  "learning_rate": 8.000000000000001e-06,
196
+ "loss": 0.6587,
197
  "step": 20
198
  },
199
  {
200
  "epoch": 0.013021237017516665,
201
+ "grad_norm": 0.4369220435619354,
202
  "learning_rate": 7.2000000000000005e-06,
203
+ "loss": 0.9372,
204
  "step": 21
205
  },
206
  {
207
  "epoch": 0.013021237017516665,
208
+ "eval_loss": 0.9746317267417908,
209
+ "eval_runtime": 26.2985,
210
+ "eval_samples_per_second": 3.803,
211
+ "eval_steps_per_second": 3.803,
212
  "step": 21
213
  }
214
  ],
checkpoint-21/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7c433cc5009675ccadca19513c8ae659f957c78885071da55c2a8d32a98d517a
3
  size 5688
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cbe2fb31cf4c04f893e5299faa77267af76b8447f922d82c2d6c0200e1fe225f
3
  size 5688