Kei5uke commited on
Commit
49d7079
·
verified ·
1 Parent(s): cacdae6

Training in progress, step 18, checkpoint

Browse files
checkpoint-18/adapter_config.json CHANGED
@@ -23,13 +23,13 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
- "k_proj",
27
- "o_proj",
28
  "v_proj",
29
- "gate_proj",
 
30
  "up_proj",
31
- "down_proj",
32
- "q_proj"
33
  ],
34
  "task_type": "CAUSAL_LM",
35
  "use_dora": false,
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
+ "down_proj",
 
27
  "v_proj",
28
+ "o_proj",
29
+ "q_proj",
30
  "up_proj",
31
+ "gate_proj",
32
+ "k_proj"
33
  ],
34
  "task_type": "CAUSAL_LM",
35
  "use_dora": false,
checkpoint-18/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:51e0438cbd1f9c905ce9451cbcd8846b4101cccb9b760ebe70e3f97958e59c50
3
  size 80013120
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:994230e521b78fea491b713abf254cc71f7605623951b9e08f3a1eebf0c18fa4
3
  size 80013120
checkpoint-18/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d511730bf1cc83896f70d4c77b7a69992fbcb1efe453cf0f82c3d5f1afeb2591
3
  size 160284754
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9fd7e0126ac475c04bd895370aef59eeb1162e8c845b85543502d5aeb48ecd6
3
  size 160284754
checkpoint-18/trainer_state.json CHANGED
@@ -10,176 +10,176 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.0006200589055960316,
13
- "grad_norm": 0.18599237501621246,
14
  "learning_rate": 4.000000000000001e-06,
15
- "loss": 0.9319,
16
  "step": 1
17
  },
18
  {
19
  "epoch": 0.0012401178111920632,
20
- "grad_norm": 0.09086523950099945,
21
  "learning_rate": 8.000000000000001e-06,
22
- "loss": 0.6463,
23
  "step": 2
24
  },
25
  {
26
  "epoch": 0.0018601767167880949,
27
- "grad_norm": 0.09325356036424637,
28
  "learning_rate": 1.2e-05,
29
- "loss": 0.575,
30
  "step": 3
31
  },
32
  {
33
  "epoch": 0.0018601767167880949,
34
- "eval_loss": 0.7272388935089111,
35
- "eval_runtime": 33.9288,
36
- "eval_samples_per_second": 2.947,
37
- "eval_steps_per_second": 2.947,
38
  "step": 3
39
  },
40
  {
41
  "epoch": 0.0024802356223841263,
42
- "grad_norm": 0.09707680344581604,
43
  "learning_rate": 1.6000000000000003e-05,
44
- "loss": 0.7038,
45
  "step": 4
46
  },
47
  {
48
  "epoch": 0.0031002945279801583,
49
- "grad_norm": 0.10810782760381699,
50
  "learning_rate": 2e-05,
51
- "loss": 0.5471,
52
  "step": 5
53
  },
54
  {
55
  "epoch": 0.0037203534335761897,
56
- "grad_norm": 0.13030411303043365,
57
  "learning_rate": 1.9200000000000003e-05,
58
- "loss": 0.6735,
59
  "step": 6
60
  },
61
  {
62
  "epoch": 0.0037203534335761897,
63
- "eval_loss": 0.7256982922554016,
64
- "eval_runtime": 33.9528,
65
- "eval_samples_per_second": 2.945,
66
- "eval_steps_per_second": 2.945,
67
  "step": 6
68
  },
69
  {
70
  "epoch": 0.004340412339172222,
71
- "grad_norm": 0.12336976826190948,
72
  "learning_rate": 1.8400000000000003e-05,
73
- "loss": 0.654,
74
  "step": 7
75
  },
76
  {
77
  "epoch": 0.004960471244768253,
78
- "grad_norm": 0.16046489775180817,
79
  "learning_rate": 1.76e-05,
80
- "loss": 0.6301,
81
  "step": 8
82
  },
83
  {
84
  "epoch": 0.005580530150364285,
85
- "grad_norm": 0.11589628458023071,
86
  "learning_rate": 1.6800000000000002e-05,
87
- "loss": 0.5326,
88
  "step": 9
89
  },
90
  {
91
  "epoch": 0.005580530150364285,
92
- "eval_loss": 0.723076343536377,
93
- "eval_runtime": 34.5303,
94
- "eval_samples_per_second": 2.896,
95
- "eval_steps_per_second": 2.896,
96
  "step": 9
97
  },
98
  {
99
  "epoch": 0.0062005890559603165,
100
- "grad_norm": 0.12244553864002228,
101
  "learning_rate": 1.6000000000000003e-05,
102
- "loss": 0.8055,
103
  "step": 10
104
  },
105
  {
106
  "epoch": 0.0068206479615563476,
107
- "grad_norm": 0.10270956158638,
108
  "learning_rate": 1.5200000000000002e-05,
109
- "loss": 0.5492,
110
  "step": 11
111
  },
112
  {
113
  "epoch": 0.0074407068671523795,
114
- "grad_norm": 0.14439155161380768,
115
  "learning_rate": 1.4400000000000001e-05,
116
- "loss": 0.8579,
117
  "step": 12
118
  },
119
  {
120
  "epoch": 0.0074407068671523795,
121
- "eval_loss": 0.7205619215965271,
122
- "eval_runtime": 33.7297,
123
- "eval_samples_per_second": 2.965,
124
- "eval_steps_per_second": 2.965,
125
  "step": 12
126
  },
127
  {
128
  "epoch": 0.008060765772748411,
129
- "grad_norm": 0.1271936148405075,
130
  "learning_rate": 1.3600000000000002e-05,
131
- "loss": 0.6923,
132
  "step": 13
133
  },
134
  {
135
  "epoch": 0.008680824678344443,
136
- "grad_norm": 0.092640720307827,
137
  "learning_rate": 1.2800000000000001e-05,
138
- "loss": 0.5983,
139
  "step": 14
140
  },
141
  {
142
  "epoch": 0.009300883583940475,
143
- "grad_norm": 0.1161455512046814,
144
  "learning_rate": 1.2e-05,
145
- "loss": 0.6386,
146
  "step": 15
147
  },
148
  {
149
  "epoch": 0.009300883583940475,
150
- "eval_loss": 0.7178443074226379,
151
- "eval_runtime": 33.5878,
152
- "eval_samples_per_second": 2.977,
153
- "eval_steps_per_second": 2.977,
154
  "step": 15
155
  },
156
  {
157
  "epoch": 0.009920942489536505,
158
- "grad_norm": 0.09220079332590103,
159
  "learning_rate": 1.1200000000000001e-05,
160
- "loss": 0.4665,
161
  "step": 16
162
  },
163
  {
164
  "epoch": 0.010541001395132537,
165
- "grad_norm": 0.08713295310735703,
166
  "learning_rate": 1.04e-05,
167
- "loss": 0.6114,
168
  "step": 17
169
  },
170
  {
171
  "epoch": 0.01116106030072857,
172
- "grad_norm": 0.11913106590509415,
173
  "learning_rate": 9.600000000000001e-06,
174
- "loss": 0.5703,
175
  "step": 18
176
  },
177
  {
178
  "epoch": 0.01116106030072857,
179
- "eval_loss": 0.7152560949325562,
180
- "eval_runtime": 33.4806,
181
- "eval_samples_per_second": 2.987,
182
- "eval_steps_per_second": 2.987,
183
  "step": 18
184
  }
185
  ],
@@ -200,7 +200,7 @@
200
  "attributes": {}
201
  }
202
  },
203
- "total_flos": 2.436141236915405e+16,
204
  "train_batch_size": 2,
205
  "trial_name": null,
206
  "trial_params": null
 
10
  "log_history": [
11
  {
12
  "epoch": 0.0006200589055960316,
13
+ "grad_norm": 0.16878783702850342,
14
  "learning_rate": 4.000000000000001e-06,
15
+ "loss": 0.9729,
16
  "step": 1
17
  },
18
  {
19
  "epoch": 0.0012401178111920632,
20
+ "grad_norm": 0.0865061804652214,
21
  "learning_rate": 8.000000000000001e-06,
22
+ "loss": 0.6661,
23
  "step": 2
24
  },
25
  {
26
  "epoch": 0.0018601767167880949,
27
+ "grad_norm": 0.0920981764793396,
28
  "learning_rate": 1.2e-05,
29
+ "loss": 0.5954,
30
  "step": 3
31
  },
32
  {
33
  "epoch": 0.0018601767167880949,
34
+ "eval_loss": 0.762992262840271,
35
+ "eval_runtime": 33.3972,
36
+ "eval_samples_per_second": 2.994,
37
+ "eval_steps_per_second": 2.994,
38
  "step": 3
39
  },
40
  {
41
  "epoch": 0.0024802356223841263,
42
+ "grad_norm": 0.10451442003250122,
43
  "learning_rate": 1.6000000000000003e-05,
44
+ "loss": 0.7227,
45
  "step": 4
46
  },
47
  {
48
  "epoch": 0.0031002945279801583,
49
+ "grad_norm": 0.11351602524518967,
50
  "learning_rate": 2e-05,
51
+ "loss": 0.5719,
52
  "step": 5
53
  },
54
  {
55
  "epoch": 0.0037203534335761897,
56
+ "grad_norm": 0.12484906613826752,
57
  "learning_rate": 1.9200000000000003e-05,
58
+ "loss": 0.6972,
59
  "step": 6
60
  },
61
  {
62
  "epoch": 0.0037203534335761897,
63
+ "eval_loss": 0.7615777850151062,
64
+ "eval_runtime": 33.3804,
65
+ "eval_samples_per_second": 2.996,
66
+ "eval_steps_per_second": 2.996,
67
  "step": 6
68
  },
69
  {
70
  "epoch": 0.004340412339172222,
71
+ "grad_norm": 0.11116621643304825,
72
  "learning_rate": 1.8400000000000003e-05,
73
+ "loss": 0.6789,
74
  "step": 7
75
  },
76
  {
77
  "epoch": 0.004960471244768253,
78
+ "grad_norm": 0.16013608872890472,
79
  "learning_rate": 1.76e-05,
80
+ "loss": 0.6612,
81
  "step": 8
82
  },
83
  {
84
  "epoch": 0.005580530150364285,
85
+ "grad_norm": 0.1211152970790863,
86
  "learning_rate": 1.6800000000000002e-05,
87
+ "loss": 0.5557,
88
  "step": 9
89
  },
90
  {
91
  "epoch": 0.005580530150364285,
92
+ "eval_loss": 0.7586225867271423,
93
+ "eval_runtime": 33.7015,
94
+ "eval_samples_per_second": 2.967,
95
+ "eval_steps_per_second": 2.967,
96
  "step": 9
97
  },
98
  {
99
  "epoch": 0.0062005890559603165,
100
+ "grad_norm": 0.1299518495798111,
101
  "learning_rate": 1.6000000000000003e-05,
102
+ "loss": 0.8339,
103
  "step": 10
104
  },
105
  {
106
  "epoch": 0.0068206479615563476,
107
+ "grad_norm": 0.11476285755634308,
108
  "learning_rate": 1.5200000000000002e-05,
109
+ "loss": 0.5786,
110
  "step": 11
111
  },
112
  {
113
  "epoch": 0.0074407068671523795,
114
+ "grad_norm": 0.18253028392791748,
115
  "learning_rate": 1.4400000000000001e-05,
116
+ "loss": 0.8889,
117
  "step": 12
118
  },
119
  {
120
  "epoch": 0.0074407068671523795,
121
+ "eval_loss": 0.7548640966415405,
122
+ "eval_runtime": 34.763,
123
+ "eval_samples_per_second": 2.877,
124
+ "eval_steps_per_second": 2.877,
125
  "step": 12
126
  },
127
  {
128
  "epoch": 0.008060765772748411,
129
+ "grad_norm": 0.14364618062973022,
130
  "learning_rate": 1.3600000000000002e-05,
131
+ "loss": 0.7251,
132
  "step": 13
133
  },
134
  {
135
  "epoch": 0.008680824678344443,
136
+ "grad_norm": 0.11001864820718765,
137
  "learning_rate": 1.2800000000000001e-05,
138
+ "loss": 0.6209,
139
  "step": 14
140
  },
141
  {
142
  "epoch": 0.009300883583940475,
143
+ "grad_norm": 0.1239691898226738,
144
  "learning_rate": 1.2e-05,
145
+ "loss": 0.6641,
146
  "step": 15
147
  },
148
  {
149
  "epoch": 0.009300883583940475,
150
+ "eval_loss": 0.7505818009376526,
151
+ "eval_runtime": 34.0273,
152
+ "eval_samples_per_second": 2.939,
153
+ "eval_steps_per_second": 2.939,
154
  "step": 15
155
  },
156
  {
157
  "epoch": 0.009920942489536505,
158
+ "grad_norm": 0.11025531589984894,
159
  "learning_rate": 1.1200000000000001e-05,
160
+ "loss": 0.4851,
161
  "step": 16
162
  },
163
  {
164
  "epoch": 0.010541001395132537,
165
+ "grad_norm": 0.10391498357057571,
166
  "learning_rate": 1.04e-05,
167
+ "loss": 0.6274,
168
  "step": 17
169
  },
170
  {
171
  "epoch": 0.01116106030072857,
172
+ "grad_norm": 0.13879302144050598,
173
  "learning_rate": 9.600000000000001e-06,
174
+ "loss": 0.5931,
175
  "step": 18
176
  },
177
  {
178
  "epoch": 0.01116106030072857,
179
+ "eval_loss": 0.7464621067047119,
180
+ "eval_runtime": 34.3301,
181
+ "eval_samples_per_second": 2.913,
182
+ "eval_steps_per_second": 2.913,
183
  "step": 18
184
  }
185
  ],
 
200
  "attributes": {}
201
  }
202
  },
203
+ "total_flos": 2.459045523131597e+16,
204
  "train_batch_size": 2,
205
  "trial_name": null,
206
  "trial_params": null
checkpoint-18/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bd165a72088a81e5441087b0ae01dd72f7f14c5dad6a2e407eebc961586b0fc2
3
  size 5752
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1fba20505ee5f941c5085404aed8036910f9919d2bc7b2ba5ba4186849388b0c
3
  size 5752