hellosindh commited on
Commit
8424953
·
verified ·
1 Parent(s): e0f6e92

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -20,7 +20,8 @@ First BERT-style model trained from scratch on Sindhi text.
20
  | S1 | 500K lines | 5 | 78.10 | from scratch |
21
  | S2 | 1.5M lines | 3 | 41.62 | continued |
22
  | S3 | 1.49M lines | 2 | 28.46 | bf16, cosine LR |
23
- | S4 | 87M words | 3 | 35.42 | grouped context |
 
24
 
25
  ## Usage
26
 
 
20
  | S1 | 500K lines | 5 | 78.10 | from scratch |
21
  | S2 | 1.5M lines | 3 | 41.62 | continued |
22
  | S3 | 1.49M lines | 2 | 28.46 | bf16, cosine LR |
23
+ | S4 | 87M words | 3 | 35.16 | grouped context, MLM=0.20 |
24
+ | S5 | 87M words | 2 | 29.45 | fine polish, MLM=0.15 |
25
 
26
  ## Usage
27
 
checkpoint-1962/config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_cross_attention": false,
3
+ "architectures": [
4
+ "RobertaForMaskedLM"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 1,
8
+ "classifier_dropout": null,
9
+ "dtype": "float32",
10
+ "eos_token_id": 2,
11
+ "hidden_act": "gelu",
12
+ "hidden_dropout_prob": 0.1,
13
+ "hidden_size": 768,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 3072,
16
+ "is_decoder": false,
17
+ "layer_norm_eps": 1e-12,
18
+ "max_position_embeddings": 514,
19
+ "model_type": "roberta",
20
+ "num_attention_heads": 12,
21
+ "num_hidden_layers": 12,
22
+ "pad_token_id": 0,
23
+ "tie_word_embeddings": true,
24
+ "transformers_version": "5.0.0",
25
+ "type_vocab_size": 1,
26
+ "use_cache": false,
27
+ "vocab_size": 32001
28
+ }
checkpoint-1962/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b408867c203ef5560d1d08d8c10f23eb9ec84b0c35800910f671bbbbc54f310
3
+ size 442633860
checkpoint-1962/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94b7e14d685848516c9ccb2ef50864ae9c59c5ddd1fdcade33822f46bb2f6140
3
+ size 885391563
checkpoint-1962/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:533785fa47d37d3c80b2d9c2a504e57c9142557c2ae85831a5276a2fd23d116b
3
+ size 14645
checkpoint-1962/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c1c18f6958e463b9bf1adf762c04149cbe5a397f12a80ca8aa0df8208fb9090
3
+ size 1465
checkpoint-1962/trainer_state.json ADDED
@@ -0,0 +1,184 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 1962,
3
+ "best_metric": 3.396049737930298,
4
+ "best_model_checkpoint": "sindhibert_session5/checkpoint-1962",
5
+ "epoch": 1.0,
6
+ "eval_steps": 1962,
7
+ "global_step": 1962,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.05098139179199592,
14
+ "grad_norm": 5.245308876037598,
15
+ "learning_rate": 2.1063829787234044e-06,
16
+ "loss": 14.123677978515625,
17
+ "step": 100
18
+ },
19
+ {
20
+ "epoch": 0.10196278358399184,
21
+ "grad_norm": 5.646098613739014,
22
+ "learning_rate": 4.23404255319149e-06,
23
+ "loss": 14.066470947265625,
24
+ "step": 200
25
+ },
26
+ {
27
+ "epoch": 0.15294417537598776,
28
+ "grad_norm": 5.477426528930664,
29
+ "learning_rate": 4.99628768445363e-06,
30
+ "loss": 14.01585693359375,
31
+ "step": 300
32
+ },
33
+ {
34
+ "epoch": 0.20392556716798368,
35
+ "grad_norm": 5.2961344718933105,
36
+ "learning_rate": 4.975657001043815e-06,
37
+ "loss": 13.965966796875,
38
+ "step": 400
39
+ },
40
+ {
41
+ "epoch": 0.2549069589599796,
42
+ "grad_norm": 5.533503532409668,
43
+ "learning_rate": 4.937082682936824e-06,
44
+ "loss": 13.97388671875,
45
+ "step": 500
46
+ },
47
+ {
48
+ "epoch": 0.3058883507519755,
49
+ "grad_norm": 5.314619064331055,
50
+ "learning_rate": 4.8808443179239025e-06,
51
+ "loss": 13.95988525390625,
52
+ "step": 600
53
+ },
54
+ {
55
+ "epoch": 0.35686974254397147,
56
+ "grad_norm": 5.3324384689331055,
57
+ "learning_rate": 4.807349523327375e-06,
58
+ "loss": 13.953388671875,
59
+ "step": 700
60
+ },
61
+ {
62
+ "epoch": 0.40785113433596737,
63
+ "grad_norm": 5.467737674713135,
64
+ "learning_rate": 4.717130991578312e-06,
65
+ "loss": 13.9411962890625,
66
+ "step": 800
67
+ },
68
+ {
69
+ "epoch": 0.45883252612796327,
70
+ "grad_norm": 5.45510721206665,
71
+ "learning_rate": 4.610842629246138e-06,
72
+ "loss": 13.934332275390625,
73
+ "step": 900
74
+ },
75
+ {
76
+ "epoch": 0.5098139179199592,
77
+ "grad_norm": 5.6341962814331055,
78
+ "learning_rate": 4.489254817504595e-06,
79
+ "loss": 13.926766357421876,
80
+ "step": 1000
81
+ },
82
+ {
83
+ "epoch": 0.5607953097119551,
84
+ "grad_norm": 5.584266662597656,
85
+ "learning_rate": 4.353248828386359e-06,
86
+ "loss": 13.91165283203125,
87
+ "step": 1100
88
+ },
89
+ {
90
+ "epoch": 0.611776701503951,
91
+ "grad_norm": 5.35775899887085,
92
+ "learning_rate": 4.203810437297447e-06,
93
+ "loss": 13.915504150390625,
94
+ "step": 1200
95
+ },
96
+ {
97
+ "epoch": 0.6627580932959469,
98
+ "grad_norm": 5.549437999725342,
99
+ "learning_rate": 4.042022778088111e-06,
100
+ "loss": 13.945462646484375,
101
+ "step": 1300
102
+ },
103
+ {
104
+ "epoch": 0.7137394850879429,
105
+ "grad_norm": 5.455720901489258,
106
+ "learning_rate": 3.869058492466847e-06,
107
+ "loss": 13.89906494140625,
108
+ "step": 1400
109
+ },
110
+ {
111
+ "epoch": 0.7647208768799388,
112
+ "grad_norm": 5.856232166290283,
113
+ "learning_rate": 3.6861712306588434e-06,
114
+ "loss": 13.905211181640626,
115
+ "step": 1500
116
+ },
117
+ {
118
+ "epoch": 0.8157022686719347,
119
+ "grad_norm": 5.481319904327393,
120
+ "learning_rate": 3.4946865649123025e-06,
121
+ "loss": 13.896219482421875,
122
+ "step": 1600
123
+ },
124
+ {
125
+ "epoch": 0.8666836604639306,
126
+ "grad_norm": 5.278250217437744,
127
+ "learning_rate": 3.2959923817118296e-06,
128
+ "loss": 13.906019287109375,
129
+ "step": 1700
130
+ },
131
+ {
132
+ "epoch": 0.9176650522559265,
133
+ "grad_norm": 5.768893718719482,
134
+ "learning_rate": 3.091528822336405e-06,
135
+ "loss": 13.93161376953125,
136
+ "step": 1800
137
+ },
138
+ {
139
+ "epoch": 0.9686464440479226,
140
+ "grad_norm": 5.366046905517578,
141
+ "learning_rate": 2.8827778446730764e-06,
142
+ "loss": 13.906478271484374,
143
+ "step": 1900
144
+ },
145
+ {
146
+ "epoch": 1.0,
147
+ "eval_loss": 3.396049737930298,
148
+ "eval_runtime": 7.952,
149
+ "eval_samples_per_second": 637.827,
150
+ "eval_steps_per_second": 10.06,
151
+ "step": 1962
152
+ }
153
+ ],
154
+ "logging_steps": 100,
155
+ "max_steps": 3924,
156
+ "num_input_tokens_seen": 0,
157
+ "num_train_epochs": 2,
158
+ "save_steps": 1962,
159
+ "stateful_callbacks": {
160
+ "EarlyStoppingCallback": {
161
+ "args": {
162
+ "early_stopping_patience": 3,
163
+ "early_stopping_threshold": 0.0
164
+ },
165
+ "attributes": {
166
+ "early_stopping_patience_counter": 0
167
+ }
168
+ },
169
+ "TrainerControl": {
170
+ "args": {
171
+ "should_epoch_stop": false,
172
+ "should_evaluate": false,
173
+ "should_log": false,
174
+ "should_save": true,
175
+ "should_training_stop": false
176
+ },
177
+ "attributes": {}
178
+ }
179
+ },
180
+ "total_flos": 1.321661037009623e+17,
181
+ "train_batch_size": 64,
182
+ "trial_name": null,
183
+ "trial_params": null
184
+ }
checkpoint-1962/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a0a0f3ce1e27fd1d47d30a38c39aa86503e7f6fb63f533cef4d2e9ac2db41014
3
+ size 5137
checkpoint-3924/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:66c9b40b4d1b2943a622be928e3f8beb231f2cf80d2acbe19352c740edfa76b9
3
  size 442633860
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f0ec985a6dec33fba0018d1ffbc1fcc4317a2291a75a8d6fc24f269454e6581
3
  size 442633860
checkpoint-3924/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8cdbb31b8e427b2d5c5d5dce127c362cb391d70f8282995b2a405651b6695774
3
  size 885391563
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ed56030f6a04a8bb3427806cc07b51579a5943f3c41ece7a44ff4210b13a4a7
3
  size 885391563
checkpoint-3924/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8eef8b1a8fe3ca13b13452c68d049d5772a114b25d47fd7c271209bdd37c174b
3
  size 1465
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e6f67f71bb87604f1d79bd9533843f4244e3f1cc9c56bc98e0df5eacc07beced
3
  size 1465
checkpoint-3924/trainer_state.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "best_global_step": 3924,
3
- "best_metric": 3.56946063041687,
4
- "best_model_checkpoint": "sindhibert_session4/checkpoint-3924",
5
  "epoch": 2.0,
6
  "eval_steps": 1962,
7
  "global_step": 3924,
@@ -11,298 +11,298 @@
11
  "log_history": [
12
  {
13
  "epoch": 0.05098139179199592,
14
- "grad_norm": 4.590001106262207,
15
- "learning_rate": 5.609065155807366e-06,
16
- "loss": 15.86372314453125,
17
  "step": 100
18
  },
19
  {
20
  "epoch": 0.10196278358399184,
21
- "grad_norm": 5.000253677368164,
22
- "learning_rate": 1.1274787535410765e-05,
23
- "loss": 15.6683056640625,
24
  "step": 200
25
  },
26
  {
27
  "epoch": 0.15294417537598776,
28
- "grad_norm": 5.164661407470703,
29
- "learning_rate": 1.6940509915014164e-05,
30
- "loss": 15.58547607421875,
31
  "step": 300
32
  },
33
  {
34
  "epoch": 0.20392556716798368,
35
- "grad_norm": 4.895200729370117,
36
- "learning_rate": 1.999658933249201e-05,
37
- "loss": 15.5261376953125,
38
  "step": 400
39
  },
40
  {
41
  "epoch": 0.2549069589599796,
42
- "grad_norm": 5.010247707366943,
43
- "learning_rate": 1.9965659596003744e-05,
44
- "loss": 15.493291015625,
45
  "step": 500
46
  },
47
  {
48
  "epoch": 0.3058883507519755,
49
- "grad_norm": 4.85853910446167,
50
- "learning_rate": 1.990261043359342e-05,
51
- "loss": 15.43971435546875,
52
  "step": 600
53
  },
54
  {
55
  "epoch": 0.35686974254397147,
56
- "grad_norm": 4.788653373718262,
57
- "learning_rate": 1.9807645053376055e-05,
58
- "loss": 15.409666748046876,
59
  "step": 700
60
  },
61
  {
62
  "epoch": 0.40785113433596737,
63
- "grad_norm": 4.742185592651367,
64
- "learning_rate": 1.968106952977309e-05,
65
- "loss": 15.346304931640624,
66
  "step": 800
67
  },
68
  {
69
  "epoch": 0.45883252612796327,
70
- "grad_norm": 4.758422374725342,
71
- "learning_rate": 1.9523291817031276e-05,
72
- "loss": 15.344024658203125,
73
  "step": 900
74
  },
75
  {
76
  "epoch": 0.5098139179199592,
77
- "grad_norm": 4.854381084442139,
78
- "learning_rate": 1.933482043438185e-05,
79
- "loss": 15.307811279296875,
80
  "step": 1000
81
  },
82
  {
83
  "epoch": 0.5607953097119551,
84
- "grad_norm": 4.7934041023254395,
85
- "learning_rate": 1.9116262827077703e-05,
86
- "loss": 15.254422607421875,
87
  "step": 1100
88
  },
89
  {
90
  "epoch": 0.611776701503951,
91
- "grad_norm": 4.670731544494629,
92
- "learning_rate": 1.88683234085909e-05,
93
- "loss": 15.23345703125,
94
  "step": 1200
95
  },
96
  {
97
  "epoch": 0.6627580932959469,
98
- "grad_norm": 4.993561267852783,
99
- "learning_rate": 1.8591801290280664e-05,
100
- "loss": 15.2450927734375,
101
  "step": 1300
102
  },
103
  {
104
  "epoch": 0.7137394850879429,
105
- "grad_norm": 4.720964431762695,
106
- "learning_rate": 1.8287587705849013e-05,
107
- "loss": 15.1839599609375,
108
  "step": 1400
109
  },
110
  {
111
  "epoch": 0.7647208768799388,
112
- "grad_norm": 5.050419330596924,
113
- "learning_rate": 1.7956663138885173e-05,
114
- "loss": 15.164833984375,
115
  "step": 1500
116
  },
117
  {
118
  "epoch": 0.8157022686719347,
119
- "grad_norm": 4.826648712158203,
120
- "learning_rate": 1.760009416275661e-05,
121
- "loss": 15.130496826171875,
122
  "step": 1600
123
  },
124
  {
125
  "epoch": 0.8666836604639306,
126
- "grad_norm": 4.858438014984131,
127
- "learning_rate": 1.721903000303185e-05,
128
- "loss": 15.125797119140625,
129
  "step": 1700
130
  },
131
  {
132
  "epoch": 0.9176650522559265,
133
- "grad_norm": 4.9611430168151855,
134
- "learning_rate": 1.6814698833514326e-05,
135
- "loss": 15.13617431640625,
136
  "step": 1800
137
  },
138
  {
139
  "epoch": 0.9686464440479226,
140
- "grad_norm": 4.663859844207764,
141
- "learning_rate": 1.63884038178253e-05,
142
- "loss": 15.072591552734375,
143
  "step": 1900
144
  },
145
  {
146
  "epoch": 1.0,
147
- "eval_loss": 3.636704444885254,
148
- "eval_runtime": 8.0138,
149
- "eval_samples_per_second": 632.91,
150
- "eval_steps_per_second": 9.983,
151
  "step": 1962
152
  },
153
  {
154
  "epoch": 1.0193729288809585,
155
- "grad_norm": 4.863068103790283,
156
- "learning_rate": 1.5941518909293737e-05,
157
- "loss": 14.968798828125,
158
  "step": 2000
159
  },
160
  {
161
  "epoch": 1.0703543206729544,
162
- "grad_norm": 5.036495685577393,
163
- "learning_rate": 1.5475484422690282e-05,
164
- "loss": 15.0290869140625,
165
  "step": 2100
166
  },
167
  {
168
  "epoch": 1.1213357124649503,
169
- "grad_norm": 5.248174667358398,
170
- "learning_rate": 1.4991802392077543e-05,
171
- "loss": 15.004036865234376,
172
  "step": 2200
173
  },
174
  {
175
  "epoch": 1.1723171042569462,
176
- "grad_norm": 4.950564384460449,
177
- "learning_rate": 1.4492031729738489e-05,
178
- "loss": 15.002611083984375,
179
  "step": 2300
180
  },
181
  {
182
  "epoch": 1.2232984960489421,
183
- "grad_norm": 4.509192943572998,
184
- "learning_rate": 1.3977783201785732e-05,
185
- "loss": 14.96060302734375,
186
  "step": 2400
187
  },
188
  {
189
  "epoch": 1.274279887840938,
190
- "grad_norm": 4.900182723999023,
191
- "learning_rate": 1.3450714236645352e-05,
192
- "loss": 14.971297607421874,
193
  "step": 2500
194
  },
195
  {
196
  "epoch": 1.325261279632934,
197
- "grad_norm": 5.138764381408691,
198
- "learning_rate": 1.2912523583147625e-05,
199
- "loss": 14.928385009765625,
200
  "step": 2600
201
  },
202
  {
203
  "epoch": 1.3762426714249298,
204
- "grad_norm": 4.894199848175049,
205
- "learning_rate": 1.2364945835441636e-05,
206
- "loss": 14.938167724609375,
207
  "step": 2700
208
  },
209
  {
210
  "epoch": 1.4272240632169257,
211
- "grad_norm": 4.8737921714782715,
212
- "learning_rate": 1.1809745842380042e-05,
213
- "loss": 14.923902587890625,
214
  "step": 2800
215
  },
216
  {
217
  "epoch": 1.4782054550089216,
218
- "grad_norm": 4.8258819580078125,
219
- "learning_rate": 1.1248713019392635e-05,
220
- "loss": 14.89677001953125,
221
  "step": 2900
222
  },
223
  {
224
  "epoch": 1.5291868468009175,
225
- "grad_norm": 4.769787788391113,
226
- "learning_rate": 1.0683655581181524e-05,
227
- "loss": 14.87692626953125,
228
  "step": 3000
229
  },
230
  {
231
  "epoch": 1.5801682385929134,
232
- "grad_norm": 4.92316198348999,
233
- "learning_rate": 1.0116394713826117e-05,
234
- "loss": 14.849693603515625,
235
  "step": 3100
236
  },
237
  {
238
  "epoch": 1.6311496303849093,
239
- "grad_norm": 4.873258590698242,
240
- "learning_rate": 9.548758705081177e-06,
241
- "loss": 14.833634033203126,
242
  "step": 3200
243
  },
244
  {
245
  "epoch": 1.6821310221769055,
246
- "grad_norm": 4.738825798034668,
247
- "learning_rate": 8.98257705178612e-06,
248
- "loss": 14.85665283203125,
249
  "step": 3300
250
  },
251
  {
252
  "epoch": 1.7331124139689014,
253
- "grad_norm": 4.907736778259277,
254
- "learning_rate": 8.419674563377416e-06,
255
- "loss": 14.8664599609375,
256
  "step": 3400
257
  },
258
  {
259
  "epoch": 1.7840938057608973,
260
- "grad_norm": 4.977413177490234,
261
- "learning_rate": 7.861865480508541e-06,
262
- "loss": 14.83008056640625,
263
  "step": 3500
264
  },
265
  {
266
  "epoch": 1.8350751975528932,
267
- "grad_norm": 4.792273044586182,
268
- "learning_rate": 7.310947627733231e-06,
269
- "loss": 14.81404541015625,
270
  "step": 3600
271
  },
272
  {
273
  "epoch": 1.886056589344889,
274
- "grad_norm": 4.84648323059082,
275
- "learning_rate": 6.768696619097996e-06,
276
- "loss": 14.831793212890625,
277
  "step": 3700
278
  },
279
  {
280
  "epoch": 1.9370379811368852,
281
- "grad_norm": 4.854404449462891,
282
- "learning_rate": 6.236860135319321e-06,
283
- "loss": 14.826976318359375,
284
  "step": 3800
285
  },
286
  {
287
  "epoch": 1.988019372928881,
288
- "grad_norm": 4.615888595581055,
289
- "learning_rate": 5.717152290990302e-06,
290
- "loss": 14.767562255859374,
291
  "step": 3900
292
  },
293
  {
294
  "epoch": 2.0,
295
- "eval_loss": 3.56946063041687,
296
- "eval_runtime": 8.0481,
297
- "eval_samples_per_second": 630.208,
298
- "eval_steps_per_second": 9.94,
299
  "step": 3924
300
  }
301
  ],
302
  "logging_steps": 100,
303
- "max_steps": 5886,
304
  "num_input_tokens_seen": 0,
305
- "num_train_epochs": 3,
306
  "save_steps": 1962,
307
  "stateful_callbacks": {
308
  "EarlyStoppingCallback": {
@@ -320,7 +320,7 @@
320
  "should_evaluate": false,
321
  "should_log": false,
322
  "should_save": true,
323
- "should_training_stop": false
324
  },
325
  "attributes": {}
326
  }
 
1
  {
2
  "best_global_step": 3924,
3
+ "best_metric": 3.390385389328003,
4
+ "best_model_checkpoint": "sindhibert_session5/checkpoint-3924",
5
  "epoch": 2.0,
6
  "eval_steps": 1962,
7
  "global_step": 3924,
 
11
  "log_history": [
12
  {
13
  "epoch": 0.05098139179199592,
14
+ "grad_norm": 5.245308876037598,
15
+ "learning_rate": 2.1063829787234044e-06,
16
+ "loss": 14.123677978515625,
17
  "step": 100
18
  },
19
  {
20
  "epoch": 0.10196278358399184,
21
+ "grad_norm": 5.646098613739014,
22
+ "learning_rate": 4.23404255319149e-06,
23
+ "loss": 14.066470947265625,
24
  "step": 200
25
  },
26
  {
27
  "epoch": 0.15294417537598776,
28
+ "grad_norm": 5.477426528930664,
29
+ "learning_rate": 4.99628768445363e-06,
30
+ "loss": 14.01585693359375,
31
  "step": 300
32
  },
33
  {
34
  "epoch": 0.20392556716798368,
35
+ "grad_norm": 5.2961344718933105,
36
+ "learning_rate": 4.975657001043815e-06,
37
+ "loss": 13.965966796875,
38
  "step": 400
39
  },
40
  {
41
  "epoch": 0.2549069589599796,
42
+ "grad_norm": 5.533503532409668,
43
+ "learning_rate": 4.937082682936824e-06,
44
+ "loss": 13.97388671875,
45
  "step": 500
46
  },
47
  {
48
  "epoch": 0.3058883507519755,
49
+ "grad_norm": 5.314619064331055,
50
+ "learning_rate": 4.8808443179239025e-06,
51
+ "loss": 13.95988525390625,
52
  "step": 600
53
  },
54
  {
55
  "epoch": 0.35686974254397147,
56
+ "grad_norm": 5.3324384689331055,
57
+ "learning_rate": 4.807349523327375e-06,
58
+ "loss": 13.953388671875,
59
  "step": 700
60
  },
61
  {
62
  "epoch": 0.40785113433596737,
63
+ "grad_norm": 5.467737674713135,
64
+ "learning_rate": 4.717130991578312e-06,
65
+ "loss": 13.9411962890625,
66
  "step": 800
67
  },
68
  {
69
  "epoch": 0.45883252612796327,
70
+ "grad_norm": 5.45510721206665,
71
+ "learning_rate": 4.610842629246138e-06,
72
+ "loss": 13.934332275390625,
73
  "step": 900
74
  },
75
  {
76
  "epoch": 0.5098139179199592,
77
+ "grad_norm": 5.6341962814331055,
78
+ "learning_rate": 4.489254817504595e-06,
79
+ "loss": 13.926766357421876,
80
  "step": 1000
81
  },
82
  {
83
  "epoch": 0.5607953097119551,
84
+ "grad_norm": 5.584266662597656,
85
+ "learning_rate": 4.353248828386359e-06,
86
+ "loss": 13.91165283203125,
87
  "step": 1100
88
  },
89
  {
90
  "epoch": 0.611776701503951,
91
+ "grad_norm": 5.35775899887085,
92
+ "learning_rate": 4.203810437297447e-06,
93
+ "loss": 13.915504150390625,
94
  "step": 1200
95
  },
96
  {
97
  "epoch": 0.6627580932959469,
98
+ "grad_norm": 5.549437999725342,
99
+ "learning_rate": 4.042022778088111e-06,
100
+ "loss": 13.945462646484375,
101
  "step": 1300
102
  },
103
  {
104
  "epoch": 0.7137394850879429,
105
+ "grad_norm": 5.455720901489258,
106
+ "learning_rate": 3.869058492466847e-06,
107
+ "loss": 13.89906494140625,
108
  "step": 1400
109
  },
110
  {
111
  "epoch": 0.7647208768799388,
112
+ "grad_norm": 5.856232166290283,
113
+ "learning_rate": 3.6861712306588434e-06,
114
+ "loss": 13.905211181640626,
115
  "step": 1500
116
  },
117
  {
118
  "epoch": 0.8157022686719347,
119
+ "grad_norm": 5.481319904327393,
120
+ "learning_rate": 3.4946865649123025e-06,
121
+ "loss": 13.896219482421875,
122
  "step": 1600
123
  },
124
  {
125
  "epoch": 0.8666836604639306,
126
+ "grad_norm": 5.278250217437744,
127
+ "learning_rate": 3.2959923817118296e-06,
128
+ "loss": 13.906019287109375,
129
  "step": 1700
130
  },
131
  {
132
  "epoch": 0.9176650522559265,
133
+ "grad_norm": 5.768893718719482,
134
+ "learning_rate": 3.091528822336405e-06,
135
+ "loss": 13.93161376953125,
136
  "step": 1800
137
  },
138
  {
139
  "epoch": 0.9686464440479226,
140
+ "grad_norm": 5.366046905517578,
141
+ "learning_rate": 2.8827778446730764e-06,
142
+ "loss": 13.906478271484374,
143
  "step": 1900
144
  },
145
  {
146
  "epoch": 1.0,
147
+ "eval_loss": 3.396049737930298,
148
+ "eval_runtime": 7.952,
149
+ "eval_samples_per_second": 637.827,
150
+ "eval_steps_per_second": 10.06,
151
  "step": 1962
152
  },
153
  {
154
  "epoch": 1.0193729288809585,
155
+ "grad_norm": 5.584351062774658,
156
+ "learning_rate": 2.6712524819426355e-06,
157
+ "loss": 13.823892822265625,
158
  "step": 2000
159
  },
160
  {
161
  "epoch": 1.0703543206729544,
162
+ "grad_norm": 5.338160037994385,
163
+ "learning_rate": 2.458485876190375e-06,
164
+ "loss": 13.90267333984375,
165
  "step": 2100
166
  },
167
  {
168
  "epoch": 1.1213357124649503,
169
+ "grad_norm": 5.378915309906006,
170
+ "learning_rate": 2.246020166027488e-06,
171
+ "loss": 13.9045458984375,
172
  "step": 2200
173
  },
174
  {
175
  "epoch": 1.1723171042569462,
176
+ "grad_norm": 5.420534133911133,
177
+ "learning_rate": 2.0353953091650912e-06,
178
+ "loss": 13.91886474609375,
179
  "step": 2300
180
  },
181
  {
182
  "epoch": 1.2232984960489421,
183
+ "grad_norm": 5.299575328826904,
184
+ "learning_rate": 1.82813792075548e-06,
185
+ "loss": 13.879603271484376,
186
  "step": 2400
187
  },
188
  {
189
  "epoch": 1.274279887840938,
190
+ "grad_norm": 5.256579875946045,
191
+ "learning_rate": 1.6257502084406458e-06,
192
+ "loss": 13.92048828125,
193
  "step": 2500
194
  },
195
  {
196
  "epoch": 1.325261279632934,
197
+ "grad_norm": 5.846710205078125,
198
+ "learning_rate": 1.4296990843071351e-06,
199
+ "loss": 13.912366943359375,
200
  "step": 2600
201
  },
202
  {
203
  "epoch": 1.3762426714249298,
204
+ "grad_norm": 5.547558307647705,
205
+ "learning_rate": 1.2414055326641378e-06,
206
+ "loss": 13.929051513671874,
207
  "step": 2700
208
  },
209
  {
210
  "epoch": 1.4272240632169257,
211
+ "grad_norm": 5.693734169006348,
212
+ "learning_rate": 1.062234310707458e-06,
213
+ "loss": 13.929571533203125,
214
  "step": 2800
215
  },
216
  {
217
  "epoch": 1.4782054550089216,
218
+ "grad_norm": 5.414492607116699,
219
+ "learning_rate": 8.934840567192496e-07,
220
+ "loss": 13.91981201171875,
221
  "step": 2900
222
  },
223
  {
224
  "epoch": 1.5291868468009175,
225
+ "grad_norm": 5.501009941101074,
226
+ "learning_rate": 7.363778774995834e-07,
227
+ "loss": 13.91615478515625,
228
  "step": 3000
229
  },
230
  {
231
  "epoch": 1.5801682385929134,
232
+ "grad_norm": 5.355737686157227,
233
+ "learning_rate": 5.920544832524166e-07,
234
+ "loss": 13.9007421875,
235
  "step": 3100
236
  },
237
  {
238
  "epoch": 1.6311496303849093,
239
+ "grad_norm": 5.376833915710449,
240
+ "learning_rate": 4.61559934180571e-07,
241
+ "loss": 13.915396728515624,
242
  "step": 3200
243
  },
244
  {
245
  "epoch": 1.6821310221769055,
246
+ "grad_norm": 5.302165985107422,
247
+ "learning_rate": 3.4584005861063974e-07,
248
+ "loss": 13.94998046875,
249
  "step": 3300
250
  },
251
  {
252
  "epoch": 1.7331124139689014,
253
+ "grad_norm": 5.571377277374268,
254
+ "learning_rate": 2.4573359760148354e-07,
255
+ "loss": 13.98015380859375,
256
  "step": 3400
257
  },
258
  {
259
  "epoch": 1.7840938057608973,
260
+ "grad_norm": 5.618070125579834,
261
+ "learning_rate": 1.619661257243843e-07,
262
+ "loss": 13.96041015625,
263
  "step": 3500
264
  },
265
  {
266
  "epoch": 1.8350751975528932,
267
+ "grad_norm": 5.216390132904053,
268
+ "learning_rate": 9.514479207723276e-08,
269
+ "loss": 13.9774560546875,
270
  "step": 3600
271
  },
272
  {
273
  "epoch": 1.886056589344889,
274
+ "grad_norm": 5.458727836608887,
275
+ "learning_rate": 4.575391965002174e-08,
276
+ "loss": 14.00027587890625,
277
  "step": 3700
278
  },
279
  {
280
  "epoch": 1.9370379811368852,
281
+ "grad_norm": 5.513509273529053,
282
+ "learning_rate": 1.4151494937582056e-08,
283
+ "loss": 14.0021728515625,
284
  "step": 3800
285
  },
286
  {
287
  "epoch": 1.988019372928881,
288
+ "grad_norm": 5.200069904327393,
289
+ "learning_rate": 5.665732429496751e-10,
290
+ "loss": 13.960343017578126,
291
  "step": 3900
292
  },
293
  {
294
  "epoch": 2.0,
295
+ "eval_loss": 3.390385389328003,
296
+ "eval_runtime": 7.9843,
297
+ "eval_samples_per_second": 635.249,
298
+ "eval_steps_per_second": 10.02,
299
  "step": 3924
300
  }
301
  ],
302
  "logging_steps": 100,
303
+ "max_steps": 3924,
304
  "num_input_tokens_seen": 0,
305
+ "num_train_epochs": 2,
306
  "save_steps": 1962,
307
  "stateful_callbacks": {
308
  "EarlyStoppingCallback": {
 
320
  "should_evaluate": false,
321
  "should_log": false,
322
  "should_save": true,
323
+ "should_training_stop": true
324
  },
325
  "attributes": {}
326
  }
checkpoint-3924/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:accc825ca2e280888c9eed825fcb7985700c1fb466ed8b16208ff9e7b14f1318
3
  size 5137
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a0a0f3ce1e27fd1d47d30a38c39aa86503e7f6fb63f533cef4d2e9ac2db41014
3
  size 5137
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:170894fbff2599922589dc645dfc871455543fe1f1fa33d3381f8353cf0b2a5b
3
  size 442633860
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f0ec985a6dec33fba0018d1ffbc1fcc4317a2291a75a8d6fc24f269454e6581
3
  size 442633860
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:accc825ca2e280888c9eed825fcb7985700c1fb466ed8b16208ff9e7b14f1318
3
  size 5137
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a0a0f3ce1e27fd1d47d30a38c39aa86503e7f6fb63f533cef4d2e9ac2db41014
3
  size 5137