Yuvrajg2107 commited on
Commit
5da5437
·
verified ·
1 Parent(s): 531cbd0

Training in progress, step 7000, checkpoint

Browse files
last-checkpoint/config.json CHANGED
@@ -19,7 +19,7 @@
19
  "num_hidden_layers": 12,
20
  "pad_token_id": 1,
21
  "position_embedding_type": "absolute",
22
- "transformers_version": "4.57.1",
23
  "type_vocab_size": 1,
24
  "use_cache": true,
25
  "vocab_size": 50265
 
19
  "num_hidden_layers": 12,
20
  "pad_token_id": 1,
21
  "position_embedding_type": "absolute",
22
+ "transformers_version": "4.57.3",
23
  "type_vocab_size": 1,
24
  "use_cache": true,
25
  "vocab_size": 50265
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:41da70138a1773e4f5cd2778674af186f8e5ad4d5bfbd72c39a71b1b96bc5185
3
  size 498612824
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b0075e6716b391b423ea614a7409f90848b1fdf865d2dd524d7a1704b32111c
3
  size 498612824
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fcb38ea326cd284005454648e073cc75673ef6d3edbb0d25347b119555de5f08
3
  size 997345931
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:108ee5c91c852120064f3994a646e206fd464f66033eb94c64a32e953543d328
3
  size 997345931
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c7bbc1b39311f89c58ff1ea8b0baf3cad68be72482b939daed1a03b8c1807179
3
  size 1465
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5503916e083484344eb90ffdd7cef8671b9c273655dd45784cda2988a30dceb9
3
  size 1465
last-checkpoint/trainer_state.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "best_global_step": 6000,
3
- "best_metric": 0.06669048964977264,
4
- "best_model_checkpoint": "./training_output/checkpoint-6000",
5
- "epoch": 0.875,
6
  "eval_steps": 1000,
7
  "global_step": 7000,
8
  "is_hyper_param_search": false,
@@ -10,169 +10,169 @@
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
- "epoch": 0.0625,
14
- "grad_norm": 33.47822570800781,
15
- "learning_rate": 1.8752500000000004e-05,
16
- "loss": 0.1408,
17
  "step": 500
18
  },
19
  {
20
- "epoch": 0.125,
21
- "grad_norm": 66.1669692993164,
22
- "learning_rate": 1.7502500000000004e-05,
23
- "loss": 0.1244,
24
  "step": 1000
25
  },
26
  {
27
- "epoch": 0.125,
28
- "eval_accuracy": 0.931,
29
- "eval_loss": 0.08424179255962372,
30
- "eval_runtime": 148.5886,
31
- "eval_samples_per_second": 53.84,
32
- "eval_steps_per_second": 3.365,
33
  "step": 1000
34
  },
35
  {
36
- "epoch": 0.1875,
37
- "grad_norm": 40.86865234375,
38
- "learning_rate": 1.62525e-05,
39
- "loss": 0.083,
40
  "step": 1500
41
  },
42
  {
43
- "epoch": 0.25,
44
- "grad_norm": 0.046407029032707214,
45
- "learning_rate": 1.5002500000000002e-05,
46
- "loss": 0.0851,
47
  "step": 2000
48
  },
49
  {
50
- "epoch": 0.25,
51
- "eval_accuracy": 0.7715,
52
- "eval_loss": 0.20412930846214294,
53
- "eval_runtime": 149.0575,
54
- "eval_samples_per_second": 53.671,
55
- "eval_steps_per_second": 3.354,
56
  "step": 2000
57
  },
58
  {
59
- "epoch": 0.3125,
60
- "grad_norm": 0.12031727284193039,
61
- "learning_rate": 1.3752500000000003e-05,
62
- "loss": 0.0689,
63
  "step": 2500
64
  },
65
  {
66
- "epoch": 0.375,
67
- "grad_norm": 0.48724281787872314,
68
- "learning_rate": 1.2502500000000003e-05,
69
- "loss": 0.0632,
70
  "step": 3000
71
  },
72
  {
73
- "epoch": 0.375,
74
- "eval_accuracy": 0.91225,
75
- "eval_loss": 0.16098278760910034,
76
- "eval_runtime": 148.8996,
77
- "eval_samples_per_second": 53.727,
78
- "eval_steps_per_second": 3.358,
79
  "step": 3000
80
  },
81
  {
82
- "epoch": 0.4375,
83
- "grad_norm": 0.004755391739308834,
84
- "learning_rate": 1.1252500000000001e-05,
85
- "loss": 0.0525,
86
  "step": 3500
87
  },
88
  {
89
- "epoch": 0.5,
90
- "grad_norm": 0.2557525336742401,
91
- "learning_rate": 1.0002500000000001e-05,
92
- "loss": 0.0521,
93
  "step": 4000
94
  },
95
  {
96
- "epoch": 0.5,
97
- "eval_accuracy": 0.922375,
98
- "eval_loss": 0.09066133201122284,
99
- "eval_runtime": 148.5612,
100
- "eval_samples_per_second": 53.85,
101
- "eval_steps_per_second": 3.366,
102
  "step": 4000
103
  },
104
  {
105
- "epoch": 0.5625,
106
- "grad_norm": 0.641652524471283,
107
- "learning_rate": 8.7525e-06,
108
- "loss": 0.0463,
109
  "step": 4500
110
  },
111
  {
112
- "epoch": 0.625,
113
- "grad_norm": 0.1168695017695427,
114
- "learning_rate": 7.502500000000001e-06,
115
- "loss": 0.0426,
116
  "step": 5000
117
  },
118
  {
119
- "epoch": 0.625,
120
- "eval_accuracy": 0.878375,
121
- "eval_loss": 0.24468091130256653,
122
- "eval_runtime": 148.7322,
123
- "eval_samples_per_second": 53.788,
124
- "eval_steps_per_second": 3.362,
125
  "step": 5000
126
  },
127
  {
128
- "epoch": 0.6875,
129
- "grad_norm": 0.001182532636448741,
130
- "learning_rate": 6.2525e-06,
131
- "loss": 0.0286,
132
  "step": 5500
133
  },
134
  {
135
- "epoch": 0.75,
136
- "grad_norm": 0.0016105002723634243,
137
- "learning_rate": 5.0025e-06,
138
- "loss": 0.0341,
139
  "step": 6000
140
  },
141
  {
142
- "epoch": 0.75,
143
- "eval_accuracy": 0.964125,
144
- "eval_loss": 0.06669048964977264,
145
- "eval_runtime": 148.8829,
146
- "eval_samples_per_second": 53.733,
147
- "eval_steps_per_second": 3.358,
148
  "step": 6000
149
  },
150
  {
151
- "epoch": 0.8125,
152
- "grad_norm": 0.00258028507232666,
153
- "learning_rate": 3.7525e-06,
154
- "loss": 0.0337,
155
  "step": 6500
156
  },
157
  {
158
- "epoch": 0.875,
159
- "grad_norm": 0.005781837739050388,
160
- "learning_rate": 2.5024999999999998e-06,
161
- "loss": 0.0252,
162
  "step": 7000
163
  },
164
  {
165
- "epoch": 0.875,
166
- "eval_accuracy": 0.957,
167
- "eval_loss": 0.0872097983956337,
168
- "eval_runtime": 148.3454,
169
- "eval_samples_per_second": 53.928,
170
- "eval_steps_per_second": 3.371,
171
  "step": 7000
172
  }
173
  ],
174
  "logging_steps": 500,
175
- "max_steps": 8000,
176
  "num_input_tokens_seen": 0,
177
  "num_train_epochs": 1,
178
  "save_steps": 1000,
 
1
  {
2
+ "best_global_step": 4000,
3
+ "best_metric": 0.07567641884088516,
4
+ "best_model_checkpoint": "./training_output/checkpoint-4000",
5
+ "epoch": 0.35,
6
  "eval_steps": 1000,
7
  "global_step": 7000,
8
  "is_hyper_param_search": false,
 
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
+ "epoch": 0.025,
14
+ "grad_norm": 0.8834348917007446,
15
+ "learning_rate": 1.9501000000000002e-05,
16
+ "loss": 0.1437,
17
  "step": 500
18
  },
19
  {
20
+ "epoch": 0.05,
21
+ "grad_norm": 9.513919830322266,
22
+ "learning_rate": 1.9001e-05,
23
+ "loss": 0.1085,
24
  "step": 1000
25
  },
26
  {
27
+ "epoch": 0.05,
28
+ "eval_accuracy": 0.91835,
29
+ "eval_loss": 0.10802757740020752,
30
+ "eval_runtime": 381.1922,
31
+ "eval_samples_per_second": 52.467,
32
+ "eval_steps_per_second": 3.279,
33
  "step": 1000
34
  },
35
  {
36
+ "epoch": 0.075,
37
+ "grad_norm": 0.013394818641245365,
38
+ "learning_rate": 1.8501e-05,
39
+ "loss": 0.0965,
40
  "step": 1500
41
  },
42
  {
43
+ "epoch": 0.1,
44
+ "grad_norm": 0.017180055379867554,
45
+ "learning_rate": 1.8001000000000003e-05,
46
+ "loss": 0.0716,
47
  "step": 2000
48
  },
49
  {
50
+ "epoch": 0.1,
51
+ "eval_accuracy": 0.8473,
52
+ "eval_loss": 0.25237375497817993,
53
+ "eval_runtime": 381.1616,
54
+ "eval_samples_per_second": 52.471,
55
+ "eval_steps_per_second": 3.279,
56
  "step": 2000
57
  },
58
  {
59
+ "epoch": 0.125,
60
+ "grad_norm": 0.017619747668504715,
61
+ "learning_rate": 1.7501e-05,
62
+ "loss": 0.0658,
63
  "step": 2500
64
  },
65
  {
66
+ "epoch": 0.15,
67
+ "grad_norm": 0.527113676071167,
68
+ "learning_rate": 1.7001000000000002e-05,
69
+ "loss": 0.0615,
70
  "step": 3000
71
  },
72
  {
73
+ "epoch": 0.15,
74
+ "eval_accuracy": 0.9299,
75
+ "eval_loss": 0.11818733811378479,
76
+ "eval_runtime": 381.4889,
77
+ "eval_samples_per_second": 52.426,
78
+ "eval_steps_per_second": 3.277,
79
  "step": 3000
80
  },
81
  {
82
+ "epoch": 0.175,
83
+ "grad_norm": 140.78282165527344,
84
+ "learning_rate": 1.6501e-05,
85
+ "loss": 0.056,
86
  "step": 3500
87
  },
88
  {
89
+ "epoch": 0.2,
90
+ "grad_norm": 0.9989501237869263,
91
+ "learning_rate": 1.6001e-05,
92
+ "loss": 0.0648,
93
  "step": 4000
94
  },
95
  {
96
+ "epoch": 0.2,
97
+ "eval_accuracy": 0.9498,
98
+ "eval_loss": 0.07567641884088516,
99
+ "eval_runtime": 380.8034,
100
+ "eval_samples_per_second": 52.521,
101
+ "eval_steps_per_second": 3.283,
102
  "step": 4000
103
  },
104
  {
105
+ "epoch": 0.225,
106
+ "grad_norm": 0.01812303625047207,
107
+ "learning_rate": 1.5501000000000003e-05,
108
+ "loss": 0.0487,
109
  "step": 4500
110
  },
111
  {
112
+ "epoch": 0.25,
113
+ "grad_norm": 0.05552659556269646,
114
+ "learning_rate": 1.5001000000000001e-05,
115
+ "loss": 0.0522,
116
  "step": 5000
117
  },
118
  {
119
+ "epoch": 0.25,
120
+ "eval_accuracy": 0.92725,
121
+ "eval_loss": 0.12006673216819763,
122
+ "eval_runtime": 380.6188,
123
+ "eval_samples_per_second": 52.546,
124
+ "eval_steps_per_second": 3.284,
125
  "step": 5000
126
  },
127
  {
128
+ "epoch": 0.275,
129
+ "grad_norm": 0.14319103956222534,
130
+ "learning_rate": 1.4501e-05,
131
+ "loss": 0.0554,
132
  "step": 5500
133
  },
134
  {
135
+ "epoch": 0.3,
136
+ "grad_norm": 0.012562028132379055,
137
+ "learning_rate": 1.4001e-05,
138
+ "loss": 0.0377,
139
  "step": 6000
140
  },
141
  {
142
+ "epoch": 0.3,
143
+ "eval_accuracy": 0.95545,
144
+ "eval_loss": 0.08464282751083374,
145
+ "eval_runtime": 380.6212,
146
+ "eval_samples_per_second": 52.546,
147
+ "eval_steps_per_second": 3.284,
148
  "step": 6000
149
  },
150
  {
151
+ "epoch": 0.325,
152
+ "grad_norm": 0.0012805273290723562,
153
+ "learning_rate": 1.3501000000000002e-05,
154
+ "loss": 0.0327,
155
  "step": 6500
156
  },
157
  {
158
+ "epoch": 0.35,
159
+ "grad_norm": 0.024555200710892677,
160
+ "learning_rate": 1.3001000000000001e-05,
161
+ "loss": 0.0447,
162
  "step": 7000
163
  },
164
  {
165
+ "epoch": 0.35,
166
+ "eval_accuracy": 0.93225,
167
+ "eval_loss": 0.10355959832668304,
168
+ "eval_runtime": 380.8232,
169
+ "eval_samples_per_second": 52.518,
170
+ "eval_steps_per_second": 3.282,
171
  "step": 7000
172
  }
173
  ],
174
  "logging_steps": 500,
175
+ "max_steps": 20000,
176
  "num_input_tokens_seen": 0,
177
  "num_train_epochs": 1,
178
  "save_steps": 1000,
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5d1dcf0aa731896a45dc92ce645a0147a079fc2db258f2967bafe27b497eba6b
3
  size 5841
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4672bef08597eaac4166af409003080f023f24990f8f58c185b2b990119dc93e
3
  size 5841