junaid1993 commited on
Commit
1221d75
·
verified ·
1 Parent(s): 57ebaf1

Upload 8 files

Browse files
output/checkpoint-34/config.json ADDED
@@ -0,0 +1,74 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "ibm-granite/granite-timeseries-ttm-r2",
3
+ "adaptive_patching_levels": 0,
4
+ "architectures": [
5
+ "TinyTimeMixerForPrediction"
6
+ ],
7
+ "categorical_vocab_size_list": [
8
+ 1
9
+ ],
10
+ "context_length": 90,
11
+ "d_model": 63,
12
+ "d_model_scale": 7,
13
+ "decoder_adaptive_patching_levels": 0,
14
+ "decoder_d_model": 18,
15
+ "decoder_d_model_scale": 2,
16
+ "decoder_mode": "mix_channel",
17
+ "decoder_num_layers": 4,
18
+ "decoder_raw_residual": false,
19
+ "distribution_output": "student_t",
20
+ "dropout": 0.4,
21
+ "enable_forecast_channel_mixing": true,
22
+ "exogenous_channel_indices": [
23
+ 1,
24
+ 2,
25
+ 3,
26
+ 4
27
+ ],
28
+ "expansion_factor": 2,
29
+ "fcm_context_length": 1,
30
+ "fcm_gated_attn": true,
31
+ "fcm_mix_layers": 2,
32
+ "fcm_prepend_past": true,
33
+ "fcm_prepend_past_offset": null,
34
+ "fcm_use_mixer": true,
35
+ "frequency_token_vocab_size": 10,
36
+ "gated_attn": true,
37
+ "head_dropout": 0.4,
38
+ "huber_delta": 1,
39
+ "init_embed": "pytorch",
40
+ "init_linear": "pytorch",
41
+ "init_processing": true,
42
+ "init_std": 0.02,
43
+ "loss": "mae",
44
+ "mask_value": 0,
45
+ "masked_context_length": null,
46
+ "mode": "common_channel",
47
+ "model_type": "tinytimemixer",
48
+ "norm_eps": 1e-05,
49
+ "norm_mlp": "LayerNorm",
50
+ "num_input_channels": 5,
51
+ "num_layers": 20,
52
+ "num_parallel_samples": 100,
53
+ "num_patches": 11,
54
+ "patch_last": true,
55
+ "patch_length": 9,
56
+ "patch_stride": 9,
57
+ "positional_encoding_type": "sincos",
58
+ "post_init": false,
59
+ "prediction_channel_indices": [
60
+ 0
61
+ ],
62
+ "prediction_filter_length": 28,
63
+ "prediction_length": 30,
64
+ "quantile": 0.5,
65
+ "resolution_prefix_tuning": true,
66
+ "scaling": "std",
67
+ "self_attn": false,
68
+ "self_attn_heads": 1,
69
+ "stride_ratio": 1,
70
+ "torch_dtype": "float32",
71
+ "transformers_version": "4.48.3",
72
+ "use_decoder": true,
73
+ "use_positional_encoding": false
74
+ }
output/checkpoint-34/generation_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "transformers_version": "4.48.3"
4
+ }
output/checkpoint-34/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:949093ec090b44d4b429b870d56c4944b3298dadd4599afa5f638186b7490078
3
+ size 1834736
output/checkpoint-34/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b0ec42ce212f54e8bac7fb7303202a526e766b2ddfda4b8757b4badff31aa5a2
3
+ size 3899848
output/checkpoint-34/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d185de07776e1c4a03e3c3fbc4d13ea5b2b45bbc93841e7cfc3e1a0ca50151c
3
+ size 14244
output/checkpoint-34/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b6b54b488d5f85ef29e763c79e982987c330419d5ac28eb029d53d5d5cbd910
3
+ size 1320
output/checkpoint-34/trainer_state.json ADDED
@@ -0,0 +1,297 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.36268436908721924,
3
+ "best_model_checkpoint": "./ttm_finetuned_models_moderna/output/checkpoint-34",
4
+ "epoch": 17.0,
5
+ "eval_steps": 500,
6
+ "global_step": 34,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 1.0,
13
+ "grad_norm": 4.056020736694336,
14
+ "learning_rate": 1.8667028184439364e-05,
15
+ "loss": 1.0024,
16
+ "step": 2
17
+ },
18
+ {
19
+ "epoch": 1.0,
20
+ "eval_loss": 0.4427066445350647,
21
+ "eval_runtime": 0.3723,
22
+ "eval_samples_per_second": 185.339,
23
+ "eval_steps_per_second": 2.686,
24
+ "step": 2
25
+ },
26
+ {
27
+ "epoch": 2.0,
28
+ "grad_norm": 2.923017740249634,
29
+ "learning_rate": 1.8968955754218564e-05,
30
+ "loss": 0.9065,
31
+ "step": 4
32
+ },
33
+ {
34
+ "epoch": 2.0,
35
+ "eval_loss": 0.40708231925964355,
36
+ "eval_runtime": 0.361,
37
+ "eval_samples_per_second": 191.131,
38
+ "eval_steps_per_second": 2.77,
39
+ "step": 4
40
+ },
41
+ {
42
+ "epoch": 3.0,
43
+ "grad_norm": 1.819862723350525,
44
+ "learning_rate": 1.9471865185452704e-05,
45
+ "loss": 0.8237,
46
+ "step": 6
47
+ },
48
+ {
49
+ "epoch": 3.0,
50
+ "eval_loss": 0.393745094537735,
51
+ "eval_runtime": 0.3519,
52
+ "eval_samples_per_second": 196.077,
53
+ "eval_steps_per_second": 2.842,
54
+ "step": 6
55
+ },
56
+ {
57
+ "epoch": 4.0,
58
+ "grad_norm": 1.4791889190673828,
59
+ "learning_rate": 2.017530198824305e-05,
60
+ "loss": 0.7897,
61
+ "step": 8
62
+ },
63
+ {
64
+ "epoch": 4.0,
65
+ "eval_loss": 0.38901111483573914,
66
+ "eval_runtime": 0.3611,
67
+ "eval_samples_per_second": 191.097,
68
+ "eval_steps_per_second": 2.77,
69
+ "step": 8
70
+ },
71
+ {
72
+ "epoch": 5.0,
73
+ "grad_norm": 1.0808687210083008,
74
+ "learning_rate": 2.1078630451860426e-05,
75
+ "loss": 0.7637,
76
+ "step": 10
77
+ },
78
+ {
79
+ "epoch": 5.0,
80
+ "eval_loss": 0.3875712752342224,
81
+ "eval_runtime": 0.3339,
82
+ "eval_samples_per_second": 206.657,
83
+ "eval_steps_per_second": 2.995,
84
+ "step": 10
85
+ },
86
+ {
87
+ "epoch": 6.0,
88
+ "grad_norm": 0.7717658877372742,
89
+ "learning_rate": 2.2181034219250504e-05,
90
+ "loss": 0.761,
91
+ "step": 12
92
+ },
93
+ {
94
+ "epoch": 6.0,
95
+ "eval_loss": 0.38537338376045227,
96
+ "eval_runtime": 0.3572,
97
+ "eval_samples_per_second": 193.195,
98
+ "eval_steps_per_second": 2.8,
99
+ "step": 12
100
+ },
101
+ {
102
+ "epoch": 7.0,
103
+ "grad_norm": 0.70772784948349,
104
+ "learning_rate": 2.3481517024792826e-05,
105
+ "loss": 0.7687,
106
+ "step": 14
107
+ },
108
+ {
109
+ "epoch": 7.0,
110
+ "eval_loss": 0.3826580345630646,
111
+ "eval_runtime": 0.3504,
112
+ "eval_samples_per_second": 196.904,
113
+ "eval_steps_per_second": 2.854,
114
+ "step": 14
115
+ },
116
+ {
117
+ "epoch": 8.0,
118
+ "grad_norm": 0.8968775272369385,
119
+ "learning_rate": 2.4978903594647175e-05,
120
+ "loss": 0.7411,
121
+ "step": 16
122
+ },
123
+ {
124
+ "epoch": 8.0,
125
+ "eval_loss": 0.37533628940582275,
126
+ "eval_runtime": 0.359,
127
+ "eval_samples_per_second": 192.223,
128
+ "eval_steps_per_second": 2.786,
129
+ "step": 16
130
+ },
131
+ {
132
+ "epoch": 9.0,
133
+ "grad_norm": 0.8349848389625549,
134
+ "learning_rate": 2.6671840708873478e-05,
135
+ "loss": 0.7507,
136
+ "step": 18
137
+ },
138
+ {
139
+ "epoch": 9.0,
140
+ "eval_loss": 0.3686024844646454,
141
+ "eval_runtime": 0.359,
142
+ "eval_samples_per_second": 192.224,
143
+ "eval_steps_per_second": 2.786,
144
+ "step": 18
145
+ },
146
+ {
147
+ "epoch": 10.0,
148
+ "grad_norm": 0.8916922807693481,
149
+ "learning_rate": 2.8558798424365177e-05,
150
+ "loss": 0.7418,
151
+ "step": 20
152
+ },
153
+ {
154
+ "epoch": 10.0,
155
+ "eval_loss": 0.36738961935043335,
156
+ "eval_runtime": 0.7302,
157
+ "eval_samples_per_second": 94.489,
158
+ "eval_steps_per_second": 1.369,
159
+ "step": 20
160
+ },
161
+ {
162
+ "epoch": 11.0,
163
+ "grad_norm": 0.632808268070221,
164
+ "learning_rate": 3.0638071457491495e-05,
165
+ "loss": 0.7092,
166
+ "step": 22
167
+ },
168
+ {
169
+ "epoch": 11.0,
170
+ "eval_loss": 0.3688909411430359,
171
+ "eval_runtime": 0.3526,
172
+ "eval_samples_per_second": 195.693,
173
+ "eval_steps_per_second": 2.836,
174
+ "step": 22
175
+ },
176
+ {
177
+ "epoch": 12.0,
178
+ "grad_norm": 0.6958911418914795,
179
+ "learning_rate": 3.290778072519786e-05,
180
+ "loss": 0.7476,
181
+ "step": 24
182
+ },
183
+ {
184
+ "epoch": 12.0,
185
+ "eval_loss": 0.3743250370025635,
186
+ "eval_runtime": 0.3634,
187
+ "eval_samples_per_second": 189.876,
188
+ "eval_steps_per_second": 2.752,
189
+ "step": 24
190
+ },
191
+ {
192
+ "epoch": 13.0,
193
+ "grad_norm": 0.8087581992149353,
194
+ "learning_rate": 3.5365875043173484e-05,
195
+ "loss": 0.7315,
196
+ "step": 26
197
+ },
198
+ {
199
+ "epoch": 13.0,
200
+ "eval_loss": 0.375684916973114,
201
+ "eval_runtime": 0.3555,
202
+ "eval_samples_per_second": 194.088,
203
+ "eval_steps_per_second": 2.813,
204
+ "step": 26
205
+ },
206
+ {
207
+ "epoch": 14.0,
208
+ "grad_norm": 0.8785252571105957,
209
+ "learning_rate": 3.801013297955e-05,
210
+ "loss": 0.7262,
211
+ "step": 28
212
+ },
213
+ {
214
+ "epoch": 14.0,
215
+ "eval_loss": 0.3696945905685425,
216
+ "eval_runtime": 0.3621,
217
+ "eval_samples_per_second": 190.574,
218
+ "eval_steps_per_second": 2.762,
219
+ "step": 28
220
+ },
221
+ {
222
+ "epoch": 15.0,
223
+ "grad_norm": 0.671592652797699,
224
+ "learning_rate": 4.083816486245606e-05,
225
+ "loss": 0.7121,
226
+ "step": 30
227
+ },
228
+ {
229
+ "epoch": 15.0,
230
+ "eval_loss": 0.3645818829536438,
231
+ "eval_runtime": 0.3601,
232
+ "eval_samples_per_second": 191.597,
233
+ "eval_steps_per_second": 2.777,
234
+ "step": 30
235
+ },
236
+ {
237
+ "epoch": 16.0,
238
+ "grad_norm": 0.6458890438079834,
239
+ "learning_rate": 4.3847414939614856e-05,
240
+ "loss": 0.737,
241
+ "step": 32
242
+ },
243
+ {
244
+ "epoch": 16.0,
245
+ "eval_loss": 0.36310243606567383,
246
+ "eval_runtime": 0.35,
247
+ "eval_samples_per_second": 197.168,
248
+ "eval_steps_per_second": 2.858,
249
+ "step": 32
250
+ },
251
+ {
252
+ "epoch": 17.0,
253
+ "grad_norm": 0.4557172954082489,
254
+ "learning_rate": 4.703516368803146e-05,
255
+ "loss": 0.707,
256
+ "step": 34
257
+ },
258
+ {
259
+ "epoch": 17.0,
260
+ "eval_loss": 0.36268436908721924,
261
+ "eval_runtime": 0.3494,
262
+ "eval_samples_per_second": 197.489,
263
+ "eval_steps_per_second": 2.862,
264
+ "step": 34
265
+ }
266
+ ],
267
+ "logging_steps": 500,
268
+ "max_steps": 100,
269
+ "num_input_tokens_seen": 0,
270
+ "num_train_epochs": 50,
271
+ "save_steps": 500,
272
+ "stateful_callbacks": {
273
+ "EarlyStoppingCallback": {
274
+ "args": {
275
+ "early_stopping_patience": 10,
276
+ "early_stopping_threshold": 0.0
277
+ },
278
+ "attributes": {
279
+ "early_stopping_patience_counter": 0
280
+ }
281
+ },
282
+ "TrainerControl": {
283
+ "args": {
284
+ "should_epoch_stop": false,
285
+ "should_evaluate": false,
286
+ "should_log": false,
287
+ "should_save": true,
288
+ "should_training_stop": false
289
+ },
290
+ "attributes": {}
291
+ }
292
+ },
293
+ "total_flos": 9002664340200.0,
294
+ "train_batch_size": 288,
295
+ "trial_name": null,
296
+ "trial_params": null
297
+ }
output/checkpoint-34/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f5c158e6e7d833b2ec1091d4b44cc99c194eae1b0e8d2bf20a020375c57091b4
3
+ size 5304