Seynro commited on
Commit
09e6e0d
·
verified ·
1 Parent(s): f0da7ff

Training in progress, step 200, checkpoint

Browse files
last-checkpoint/config.json CHANGED
@@ -1,62 +1,39 @@
1
  {
2
- "_name_or_path": "google/flan-t5-base",
 
3
  "architectures": [
4
- "T5ForConditionalGeneration"
5
  ],
6
- "classifier_dropout": 0.0,
7
- "d_ff": 2048,
8
- "d_kv": 64,
9
- "d_model": 768,
10
- "decoder_start_token_id": 0,
11
- "dense_act_fn": "gelu_new",
12
- "dropout_rate": 0.1,
13
- "eos_token_id": 1,
14
- "feed_forward_proj": "gated-gelu",
15
- "initializer_factor": 1.0,
16
- "is_encoder_decoder": true,
17
- "is_gated_act": true,
18
- "layer_norm_epsilon": 1e-06,
19
- "model_type": "t5",
20
- "n_positions": 512,
21
- "num_decoder_layers": 12,
22
- "num_heads": 12,
23
- "num_layers": 12,
24
- "output_past": true,
25
- "pad_token_id": 0,
26
- "relative_attention_max_distance": 128,
27
- "relative_attention_num_buckets": 32,
28
  "task_specific_params": {
29
- "summarization": {
30
- "early_stopping": true,
31
- "length_penalty": 2.0,
32
- "max_length": 200,
33
- "min_length": 30,
34
- "no_repeat_ngram_size": 3,
35
- "num_beams": 4,
36
- "prefix": "summarize: "
37
- },
38
- "translation_en_to_de": {
39
- "early_stopping": true,
40
- "max_length": 300,
41
- "num_beams": 4,
42
- "prefix": "translate English to German: "
43
- },
44
- "translation_en_to_fr": {
45
- "early_stopping": true,
46
- "max_length": 300,
47
- "num_beams": 4,
48
- "prefix": "translate English to French: "
49
- },
50
- "translation_en_to_ro": {
51
- "early_stopping": true,
52
- "max_length": 300,
53
- "num_beams": 4,
54
- "prefix": "translate English to Romanian: "
55
  }
56
  },
57
- "tie_word_embeddings": false,
58
  "torch_dtype": "float32",
59
  "transformers_version": "4.49.0",
60
  "use_cache": true,
61
- "vocab_size": 32103
62
  }
 
1
  {
2
+ "_name_or_path": "allmalab/gpt2-aze",
3
+ "activation_function": "gelu_new",
4
  "architectures": [
5
+ "GPT2LMHeadModel"
6
  ],
7
+ "attn_pdrop": 0,
8
+ "bos_token_id": 11,
9
+ "embd_pdrop": 0,
10
+ "eos_token_id": 12,
11
+ "initializer_range": 0.02,
12
+ "layer_norm_epsilon": 1e-05,
13
+ "model_type": "gpt2",
14
+ "n_ctx": 1024,
15
+ "n_embd": 768,
16
+ "n_head": 12,
17
+ "n_inner": null,
18
+ "n_layer": 12,
19
+ "n_positions": 1024,
20
+ "reorder_and_upcast_attn": false,
21
+ "resid_pdrop": 0,
22
+ "scale_attn_by_inverse_layer_idx": false,
23
+ "scale_attn_weights": true,
24
+ "summary_activation": null,
25
+ "summary_first_dropout": 0,
26
+ "summary_proj_to_labels": true,
27
+ "summary_type": "cls_index",
28
+ "summary_use_proj": true,
29
  "task_specific_params": {
30
+ "text-generation": {
31
+ "do_sample": true,
32
+ "max_length": 50
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  }
34
  },
 
35
  "torch_dtype": "float32",
36
  "transformers_version": "4.49.0",
37
  "use_cache": true,
38
+ "vocab_size": 64003
39
  }
last-checkpoint/generation_config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
- "decoder_start_token_id": 0,
3
- "eos_token_id": 1,
4
- "pad_token_id": 0,
5
  "transformers_version": "4.49.0"
6
  }
 
1
  {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 11,
4
+ "eos_token_id": 12,
5
  "transformers_version": "4.49.0"
6
  }
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:779549eacf5a18b37319e66bf14475b161dd4433549e850971a454b5266d4eb0
3
- size 990191464
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:757ad62f4ea26e898b3b9b185d741bcfc0f7b27dc79d116ce13e3d2182299eb4
3
+ size 540001920
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8dccba7d5cd5e9593985197edaea56591d66d860b7cc48b065cdea5202db56bd
3
- size 1980553210
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:130c575e1cb9ab89e4dc38eac6917525ce5de05624065a5f7c77ffcf4ad891ca
3
+ size 1080097722
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:de7fb8a41dbba15b3b8bc11390e2d7ddfe33c799d6cf98b8039e15d400789fe3
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ffcb25428d146633ce210a40c9c5329ef07f7e8ca498959bffcf2c559d69eea
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:72c1b866fa528cbddaa344b3307a55102bdc990e6cf82a6d36d368c2a421c816
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae4ecdb1b8aee070470f06546bc65633fa4bdcb826c1df9b2e617eb80bcc34ab
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,177 +1,40 @@
1
  {
2
- "best_metric": 0.6030490398406982,
3
- "best_model_checkpoint": "./aztelecom_t5_finetuned/checkpoint-1000",
4
- "epoch": 10.0,
5
  "eval_steps": 500,
6
- "global_step": 1070,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.4672897196261682,
13
- "grad_norm": 95.7200927734375,
14
  "learning_rate": 1.4018691588785047e-05,
15
- "loss": 24.3127,
16
  "step": 50
17
  },
18
  {
19
  "epoch": 0.9345794392523364,
20
- "grad_norm": 8.241812705993652,
21
  "learning_rate": 2.8037383177570094e-05,
22
- "loss": 8.7035,
23
  "step": 100
24
  },
25
  {
26
  "epoch": 1.4018691588785046,
27
- "grad_norm": 6.154674530029297,
28
  "learning_rate": 2.7699933091326993e-05,
29
- "loss": 2.6262,
30
  "step": 150
31
  },
32
  {
33
  "epoch": 1.8691588785046729,
34
- "grad_norm": 1.0018048286437988,
35
  "learning_rate": 2.0245088641574458e-05,
36
- "loss": 1.4122,
37
  "step": 200
38
- },
39
- {
40
- "epoch": 2.336448598130841,
41
- "grad_norm": 0.5541874170303345,
42
- "learning_rate": 1.0635487465051559e-05,
43
- "loss": 1.0114,
44
- "step": 250
45
- },
46
- {
47
- "epoch": 2.803738317757009,
48
- "grad_norm": 0.4578002691268921,
49
- "learning_rate": 2.8188898746283444e-06,
50
- "loss": 0.9539,
51
- "step": 300
52
- },
53
- {
54
- "epoch": 3.2710280373831777,
55
- "grad_norm": 0.5540403723716736,
56
- "learning_rate": 6.464906595023967e-09,
57
- "loss": 0.9415,
58
- "step": 350
59
- },
60
- {
61
- "epoch": 3.7383177570093458,
62
- "grad_norm": 6.799459457397461,
63
- "learning_rate": 3.3535965844688068e-06,
64
- "loss": 0.9212,
65
- "step": 400
66
- },
67
- {
68
- "epoch": 4.205607476635514,
69
- "grad_norm": 0.5957446098327637,
70
- "learning_rate": 1.1485235792565021e-05,
71
- "loss": 0.8961,
72
- "step": 450
73
- },
74
- {
75
- "epoch": 4.672897196261682,
76
- "grad_norm": 0.44224458932876587,
77
- "learning_rate": 2.1060789934519968e-05,
78
- "loss": 0.8492,
79
- "step": 500
80
- },
81
- {
82
- "epoch": 4.672897196261682,
83
- "eval_generation_quality": 1.4625850340136055,
84
- "eval_loss": 0.7530518174171448,
85
- "eval_runtime": 287.5608,
86
- "eval_samples_per_second": 2.977,
87
- "eval_steps_per_second": 0.372,
88
- "step": 500
89
- },
90
- {
91
- "epoch": 5.140186915887851,
92
- "grad_norm": 0.5181770920753479,
93
- "learning_rate": 2.814648570164509e-05,
94
- "loss": 0.8135,
95
- "step": 550
96
- },
97
- {
98
- "epoch": 5.607476635514018,
99
- "grad_norm": 0.47956785559654236,
100
- "learning_rate": 2.9831418882133196e-05,
101
- "loss": 0.7584,
102
- "step": 600
103
- },
104
- {
105
- "epoch": 6.074766355140187,
106
- "grad_norm": 0.584481954574585,
107
- "learning_rate": 2.542339505317937e-05,
108
- "loss": 0.7438,
109
- "step": 650
110
- },
111
- {
112
- "epoch": 6.542056074766355,
113
- "grad_norm": 0.5437303781509399,
114
- "learning_rate": 1.673329285693197e-05,
115
- "loss": 0.6918,
116
- "step": 700
117
- },
118
- {
119
- "epoch": 7.009345794392523,
120
- "grad_norm": 0.5896743535995483,
121
- "learning_rate": 7.331129397632949e-06,
122
- "loss": 0.6872,
123
- "step": 750
124
- },
125
- {
126
- "epoch": 7.4766355140186915,
127
- "grad_norm": 0.37604203820228577,
128
- "learning_rate": 1.0794466349235933e-06,
129
- "loss": 0.6681,
130
- "step": 800
131
- },
132
- {
133
- "epoch": 7.94392523364486,
134
- "grad_norm": 0.4640524983406067,
135
- "learning_rate": 5.465244220378685e-07,
136
- "loss": 0.6794,
137
- "step": 850
138
- },
139
- {
140
- "epoch": 8.411214953271028,
141
- "grad_norm": 0.3652268052101135,
142
- "learning_rate": 5.951294754511376e-06,
143
- "loss": 0.669,
144
- "step": 900
145
- },
146
- {
147
- "epoch": 8.878504672897197,
148
- "grad_norm": 1.2286640405654907,
149
- "learning_rate": 1.5073401404570173e-05,
150
- "loss": 0.6918,
151
- "step": 950
152
- },
153
- {
154
- "epoch": 9.345794392523365,
155
- "grad_norm": 0.3932737708091736,
156
- "learning_rate": 2.4165353717674908e-05,
157
- "loss": 0.6632,
158
- "step": 1000
159
- },
160
- {
161
- "epoch": 9.345794392523365,
162
- "eval_generation_quality": 0.0,
163
- "eval_loss": 0.6030490398406982,
164
- "eval_runtime": 238.7271,
165
- "eval_samples_per_second": 3.586,
166
- "eval_steps_per_second": 0.448,
167
- "step": 1000
168
- },
169
- {
170
- "epoch": 9.813084112149532,
171
- "grad_norm": 0.5340766906738281,
172
- "learning_rate": 2.9492048868135868e-05,
173
- "loss": 0.6104,
174
- "step": 1050
175
  }
176
  ],
177
  "logging_steps": 50,
@@ -195,12 +58,12 @@
195
  "should_evaluate": false,
196
  "should_log": false,
197
  "should_save": true,
198
- "should_training_stop": true
199
  },
200
  "attributes": {}
201
  }
202
  },
203
- "total_flos": 1.757791890505728e+16,
204
  "train_batch_size": 8,
205
  "trial_name": null,
206
  "trial_params": null
 
1
  {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 1.8691588785046729,
5
  "eval_steps": 500,
6
+ "global_step": 200,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.4672897196261682,
13
+ "grad_norm": 212.5613250732422,
14
  "learning_rate": 1.4018691588785047e-05,
15
+ "loss": 53.6674,
16
  "step": 50
17
  },
18
  {
19
  "epoch": 0.9345794392523364,
20
+ "grad_norm": 208.6332550048828,
21
  "learning_rate": 2.8037383177570094e-05,
22
+ "loss": 22.0835,
23
  "step": 100
24
  },
25
  {
26
  "epoch": 1.4018691588785046,
27
+ "grad_norm": 40.16309356689453,
28
  "learning_rate": 2.7699933091326993e-05,
29
+ "loss": 6.2918,
30
  "step": 150
31
  },
32
  {
33
  "epoch": 1.8691588785046729,
34
+ "grad_norm": 12.063589096069336,
35
  "learning_rate": 2.0245088641574458e-05,
36
+ "loss": 5.0856,
37
  "step": 200
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
38
  }
39
  ],
40
  "logging_steps": 50,
 
58
  "should_evaluate": false,
59
  "should_log": false,
60
  "should_save": true,
61
+ "should_training_stop": false
62
  },
63
  "attributes": {}
64
  }
65
  },
66
+ "total_flos": 1254005784576000.0,
67
  "train_batch_size": 8,
68
  "trial_name": null,
69
  "trial_params": null