Training in progress, epoch 1
Browse files
config.json
CHANGED
|
@@ -3,9 +3,9 @@
|
|
| 3 |
"T5ForConditionalGeneration"
|
| 4 |
],
|
| 5 |
"classifier_dropout": 0.0,
|
| 6 |
-
"d_ff":
|
| 7 |
-
"d_kv":
|
| 8 |
-
"d_model":
|
| 9 |
"decoder_start_token_id": 259,
|
| 10 |
"dense_act_fn": "relu",
|
| 11 |
"dropout_rate": 0.3,
|
|
@@ -16,9 +16,9 @@
|
|
| 16 |
"is_gated_act": false,
|
| 17 |
"layer_norm_epsilon": 1e-06,
|
| 18 |
"model_type": "t5",
|
| 19 |
-
"num_decoder_layers":
|
| 20 |
-
"num_heads":
|
| 21 |
-
"num_layers":
|
| 22 |
"pad_token_id": 0,
|
| 23 |
"relative_attention_max_distance": 64,
|
| 24 |
"relative_attention_num_buckets": 16,
|
|
|
|
| 3 |
"T5ForConditionalGeneration"
|
| 4 |
],
|
| 5 |
"classifier_dropout": 0.0,
|
| 6 |
+
"d_ff": 1024,
|
| 7 |
+
"d_kv": 64,
|
| 8 |
+
"d_model": 256,
|
| 9 |
"decoder_start_token_id": 259,
|
| 10 |
"dense_act_fn": "relu",
|
| 11 |
"dropout_rate": 0.3,
|
|
|
|
| 16 |
"is_gated_act": false,
|
| 17 |
"layer_norm_epsilon": 1e-06,
|
| 18 |
"model_type": "t5",
|
| 19 |
+
"num_decoder_layers": 4,
|
| 20 |
+
"num_heads": 4,
|
| 21 |
+
"num_layers": 4,
|
| 22 |
"pad_token_id": 0,
|
| 23 |
"relative_attention_max_distance": 64,
|
| 24 |
"relative_attention_num_buckets": 16,
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:464ddd05bf59e3211072364dc59ff0cfa4ed1f7dd9bcbe06568bcd5c3fbf4b58
|
| 3 |
+
size 29787736
|
runs/Nov23_06-52-32_3a71d3232b61/events.out.tfevents.1700722361.3a71d3232b61.1811.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7f685c6432689a28c22acf2337e2b50b2ad81dec6bec09caf5fa34c9ec7bb1ee
|
| 3 |
+
size 4780
|
training_args.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4728
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:12424ea04b3b636595515d470a771cc36ba4c1a2d1fbfc16913bebc1d97e90d3
|
| 3 |
size 4728
|