Training in progress, step 16800
Browse files- config.json +1 -1
- model.safetensors +1 -1
- tokenizer_config.json +1 -1
- training_args.bin +1 -1
- vocab.json +1 -1
config.json
CHANGED
|
@@ -8,7 +8,7 @@
|
|
| 8 |
"d_model": 256,
|
| 9 |
"decoder_start_token_id": 1,
|
| 10 |
"dense_act_fn": "relu",
|
| 11 |
-
"dropout_rate": 0.
|
| 12 |
"eos_token_id": 1,
|
| 13 |
"feed_forward_proj": "relu",
|
| 14 |
"initializer_factor": 1.0,
|
|
|
|
| 8 |
"d_model": 256,
|
| 9 |
"decoder_start_token_id": 1,
|
| 10 |
"dense_act_fn": "relu",
|
| 11 |
+
"dropout_rate": 0.3,
|
| 12 |
"eos_token_id": 1,
|
| 13 |
"feed_forward_proj": "relu",
|
| 14 |
"initializer_factor": 1.0,
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 29454424
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0db0436f589b07b05c135a54252dfdf9c797e66ba1b4711ab3726d72fac31b4c
|
| 3 |
size 29454424
|
tokenizer_config.json
CHANGED
|
@@ -39,6 +39,6 @@
|
|
| 39 |
"max_len": 128,
|
| 40 |
"model_max_length": 128,
|
| 41 |
"pad_token": "<pad>",
|
| 42 |
-
"tokenizer_class": "
|
| 43 |
"unk_token": "<unk>"
|
| 44 |
}
|
|
|
|
| 39 |
"max_len": 128,
|
| 40 |
"model_max_length": 128,
|
| 41 |
"pad_token": "<pad>",
|
| 42 |
+
"tokenizer_class": "MyByT5Tokenizer",
|
| 43 |
"unk_token": "<unk>"
|
| 44 |
}
|
training_args.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4728
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b4ce67ebca446b6b6caba45b9fc27fdf79fdcab8443377ca428ec7e56a445621
|
| 3 |
size 4728
|
vocab.json
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
{"
|
|
|
|
| 1 |
+
{"F": 0, "\u0144": 1, "V": 2, "i": 3, "j": 4, "\u00e9": 5, "n": 6, ")": 7, "g": 8, "\u0119": 9, "1": 10, "(": 11, "\u02bc": 12, "\u0301": 13, "l": 14, "\u01eb": 15, "\u00e1": 16, "b": 17, "c": 18, "4": 19, "T": 20, "x": 21, "\u00ed": 22, "o": 23, "M": 24, "y": 25, "\u00f3": 26, "D": 27, "h": 28, "a": 29, "\u012f": 30, "3": 31, ",": 32, "z": 33, "N": 34, "\u0142": 35, "P": 36, "s": 37, "e": 38, "L": 39, "d": 40, "O": 41, "k": 42, "w": 43, "G": 44, "E": 45, "R": 46, " ": 47, "2": 48, "\u0105": 49, "m": 50, "S": 51, "I": 52, ";": 53, "t": 54}
|