Commit ·
bea3f57
1
Parent(s): d9bcaae
End of training
Browse files- .gitattributes +0 -1
- last-checkpoint/generation_config.json → generation_config.json +0 -0
- last-checkpoint/config.json +0 -35
- last-checkpoint/optimizer.pt +0 -3
- last-checkpoint/pytorch_model.bin +0 -3
- last-checkpoint/rng_state.pth +0 -3
- last-checkpoint/scheduler.pt +0 -3
- last-checkpoint/sentencepiece.bpe.model +0 -3
- last-checkpoint/special_tokens_map.json +0 -219
- last-checkpoint/tokenizer.json +0 -3
- last-checkpoint/tokenizer_config.json +0 -24
- last-checkpoint/trainer_state.json +0 -166
- last-checkpoint/training_args.bin +0 -3
.gitattributes
CHANGED
|
@@ -32,5 +32,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 32 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 33 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 35 |
-
last-checkpoint/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 36 |
tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
|
|
|
| 32 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 33 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
| 35 |
tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
last-checkpoint/generation_config.json → generation_config.json
RENAMED
|
File without changes
|
last-checkpoint/config.json
DELETED
|
@@ -1,35 +0,0 @@
|
|
| 1 |
-
{
|
| 2 |
-
"_name_or_path": "./models/nllb-en-id",
|
| 3 |
-
"activation_dropout": 0.0,
|
| 4 |
-
"activation_function": "relu",
|
| 5 |
-
"architectures": [
|
| 6 |
-
"M2M100ForConditionalGeneration"
|
| 7 |
-
],
|
| 8 |
-
"attention_dropout": 0.1,
|
| 9 |
-
"bos_token_id": 0,
|
| 10 |
-
"d_model": 1024,
|
| 11 |
-
"decoder_attention_heads": 16,
|
| 12 |
-
"decoder_ffn_dim": 4096,
|
| 13 |
-
"decoder_layerdrop": 0,
|
| 14 |
-
"decoder_layers": 12,
|
| 15 |
-
"decoder_start_token_id": 2,
|
| 16 |
-
"dropout": 0.1,
|
| 17 |
-
"encoder_attention_heads": 16,
|
| 18 |
-
"encoder_ffn_dim": 4096,
|
| 19 |
-
"encoder_layerdrop": 0,
|
| 20 |
-
"encoder_layers": 12,
|
| 21 |
-
"eos_token_id": 2,
|
| 22 |
-
"init_std": 0.02,
|
| 23 |
-
"is_encoder_decoder": true,
|
| 24 |
-
"max_length": 200,
|
| 25 |
-
"max_position_embeddings": 1024,
|
| 26 |
-
"model_type": "m2m_100",
|
| 27 |
-
"num_hidden_layers": 12,
|
| 28 |
-
"pad_token_id": 1,
|
| 29 |
-
"scale_embedding": true,
|
| 30 |
-
"tokenizer_class": "NllbTokenizer",
|
| 31 |
-
"torch_dtype": "float32",
|
| 32 |
-
"transformers_version": "4.26.1",
|
| 33 |
-
"use_cache": true,
|
| 34 |
-
"vocab_size": 256206
|
| 35 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
last-checkpoint/optimizer.pt
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:cde4ef0d03921cf301b7e4277f3075c28c45fe455ef24b0a627014198d3c94ff
|
| 3 |
-
size 4920898947
|
|
|
|
|
|
|
|
|
|
|
|
last-checkpoint/pytorch_model.bin
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:399e4ee176be40862a03b661a1ead63796f58728991959abc071fd0eb140a6c9
|
| 3 |
-
size 2468874377
|
|
|
|
|
|
|
|
|
|
|
|
last-checkpoint/rng_state.pth
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:00942863925632570829ac94b5d27075d207d665dd053a5a65d5cd92501a9c4d
|
| 3 |
-
size 14575
|
|
|
|
|
|
|
|
|
|
|
|
last-checkpoint/scheduler.pt
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:b853d56fd50fc738f70b55e5a38b11254561d0eeeea3283e462a1b133bae651e
|
| 3 |
-
size 627
|
|
|
|
|
|
|
|
|
|
|
|
last-checkpoint/sentencepiece.bpe.model
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:14bb8dfb35c0ffdea7bc01e56cea38b9e3d5efcdcb9c251d6b40538e1aab555a
|
| 3 |
-
size 4852054
|
|
|
|
|
|
|
|
|
|
|
|
last-checkpoint/special_tokens_map.json
DELETED
|
@@ -1,219 +0,0 @@
|
|
| 1 |
-
{
|
| 2 |
-
"additional_special_tokens": [
|
| 3 |
-
"ace_Arab",
|
| 4 |
-
"ace_Latn",
|
| 5 |
-
"acm_Arab",
|
| 6 |
-
"acq_Arab",
|
| 7 |
-
"aeb_Arab",
|
| 8 |
-
"afr_Latn",
|
| 9 |
-
"ajp_Arab",
|
| 10 |
-
"aka_Latn",
|
| 11 |
-
"amh_Ethi",
|
| 12 |
-
"apc_Arab",
|
| 13 |
-
"arb_Arab",
|
| 14 |
-
"ars_Arab",
|
| 15 |
-
"ary_Arab",
|
| 16 |
-
"arz_Arab",
|
| 17 |
-
"asm_Beng",
|
| 18 |
-
"ast_Latn",
|
| 19 |
-
"awa_Deva",
|
| 20 |
-
"ayr_Latn",
|
| 21 |
-
"azb_Arab",
|
| 22 |
-
"azj_Latn",
|
| 23 |
-
"bak_Cyrl",
|
| 24 |
-
"bam_Latn",
|
| 25 |
-
"ban_Latn",
|
| 26 |
-
"bel_Cyrl",
|
| 27 |
-
"bem_Latn",
|
| 28 |
-
"ben_Beng",
|
| 29 |
-
"bho_Deva",
|
| 30 |
-
"bjn_Arab",
|
| 31 |
-
"bjn_Latn",
|
| 32 |
-
"bod_Tibt",
|
| 33 |
-
"bos_Latn",
|
| 34 |
-
"bug_Latn",
|
| 35 |
-
"bul_Cyrl",
|
| 36 |
-
"cat_Latn",
|
| 37 |
-
"ceb_Latn",
|
| 38 |
-
"ces_Latn",
|
| 39 |
-
"cjk_Latn",
|
| 40 |
-
"ckb_Arab",
|
| 41 |
-
"crh_Latn",
|
| 42 |
-
"cym_Latn",
|
| 43 |
-
"dan_Latn",
|
| 44 |
-
"deu_Latn",
|
| 45 |
-
"dik_Latn",
|
| 46 |
-
"dyu_Latn",
|
| 47 |
-
"dzo_Tibt",
|
| 48 |
-
"ell_Grek",
|
| 49 |
-
"eng_Latn",
|
| 50 |
-
"epo_Latn",
|
| 51 |
-
"est_Latn",
|
| 52 |
-
"eus_Latn",
|
| 53 |
-
"ewe_Latn",
|
| 54 |
-
"fao_Latn",
|
| 55 |
-
"pes_Arab",
|
| 56 |
-
"fij_Latn",
|
| 57 |
-
"fin_Latn",
|
| 58 |
-
"fon_Latn",
|
| 59 |
-
"fra_Latn",
|
| 60 |
-
"fur_Latn",
|
| 61 |
-
"fuv_Latn",
|
| 62 |
-
"gla_Latn",
|
| 63 |
-
"gle_Latn",
|
| 64 |
-
"glg_Latn",
|
| 65 |
-
"grn_Latn",
|
| 66 |
-
"guj_Gujr",
|
| 67 |
-
"hat_Latn",
|
| 68 |
-
"hau_Latn",
|
| 69 |
-
"heb_Hebr",
|
| 70 |
-
"hin_Deva",
|
| 71 |
-
"hne_Deva",
|
| 72 |
-
"hrv_Latn",
|
| 73 |
-
"hun_Latn",
|
| 74 |
-
"hye_Armn",
|
| 75 |
-
"ibo_Latn",
|
| 76 |
-
"ilo_Latn",
|
| 77 |
-
"ind_Latn",
|
| 78 |
-
"isl_Latn",
|
| 79 |
-
"ita_Latn",
|
| 80 |
-
"jav_Latn",
|
| 81 |
-
"jpn_Jpan",
|
| 82 |
-
"kab_Latn",
|
| 83 |
-
"kac_Latn",
|
| 84 |
-
"kam_Latn",
|
| 85 |
-
"kan_Knda",
|
| 86 |
-
"kas_Arab",
|
| 87 |
-
"kas_Deva",
|
| 88 |
-
"kat_Geor",
|
| 89 |
-
"knc_Arab",
|
| 90 |
-
"knc_Latn",
|
| 91 |
-
"kaz_Cyrl",
|
| 92 |
-
"kbp_Latn",
|
| 93 |
-
"kea_Latn",
|
| 94 |
-
"khm_Khmr",
|
| 95 |
-
"kik_Latn",
|
| 96 |
-
"kin_Latn",
|
| 97 |
-
"kir_Cyrl",
|
| 98 |
-
"kmb_Latn",
|
| 99 |
-
"kon_Latn",
|
| 100 |
-
"kor_Hang",
|
| 101 |
-
"kmr_Latn",
|
| 102 |
-
"lao_Laoo",
|
| 103 |
-
"lvs_Latn",
|
| 104 |
-
"lij_Latn",
|
| 105 |
-
"lim_Latn",
|
| 106 |
-
"lin_Latn",
|
| 107 |
-
"lit_Latn",
|
| 108 |
-
"lmo_Latn",
|
| 109 |
-
"ltg_Latn",
|
| 110 |
-
"ltz_Latn",
|
| 111 |
-
"lua_Latn",
|
| 112 |
-
"lug_Latn",
|
| 113 |
-
"luo_Latn",
|
| 114 |
-
"lus_Latn",
|
| 115 |
-
"mag_Deva",
|
| 116 |
-
"mai_Deva",
|
| 117 |
-
"mal_Mlym",
|
| 118 |
-
"mar_Deva",
|
| 119 |
-
"min_Latn",
|
| 120 |
-
"mkd_Cyrl",
|
| 121 |
-
"plt_Latn",
|
| 122 |
-
"mlt_Latn",
|
| 123 |
-
"mni_Beng",
|
| 124 |
-
"khk_Cyrl",
|
| 125 |
-
"mos_Latn",
|
| 126 |
-
"mri_Latn",
|
| 127 |
-
"zsm_Latn",
|
| 128 |
-
"mya_Mymr",
|
| 129 |
-
"nld_Latn",
|
| 130 |
-
"nno_Latn",
|
| 131 |
-
"nob_Latn",
|
| 132 |
-
"npi_Deva",
|
| 133 |
-
"nso_Latn",
|
| 134 |
-
"nus_Latn",
|
| 135 |
-
"nya_Latn",
|
| 136 |
-
"oci_Latn",
|
| 137 |
-
"gaz_Latn",
|
| 138 |
-
"ory_Orya",
|
| 139 |
-
"pag_Latn",
|
| 140 |
-
"pan_Guru",
|
| 141 |
-
"pap_Latn",
|
| 142 |
-
"pol_Latn",
|
| 143 |
-
"por_Latn",
|
| 144 |
-
"prs_Arab",
|
| 145 |
-
"pbt_Arab",
|
| 146 |
-
"quy_Latn",
|
| 147 |
-
"ron_Latn",
|
| 148 |
-
"run_Latn",
|
| 149 |
-
"rus_Cyrl",
|
| 150 |
-
"sag_Latn",
|
| 151 |
-
"san_Deva",
|
| 152 |
-
"sat_Beng",
|
| 153 |
-
"scn_Latn",
|
| 154 |
-
"shn_Mymr",
|
| 155 |
-
"sin_Sinh",
|
| 156 |
-
"slk_Latn",
|
| 157 |
-
"slv_Latn",
|
| 158 |
-
"smo_Latn",
|
| 159 |
-
"sna_Latn",
|
| 160 |
-
"snd_Arab",
|
| 161 |
-
"som_Latn",
|
| 162 |
-
"sot_Latn",
|
| 163 |
-
"spa_Latn",
|
| 164 |
-
"als_Latn",
|
| 165 |
-
"srd_Latn",
|
| 166 |
-
"srp_Cyrl",
|
| 167 |
-
"ssw_Latn",
|
| 168 |
-
"sun_Latn",
|
| 169 |
-
"swe_Latn",
|
| 170 |
-
"swh_Latn",
|
| 171 |
-
"szl_Latn",
|
| 172 |
-
"tam_Taml",
|
| 173 |
-
"tat_Cyrl",
|
| 174 |
-
"tel_Telu",
|
| 175 |
-
"tgk_Cyrl",
|
| 176 |
-
"tgl_Latn",
|
| 177 |
-
"tha_Thai",
|
| 178 |
-
"tir_Ethi",
|
| 179 |
-
"taq_Latn",
|
| 180 |
-
"taq_Tfng",
|
| 181 |
-
"tpi_Latn",
|
| 182 |
-
"tsn_Latn",
|
| 183 |
-
"tso_Latn",
|
| 184 |
-
"tuk_Latn",
|
| 185 |
-
"tum_Latn",
|
| 186 |
-
"tur_Latn",
|
| 187 |
-
"twi_Latn",
|
| 188 |
-
"tzm_Tfng",
|
| 189 |
-
"uig_Arab",
|
| 190 |
-
"ukr_Cyrl",
|
| 191 |
-
"umb_Latn",
|
| 192 |
-
"urd_Arab",
|
| 193 |
-
"uzn_Latn",
|
| 194 |
-
"vec_Latn",
|
| 195 |
-
"vie_Latn",
|
| 196 |
-
"war_Latn",
|
| 197 |
-
"wol_Latn",
|
| 198 |
-
"xho_Latn",
|
| 199 |
-
"ydd_Hebr",
|
| 200 |
-
"yor_Latn",
|
| 201 |
-
"yue_Hant",
|
| 202 |
-
"zho_Hans",
|
| 203 |
-
"zho_Hant",
|
| 204 |
-
"zul_Latn"
|
| 205 |
-
],
|
| 206 |
-
"bos_token": "<s>",
|
| 207 |
-
"cls_token": "<s>",
|
| 208 |
-
"eos_token": "</s>",
|
| 209 |
-
"mask_token": {
|
| 210 |
-
"content": "<mask>",
|
| 211 |
-
"lstrip": true,
|
| 212 |
-
"normalized": true,
|
| 213 |
-
"rstrip": false,
|
| 214 |
-
"single_word": false
|
| 215 |
-
},
|
| 216 |
-
"pad_token": "<pad>",
|
| 217 |
-
"sep_token": "</s>",
|
| 218 |
-
"unk_token": "<unk>"
|
| 219 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
last-checkpoint/tokenizer.json
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:8dec87343dd7b147b84072891b984b69471e9b7539eee4e544a0450ffc88e0e8
|
| 3 |
-
size 17331294
|
|
|
|
|
|
|
|
|
|
|
|
last-checkpoint/tokenizer_config.json
DELETED
|
@@ -1,24 +0,0 @@
|
|
| 1 |
-
{
|
| 2 |
-
"additional_special_tokens": null,
|
| 3 |
-
"bos_token": "<s>",
|
| 4 |
-
"cls_token": "<s>",
|
| 5 |
-
"eos_token": "</s>",
|
| 6 |
-
"mask_token": {
|
| 7 |
-
"__type": "AddedToken",
|
| 8 |
-
"content": "<mask>",
|
| 9 |
-
"lstrip": true,
|
| 10 |
-
"normalized": true,
|
| 11 |
-
"rstrip": false,
|
| 12 |
-
"single_word": false
|
| 13 |
-
},
|
| 14 |
-
"model_max_length": 1024,
|
| 15 |
-
"name_or_path": "./models/nllb-en-id",
|
| 16 |
-
"pad_token": "<pad>",
|
| 17 |
-
"sep_token": "</s>",
|
| 18 |
-
"sp_model_kwargs": {},
|
| 19 |
-
"special_tokens_map_file": null,
|
| 20 |
-
"src_lang": "eng_Latn",
|
| 21 |
-
"tgt_lang": "ind_Latn",
|
| 22 |
-
"tokenizer_class": "NllbTokenizer",
|
| 23 |
-
"unk_token": "<unk>"
|
| 24 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
last-checkpoint/trainer_state.json
DELETED
|
@@ -1,166 +0,0 @@
|
|
| 1 |
-
{
|
| 2 |
-
"best_metric": 0.8704027533531189,
|
| 3 |
-
"best_model_checkpoint": "models/nllb-ecolindo/checkpoint-210190",
|
| 4 |
-
"epoch": 10.0,
|
| 5 |
-
"global_step": 210190,
|
| 6 |
-
"is_hyper_param_search": false,
|
| 7 |
-
"is_local_process_zero": true,
|
| 8 |
-
"is_world_process_zero": true,
|
| 9 |
-
"log_history": [
|
| 10 |
-
{
|
| 11 |
-
"epoch": 1.0,
|
| 12 |
-
"learning_rate": 9.174596246180708e-06,
|
| 13 |
-
"loss": 1.3017,
|
| 14 |
-
"step": 21019
|
| 15 |
-
},
|
| 16 |
-
{
|
| 17 |
-
"epoch": 1.0,
|
| 18 |
-
"eval_bleu": 35.0676,
|
| 19 |
-
"eval_loss": 0.9831313490867615,
|
| 20 |
-
"eval_runtime": 94.4525,
|
| 21 |
-
"eval_samples_per_second": 21.175,
|
| 22 |
-
"eval_steps_per_second": 0.667,
|
| 23 |
-
"step": 21019
|
| 24 |
-
},
|
| 25 |
-
{
|
| 26 |
-
"epoch": 2.0,
|
| 27 |
-
"learning_rate": 8.15519666327174e-06,
|
| 28 |
-
"loss": 1.0088,
|
| 29 |
-
"step": 42038
|
| 30 |
-
},
|
| 31 |
-
{
|
| 32 |
-
"epoch": 2.0,
|
| 33 |
-
"eval_bleu": 36.3191,
|
| 34 |
-
"eval_loss": 0.9318408966064453,
|
| 35 |
-
"eval_runtime": 72.147,
|
| 36 |
-
"eval_samples_per_second": 27.721,
|
| 37 |
-
"eval_steps_per_second": 0.873,
|
| 38 |
-
"step": 42038
|
| 39 |
-
},
|
| 40 |
-
{
|
| 41 |
-
"epoch": 3.0,
|
| 42 |
-
"learning_rate": 7.135797080362773e-06,
|
| 43 |
-
"loss": 0.9472,
|
| 44 |
-
"step": 63057
|
| 45 |
-
},
|
| 46 |
-
{
|
| 47 |
-
"epoch": 3.0,
|
| 48 |
-
"eval_bleu": 36.5221,
|
| 49 |
-
"eval_loss": 0.9090332388877869,
|
| 50 |
-
"eval_runtime": 43.552,
|
| 51 |
-
"eval_samples_per_second": 45.922,
|
| 52 |
-
"eval_steps_per_second": 1.447,
|
| 53 |
-
"step": 63057
|
| 54 |
-
},
|
| 55 |
-
{
|
| 56 |
-
"epoch": 4.0,
|
| 57 |
-
"learning_rate": 6.116397497453806e-06,
|
| 58 |
-
"loss": 0.9078,
|
| 59 |
-
"step": 84076
|
| 60 |
-
},
|
| 61 |
-
{
|
| 62 |
-
"epoch": 4.0,
|
| 63 |
-
"eval_bleu": 36.3949,
|
| 64 |
-
"eval_loss": 0.8919472694396973,
|
| 65 |
-
"eval_runtime": 57.3849,
|
| 66 |
-
"eval_samples_per_second": 34.852,
|
| 67 |
-
"eval_steps_per_second": 1.098,
|
| 68 |
-
"step": 84076
|
| 69 |
-
},
|
| 70 |
-
{
|
| 71 |
-
"epoch": 5.0,
|
| 72 |
-
"learning_rate": 5.096997914544838e-06,
|
| 73 |
-
"loss": 0.8789,
|
| 74 |
-
"step": 105095
|
| 75 |
-
},
|
| 76 |
-
{
|
| 77 |
-
"epoch": 5.0,
|
| 78 |
-
"eval_bleu": 37.3689,
|
| 79 |
-
"eval_loss": 0.8832682967185974,
|
| 80 |
-
"eval_runtime": 56.9673,
|
| 81 |
-
"eval_samples_per_second": 35.108,
|
| 82 |
-
"eval_steps_per_second": 1.106,
|
| 83 |
-
"step": 105095
|
| 84 |
-
},
|
| 85 |
-
{
|
| 86 |
-
"epoch": 6.0,
|
| 87 |
-
"learning_rate": 4.07759833163587e-06,
|
| 88 |
-
"loss": 0.8576,
|
| 89 |
-
"step": 126114
|
| 90 |
-
},
|
| 91 |
-
{
|
| 92 |
-
"epoch": 6.0,
|
| 93 |
-
"eval_bleu": 37.2262,
|
| 94 |
-
"eval_loss": 0.8784551620483398,
|
| 95 |
-
"eval_runtime": 55.8811,
|
| 96 |
-
"eval_samples_per_second": 35.79,
|
| 97 |
-
"eval_steps_per_second": 1.127,
|
| 98 |
-
"step": 126114
|
| 99 |
-
},
|
| 100 |
-
{
|
| 101 |
-
"epoch": 7.0,
|
| 102 |
-
"learning_rate": 3.058198748726903e-06,
|
| 103 |
-
"loss": 0.8403,
|
| 104 |
-
"step": 147133
|
| 105 |
-
},
|
| 106 |
-
{
|
| 107 |
-
"epoch": 7.0,
|
| 108 |
-
"eval_bleu": 37.2933,
|
| 109 |
-
"eval_loss": 0.8747518062591553,
|
| 110 |
-
"eval_runtime": 44.6649,
|
| 111 |
-
"eval_samples_per_second": 44.778,
|
| 112 |
-
"eval_steps_per_second": 1.411,
|
| 113 |
-
"step": 147133
|
| 114 |
-
},
|
| 115 |
-
{
|
| 116 |
-
"epoch": 8.0,
|
| 117 |
-
"learning_rate": 2.038799165817935e-06,
|
| 118 |
-
"loss": 0.8281,
|
| 119 |
-
"step": 168152
|
| 120 |
-
},
|
| 121 |
-
{
|
| 122 |
-
"epoch": 8.0,
|
| 123 |
-
"eval_bleu": 37.2952,
|
| 124 |
-
"eval_loss": 0.8716793656349182,
|
| 125 |
-
"eval_runtime": 43.4982,
|
| 126 |
-
"eval_samples_per_second": 45.979,
|
| 127 |
-
"eval_steps_per_second": 1.448,
|
| 128 |
-
"step": 168152
|
| 129 |
-
},
|
| 130 |
-
{
|
| 131 |
-
"epoch": 9.0,
|
| 132 |
-
"learning_rate": 1.0193995829089675e-06,
|
| 133 |
-
"loss": 0.8186,
|
| 134 |
-
"step": 189171
|
| 135 |
-
},
|
| 136 |
-
{
|
| 137 |
-
"epoch": 9.0,
|
| 138 |
-
"eval_bleu": 37.2801,
|
| 139 |
-
"eval_loss": 0.8708682656288147,
|
| 140 |
-
"eval_runtime": 41.2854,
|
| 141 |
-
"eval_samples_per_second": 48.443,
|
| 142 |
-
"eval_steps_per_second": 1.526,
|
| 143 |
-
"step": 189171
|
| 144 |
-
},
|
| 145 |
-
{
|
| 146 |
-
"epoch": 10.0,
|
| 147 |
-
"learning_rate": 0.0,
|
| 148 |
-
"loss": 0.813,
|
| 149 |
-
"step": 210190
|
| 150 |
-
},
|
| 151 |
-
{
|
| 152 |
-
"epoch": 10.0,
|
| 153 |
-
"eval_bleu": 37.2396,
|
| 154 |
-
"eval_loss": 0.8704027533531189,
|
| 155 |
-
"eval_runtime": 44.7957,
|
| 156 |
-
"eval_samples_per_second": 44.647,
|
| 157 |
-
"eval_steps_per_second": 1.406,
|
| 158 |
-
"step": 210190
|
| 159 |
-
}
|
| 160 |
-
],
|
| 161 |
-
"max_steps": 210190,
|
| 162 |
-
"num_train_epochs": 10,
|
| 163 |
-
"total_flos": 3.338536761820938e+17,
|
| 164 |
-
"trial_name": null,
|
| 165 |
-
"trial_params": null
|
| 166 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
last-checkpoint/training_args.bin
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:4d9db2c86ed401d06f5076171976f58cd756593b824f876e75dd4564d3c0960e
|
| 3 |
-
size 3643
|
|
|
|
|
|
|
|
|
|
|
|