azawahry
commited on
Commit
·
981a23b
1
Parent(s):
4a71f29
Initial commit
Browse files- checkpoint-150/added_tokens.json +44 -0
- checkpoint-150/config.json +61 -0
- checkpoint-150/generation_config.json +15 -0
- checkpoint-150/optimizer.pt +3 -0
- checkpoint-150/pytorch_model.bin +3 -0
- checkpoint-150/rng_state.pth +3 -0
- checkpoint-150/scaler.pt +3 -0
- checkpoint-150/scheduler.pt +3 -0
- checkpoint-150/source.spm +0 -0
- checkpoint-150/special_tokens_map.json +5 -0
- checkpoint-150/target.spm +0 -0
- checkpoint-150/tokenizer_config.json +12 -0
- checkpoint-150/trainer_state.json +30 -0
- checkpoint-150/training_args.bin +3 -0
- checkpoint-150/vocab.json +0 -0
- checkpoint-450/added_tokens.json +44 -0
- checkpoint-450/config.json +61 -0
- checkpoint-450/generation_config.json +15 -0
- checkpoint-450/optimizer.pt +3 -0
- checkpoint-450/pytorch_model.bin +3 -0
- checkpoint-450/rng_state.pth +3 -0
- checkpoint-450/scaler.pt +3 -0
- checkpoint-450/scheduler.pt +3 -0
- checkpoint-450/source.spm +0 -0
- checkpoint-450/special_tokens_map.json +5 -0
- checkpoint-450/target.spm +0 -0
- checkpoint-450/tokenizer_config.json +12 -0
- checkpoint-450/trainer_state.json +58 -0
- checkpoint-450/training_args.bin +3 -0
- checkpoint-450/vocab.json +0 -0
- checkpoint-600/added_tokens.json +44 -0
- checkpoint-600/config.json +61 -0
- checkpoint-600/generation_config.json +15 -0
- checkpoint-600/optimizer.pt +3 -0
- checkpoint-600/pytorch_model.bin +3 -0
- checkpoint-600/rng_state.pth +3 -0
- checkpoint-600/scaler.pt +3 -0
- checkpoint-600/scheduler.pt +3 -0
- checkpoint-600/source.spm +0 -0
- checkpoint-600/special_tokens_map.json +5 -0
- checkpoint-600/target.spm +0 -0
- checkpoint-600/tokenizer_config.json +12 -0
- checkpoint-600/trainer_state.json +78 -0
- checkpoint-600/training_args.bin +3 -0
- checkpoint-600/vocab.json +0 -0
checkpoint-150/added_tokens.json
ADDED
|
@@ -0,0 +1,44 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"|ach_aug_bt|": 64199,
|
| 3 |
+
"|ach_aug_ggl|": 64197,
|
| 4 |
+
"|ach_aug_hq|": 64198,
|
| 5 |
+
"|ach_bt|": 64196,
|
| 6 |
+
"|ach_ggl|": 64194,
|
| 7 |
+
"|ach_hq|": 64195,
|
| 8 |
+
"|ach_ood|": 64193,
|
| 9 |
+
"|lgg_aug_bt|": 64191,
|
| 10 |
+
"|lgg_aug_ggl|": 64190,
|
| 11 |
+
"|lgg_aug_hq|": 64192,
|
| 12 |
+
"|lgg_bt|": 64188,
|
| 13 |
+
"|lgg_ggl|": 64187,
|
| 14 |
+
"|lgg_hq|": 64189,
|
| 15 |
+
"|lgg_ood|": 64186,
|
| 16 |
+
"|lug_aug_bt|": 64206,
|
| 17 |
+
"|lug_aug_ggl|": 64204,
|
| 18 |
+
"|lug_aug_hq|": 64205,
|
| 19 |
+
"|lug_bt|": 64203,
|
| 20 |
+
"|lug_ggl|": 64201,
|
| 21 |
+
"|lug_hq|": 64202,
|
| 22 |
+
"|lug_ood|": 64200,
|
| 23 |
+
"|luo_aug_bt|": 64213,
|
| 24 |
+
"|luo_aug_ggl|": 64211,
|
| 25 |
+
"|luo_aug_hq|": 64212,
|
| 26 |
+
"|luo_bt|": 64210,
|
| 27 |
+
"|luo_ggl|": 64208,
|
| 28 |
+
"|luo_hq|": 64209,
|
| 29 |
+
"|luo_ood|": 64207,
|
| 30 |
+
"|nyn_aug_bt|": 64178,
|
| 31 |
+
"|nyn_aug_ggl|": 64176,
|
| 32 |
+
"|nyn_aug_hq|": 64177,
|
| 33 |
+
"|nyn_bt|": 64175,
|
| 34 |
+
"|nyn_ggl|": 64173,
|
| 35 |
+
"|nyn_hq|": 64174,
|
| 36 |
+
"|nyn_ood|": 64172,
|
| 37 |
+
"|teo_aug_bt|": 64185,
|
| 38 |
+
"|teo_aug_ggl|": 64183,
|
| 39 |
+
"|teo_aug_hq|": 64184,
|
| 40 |
+
"|teo_bt|": 64182,
|
| 41 |
+
"|teo_ggl|": 64180,
|
| 42 |
+
"|teo_hq|": 64181,
|
| 43 |
+
"|teo_ood|": 64179
|
| 44 |
+
}
|
checkpoint-150/config.json
ADDED
|
@@ -0,0 +1,61 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_name_or_path": "/home/azawahry_sunbird_ai/models/m2e/dataset-tags-expanded-tokenizer/checkpoint-1050",
|
| 3 |
+
"activation_dropout": 0.0,
|
| 4 |
+
"activation_function": "swish",
|
| 5 |
+
"add_bias_logits": false,
|
| 6 |
+
"add_final_layer_norm": false,
|
| 7 |
+
"architectures": [
|
| 8 |
+
"MarianMTModel"
|
| 9 |
+
],
|
| 10 |
+
"attention_dropout": 0.0,
|
| 11 |
+
"bad_words_ids": [
|
| 12 |
+
[
|
| 13 |
+
64171
|
| 14 |
+
]
|
| 15 |
+
],
|
| 16 |
+
"bos_token_id": 0,
|
| 17 |
+
"classif_dropout": 0.0,
|
| 18 |
+
"classifier_dropout": 0.0,
|
| 19 |
+
"d_model": 512,
|
| 20 |
+
"decoder_attention_heads": 8,
|
| 21 |
+
"decoder_ffn_dim": 2048,
|
| 22 |
+
"decoder_layerdrop": 0.0,
|
| 23 |
+
"decoder_layers": 6,
|
| 24 |
+
"decoder_start_token_id": 64171,
|
| 25 |
+
"decoder_vocab_size": 64214,
|
| 26 |
+
"dropout": 0.1,
|
| 27 |
+
"encoder_attention_heads": 8,
|
| 28 |
+
"encoder_ffn_dim": 2048,
|
| 29 |
+
"encoder_layerdrop": 0.0,
|
| 30 |
+
"encoder_layers": 6,
|
| 31 |
+
"eos_token_id": 0,
|
| 32 |
+
"extra_pos_embeddings": 64172,
|
| 33 |
+
"forced_eos_token_id": 0,
|
| 34 |
+
"id2label": {
|
| 35 |
+
"0": "LABEL_0",
|
| 36 |
+
"1": "LABEL_1",
|
| 37 |
+
"2": "LABEL_2"
|
| 38 |
+
},
|
| 39 |
+
"init_std": 0.02,
|
| 40 |
+
"is_encoder_decoder": true,
|
| 41 |
+
"label2id": {
|
| 42 |
+
"LABEL_0": 0,
|
| 43 |
+
"LABEL_1": 1,
|
| 44 |
+
"LABEL_2": 2
|
| 45 |
+
},
|
| 46 |
+
"max_length": 512,
|
| 47 |
+
"max_position_embeddings": 512,
|
| 48 |
+
"model_type": "marian",
|
| 49 |
+
"normalize_before": false,
|
| 50 |
+
"normalize_embedding": false,
|
| 51 |
+
"num_beams": 6,
|
| 52 |
+
"num_hidden_layers": 6,
|
| 53 |
+
"pad_token_id": 64171,
|
| 54 |
+
"scale_embedding": true,
|
| 55 |
+
"share_encoder_decoder_embeddings": true,
|
| 56 |
+
"static_position_embeddings": true,
|
| 57 |
+
"torch_dtype": "float32",
|
| 58 |
+
"transformers_version": "4.27.3",
|
| 59 |
+
"use_cache": true,
|
| 60 |
+
"vocab_size": 64214
|
| 61 |
+
}
|
checkpoint-150/generation_config.json
ADDED
|
@@ -0,0 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"bad_words_ids": [
|
| 3 |
+
[
|
| 4 |
+
64171
|
| 5 |
+
]
|
| 6 |
+
],
|
| 7 |
+
"bos_token_id": 0,
|
| 8 |
+
"decoder_start_token_id": 64171,
|
| 9 |
+
"eos_token_id": 0,
|
| 10 |
+
"forced_eos_token_id": 0,
|
| 11 |
+
"max_length": 512,
|
| 12 |
+
"num_beams": 6,
|
| 13 |
+
"pad_token_id": 64171,
|
| 14 |
+
"transformers_version": "4.27.3"
|
| 15 |
+
}
|
checkpoint-150/optimizer.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3dae2189b093b86ba289af4f005a508ae538bb31354719e3db9116d35f245e32
|
| 3 |
+
size 616276741
|
checkpoint-150/pytorch_model.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2270bc13e810caca72693d79f82753223c8de32941ed9dc649902821d62a39ab
|
| 3 |
+
size 308407621
|
checkpoint-150/rng_state.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:42a80f6998d2d4ab9eb7566bcf48c742bb00e92d4722e6ad00fe328123bdb3e9
|
| 3 |
+
size 14575
|
checkpoint-150/scaler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c51c8702e03bc24d0566e4b37e893acfcfe0af87dcc3883f22471d07682d7ac5
|
| 3 |
+
size 557
|
checkpoint-150/scheduler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f7c7bf40a66ebd048fdd4239a501cedcc9711d3fdb1b1d9a7f4d677727d2347a
|
| 3 |
+
size 627
|
checkpoint-150/source.spm
ADDED
|
Binary file (707 kB). View file
|
|
|
checkpoint-150/special_tokens_map.json
ADDED
|
@@ -0,0 +1,5 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eos_token": "</s>",
|
| 3 |
+
"pad_token": "<pad>",
|
| 4 |
+
"unk_token": "<unk>"
|
| 5 |
+
}
|
checkpoint-150/target.spm
ADDED
|
Binary file (791 kB). View file
|
|
|
checkpoint-150/tokenizer_config.json
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eos_token": "</s>",
|
| 3 |
+
"model_max_length": 512,
|
| 4 |
+
"pad_token": "<pad>",
|
| 5 |
+
"separate_vocabs": false,
|
| 6 |
+
"source_lang": "mul",
|
| 7 |
+
"sp_model_kwargs": {},
|
| 8 |
+
"special_tokens_map_file": null,
|
| 9 |
+
"target_lang": "eng",
|
| 10 |
+
"tokenizer_class": "MarianTokenizer",
|
| 11 |
+
"unk_token": "<unk>"
|
| 12 |
+
}
|
checkpoint-150/trainer_state.json
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"best_metric": 2.9462931156158447,
|
| 3 |
+
"best_model_checkpoint": "/home/azawahry_sunbird_ai/models/m2e/dataset-tags-expanded-tokenizer/checkpoint-150",
|
| 4 |
+
"epoch": 0.2562481849086902,
|
| 5 |
+
"global_step": 150,
|
| 6 |
+
"is_hyper_param_search": false,
|
| 7 |
+
"is_local_process_zero": true,
|
| 8 |
+
"is_world_process_zero": true,
|
| 9 |
+
"log_history": [
|
| 10 |
+
{
|
| 11 |
+
"epoch": 0.26,
|
| 12 |
+
"eval_BLEU_ach": 21.4505,
|
| 13 |
+
"eval_BLEU_lgg": 24.4794,
|
| 14 |
+
"eval_BLEU_lug": 35.1488,
|
| 15 |
+
"eval_BLEU_mean": 26.3376,
|
| 16 |
+
"eval_BLEU_nyn": 25.5547,
|
| 17 |
+
"eval_BLEU_teo": 25.0545,
|
| 18 |
+
"eval_loss": 2.9462931156158447,
|
| 19 |
+
"eval_runtime": 59.9936,
|
| 20 |
+
"eval_samples_per_second": 41.671,
|
| 21 |
+
"eval_steps_per_second": 0.833,
|
| 22 |
+
"step": 150
|
| 23 |
+
}
|
| 24 |
+
],
|
| 25 |
+
"max_steps": 1755,
|
| 26 |
+
"num_train_epochs": 3,
|
| 27 |
+
"total_flos": 2.16451520888832e+16,
|
| 28 |
+
"trial_name": null,
|
| 29 |
+
"trial_params": null
|
| 30 |
+
}
|
checkpoint-150/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e46a42e4fff4eb7e4702f865fd1022499e6ad9fc3cf4d3f71aef8c990651b558
|
| 3 |
+
size 3707
|
checkpoint-150/vocab.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
checkpoint-450/added_tokens.json
ADDED
|
@@ -0,0 +1,44 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"|ach_aug_bt|": 64199,
|
| 3 |
+
"|ach_aug_ggl|": 64197,
|
| 4 |
+
"|ach_aug_hq|": 64198,
|
| 5 |
+
"|ach_bt|": 64196,
|
| 6 |
+
"|ach_ggl|": 64194,
|
| 7 |
+
"|ach_hq|": 64195,
|
| 8 |
+
"|ach_ood|": 64193,
|
| 9 |
+
"|lgg_aug_bt|": 64191,
|
| 10 |
+
"|lgg_aug_ggl|": 64190,
|
| 11 |
+
"|lgg_aug_hq|": 64192,
|
| 12 |
+
"|lgg_bt|": 64188,
|
| 13 |
+
"|lgg_ggl|": 64187,
|
| 14 |
+
"|lgg_hq|": 64189,
|
| 15 |
+
"|lgg_ood|": 64186,
|
| 16 |
+
"|lug_aug_bt|": 64206,
|
| 17 |
+
"|lug_aug_ggl|": 64204,
|
| 18 |
+
"|lug_aug_hq|": 64205,
|
| 19 |
+
"|lug_bt|": 64203,
|
| 20 |
+
"|lug_ggl|": 64201,
|
| 21 |
+
"|lug_hq|": 64202,
|
| 22 |
+
"|lug_ood|": 64200,
|
| 23 |
+
"|luo_aug_bt|": 64213,
|
| 24 |
+
"|luo_aug_ggl|": 64211,
|
| 25 |
+
"|luo_aug_hq|": 64212,
|
| 26 |
+
"|luo_bt|": 64210,
|
| 27 |
+
"|luo_ggl|": 64208,
|
| 28 |
+
"|luo_hq|": 64209,
|
| 29 |
+
"|luo_ood|": 64207,
|
| 30 |
+
"|nyn_aug_bt|": 64178,
|
| 31 |
+
"|nyn_aug_ggl|": 64176,
|
| 32 |
+
"|nyn_aug_hq|": 64177,
|
| 33 |
+
"|nyn_bt|": 64175,
|
| 34 |
+
"|nyn_ggl|": 64173,
|
| 35 |
+
"|nyn_hq|": 64174,
|
| 36 |
+
"|nyn_ood|": 64172,
|
| 37 |
+
"|teo_aug_bt|": 64185,
|
| 38 |
+
"|teo_aug_ggl|": 64183,
|
| 39 |
+
"|teo_aug_hq|": 64184,
|
| 40 |
+
"|teo_bt|": 64182,
|
| 41 |
+
"|teo_ggl|": 64180,
|
| 42 |
+
"|teo_hq|": 64181,
|
| 43 |
+
"|teo_ood|": 64179
|
| 44 |
+
}
|
checkpoint-450/config.json
ADDED
|
@@ -0,0 +1,61 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_name_or_path": "/home/azawahry_sunbird_ai/models/m2e/dataset-tags-expanded-tokenizer/checkpoint-1050",
|
| 3 |
+
"activation_dropout": 0.0,
|
| 4 |
+
"activation_function": "swish",
|
| 5 |
+
"add_bias_logits": false,
|
| 6 |
+
"add_final_layer_norm": false,
|
| 7 |
+
"architectures": [
|
| 8 |
+
"MarianMTModel"
|
| 9 |
+
],
|
| 10 |
+
"attention_dropout": 0.0,
|
| 11 |
+
"bad_words_ids": [
|
| 12 |
+
[
|
| 13 |
+
64171
|
| 14 |
+
]
|
| 15 |
+
],
|
| 16 |
+
"bos_token_id": 0,
|
| 17 |
+
"classif_dropout": 0.0,
|
| 18 |
+
"classifier_dropout": 0.0,
|
| 19 |
+
"d_model": 512,
|
| 20 |
+
"decoder_attention_heads": 8,
|
| 21 |
+
"decoder_ffn_dim": 2048,
|
| 22 |
+
"decoder_layerdrop": 0.0,
|
| 23 |
+
"decoder_layers": 6,
|
| 24 |
+
"decoder_start_token_id": 64171,
|
| 25 |
+
"decoder_vocab_size": 64214,
|
| 26 |
+
"dropout": 0.1,
|
| 27 |
+
"encoder_attention_heads": 8,
|
| 28 |
+
"encoder_ffn_dim": 2048,
|
| 29 |
+
"encoder_layerdrop": 0.0,
|
| 30 |
+
"encoder_layers": 6,
|
| 31 |
+
"eos_token_id": 0,
|
| 32 |
+
"extra_pos_embeddings": 64172,
|
| 33 |
+
"forced_eos_token_id": 0,
|
| 34 |
+
"id2label": {
|
| 35 |
+
"0": "LABEL_0",
|
| 36 |
+
"1": "LABEL_1",
|
| 37 |
+
"2": "LABEL_2"
|
| 38 |
+
},
|
| 39 |
+
"init_std": 0.02,
|
| 40 |
+
"is_encoder_decoder": true,
|
| 41 |
+
"label2id": {
|
| 42 |
+
"LABEL_0": 0,
|
| 43 |
+
"LABEL_1": 1,
|
| 44 |
+
"LABEL_2": 2
|
| 45 |
+
},
|
| 46 |
+
"max_length": 512,
|
| 47 |
+
"max_position_embeddings": 512,
|
| 48 |
+
"model_type": "marian",
|
| 49 |
+
"normalize_before": false,
|
| 50 |
+
"normalize_embedding": false,
|
| 51 |
+
"num_beams": 6,
|
| 52 |
+
"num_hidden_layers": 6,
|
| 53 |
+
"pad_token_id": 64171,
|
| 54 |
+
"scale_embedding": true,
|
| 55 |
+
"share_encoder_decoder_embeddings": true,
|
| 56 |
+
"static_position_embeddings": true,
|
| 57 |
+
"torch_dtype": "float32",
|
| 58 |
+
"transformers_version": "4.27.3",
|
| 59 |
+
"use_cache": true,
|
| 60 |
+
"vocab_size": 64214
|
| 61 |
+
}
|
checkpoint-450/generation_config.json
ADDED
|
@@ -0,0 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"bad_words_ids": [
|
| 3 |
+
[
|
| 4 |
+
64171
|
| 5 |
+
]
|
| 6 |
+
],
|
| 7 |
+
"bos_token_id": 0,
|
| 8 |
+
"decoder_start_token_id": 64171,
|
| 9 |
+
"eos_token_id": 0,
|
| 10 |
+
"forced_eos_token_id": 0,
|
| 11 |
+
"max_length": 512,
|
| 12 |
+
"num_beams": 6,
|
| 13 |
+
"pad_token_id": 64171,
|
| 14 |
+
"transformers_version": "4.27.3"
|
| 15 |
+
}
|
checkpoint-450/optimizer.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0b18e3590002b877b6e3670c8f8c7ef25095ceb720573dd2db865e5e771cd1d6
|
| 3 |
+
size 616276997
|
checkpoint-450/pytorch_model.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8f98f0032593bbf0363c8517501389a11c533da912a79454abcd692f771f0c8e
|
| 3 |
+
size 308407621
|
checkpoint-450/rng_state.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0f357c535c5b43fc8733e6460972e270983e96db90cb51bf9b46ee6d2b7c4965
|
| 3 |
+
size 14639
|
checkpoint-450/scaler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:206389f090abcc3222507a1886e7890301684ad43da4a05adfa7cac210314c85
|
| 3 |
+
size 557
|
checkpoint-450/scheduler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1cf1d48671f6e11f1aa6e8904b2ccf053a76b8f1676467007e8acdff715fc054
|
| 3 |
+
size 627
|
checkpoint-450/source.spm
ADDED
|
Binary file (707 kB). View file
|
|
|
checkpoint-450/special_tokens_map.json
ADDED
|
@@ -0,0 +1,5 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eos_token": "</s>",
|
| 3 |
+
"pad_token": "<pad>",
|
| 4 |
+
"unk_token": "<unk>"
|
| 5 |
+
}
|
checkpoint-450/target.spm
ADDED
|
Binary file (791 kB). View file
|
|
|
checkpoint-450/tokenizer_config.json
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eos_token": "</s>",
|
| 3 |
+
"model_max_length": 512,
|
| 4 |
+
"pad_token": "<pad>",
|
| 5 |
+
"separate_vocabs": false,
|
| 6 |
+
"source_lang": "mul",
|
| 7 |
+
"sp_model_kwargs": {},
|
| 8 |
+
"special_tokens_map_file": null,
|
| 9 |
+
"target_lang": "eng",
|
| 10 |
+
"tokenizer_class": "MarianTokenizer",
|
| 11 |
+
"unk_token": "<unk>"
|
| 12 |
+
}
|
checkpoint-450/trainer_state.json
ADDED
|
@@ -0,0 +1,58 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"best_metric": 2.9462931156158447,
|
| 3 |
+
"best_model_checkpoint": "/home/azawahry_sunbird_ai/models/m2e/dataset-tags-expanded-tokenizer/checkpoint-150",
|
| 4 |
+
"epoch": 0.7687445547260707,
|
| 5 |
+
"global_step": 450,
|
| 6 |
+
"is_hyper_param_search": false,
|
| 7 |
+
"is_local_process_zero": true,
|
| 8 |
+
"is_world_process_zero": true,
|
| 9 |
+
"log_history": [
|
| 10 |
+
{
|
| 11 |
+
"epoch": 0.26,
|
| 12 |
+
"eval_BLEU_ach": 21.4505,
|
| 13 |
+
"eval_BLEU_lgg": 24.4794,
|
| 14 |
+
"eval_BLEU_lug": 35.1488,
|
| 15 |
+
"eval_BLEU_mean": 26.3376,
|
| 16 |
+
"eval_BLEU_nyn": 25.5547,
|
| 17 |
+
"eval_BLEU_teo": 25.0545,
|
| 18 |
+
"eval_loss": 2.9462931156158447,
|
| 19 |
+
"eval_runtime": 59.9936,
|
| 20 |
+
"eval_samples_per_second": 41.671,
|
| 21 |
+
"eval_steps_per_second": 0.833,
|
| 22 |
+
"step": 150
|
| 23 |
+
},
|
| 24 |
+
{
|
| 25 |
+
"epoch": 0.51,
|
| 26 |
+
"eval_BLEU_ach": 22.0556,
|
| 27 |
+
"eval_BLEU_lgg": 23.8502,
|
| 28 |
+
"eval_BLEU_lug": 34.6592,
|
| 29 |
+
"eval_BLEU_mean": 26.1806,
|
| 30 |
+
"eval_BLEU_nyn": 25.075,
|
| 31 |
+
"eval_BLEU_teo": 25.2631,
|
| 32 |
+
"eval_loss": 2.9494848251342773,
|
| 33 |
+
"eval_runtime": 58.1523,
|
| 34 |
+
"eval_samples_per_second": 42.991,
|
| 35 |
+
"eval_steps_per_second": 0.86,
|
| 36 |
+
"step": 300
|
| 37 |
+
},
|
| 38 |
+
{
|
| 39 |
+
"epoch": 0.77,
|
| 40 |
+
"eval_BLEU_ach": 21.6851,
|
| 41 |
+
"eval_BLEU_lgg": 24.2555,
|
| 42 |
+
"eval_BLEU_lug": 34.7523,
|
| 43 |
+
"eval_BLEU_mean": 26.1102,
|
| 44 |
+
"eval_BLEU_nyn": 24.5712,
|
| 45 |
+
"eval_BLEU_teo": 25.2871,
|
| 46 |
+
"eval_loss": 2.9563159942626953,
|
| 47 |
+
"eval_runtime": 57.8967,
|
| 48 |
+
"eval_samples_per_second": 43.18,
|
| 49 |
+
"eval_steps_per_second": 0.864,
|
| 50 |
+
"step": 450
|
| 51 |
+
}
|
| 52 |
+
],
|
| 53 |
+
"max_steps": 1755,
|
| 54 |
+
"num_train_epochs": 3,
|
| 55 |
+
"total_flos": 6.48118331670528e+16,
|
| 56 |
+
"trial_name": null,
|
| 57 |
+
"trial_params": null
|
| 58 |
+
}
|
checkpoint-450/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e46a42e4fff4eb7e4702f865fd1022499e6ad9fc3cf4d3f71aef8c990651b558
|
| 3 |
+
size 3707
|
checkpoint-450/vocab.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
checkpoint-600/added_tokens.json
ADDED
|
@@ -0,0 +1,44 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"|ach_aug_bt|": 64199,
|
| 3 |
+
"|ach_aug_ggl|": 64197,
|
| 4 |
+
"|ach_aug_hq|": 64198,
|
| 5 |
+
"|ach_bt|": 64196,
|
| 6 |
+
"|ach_ggl|": 64194,
|
| 7 |
+
"|ach_hq|": 64195,
|
| 8 |
+
"|ach_ood|": 64193,
|
| 9 |
+
"|lgg_aug_bt|": 64191,
|
| 10 |
+
"|lgg_aug_ggl|": 64190,
|
| 11 |
+
"|lgg_aug_hq|": 64192,
|
| 12 |
+
"|lgg_bt|": 64188,
|
| 13 |
+
"|lgg_ggl|": 64187,
|
| 14 |
+
"|lgg_hq|": 64189,
|
| 15 |
+
"|lgg_ood|": 64186,
|
| 16 |
+
"|lug_aug_bt|": 64206,
|
| 17 |
+
"|lug_aug_ggl|": 64204,
|
| 18 |
+
"|lug_aug_hq|": 64205,
|
| 19 |
+
"|lug_bt|": 64203,
|
| 20 |
+
"|lug_ggl|": 64201,
|
| 21 |
+
"|lug_hq|": 64202,
|
| 22 |
+
"|lug_ood|": 64200,
|
| 23 |
+
"|luo_aug_bt|": 64213,
|
| 24 |
+
"|luo_aug_ggl|": 64211,
|
| 25 |
+
"|luo_aug_hq|": 64212,
|
| 26 |
+
"|luo_bt|": 64210,
|
| 27 |
+
"|luo_ggl|": 64208,
|
| 28 |
+
"|luo_hq|": 64209,
|
| 29 |
+
"|luo_ood|": 64207,
|
| 30 |
+
"|nyn_aug_bt|": 64178,
|
| 31 |
+
"|nyn_aug_ggl|": 64176,
|
| 32 |
+
"|nyn_aug_hq|": 64177,
|
| 33 |
+
"|nyn_bt|": 64175,
|
| 34 |
+
"|nyn_ggl|": 64173,
|
| 35 |
+
"|nyn_hq|": 64174,
|
| 36 |
+
"|nyn_ood|": 64172,
|
| 37 |
+
"|teo_aug_bt|": 64185,
|
| 38 |
+
"|teo_aug_ggl|": 64183,
|
| 39 |
+
"|teo_aug_hq|": 64184,
|
| 40 |
+
"|teo_bt|": 64182,
|
| 41 |
+
"|teo_ggl|": 64180,
|
| 42 |
+
"|teo_hq|": 64181,
|
| 43 |
+
"|teo_ood|": 64179
|
| 44 |
+
}
|
checkpoint-600/config.json
ADDED
|
@@ -0,0 +1,61 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_name_or_path": "/home/azawahry_sunbird_ai/models/m2e/dataset-tags-expanded-tokenizer/checkpoint-1050",
|
| 3 |
+
"activation_dropout": 0.0,
|
| 4 |
+
"activation_function": "swish",
|
| 5 |
+
"add_bias_logits": false,
|
| 6 |
+
"add_final_layer_norm": false,
|
| 7 |
+
"architectures": [
|
| 8 |
+
"MarianMTModel"
|
| 9 |
+
],
|
| 10 |
+
"attention_dropout": 0.0,
|
| 11 |
+
"bad_words_ids": [
|
| 12 |
+
[
|
| 13 |
+
64171
|
| 14 |
+
]
|
| 15 |
+
],
|
| 16 |
+
"bos_token_id": 0,
|
| 17 |
+
"classif_dropout": 0.0,
|
| 18 |
+
"classifier_dropout": 0.0,
|
| 19 |
+
"d_model": 512,
|
| 20 |
+
"decoder_attention_heads": 8,
|
| 21 |
+
"decoder_ffn_dim": 2048,
|
| 22 |
+
"decoder_layerdrop": 0.0,
|
| 23 |
+
"decoder_layers": 6,
|
| 24 |
+
"decoder_start_token_id": 64171,
|
| 25 |
+
"decoder_vocab_size": 64214,
|
| 26 |
+
"dropout": 0.1,
|
| 27 |
+
"encoder_attention_heads": 8,
|
| 28 |
+
"encoder_ffn_dim": 2048,
|
| 29 |
+
"encoder_layerdrop": 0.0,
|
| 30 |
+
"encoder_layers": 6,
|
| 31 |
+
"eos_token_id": 0,
|
| 32 |
+
"extra_pos_embeddings": 64172,
|
| 33 |
+
"forced_eos_token_id": 0,
|
| 34 |
+
"id2label": {
|
| 35 |
+
"0": "LABEL_0",
|
| 36 |
+
"1": "LABEL_1",
|
| 37 |
+
"2": "LABEL_2"
|
| 38 |
+
},
|
| 39 |
+
"init_std": 0.02,
|
| 40 |
+
"is_encoder_decoder": true,
|
| 41 |
+
"label2id": {
|
| 42 |
+
"LABEL_0": 0,
|
| 43 |
+
"LABEL_1": 1,
|
| 44 |
+
"LABEL_2": 2
|
| 45 |
+
},
|
| 46 |
+
"max_length": 512,
|
| 47 |
+
"max_position_embeddings": 512,
|
| 48 |
+
"model_type": "marian",
|
| 49 |
+
"normalize_before": false,
|
| 50 |
+
"normalize_embedding": false,
|
| 51 |
+
"num_beams": 6,
|
| 52 |
+
"num_hidden_layers": 6,
|
| 53 |
+
"pad_token_id": 64171,
|
| 54 |
+
"scale_embedding": true,
|
| 55 |
+
"share_encoder_decoder_embeddings": true,
|
| 56 |
+
"static_position_embeddings": true,
|
| 57 |
+
"torch_dtype": "float32",
|
| 58 |
+
"transformers_version": "4.27.3",
|
| 59 |
+
"use_cache": true,
|
| 60 |
+
"vocab_size": 64214
|
| 61 |
+
}
|
checkpoint-600/generation_config.json
ADDED
|
@@ -0,0 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"bad_words_ids": [
|
| 3 |
+
[
|
| 4 |
+
64171
|
| 5 |
+
]
|
| 6 |
+
],
|
| 7 |
+
"bos_token_id": 0,
|
| 8 |
+
"decoder_start_token_id": 64171,
|
| 9 |
+
"eos_token_id": 0,
|
| 10 |
+
"forced_eos_token_id": 0,
|
| 11 |
+
"max_length": 512,
|
| 12 |
+
"num_beams": 6,
|
| 13 |
+
"pad_token_id": 64171,
|
| 14 |
+
"transformers_version": "4.27.3"
|
| 15 |
+
}
|
checkpoint-600/optimizer.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2a0e4755674f056eacf97d924962a2e5d841acdb773975e2a4bedd4f5bf6d75b
|
| 3 |
+
size 616276997
|
checkpoint-600/pytorch_model.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3eea04476d9697fa71962473976b3416e5248309674ea553768c0f389a3abe77
|
| 3 |
+
size 308407621
|
checkpoint-600/rng_state.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:47ac1c0f8b136e6cdcb221a075d1f58aff1fff607dc25f74f47fb8a74d814335
|
| 3 |
+
size 14575
|
checkpoint-600/scaler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5dc9eacfeb00bd0bfeb98934a2309be01be65b288e0d747bbfc423b32679169f
|
| 3 |
+
size 557
|
checkpoint-600/scheduler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d2ae1c671816119b4d45adcd15ae17d1daa9fa8260c3798bcd3486935e995c78
|
| 3 |
+
size 627
|
checkpoint-600/source.spm
ADDED
|
Binary file (707 kB). View file
|
|
|
checkpoint-600/special_tokens_map.json
ADDED
|
@@ -0,0 +1,5 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eos_token": "</s>",
|
| 3 |
+
"pad_token": "<pad>",
|
| 4 |
+
"unk_token": "<unk>"
|
| 5 |
+
}
|
checkpoint-600/target.spm
ADDED
|
Binary file (791 kB). View file
|
|
|
checkpoint-600/tokenizer_config.json
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"eos_token": "</s>",
|
| 3 |
+
"model_max_length": 512,
|
| 4 |
+
"pad_token": "<pad>",
|
| 5 |
+
"separate_vocabs": false,
|
| 6 |
+
"source_lang": "mul",
|
| 7 |
+
"sp_model_kwargs": {},
|
| 8 |
+
"special_tokens_map_file": null,
|
| 9 |
+
"target_lang": "eng",
|
| 10 |
+
"tokenizer_class": "MarianTokenizer",
|
| 11 |
+
"unk_token": "<unk>"
|
| 12 |
+
}
|
checkpoint-600/trainer_state.json
ADDED
|
@@ -0,0 +1,78 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"best_metric": 2.9462931156158447,
|
| 3 |
+
"best_model_checkpoint": "/home/azawahry_sunbird_ai/models/m2e/dataset-tags-expanded-tokenizer/checkpoint-150",
|
| 4 |
+
"epoch": 1.024992739634761,
|
| 5 |
+
"global_step": 600,
|
| 6 |
+
"is_hyper_param_search": false,
|
| 7 |
+
"is_local_process_zero": true,
|
| 8 |
+
"is_world_process_zero": true,
|
| 9 |
+
"log_history": [
|
| 10 |
+
{
|
| 11 |
+
"epoch": 0.26,
|
| 12 |
+
"eval_BLEU_ach": 21.4505,
|
| 13 |
+
"eval_BLEU_lgg": 24.4794,
|
| 14 |
+
"eval_BLEU_lug": 35.1488,
|
| 15 |
+
"eval_BLEU_mean": 26.3376,
|
| 16 |
+
"eval_BLEU_nyn": 25.5547,
|
| 17 |
+
"eval_BLEU_teo": 25.0545,
|
| 18 |
+
"eval_loss": 2.9462931156158447,
|
| 19 |
+
"eval_runtime": 59.9936,
|
| 20 |
+
"eval_samples_per_second": 41.671,
|
| 21 |
+
"eval_steps_per_second": 0.833,
|
| 22 |
+
"step": 150
|
| 23 |
+
},
|
| 24 |
+
{
|
| 25 |
+
"epoch": 0.51,
|
| 26 |
+
"eval_BLEU_ach": 22.0556,
|
| 27 |
+
"eval_BLEU_lgg": 23.8502,
|
| 28 |
+
"eval_BLEU_lug": 34.6592,
|
| 29 |
+
"eval_BLEU_mean": 26.1806,
|
| 30 |
+
"eval_BLEU_nyn": 25.075,
|
| 31 |
+
"eval_BLEU_teo": 25.2631,
|
| 32 |
+
"eval_loss": 2.9494848251342773,
|
| 33 |
+
"eval_runtime": 58.1523,
|
| 34 |
+
"eval_samples_per_second": 42.991,
|
| 35 |
+
"eval_steps_per_second": 0.86,
|
| 36 |
+
"step": 300
|
| 37 |
+
},
|
| 38 |
+
{
|
| 39 |
+
"epoch": 0.77,
|
| 40 |
+
"eval_BLEU_ach": 21.6851,
|
| 41 |
+
"eval_BLEU_lgg": 24.2555,
|
| 42 |
+
"eval_BLEU_lug": 34.7523,
|
| 43 |
+
"eval_BLEU_mean": 26.1102,
|
| 44 |
+
"eval_BLEU_nyn": 24.5712,
|
| 45 |
+
"eval_BLEU_teo": 25.2871,
|
| 46 |
+
"eval_loss": 2.9563159942626953,
|
| 47 |
+
"eval_runtime": 57.8967,
|
| 48 |
+
"eval_samples_per_second": 43.18,
|
| 49 |
+
"eval_steps_per_second": 0.864,
|
| 50 |
+
"step": 450
|
| 51 |
+
},
|
| 52 |
+
{
|
| 53 |
+
"epoch": 0.85,
|
| 54 |
+
"learning_rate": 7.150997150997152e-05,
|
| 55 |
+
"loss": 2.4257,
|
| 56 |
+
"step": 500
|
| 57 |
+
},
|
| 58 |
+
{
|
| 59 |
+
"epoch": 1.02,
|
| 60 |
+
"eval_BLEU_ach": 21.8237,
|
| 61 |
+
"eval_BLEU_lgg": 23.3647,
|
| 62 |
+
"eval_BLEU_lug": 35.1063,
|
| 63 |
+
"eval_BLEU_mean": 25.832,
|
| 64 |
+
"eval_BLEU_nyn": 24.8711,
|
| 65 |
+
"eval_BLEU_teo": 23.9942,
|
| 66 |
+
"eval_loss": 2.9721972942352295,
|
| 67 |
+
"eval_runtime": 57.6727,
|
| 68 |
+
"eval_samples_per_second": 43.348,
|
| 69 |
+
"eval_steps_per_second": 0.867,
|
| 70 |
+
"step": 600
|
| 71 |
+
}
|
| 72 |
+
],
|
| 73 |
+
"max_steps": 1755,
|
| 74 |
+
"num_train_epochs": 3,
|
| 75 |
+
"total_flos": 8.641348729778995e+16,
|
| 76 |
+
"trial_name": null,
|
| 77 |
+
"trial_params": null
|
| 78 |
+
}
|
checkpoint-600/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e46a42e4fff4eb7e4702f865fd1022499e6ad9fc3cf4d3f71aef8c990651b558
|
| 3 |
+
size 3707
|
checkpoint-600/vocab.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|