ladoza03 commited on
Commit
8459dfc
·
verified ·
1 Parent(s): 077e491

End of training

Browse files
README.md CHANGED
@@ -13,6 +13,7 @@ model-index:
13
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
  should probably proofread and complete it, then remove this comment. -->
15
 
 
16
  # tst-translation
17
 
18
  This model is a fine-tuned version of [ladoza03/tst-translation](https://huggingface.co/ladoza03/tst-translation) on an unknown dataset.
@@ -20,9 +21,9 @@ It achieves the following results on the evaluation set:
20
  - eval_loss: 1.1219
21
  - eval_bleu: 30.1585
22
  - eval_gen_len: 55.5424
23
- - eval_runtime: 820.9592
24
- - eval_samples_per_second: 2.457
25
- - eval_steps_per_second: 0.308
26
  - step: 0
27
 
28
  ## Model description
@@ -52,7 +53,7 @@ The following hyperparameters were used during training:
52
 
53
  ### Framework versions
54
 
55
- - Transformers 4.40.0.dev0
56
  - Pytorch 2.1.2
57
  - Datasets 2.1.0
58
- - Tokenizers 0.15.2
 
13
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
  should probably proofread and complete it, then remove this comment. -->
15
 
16
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/ladoza03/huggingface/runs/cuqrcxzo)
17
  # tst-translation
18
 
19
  This model is a fine-tuned version of [ladoza03/tst-translation](https://huggingface.co/ladoza03/tst-translation) on an unknown dataset.
 
21
  - eval_loss: 1.1219
22
  - eval_bleu: 30.1585
23
  - eval_gen_len: 55.5424
24
+ - eval_runtime: 826.2423
25
+ - eval_samples_per_second: 2.441
26
+ - eval_steps_per_second: 0.306
27
  - step: 0
28
 
29
  ## Model description
 
53
 
54
  ### Framework versions
55
 
56
+ - Transformers 4.41.0.dev0
57
  - Pytorch 2.1.2
58
  - Datasets 2.1.0
59
+ - Tokenizers 0.19.1
all_results.json CHANGED
@@ -2,15 +2,15 @@
2
  "eval_bleu": 30.1585,
3
  "eval_gen_len": 55.5424,
4
  "eval_loss": 1.121874213218689,
5
- "eval_runtime": 820.9592,
6
  "eval_samples": 2017,
7
- "eval_samples_per_second": 2.457,
8
- "eval_steps_per_second": 0.308,
9
- "predict_bleu": 29.7226,
10
- "predict_gen_len": 55.3034,
11
- "predict_loss": 1.1471892595291138,
12
- "predict_runtime": 830.1491,
13
- "predict_samples": 2017,
14
- "predict_samples_per_second": 2.43,
15
- "predict_steps_per_second": 0.305
16
  }
 
2
  "eval_bleu": 30.1585,
3
  "eval_gen_len": 55.5424,
4
  "eval_loss": 1.121874213218689,
5
+ "eval_runtime": 826.2423,
6
  "eval_samples": 2017,
7
+ "eval_samples_per_second": 2.441,
8
+ "eval_steps_per_second": 0.306,
9
+ "predict_bleu": 35.5848,
10
+ "predict_gen_len": 59.208,
11
+ "predict_loss": 1.1299843788146973,
12
+ "predict_runtime": 385.8553,
13
+ "predict_samples": 1000,
14
+ "predict_samples_per_second": 2.592,
15
+ "predict_steps_per_second": 0.324
16
  }
config.json CHANGED
@@ -27,7 +27,7 @@
27
  "tie_word_embeddings": false,
28
  "tokenizer_class": "T5Tokenizer",
29
  "torch_dtype": "float32",
30
- "transformers_version": "4.40.0.dev0",
31
  "use_cache": true,
32
  "vocab_size": 250112
33
  }
 
27
  "tie_word_embeddings": false,
28
  "tokenizer_class": "T5Tokenizer",
29
  "torch_dtype": "float32",
30
+ "transformers_version": "4.41.0.dev0",
31
  "use_cache": true,
32
  "vocab_size": 250112
33
  }
eval_results.json CHANGED
@@ -2,8 +2,8 @@
2
  "eval_bleu": 30.1585,
3
  "eval_gen_len": 55.5424,
4
  "eval_loss": 1.121874213218689,
5
- "eval_runtime": 820.9592,
6
  "eval_samples": 2017,
7
- "eval_samples_per_second": 2.457,
8
- "eval_steps_per_second": 0.308
9
  }
 
2
  "eval_bleu": 30.1585,
3
  "eval_gen_len": 55.5424,
4
  "eval_loss": 1.121874213218689,
5
+ "eval_runtime": 826.2423,
6
  "eval_samples": 2017,
7
+ "eval_samples_per_second": 2.441,
8
+ "eval_steps_per_second": 0.306
9
  }
generated_predictions.txt CHANGED
The diff for this file is too large to render. See raw diff
 
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "decoder_start_token_id": 0,
3
  "eos_token_id": 1,
4
  "pad_token_id": 0,
5
- "transformers_version": "4.40.0.dev0"
6
  }
 
2
  "decoder_start_token_id": 0,
3
  "eos_token_id": 1,
4
  "pad_token_id": 0,
5
+ "transformers_version": "4.41.0.dev0"
6
  }
predict_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "predict_bleu": 29.7226,
3
- "predict_gen_len": 55.3034,
4
- "predict_loss": 1.1471892595291138,
5
- "predict_runtime": 830.1491,
6
- "predict_samples": 2017,
7
- "predict_samples_per_second": 2.43,
8
- "predict_steps_per_second": 0.305
9
  }
 
1
  {
2
+ "predict_bleu": 35.5848,
3
+ "predict_gen_len": 59.208,
4
+ "predict_loss": 1.1299843788146973,
5
+ "predict_runtime": 385.8553,
6
+ "predict_samples": 1000,
7
+ "predict_samples_per_second": 2.592,
8
+ "predict_steps_per_second": 0.324
9
  }
runs/May10_15-50-18_69933a5bff83/events.out.tfevents.1715357146.69933a5bff83.147.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8bb015d875f444e4df82fa6f53089035dd3915e3ae378f3c1624b1cbb87f9c77
3
+ size 390
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a0910f648726ceada34086ae80066cd253863e183cbb52ae566659a2d37716f0
3
- size 16330660
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84ec7ac09e74719df0d7ac26684f6bb9939553133a2b7916d91c08ff9d959a2f
3
+ size 16330638
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b21ad9438d948785ae0b1c13cb76a669135fec8766dea2bc2b199ac5c6a644ba
3
- size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c62ab815e9eb54bc24ed0c1881b2a1a2de23a5c118a88cc1d3a4991fda38ca1f
3
+ size 5304