hrezaei commited on
Commit
a45e432
·
1 Parent(s): 21bb5a7

End of training

Browse files
README.md CHANGED
@@ -3,25 +3,25 @@ library_name: transformers
3
  tags:
4
  - generated_from_trainer
5
  datasets:
6
- - generator
7
  metrics:
8
  - accuracy
9
  model-index:
10
  - name: T5Laa-Large-WeightedLoss
11
  results:
12
  - task:
13
- name: Sequence-to-sequence Language Modeling
14
- type: text2text-generation
15
  dataset:
16
- name: generator
17
- type: generator
18
  config: default
19
  split: train
20
- args: default
21
  metrics:
22
  - name: Accuracy
23
  type: accuracy
24
- value: 0.03652915851272016
25
  ---
26
 
27
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -29,13 +29,13 @@ should probably proofread and complete it, then remove this comment. -->
29
 
30
  # T5Laa-Large-WeightedLoss
31
 
32
- This model is a fine-tuned version of [](https://huggingface.co/) on the generator dataset.
33
  It achieves the following results on the evaluation set:
34
- - Perplexity: 186.2719
35
- - Loss: 5.2272
36
  - Accuracy: 0.0365
37
- - Lookahead Perplexity: 1785.8340
38
- - Lookahead Loss: 7.4876
39
 
40
  ## Model description
41
 
 
3
  tags:
4
  - generated_from_trainer
5
  datasets:
6
+ - HuggingFaceFW/fineweb
7
  metrics:
8
  - accuracy
9
  model-index:
10
  - name: T5Laa-Large-WeightedLoss
11
  results:
12
  - task:
13
+ name: Causal Language Modeling
14
+ type: text-generation
15
  dataset:
16
+ name: HuggingFaceFW/fineweb sample-350BT
17
+ type: HuggingFaceFW/fineweb
18
  config: default
19
  split: train
20
+ args: sample-350BT
21
  metrics:
22
  - name: Accuracy
23
  type: accuracy
24
+ value: 0.03652074363992172
25
  ---
26
 
27
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
29
 
30
  # T5Laa-Large-WeightedLoss
31
 
32
+ This model is a fine-tuned version of [](https://huggingface.co/) on the HuggingFaceFW/fineweb sample-350BT dataset.
33
  It achieves the following results on the evaluation set:
34
+ - Perplexity: 186.1024
35
+ - Loss: 5.2263
36
  - Accuracy: 0.0365
37
+ - Lookahead Perplexity: 1785.6071
38
+ - Lookahead Loss: 7.4875
39
 
40
  ## Model description
41
 
all_results.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_accuracy": 0.03652074363992172,
3
+ "eval_lookahead_loss": 7.4875137342453,
4
+ "eval_loss": 5.226296901702881,
5
+ "eval_perplexity": 186.10237056707334,
6
+ "eval_runtime": 490.6633,
7
+ "eval_samples": 10000,
8
+ "eval_samples_per_second": 20.381,
9
+ "eval_steps_per_second": 5.095,
10
+ "total_flos": 4.833448717656785e+18,
11
+ "train_loss": 1.2404409057926387,
12
+ "train_runtime": 78461.8603,
13
+ "train_samples": 2000000,
14
+ "train_samples_per_second": 26.728,
15
+ "train_steps_per_second": 6.682
16
+ }
eval_results.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_accuracy": 0.03652074363992172,
3
+ "eval_lookahead_loss": 7.4875137342453,
4
+ "eval_loss": 5.226296901702881,
5
+ "eval_perplexity": 186.10237056707334,
6
+ "eval_runtime": 490.6633,
7
+ "eval_samples": 10000,
8
+ "eval_samples_per_second": 20.381,
9
+ "eval_steps_per_second": 5.095
10
+ }
runs/Oct04_12-19-27_gpu23.viking2.yor.alces.network/events.out.tfevents.1759656120.gpu23.viking2.yor.alces.network.3198493.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:682284128fd937e7a16b3e6c46683778515aed16c28258b8962c0917d541a79c
3
+ size 596
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "total_flos": 4.833448717656785e+18,
3
+ "train_loss": 1.2404409057926387,
4
+ "train_runtime": 78461.8603,
5
+ "train_samples": 2000000,
6
+ "train_samples_per_second": 26.728,
7
+ "train_steps_per_second": 6.682
8
+ }
trainer_state.json ADDED
The diff for this file is too large to render. See raw diff