hrezaei commited on
Commit
bd4356c
·
1 Parent(s): 512cfc3

End of training

Browse files
README.md CHANGED
@@ -3,25 +3,25 @@ library_name: transformers
3
  tags:
4
  - generated_from_trainer
5
  datasets:
6
- - generator
7
  metrics:
8
  - accuracy
9
  model-index:
10
  - name: T5Lae-Large-WeightedLoss
11
  results:
12
  - task:
13
- name: Sequence-to-sequence Language Modeling
14
- type: text2text-generation
15
  dataset:
16
- name: generator
17
- type: generator
18
  config: default
19
  split: train
20
- args: default
21
  metrics:
22
  - name: Accuracy
23
  type: accuracy
24
- value: 0.02610743639921722
25
  ---
26
 
27
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -29,13 +29,13 @@ should probably proofread and complete it, then remove this comment. -->
29
 
30
  # T5Lae-Large-WeightedLoss
31
 
32
- This model is a fine-tuned version of [](https://huggingface.co/) on the generator dataset.
33
  It achieves the following results on the evaluation set:
34
- - Perplexity: 56.6355
35
- - Loss: 4.0366
36
- - Accuracy: 0.0261
37
- - Lookahead Perplexity: 672.8140
38
- - Lookahead Loss: 6.5115
39
 
40
  ## Model description
41
 
 
3
  tags:
4
  - generated_from_trainer
5
  datasets:
6
+ - HuggingFaceFW/fineweb
7
  metrics:
8
  - accuracy
9
  model-index:
10
  - name: T5Lae-Large-WeightedLoss
11
  results:
12
  - task:
13
+ name: Causal Language Modeling
14
+ type: text-generation
15
  dataset:
16
+ name: HuggingFaceFW/fineweb sample-350BT
17
+ type: HuggingFaceFW/fineweb
18
  config: default
19
  split: train
20
+ args: sample-350BT
21
  metrics:
22
  - name: Accuracy
23
  type: accuracy
24
+ value: 0.026029158512720158
25
  ---
26
 
27
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
29
 
30
  # T5Lae-Large-WeightedLoss
31
 
32
+ This model is a fine-tuned version of [](https://huggingface.co/) on the HuggingFaceFW/fineweb sample-350BT dataset.
33
  It achieves the following results on the evaluation set:
34
+ - Perplexity: 56.6106
35
+ - Loss: 4.0362
36
+ - Accuracy: 0.0260
37
+ - Lookahead Perplexity: 672.4668
38
+ - Lookahead Loss: 6.5110
39
 
40
  ## Model description
41
 
all_results.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_accuracy": 0.026029158512720158,
3
+ "eval_lookahead_loss": 6.510952686786651,
4
+ "eval_loss": 4.036196708679199,
5
+ "eval_perplexity": 56.61062614576278,
6
+ "eval_runtime": 502.5542,
7
+ "eval_samples": 10000,
8
+ "eval_samples_per_second": 19.898,
9
+ "eval_steps_per_second": 4.975,
10
+ "total_flos": 4.621498060193661e+18,
11
+ "train_loss": 0.2757485848851502,
12
+ "train_runtime": 24028.7714,
13
+ "train_samples": 2000000,
14
+ "train_samples_per_second": 87.277,
15
+ "train_steps_per_second": 21.819
16
+ }
eval_results.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_accuracy": 0.026029158512720158,
3
+ "eval_lookahead_loss": 6.510952686786651,
4
+ "eval_loss": 4.036196708679199,
5
+ "eval_perplexity": 56.61062614576278,
6
+ "eval_runtime": 502.5542,
7
+ "eval_samples": 10000,
8
+ "eval_samples_per_second": 19.898,
9
+ "eval_steps_per_second": 4.975
10
+ }
runs/Oct02_11-42-07_gpu23.viking2.yor.alces.network/events.out.tfevents.1759426779.gpu23.viking2.yor.alces.network.1856724.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:39410f128a4af8b7192c39f4b305961b7e63876faed0aeff90d1541ec58ed2d4
3
+ size 596
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "total_flos": 4.621498060193661e+18,
3
+ "train_loss": 0.2757485848851502,
4
+ "train_runtime": 24028.7714,
5
+ "train_samples": 2000000,
6
+ "train_samples_per_second": 87.277,
7
+ "train_steps_per_second": 21.819
8
+ }
trainer_state.json ADDED
The diff for this file is too large to render. See raw diff