hrezaei commited on
Commit
d6826a9
·
verified ·
1 Parent(s): af6ef19

Model save

Browse files
Files changed (3) hide show
  1. README.md +3 -17
  2. config.json +1 -0
  3. training_args.bin +1 -1
README.md CHANGED
@@ -2,36 +2,22 @@
2
  library_name: transformers
3
  tags:
4
  - generated_from_trainer
5
- datasets:
6
- - HuggingFaceFW/fineweb
7
  metrics:
8
  - accuracy
9
  model-index:
10
  - name: T5LA
11
- results:
12
- - task:
13
- name: Causal Language Modeling
14
- type: text-generation
15
- dataset:
16
- name: HuggingFaceFW/fineweb sample-10BT
17
- type: HuggingFaceFW/fineweb
18
- args: sample-10BT
19
- metrics:
20
- - name: Accuracy
21
- type: accuracy
22
- value: 0.0322300343763811
23
  ---
24
 
25
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
26
  should probably proofread and complete it, then remove this comment. -->
27
 
28
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/uoy/llm_training/runs/pzcq293g)
29
  # T5LA
30
 
31
- This model is a fine-tuned version of [](https://huggingface.co/) on the HuggingFaceFW/fineweb sample-10BT dataset.
32
  It achieves the following results on the evaluation set:
33
- - Loss: 5.5470
34
  - Accuracy: 0.0322
 
35
 
36
  ## Model description
37
 
 
2
  library_name: transformers
3
  tags:
4
  - generated_from_trainer
 
 
5
  metrics:
6
  - accuracy
7
  model-index:
8
  - name: T5LA
9
+ results: []
 
 
 
 
 
 
 
 
 
 
 
10
  ---
11
 
12
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
  should probably proofread and complete it, then remove this comment. -->
14
 
 
15
  # T5LA
16
 
17
+ This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
 
19
  - Accuracy: 0.0322
20
+ - Loss: 5.5470
21
 
22
  ## Model description
23
 
config.json CHANGED
@@ -16,6 +16,7 @@
16
  "is_gated_act": false,
17
  "layer_norm_epsilon": 1e-06,
18
  "lookahead_size": 2,
 
19
  "model_type": "t5la",
20
  "num_decoder_layers": 6,
21
  "num_heads": 8,
 
16
  "is_gated_act": false,
17
  "layer_norm_epsilon": 1e-06,
18
  "lookahead_size": 2,
19
+ "lookahead_type": "la",
20
  "model_type": "t5la",
21
  "num_decoder_layers": 6,
22
  "num_heads": 8,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:250fa1af03ffaa8bda0e2278749102dd8574803e5d42d069bf5be28611ad9412
3
  size 5432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b0f480ede6b17d1fc31d60ba9c13c539402fa821bdb403c1b911e315bfca608
3
  size 5432