williamplacroix commited on
Commit
08ebb98
·
verified ·
1 Parent(s): fc90626

Finished finetuning grade 2

Browse files
README.md CHANGED
@@ -12,13 +12,12 @@ model-index:
12
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
  should probably proofread and complete it, then remove this comment. -->
14
 
15
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/uds/Graded%20text%20simplification%20training/runs/1mkbeo5j)
16
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/uds/Graded%20text%20simplification%20training/runs/vd7wa4iw)
17
  # text-simplification
18
 
19
  This model is a fine-tuned version of [openai-community/gpt2](https://huggingface.co/openai-community/gpt2) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 0.5178
22
 
23
  ## Model description
24
 
@@ -49,16 +48,16 @@ The following hyperparameters were used during training:
49
 
50
  | Training Loss | Epoch | Step | Validation Loss |
51
  |:-------------:|:-----:|:----:|:---------------:|
52
- | 0.546 | 1.0 | 775 | 0.5205 |
53
- | 0.5405 | 2.0 | 1550 | 0.5186 |
54
- | 0.5368 | 3.0 | 2325 | 0.5162 |
55
- | 0.5338 | 4.0 | 3100 | 0.5176 |
56
- | 0.5322 | 5.0 | 3875 | 0.5189 |
57
- | 0.5304 | 6.0 | 4650 | 0.5179 |
58
- | 0.5284 | 7.0 | 5425 | 0.5184 |
59
- | 0.5278 | 8.0 | 6200 | 0.5176 |
60
- | 0.5276 | 9.0 | 6975 | 0.5177 |
61
- | 0.527 | 10.0 | 7750 | 0.5178 |
62
 
63
 
64
  ### Framework versions
 
12
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
  should probably proofread and complete it, then remove this comment. -->
14
 
15
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/uds/Graded%20text%20simplification%20training/runs/42n7wyjl)
 
16
  # text-simplification
17
 
18
  This model is a fine-tuned version of [openai-community/gpt2](https://huggingface.co/openai-community/gpt2) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 0.3608
21
 
22
  ## Model description
23
 
 
48
 
49
  | Training Loss | Epoch | Step | Validation Loss |
50
  |:-------------:|:-----:|:----:|:---------------:|
51
+ | 0.4017 | 1.0 | 469 | 0.3665 |
52
+ | 0.3951 | 2.0 | 938 | 0.3640 |
53
+ | 0.3907 | 3.0 | 1407 | 0.3641 |
54
+ | 0.3892 | 4.0 | 1876 | 0.3623 |
55
+ | 0.3864 | 5.0 | 2345 | 0.3606 |
56
+ | 0.3847 | 6.0 | 2814 | 0.3614 |
57
+ | 0.383 | 7.0 | 3283 | 0.3606 |
58
+ | 0.3821 | 8.0 | 3752 | 0.3609 |
59
+ | 0.3815 | 9.0 | 4221 | 0.3610 |
60
+ | 0.3806 | 10.0 | 4690 | 0.3608 |
61
 
62
 
63
  ### Framework versions
gpt2-grade-2-finetuned/adapter_config.json CHANGED
@@ -3,6 +3,8 @@
3
  "auto_mapping": null,
4
  "base_model_name_or_path": "openai-community/gpt2",
5
  "bias": "none",
 
 
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
8
  "init_lora_weights": true,
@@ -11,6 +13,7 @@
11
  "layers_to_transform": null,
12
  "loftq_config": {},
13
  "lora_alpha": 32,
 
14
  "lora_dropout": 0.01,
15
  "megatron_config": null,
16
  "megatron_core": "megatron.core",
@@ -20,10 +23,10 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "c_proj",
24
- "c_attn",
25
  "c_fc",
26
- "lm_head"
 
 
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
 
3
  "auto_mapping": null,
4
  "base_model_name_or_path": "openai-community/gpt2",
5
  "bias": "none",
6
+ "eva_config": null,
7
+ "exclude_modules": null,
8
  "fan_in_fan_out": false,
9
  "inference_mode": true,
10
  "init_lora_weights": true,
 
13
  "layers_to_transform": null,
14
  "loftq_config": {},
15
  "lora_alpha": 32,
16
+ "lora_bias": false,
17
  "lora_dropout": 0.01,
18
  "megatron_config": null,
19
  "megatron_core": "megatron.core",
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
 
 
26
  "c_fc",
27
+ "c_attn",
28
+ "lm_head",
29
+ "c_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
gpt2-grade-2-finetuned/adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0192c65a91bf678670e23d3d4f7dded3020605d04f62c90431a9971fb531a4de
3
- size 83581271
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4befae71d3ecf37f12917a3a0825b40c7f14e9517932cb7691543736d133d327
3
+ size 160776023
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:73d8229af7e9a6c1c4932cebc80c86e66cb19d5f222f394f37cbd4a55f5c82de
3
- size 5368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b5faedaca63599a9ee5c38b36992fc6e39d2ddac460d824c93451e91a85592a3
3
+ size 5496