KotshinZ commited on
Commit
c657b71
·
verified ·
1 Parent(s): 9df87bc

Model save

Browse files
README.md CHANGED
@@ -1,17 +1,17 @@
1
  ---
2
  base_model: openai-community/gpt2
3
- datasets: HuggingFaceFW/fineweb-edu
4
  library_name: transformers
5
  model_name: gpt2-RMT-2
6
  tags:
7
  - generated_from_trainer
8
- - open-r1
 
9
  licence: license
10
  ---
11
 
12
  # Model Card for gpt2-RMT-2
13
 
14
- This model is a fine-tuned version of [openai-community/gpt2](https://huggingface.co/openai-community/gpt2) on the [HuggingFaceFW/fineweb-edu](https://huggingface.co/datasets/HuggingFaceFW/fineweb-edu) dataset.
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
@@ -27,7 +27,7 @@ print(output["generated_text"])
27
 
28
  ## Training procedure
29
 
30
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/s18574s18574-/huggingface/runs/qz7kfrzp)
31
 
32
 
33
  This model was trained with SFT.
 
1
  ---
2
  base_model: openai-community/gpt2
 
3
  library_name: transformers
4
  model_name: gpt2-RMT-2
5
  tags:
6
  - generated_from_trainer
7
+ - trl
8
+ - sft
9
  licence: license
10
  ---
11
 
12
  # Model Card for gpt2-RMT-2
13
 
14
+ This model is a fine-tuned version of [openai-community/gpt2](https://huggingface.co/openai-community/gpt2).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
 
27
 
28
  ## Training procedure
29
 
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/s18574s18574-/huggingface/runs/qothcfg8)
31
 
32
 
33
  This model was trained with SFT.
all_results.json CHANGED
@@ -3,10 +3,10 @@
3
  "eval_samples": 100,
4
  "eval_samples_per_second": 99.696,
5
  "eval_steps_per_second": 12.462,
6
- "total_flos": 0.0,
7
- "train_loss": 4.772621268656716,
8
- "train_runtime": 29.2884,
9
  "train_samples": 19883,
10
- "train_samples_per_second": 36.093,
11
- "train_steps_per_second": 2.288
12
  }
 
3
  "eval_samples": 100,
4
  "eval_samples_per_second": 99.696,
5
  "eval_steps_per_second": 12.462,
6
+ "total_flos": 550151290617856.0,
7
+ "train_loss": 4.790811567164179,
8
+ "train_runtime": 29.3533,
9
  "train_samples": 19883,
10
+ "train_samples_per_second": 36.013,
11
+ "train_steps_per_second": 2.283
12
  }
config.json CHANGED
@@ -1,8 +1,7 @@
1
  {
2
- "_attn_implementation_autoset": true,
3
  "activation_function": "gelu_new",
4
  "architectures": [
5
- "GPT2LMHeadModel"
6
  ],
7
  "attn_pdrop": 0.1,
8
  "bos_token_id": 50256,
@@ -32,7 +31,8 @@
32
  "max_length": 50
33
  }
34
  },
 
35
  "transformers_version": "4.50.0.dev0",
36
- "use_cache": true,
37
  "vocab_size": 50257
38
  }
 
1
  {
 
2
  "activation_function": "gelu_new",
3
  "architectures": [
4
+ "RecurrentWrapper"
5
  ],
6
  "attn_pdrop": 0.1,
7
  "bos_token_id": 50256,
 
31
  "max_length": 50
32
  }
33
  },
34
+ "torch_dtype": "bfloat16",
35
  "transformers_version": "4.50.0.dev0",
36
+ "use_cache": false,
37
  "vocab_size": 50257
38
  }
generation_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "transformers_version": "4.50.0.dev0"
4
+ }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2e3c2740535e03f24f5cb574ba61bf881f6b6c33883fc13ee4948a731051a5ab
3
- size 326110968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8eead2c9b2060803ef718a002144657eb03b943597178519430d4167776a9619
3
+ size 248912768
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "total_flos": 0.0,
3
- "train_loss": 4.772621268656716,
4
- "train_runtime": 29.2884,
5
  "train_samples": 19883,
6
- "train_samples_per_second": 36.093,
7
- "train_steps_per_second": 2.288
8
  }
 
1
  {
2
+ "total_flos": 550151290617856.0,
3
+ "train_loss": 4.790811567164179,
4
+ "train_runtime": 29.3533,
5
  "train_samples": 19883,
6
+ "train_samples_per_second": 36.013,
7
+ "train_steps_per_second": 2.283
8
  }
trainer_state.json CHANGED
@@ -10,61 +10,61 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.015128593040847202,
13
- "grad_norm": 33.7647987029562,
14
  "learning_rate": 1.9876883405951378e-05,
15
- "loss": 7.125,
16
- "mean_token_accuracy": 0.14012572318315505,
17
  "step": 10
18
  },
19
  {
20
  "epoch": 0.030257186081694403,
21
- "grad_norm": 10.612758737576552,
22
  "learning_rate": 1.777145961456971e-05,
23
- "loss": 5.5062,
24
- "mean_token_accuracy": 0.1818697050213814,
25
  "step": 20
26
  },
27
  {
28
  "epoch": 0.0453857791225416,
29
- "grad_norm": 5.133709750875046,
30
  "learning_rate": 1.3583679495453e-05,
31
- "loss": 4.6188,
32
- "mean_token_accuracy": 0.24615719988942147,
33
  "step": 30
34
  },
35
  {
36
  "epoch": 0.060514372163388806,
37
- "grad_norm": 3.7566168018221653,
38
  "learning_rate": 8.43565534959769e-06,
39
- "loss": 4.1492,
40
- "mean_token_accuracy": 0.2883812889456749,
41
  "step": 40
42
  },
43
  {
44
  "epoch": 0.07564296520423601,
45
- "grad_norm": 3.0972835436341253,
46
  "learning_rate": 3.7067960895016277e-06,
47
- "loss": 3.9453,
48
- "mean_token_accuracy": 0.30843553096055987,
49
  "step": 50
50
  },
51
  {
52
  "epoch": 0.0907715582450832,
53
- "grad_norm": 2.978044733724223,
54
  "learning_rate": 6.641957350279838e-07,
55
- "loss": 3.8914,
56
- "mean_token_accuracy": 0.3149686068296432,
57
  "step": 60
58
  },
59
  {
60
  "epoch": 0.10136157337367625,
61
- "mean_token_accuracy": 0.3118449641125543,
62
  "step": 67,
63
- "total_flos": 0.0,
64
- "train_loss": 4.772621268656716,
65
- "train_runtime": 29.2884,
66
- "train_samples_per_second": 36.093,
67
- "train_steps_per_second": 2.288
68
  }
69
  ],
70
  "logging_steps": 10,
@@ -84,7 +84,7 @@
84
  "attributes": {}
85
  }
86
  },
87
- "total_flos": 0.0,
88
  "train_batch_size": 8,
89
  "trial_name": null,
90
  "trial_params": null
 
10
  "log_history": [
11
  {
12
  "epoch": 0.015128593040847202,
13
+ "grad_norm": 31.61330947025634,
14
  "learning_rate": 1.9876883405951378e-05,
15
+ "loss": 7.1484,
16
+ "mean_token_accuracy": 0.1398914076387882,
17
  "step": 10
18
  },
19
  {
20
  "epoch": 0.030257186081694403,
21
+ "grad_norm": 10.03732149051733,
22
  "learning_rate": 1.777145961456971e-05,
23
+ "loss": 5.5297,
24
+ "mean_token_accuracy": 0.18087630867958068,
25
  "step": 20
26
  },
27
  {
28
  "epoch": 0.0453857791225416,
29
+ "grad_norm": 5.344272287258441,
30
  "learning_rate": 1.3583679495453e-05,
31
+ "loss": 4.6375,
32
+ "mean_token_accuracy": 0.24480567798018454,
33
  "step": 30
34
  },
35
  {
36
  "epoch": 0.060514372163388806,
37
+ "grad_norm": 3.8172946073830114,
38
  "learning_rate": 8.43565534959769e-06,
39
+ "loss": 4.1672,
40
+ "mean_token_accuracy": 0.28589935302734376,
41
  "step": 40
42
  },
43
  {
44
  "epoch": 0.07564296520423601,
45
+ "grad_norm": 3.077300488264507,
46
  "learning_rate": 3.7067960895016277e-06,
47
+ "loss": 3.9633,
48
+ "mean_token_accuracy": 0.3073825791478157,
49
  "step": 50
50
  },
51
  {
52
  "epoch": 0.0907715582450832,
53
+ "grad_norm": 2.9692193643977336,
54
  "learning_rate": 6.641957350279838e-07,
55
+ "loss": 3.9047,
56
+ "mean_token_accuracy": 0.31336415857076644,
57
  "step": 60
58
  },
59
  {
60
  "epoch": 0.10136157337367625,
61
+ "mean_token_accuracy": 0.31038159344877514,
62
  "step": 67,
63
+ "total_flos": 550151290617856.0,
64
+ "train_loss": 4.790811567164179,
65
+ "train_runtime": 29.3533,
66
+ "train_samples_per_second": 36.013,
67
+ "train_steps_per_second": 2.283
68
  }
69
  ],
70
  "logging_steps": 10,
 
84
  "attributes": {}
85
  }
86
  },
87
+ "total_flos": 550151290617856.0,
88
  "train_batch_size": 8,
89
  "trial_name": null,
90
  "trial_params": null
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:be3f4e5fc95a8a097fc233c1e01227f2b9fb455b30c25f6543ddfec1b328f220
3
  size 7352
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:68fc74c433f058a40ea1103a36e974aec030be05fef6c7bef96ba33c56290c81
3
  size 7352