KotshinZ commited on
Commit
b585541
·
verified ·
1 Parent(s): af71d05

Model save

Browse files
README.md CHANGED
@@ -1,11 +1,9 @@
1
  ---
2
  base_model: openai-community/gpt2
3
- datasets: HuggingFaceFW/fineweb-edu
4
  library_name: transformers
5
  model_name: gpt2-RMT-2
6
  tags:
7
  - generated_from_trainer
8
- - open-r1
9
  - trl
10
  - sft
11
  licence: license
@@ -13,7 +11,7 @@ licence: license
13
 
14
  # Model Card for gpt2-RMT-2
15
 
16
- This model is a fine-tuned version of [openai-community/gpt2](https://huggingface.co/openai-community/gpt2) on the [HuggingFaceFW/fineweb-edu](https://huggingface.co/datasets/HuggingFaceFW/fineweb-edu) dataset.
17
  It has been trained using [TRL](https://github.com/huggingface/trl).
18
 
19
  ## Quick start
@@ -29,7 +27,7 @@ print(output["generated_text"])
29
 
30
  ## Training procedure
31
 
32
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/s18574s18574-/huggingface/runs/qothcfg8)
33
 
34
 
35
  This model was trained with SFT.
 
1
  ---
2
  base_model: openai-community/gpt2
 
3
  library_name: transformers
4
  model_name: gpt2-RMT-2
5
  tags:
6
  - generated_from_trainer
 
7
  - trl
8
  - sft
9
  licence: license
 
11
 
12
  # Model Card for gpt2-RMT-2
13
 
14
+ This model is a fine-tuned version of [openai-community/gpt2](https://huggingface.co/openai-community/gpt2).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
 
27
 
28
  ## Training procedure
29
 
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/s18574s18574-/huggingface/runs/1xooe0sm)
31
 
32
 
33
  This model was trained with SFT.
all_results.json CHANGED
@@ -3,10 +3,10 @@
3
  "eval_samples": 100,
4
  "eval_samples_per_second": 100.273,
5
  "eval_steps_per_second": 12.534,
6
- "total_flos": 550151290617856.0,
7
- "train_loss": 4.790811567164179,
8
- "train_runtime": 29.3533,
9
  "train_samples": 19883,
10
- "train_samples_per_second": 36.013,
11
- "train_steps_per_second": 2.283
12
  }
 
3
  "eval_samples": 100,
4
  "eval_samples_per_second": 100.273,
5
  "eval_steps_per_second": 12.534,
6
+ "total_flos": 541940076511232.0,
7
+ "train_loss": 5.546875,
8
+ "train_runtime": 36.3584,
9
  "train_samples": 19883,
10
+ "train_samples_per_second": 29.074,
11
+ "train_steps_per_second": 0.908
12
  }
config.json CHANGED
@@ -33,6 +33,6 @@
33
  },
34
  "torch_dtype": "bfloat16",
35
  "transformers_version": "4.50.0.dev0",
36
- "use_cache": true,
37
  "vocab_size": 50257
38
  }
 
33
  },
34
  "torch_dtype": "bfloat16",
35
  "transformers_version": "4.50.0.dev0",
36
+ "use_cache": false,
37
  "vocab_size": 50257
38
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8eead2c9b2060803ef718a002144657eb03b943597178519430d4167776a9619
3
  size 248912768
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ef6b24bad9283be8a3270f3898cb6e1d27404fa045313443fde712995a1cabc
3
  size 248912768
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "total_flos": 550151290617856.0,
3
- "train_loss": 4.790811567164179,
4
- "train_runtime": 29.3533,
5
  "train_samples": 19883,
6
- "train_samples_per_second": 36.013,
7
- "train_steps_per_second": 2.283
8
  }
 
1
  {
2
+ "total_flos": 541940076511232.0,
3
+ "train_loss": 5.546875,
4
+ "train_runtime": 36.3584,
5
  "train_samples": 19883,
6
+ "train_samples_per_second": 29.074,
7
+ "train_steps_per_second": 0.908
8
  }
trainer_state.json CHANGED
@@ -1,74 +1,50 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.10136157337367625,
5
  "eval_steps": 100,
6
- "global_step": 67,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
- {
12
- "epoch": 0.015128593040847202,
13
- "grad_norm": 31.61330947025634,
14
- "learning_rate": 1.9876883405951378e-05,
15
- "loss": 7.1484,
16
- "mean_token_accuracy": 0.1398914076387882,
17
- "step": 10
18
- },
19
  {
20
  "epoch": 0.030257186081694403,
21
- "grad_norm": 10.03732149051733,
22
- "learning_rate": 1.777145961456971e-05,
23
- "loss": 5.5297,
24
- "mean_token_accuracy": 0.18087630867958068,
25
- "step": 20
26
- },
27
- {
28
- "epoch": 0.0453857791225416,
29
- "grad_norm": 5.344272287258441,
30
- "learning_rate": 1.3583679495453e-05,
31
- "loss": 4.6375,
32
- "mean_token_accuracy": 0.24480567798018454,
33
- "step": 30
34
  },
35
  {
36
  "epoch": 0.060514372163388806,
37
- "grad_norm": 3.8172946073830114,
38
- "learning_rate": 8.43565534959769e-06,
39
- "loss": 4.1672,
40
- "mean_token_accuracy": 0.28589935302734376,
41
- "step": 40
42
- },
43
- {
44
- "epoch": 0.07564296520423601,
45
- "grad_norm": 3.077300488264507,
46
- "learning_rate": 3.7067960895016277e-06,
47
- "loss": 3.9633,
48
- "mean_token_accuracy": 0.3073825791478157,
49
- "step": 50
50
  },
51
  {
52
  "epoch": 0.0907715582450832,
53
- "grad_norm": 2.9692193643977336,
54
- "learning_rate": 6.641957350279838e-07,
55
- "loss": 3.9047,
56
- "mean_token_accuracy": 0.31336415857076644,
57
- "step": 60
58
  },
59
  {
60
- "epoch": 0.10136157337367625,
61
- "mean_token_accuracy": 0.31038159344877514,
62
- "step": 67,
63
- "total_flos": 550151290617856.0,
64
- "train_loss": 4.790811567164179,
65
- "train_runtime": 29.3533,
66
- "train_samples_per_second": 36.013,
67
- "train_steps_per_second": 2.283
68
  }
69
  ],
70
  "logging_steps": 10,
71
- "max_steps": 67,
72
  "num_input_tokens_seen": 0,
73
  "num_train_epochs": 1,
74
  "save_steps": 500,
@@ -84,7 +60,7 @@
84
  "attributes": {}
85
  }
86
  },
87
- "total_flos": 550151290617856.0,
88
  "train_batch_size": 8,
89
  "trial_name": null,
90
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.09984871406959153,
5
  "eval_steps": 100,
6
+ "global_step": 33,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
 
 
 
 
 
 
 
 
11
  {
12
  "epoch": 0.030257186081694403,
13
+ "grad_norm": 18.3522235886847,
14
+ "learning_rate": 1.796093065705644e-05,
15
+ "loss": 6.8937,
16
+ "mean_token_accuracy": 0.1434693343937397,
17
+ "step": 10
 
 
 
 
 
 
 
 
18
  },
19
  {
20
  "epoch": 0.060514372163388806,
21
+ "grad_norm": 10.614870433655174,
22
+ "learning_rate": 8.382180034472353e-06,
23
+ "loss": 5.3195,
24
+ "mean_token_accuracy": 0.1919048972427845,
25
+ "step": 20
 
 
 
 
 
 
 
 
26
  },
27
  {
28
  "epoch": 0.0907715582450832,
29
+ "grad_norm": 7.571339027683935,
30
+ "learning_rate": 5.234682881719766e-07,
31
+ "loss": 4.707,
32
+ "mean_token_accuracy": 0.23758598268032075,
33
+ "step": 30
34
  },
35
  {
36
+ "epoch": 0.09984871406959153,
37
+ "mean_token_accuracy": 0.24127203722794852,
38
+ "step": 33,
39
+ "total_flos": 541940076511232.0,
40
+ "train_loss": 5.546875,
41
+ "train_runtime": 36.3584,
42
+ "train_samples_per_second": 29.074,
43
+ "train_steps_per_second": 0.908
44
  }
45
  ],
46
  "logging_steps": 10,
47
+ "max_steps": 33,
48
  "num_input_tokens_seen": 0,
49
  "num_train_epochs": 1,
50
  "save_steps": 500,
 
60
  "attributes": {}
61
  }
62
  },
63
+ "total_flos": 541940076511232.0,
64
  "train_batch_size": 8,
65
  "trial_name": null,
66
  "trial_params": null
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:68fc74c433f058a40ea1103a36e974aec030be05fef6c7bef96ba33c56290c81
3
  size 7352
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:baf06d605ca2ac794fb0cb09d28d4e25826d2c1b68fdaaefca458cb0a917d847
3
  size 7352