KotshinZ commited on
Commit
10e6aea
·
verified ·
1 Parent(s): 59c0551

Model save

Browse files
README.md CHANGED
@@ -1,17 +1,15 @@
1
  ---
2
  base_model: openai-community/gpt2
3
- datasets: HuggingFaceH4/Bespoke-Stratos-17k
4
  library_name: transformers
5
  model_name: gpt2-RMT-2
6
  tags:
7
  - generated_from_trainer
8
- - open-r1
9
  licence: license
10
  ---
11
 
12
  # Model Card for gpt2-RMT-2
13
 
14
- This model is a fine-tuned version of [openai-community/gpt2](https://huggingface.co/openai-community/gpt2) on the [HuggingFaceH4/Bespoke-Stratos-17k](https://huggingface.co/datasets/HuggingFaceH4/Bespoke-Stratos-17k) dataset.
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
@@ -27,14 +25,14 @@ print(output["generated_text"])
27
 
28
  ## Training procedure
29
 
30
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/s18574s18574-/huggingface/runs/6c36vz5o)
31
 
32
 
33
  This model was trained with SFT.
34
 
35
  ### Framework versions
36
 
37
- - TRL: 0.16.0.dev0
38
  - Transformers: 4.50.0.dev0
39
  - Pytorch: 2.5.1
40
  - Datasets: 3.3.2
 
1
  ---
2
  base_model: openai-community/gpt2
 
3
  library_name: transformers
4
  model_name: gpt2-RMT-2
5
  tags:
6
  - generated_from_trainer
 
7
  licence: license
8
  ---
9
 
10
  # Model Card for gpt2-RMT-2
11
 
12
+ This model is a fine-tuned version of [openai-community/gpt2](https://huggingface.co/openai-community/gpt2).
13
  It has been trained using [TRL](https://github.com/huggingface/trl).
14
 
15
  ## Quick start
 
25
 
26
  ## Training procedure
27
 
28
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/s18574s18574-/huggingface/runs/qz7kfrzp)
29
 
30
 
31
  This model was trained with SFT.
32
 
33
  ### Framework versions
34
 
35
+ - TRL: 0.15.2
36
  - Transformers: 4.50.0.dev0
37
  - Pytorch: 2.5.1
38
  - Datasets: 3.3.2
all_results.json CHANGED
@@ -1,12 +1,12 @@
1
  {
2
- "eval_runtime": 4.1108,
3
  "eval_samples": 100,
4
- "eval_samples_per_second": 266.861,
5
- "eval_steps_per_second": 66.898,
6
  "total_flos": 0.0,
7
- "train_loss": 2.0137617991367285,
8
- "train_runtime": 996.2065,
9
- "train_samples": 16610,
10
- "train_samples_per_second": 92.083,
11
- "train_steps_per_second": 5.756
12
  }
 
1
  {
2
+ "eval_runtime": 0.22,
3
  "eval_samples": 100,
4
+ "eval_samples_per_second": 218.143,
5
+ "eval_steps_per_second": 13.634,
6
  "total_flos": 0.0,
7
+ "train_loss": 4.772621268656716,
8
+ "train_runtime": 29.2884,
9
+ "train_samples": 19883,
10
+ "train_samples_per_second": 36.093,
11
+ "train_steps_per_second": 2.288
12
  }
config.json CHANGED
@@ -1,6 +1,5 @@
1
  {
2
  "_attn_implementation_autoset": true,
3
- "_name_or_path": "openai-community/gpt2",
4
  "activation_function": "gelu_new",
5
  "architectures": [
6
  "GPT2LMHeadModel"
@@ -35,6 +34,6 @@
35
  },
36
  "torch_dtype": "bfloat16",
37
  "transformers_version": "4.50.0.dev0",
38
- "use_cache": true,
39
  "vocab_size": 50257
40
  }
 
1
  {
2
  "_attn_implementation_autoset": true,
 
3
  "activation_function": "gelu_new",
4
  "architectures": [
5
  "GPT2LMHeadModel"
 
34
  },
35
  "torch_dtype": "bfloat16",
36
  "transformers_version": "4.50.0.dev0",
37
+ "use_cache": false,
38
  "vocab_size": 50257
39
  }
eval_results.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
- "eval_runtime": 4.1108,
3
  "eval_samples": 100,
4
- "eval_samples_per_second": 266.861,
5
- "eval_steps_per_second": 66.898
6
  }
 
1
  {
2
+ "eval_runtime": 0.22,
3
  "eval_samples": 100,
4
+ "eval_samples_per_second": 218.143,
5
+ "eval_steps_per_second": 13.634
6
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:992d4bcdb3356997a0a168440f993d5c27b000e48ba6ea568caf99b214bf322f
3
  size 326110968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e3c2740535e03f24f5cb574ba61bf881f6b6c33883fc13ee4948a731051a5ab
3
  size 326110968
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e07ea7caa7f567049c6ac949b18c3c222e52a15c44b05ed9bc048bb7eda52472
3
+ size 248944286
tokenizer_config.json CHANGED
@@ -11,7 +11,6 @@
11
  }
12
  },
13
  "bos_token": "<|endoftext|>",
14
- "chat_template": "{% for message in messages %}{{ message['role'] }}: {{ message['content'] }}\n{% endfor %}",
15
  "clean_up_tokenization_spaces": false,
16
  "eos_token": "<|endoftext|>",
17
  "extra_special_tokens": {},
 
11
  }
12
  },
13
  "bos_token": "<|endoftext|>",
 
14
  "clean_up_tokenization_spaces": false,
15
  "eos_token": "<|endoftext|>",
16
  "extra_special_tokens": {},
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "total_flos": 0.0,
3
- "train_loss": 2.0137617991367285,
4
- "train_runtime": 996.2065,
5
- "train_samples": 16610,
6
- "train_samples_per_second": 92.083,
7
- "train_steps_per_second": 5.756
8
  }
 
1
  {
2
  "total_flos": 0.0,
3
+ "train_loss": 4.772621268656716,
4
+ "train_runtime": 29.2884,
5
+ "train_samples": 19883,
6
+ "train_samples_per_second": 36.093,
7
+ "train_steps_per_second": 2.288
8
  }
trainer_state.json CHANGED
The diff for this file is too large to render. See raw diff
 
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cdef709c27929e8ae9c8e3a5424f8df2c8a0dc367fd631551cc0f01060414d18
3
  size 7352
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be3f4e5fc95a8a097fc233c1e01227f2b9fb455b30c25f6543ddfec1b328f220
3
  size 7352