gaodrew commited on
Commit
3d738db
·
verified ·
1 Parent(s): e21f316

Model save

Browse files
README.md CHANGED
@@ -13,7 +13,7 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  This model is a fine-tuned version of [](https://huggingface.co/) on the None dataset.
15
  It achieves the following results on the evaluation set:
16
- - Loss: 6.8682
17
 
18
  ## Model description
19
 
@@ -46,8 +46,8 @@ The following hyperparameters were used during training:
46
 
47
  | Training Loss | Epoch | Step | Validation Loss |
48
  |:-------------:|:------:|:----:|:---------------:|
49
- | 7.2874 | 0.5214 | 500 | 7.1310 |
50
- | 6.8115 | 1.0428 | 1000 | 6.8682 |
51
 
52
 
53
  ### Framework versions
 
13
 
14
  This model is a fine-tuned version of [](https://huggingface.co/) on the None dataset.
15
  It achieves the following results on the evaluation set:
16
+ - Loss: 6.9015
17
 
18
  ## Model description
19
 
 
46
 
47
  | Training Loss | Epoch | Step | Validation Loss |
48
  |:-------------:|:------:|:----:|:---------------:|
49
+ | 7.2863 | 0.5214 | 500 | 7.1323 |
50
+ | 6.8168 | 1.0428 | 1000 | 6.9015 |
51
 
52
 
53
  ### Framework versions
config.json CHANGED
@@ -4,9 +4,9 @@
4
  "GPT2LMHeadModel"
5
  ],
6
  "attn_pdrop": 0.1,
7
- "bos_token_id": 50256,
8
  "embd_pdrop": 0.1,
9
- "eos_token_id": 50256,
10
  "initializer_range": 0.02,
11
  "layer_norm_epsilon": 1e-05,
12
  "model_type": "gpt2",
@@ -16,6 +16,7 @@
16
  "n_inner": null,
17
  "n_layer": 4,
18
  "n_positions": 64,
 
19
  "reorder_and_upcast_attn": false,
20
  "resid_pdrop": 0.1,
21
  "scale_attn_by_inverse_layer_idx": false,
@@ -27,6 +28,7 @@
27
  "summary_use_proj": true,
28
  "torch_dtype": "float32",
29
  "transformers_version": "4.41.2",
 
30
  "use_cache": true,
31
  "vocab_size": 52000
32
  }
 
4
  "GPT2LMHeadModel"
5
  ],
6
  "attn_pdrop": 0.1,
7
+ "bos_token_id": 3,
8
  "embd_pdrop": 0.1,
9
+ "eos_token_id": 2,
10
  "initializer_range": 0.02,
11
  "layer_norm_epsilon": 1e-05,
12
  "model_type": "gpt2",
 
16
  "n_inner": null,
17
  "n_layer": 4,
18
  "n_positions": 64,
19
+ "pad_token_id": 1,
20
  "reorder_and_upcast_attn": false,
21
  "resid_pdrop": 0.1,
22
  "scale_attn_by_inverse_layer_idx": false,
 
28
  "summary_use_proj": true,
29
  "torch_dtype": "float32",
30
  "transformers_version": "4.41.2",
31
+ "unk_token_id": 0,
32
  "use_cache": true,
33
  "vocab_size": 52000
34
  }
generation_config.json CHANGED
@@ -1,6 +1,7 @@
1
  {
2
  "_from_model_config": true,
3
- "bos_token_id": 50256,
4
- "eos_token_id": 50256,
 
5
  "transformers_version": "4.41.2"
6
  }
 
1
  {
2
  "_from_model_config": true,
3
+ "bos_token_id": 3,
4
+ "eos_token_id": 2,
5
+ "pad_token_id": 1,
6
  "transformers_version": "4.41.2"
7
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f9680709df71ae71c1ab82ae988403f239eb1b8e990e1606aa496d1a32b1a4d0
3
  size 273357888
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ae279fe143d3a72aad4391e032d8872966f5729862b3d5a4daa690e25541963
3
  size 273357888
special_tokens_map.json CHANGED
@@ -1,3 +1,6 @@
1
  {
2
- "pad_token": "[PAD]"
 
 
 
3
  }
 
1
  {
2
+ "bos_token": "[BOS]",
3
+ "eos_token": "[EOS]",
4
+ "pad_token": "[PAD]",
5
+ "unk_token": "[UNK]"
6
  }
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -23,10 +23,21 @@
23
  "rstrip": false,
24
  "single_word": false,
25
  "special": true
 
 
 
 
 
 
 
 
26
  }
27
  },
 
28
  "clean_up_tokenization_spaces": true,
 
29
  "model_max_length": 1000000000000000019884624838656,
30
  "pad_token": "[PAD]",
31
- "tokenizer_class": "PreTrainedTokenizerFast"
 
32
  }
 
23
  "rstrip": false,
24
  "single_word": false,
25
  "special": true
26
+ },
27
+ "3": {
28
+ "content": "[BOS]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
  }
35
  },
36
+ "bos_token": "[BOS]",
37
  "clean_up_tokenization_spaces": true,
38
+ "eos_token": "[EOS]",
39
  "model_max_length": 1000000000000000019884624838656,
40
  "pad_token": "[PAD]",
41
+ "tokenizer_class": "PreTrainedTokenizerFast",
42
+ "unk_token": "[UNK]"
43
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:123930e80256c46af25f67796c70eec2e28270c84a966d973bbd9708a0eee480
3
  size 5048
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a92b46728c53ce5b964c094407071a099d8f41d18b50e1060a0ea7bdf536a98f
3
  size 5048