augustocsc commited on
Commit
288be47
·
verified ·
1 Parent(s): e92b261

Training in progress, epoch 1

Browse files
README.md CHANGED
@@ -5,18 +5,18 @@ base_model: gpt2
5
  tags:
6
  - generated_from_trainer
7
  model-index:
8
- - name: Se124M100KInfPrompt_endtoken
9
  results: []
10
  ---
11
 
12
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
  should probably proofread and complete it, then remove this comment. -->
14
 
15
- # Se124M100KInfPrompt_endtoken
16
 
17
  This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
- - Loss: 0.6957
20
 
21
  ## Model description
22
 
@@ -35,12 +35,13 @@ More information needed
35
  ### Training hyperparameters
36
 
37
  The following hyperparameters were used during training:
38
- - learning_rate: 5e-05
39
- - train_batch_size: 32
40
- - eval_batch_size: 32
41
  - seed: 42
42
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
43
- - lr_scheduler_type: linear
 
44
  - num_epochs: 50
45
  - mixed_precision_training: Native AMP
46
 
@@ -48,56 +49,30 @@ The following hyperparameters were used during training:
48
 
49
  | Training Loss | Epoch | Step | Validation Loss |
50
  |:-------------:|:-----:|:-----:|:---------------:|
51
- | 0.2129 | 1.0 | 1430 | 0.8043 |
52
- | 0.2018 | 2.0 | 2860 | 0.7705 |
53
- | 0.1949 | 3.0 | 4290 | 0.7588 |
54
- | 0.1913 | 4.0 | 5720 | 0.7498 |
55
- | 0.1921 | 5.0 | 7150 | 0.7435 |
56
- | 0.1903 | 6.0 | 8580 | 0.7371 |
57
- | 0.1888 | 7.0 | 10010 | 0.7339 |
58
- | 0.1881 | 8.0 | 11440 | 0.7299 |
59
- | 0.1872 | 9.0 | 12870 | 0.7267 |
60
- | 0.187 | 10.0 | 14300 | 0.7251 |
61
- | 0.184 | 11.0 | 15730 | 0.7229 |
62
- | 0.1846 | 12.0 | 17160 | 0.7212 |
63
- | 0.1851 | 13.0 | 18590 | 0.7182 |
64
- | 0.1804 | 14.0 | 20020 | 0.7153 |
65
- | 0.1848 | 15.0 | 21450 | 0.7141 |
66
- | 0.1824 | 16.0 | 22880 | 0.7144 |
67
- | 0.1796 | 17.0 | 24310 | 0.7116 |
68
- | 0.18 | 18.0 | 25740 | 0.7108 |
69
- | 0.1825 | 19.0 | 27170 | 0.7082 |
70
- | 0.1852 | 20.0 | 28600 | 0.7082 |
71
- | 0.1785 | 21.0 | 30030 | 0.7072 |
72
- | 0.1811 | 22.0 | 31460 | 0.7057 |
73
- | 0.178 | 23.0 | 32890 | 0.7059 |
74
- | 0.1827 | 24.0 | 34320 | 0.7046 |
75
- | 0.1813 | 25.0 | 35750 | 0.7033 |
76
- | 0.1825 | 26.0 | 37180 | 0.7039 |
77
- | 0.1795 | 27.0 | 38610 | 0.7032 |
78
- | 0.1801 | 28.0 | 40040 | 0.7017 |
79
- | 0.1781 | 29.0 | 41470 | 0.7013 |
80
- | 0.1823 | 30.0 | 42900 | 0.7010 |
81
- | 0.1781 | 31.0 | 44330 | 0.7012 |
82
- | 0.1809 | 32.0 | 45760 | 0.6999 |
83
- | 0.1764 | 33.0 | 47190 | 0.6996 |
84
- | 0.1791 | 34.0 | 48620 | 0.6983 |
85
- | 0.1793 | 35.0 | 50050 | 0.6988 |
86
- | 0.1785 | 36.0 | 51480 | 0.6980 |
87
- | 0.1777 | 37.0 | 52910 | 0.6980 |
88
- | 0.1774 | 38.0 | 54340 | 0.6980 |
89
- | 0.1795 | 39.0 | 55770 | 0.6976 |
90
- | 0.1772 | 40.0 | 57200 | 0.6974 |
91
- | 0.1793 | 41.0 | 58630 | 0.6974 |
92
- | 0.1777 | 42.0 | 60060 | 0.6968 |
93
- | 0.1777 | 43.0 | 61490 | 0.6965 |
94
- | 0.1779 | 44.0 | 62920 | 0.6965 |
95
- | 0.1782 | 45.0 | 64350 | 0.6964 |
96
- | 0.1765 | 46.0 | 65780 | 0.6961 |
97
- | 0.1758 | 47.0 | 67210 | 0.6962 |
98
- | 0.1763 | 48.0 | 68640 | 0.6960 |
99
- | 0.1788 | 49.0 | 70070 | 0.6958 |
100
- | 0.1776 | 50.0 | 71500 | 0.6957 |
101
 
102
 
103
  ### Framework versions
 
5
  tags:
6
  - generated_from_trainer
7
  model-index:
8
+ - name: Se124M10KInfPrompt_endtoken
9
  results: []
10
  ---
11
 
12
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
  should probably proofread and complete it, then remove this comment. -->
14
 
15
+ # Se124M10KInfPrompt_endtoken
16
 
17
  This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
+ - Loss: 0.6872
20
 
21
  ## Model description
22
 
 
35
  ### Training hyperparameters
36
 
37
  The following hyperparameters were used during training:
38
+ - learning_rate: 0.0005
39
+ - train_batch_size: 8
40
+ - eval_batch_size: 8
41
  - seed: 42
42
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
43
+ - lr_scheduler_type: cosine
44
+ - lr_scheduler_warmup_steps: 200
45
  - num_epochs: 50
46
  - mixed_precision_training: Native AMP
47
 
 
49
 
50
  | Training Loss | Epoch | Step | Validation Loss |
51
  |:-------------:|:-----:|:-----:|:---------------:|
52
+ | 0.8085 | 1.0 | 610 | 0.7760 |
53
+ | 0.7801 | 2.0 | 1220 | 0.7436 |
54
+ | 0.7608 | 3.0 | 1830 | 0.7269 |
55
+ | 0.7438 | 4.0 | 2440 | 0.7199 |
56
+ | 0.7413 | 5.0 | 3050 | 0.7118 |
57
+ | 0.7343 | 6.0 | 3660 | 0.7121 |
58
+ | 0.7332 | 7.0 | 4270 | 0.7089 |
59
+ | 0.7319 | 8.0 | 4880 | 0.7025 |
60
+ | 0.7289 | 9.0 | 5490 | 0.7001 |
61
+ | 0.7236 | 10.0 | 6100 | 0.6965 |
62
+ | 0.7147 | 11.0 | 6710 | 0.6970 |
63
+ | 0.7126 | 12.0 | 7320 | 0.6973 |
64
+ | 0.7167 | 13.0 | 7930 | 0.6935 |
65
+ | 0.711 | 14.0 | 8540 | 0.6927 |
66
+ | 0.7057 | 15.0 | 9150 | 0.6940 |
67
+ | 0.7109 | 16.0 | 9760 | 0.6924 |
68
+ | 0.7117 | 17.0 | 10370 | 0.6928 |
69
+ | 0.7086 | 18.0 | 10980 | 0.6882 |
70
+ | 0.7004 | 19.0 | 11590 | 0.6872 |
71
+ | 0.7016 | 20.0 | 12200 | 0.6895 |
72
+ | 0.7027 | 21.0 | 12810 | 0.6884 |
73
+ | 0.6928 | 22.0 | 13420 | 0.6885 |
74
+ | 0.7059 | 23.0 | 14030 | 0.6894 |
75
+ | 0.6916 | 24.0 | 14640 | 0.6875 |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
76
 
77
 
78
  ### Framework versions
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7c8c43647d357635b454fd82af24fa47aad56af1560c4890ba23b7f6187fa3f6
3
- size 309974336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:118e5a3133e42214d1f7eeb3e731adb33f4fe5afa73d16c1bb82263e1c0a682a
3
+ size 309980480
added_tokens.json CHANGED
@@ -1,4 +1,5 @@
1
  {
2
- "<endofex>": 50258,
3
- "<startofex>": 50257
 
4
  }
 
1
  {
2
+ "<endofex>": 50257,
3
+ "<pad>": 50258,
4
+ "<startofex>": 50259
5
  }
all_results.json CHANGED
@@ -1,13 +1,13 @@
1
  {
2
- "epoch": 50.0,
3
- "eval_loss": 0.6956692934036255,
4
- "eval_runtime": 59.8322,
5
- "eval_samples_per_second": 163.257,
6
- "eval_steps_per_second": 5.114,
7
- "perplexity": 2.005050592091695,
8
- "total_flos": 1.49878932701184e+17,
9
- "train_loss": 0.18437957987751993,
10
- "train_runtime": 7319.2549,
11
- "train_samples_per_second": 312.395,
12
- "train_steps_per_second": 9.769
13
  }
 
1
  {
2
+ "epoch": 24.0,
3
+ "eval_loss": 0.6872262954711914,
4
+ "eval_runtime": 1.8886,
5
+ "eval_samples_per_second": 559.678,
6
+ "eval_steps_per_second": 70.423,
7
+ "perplexity": 1.9881932176157675,
8
+ "total_flos": 7672437924691968.0,
9
+ "train_loss": 0.7372308338926138,
10
+ "train_runtime": 572.2889,
11
+ "train_samples_per_second": 426.096,
12
+ "train_steps_per_second": 53.295
13
  }
eval_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "epoch": 50.0,
3
- "eval_loss": 0.6956692934036255,
4
- "eval_runtime": 59.8322,
5
- "eval_samples_per_second": 163.257,
6
- "eval_steps_per_second": 5.114,
7
- "perplexity": 2.005050592091695
8
  }
 
1
  {
2
+ "epoch": 24.0,
3
+ "eval_loss": 0.6872262954711914,
4
+ "eval_runtime": 1.8886,
5
+ "eval_samples_per_second": 559.678,
6
+ "eval_steps_per_second": 70.423,
7
+ "perplexity": 1.9881932176157675
8
  }
special_tokens_map.json CHANGED
@@ -6,23 +6,10 @@
6
  "normalized": false,
7
  "rstrip": false,
8
  "single_word": false
9
- },
10
- {
11
- "content": "<endofex>",
12
- "lstrip": false,
13
- "normalized": false,
14
- "rstrip": false,
15
- "single_word": false
16
  }
17
  ],
18
  "bos_token": "<|endoftext|>",
19
  "eos_token": "<endofex>",
20
- "pad_token": {
21
- "content": "<|endoftext|>",
22
- "lstrip": false,
23
- "normalized": false,
24
- "rstrip": false,
25
- "single_word": false
26
- },
27
  "unk_token": "<|endoftext|>"
28
  }
 
6
  "normalized": false,
7
  "rstrip": false,
8
  "single_word": false
 
 
 
 
 
 
 
9
  }
10
  ],
11
  "bos_token": "<|endoftext|>",
12
  "eos_token": "<endofex>",
13
+ "pad_token": "<pad>",
 
 
 
 
 
 
14
  "unk_token": "<|endoftext|>"
15
  }
tokenizer.json CHANGED
@@ -14,12 +14,12 @@
14
  "single_word": false,
15
  "lstrip": false,
16
  "rstrip": false,
17
- "normalized": false,
18
  "special": true
19
  },
20
  {
21
  "id": 50257,
22
- "content": "<startofex>",
23
  "single_word": false,
24
  "lstrip": false,
25
  "rstrip": false,
@@ -28,7 +28,16 @@
28
  },
29
  {
30
  "id": 50258,
31
- "content": "<endofex>",
 
 
 
 
 
 
 
 
 
32
  "single_word": false,
33
  "lstrip": false,
34
  "rstrip": false,
 
14
  "single_word": false,
15
  "lstrip": false,
16
  "rstrip": false,
17
+ "normalized": true,
18
  "special": true
19
  },
20
  {
21
  "id": 50257,
22
+ "content": "<endofex>",
23
  "single_word": false,
24
  "lstrip": false,
25
  "rstrip": false,
 
28
  },
29
  {
30
  "id": 50258,
31
+ "content": "<pad>",
32
+ "single_word": false,
33
+ "lstrip": false,
34
+ "rstrip": false,
35
+ "normalized": false,
36
+ "special": true
37
+ },
38
+ {
39
+ "id": 50259,
40
+ "content": "<startofex>",
41
  "single_word": false,
42
  "lstrip": false,
43
  "rstrip": false,
tokenizer_config.json CHANGED
@@ -4,13 +4,13 @@
4
  "50256": {
5
  "content": "<|endoftext|>",
6
  "lstrip": false,
7
- "normalized": false,
8
  "rstrip": false,
9
  "single_word": false,
10
  "special": true
11
  },
12
  "50257": {
13
- "content": "<startofex>",
14
  "lstrip": false,
15
  "normalized": false,
16
  "rstrip": false,
@@ -18,7 +18,15 @@
18
  "special": true
19
  },
20
  "50258": {
21
- "content": "<endofex>",
 
 
 
 
 
 
 
 
22
  "lstrip": false,
23
  "normalized": false,
24
  "rstrip": false,
@@ -27,15 +35,14 @@
27
  }
28
  },
29
  "additional_special_tokens": [
30
- "<startofex>",
31
- "<endofex>"
32
  ],
33
  "bos_token": "<|endoftext|>",
34
  "clean_up_tokenization_spaces": false,
35
  "eos_token": "<endofex>",
36
  "extra_special_tokens": {},
37
  "model_max_length": 1024,
38
- "pad_token": "<|endoftext|>",
39
  "tokenizer_class": "GPT2Tokenizer",
40
  "unk_token": "<|endoftext|>"
41
  }
 
4
  "50256": {
5
  "content": "<|endoftext|>",
6
  "lstrip": false,
7
+ "normalized": true,
8
  "rstrip": false,
9
  "single_word": false,
10
  "special": true
11
  },
12
  "50257": {
13
+ "content": "<endofex>",
14
  "lstrip": false,
15
  "normalized": false,
16
  "rstrip": false,
 
18
  "special": true
19
  },
20
  "50258": {
21
+ "content": "<pad>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "50259": {
29
+ "content": "<startofex>",
30
  "lstrip": false,
31
  "normalized": false,
32
  "rstrip": false,
 
35
  }
36
  },
37
  "additional_special_tokens": [
38
+ "<startofex>"
 
39
  ],
40
  "bos_token": "<|endoftext|>",
41
  "clean_up_tokenization_spaces": false,
42
  "eos_token": "<endofex>",
43
  "extra_special_tokens": {},
44
  "model_max_length": 1024,
45
+ "pad_token": "<pad>",
46
  "tokenizer_class": "GPT2Tokenizer",
47
  "unk_token": "<|endoftext|>"
48
  }
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "epoch": 50.0,
3
- "total_flos": 1.49878932701184e+17,
4
- "train_loss": 0.18437957987751993,
5
- "train_runtime": 7319.2549,
6
- "train_samples_per_second": 312.395,
7
- "train_steps_per_second": 9.769
8
  }
 
1
  {
2
+ "epoch": 24.0,
3
+ "total_flos": 7672437924691968.0,
4
+ "train_loss": 0.7372308338926138,
5
+ "train_runtime": 572.2889,
6
+ "train_samples_per_second": 426.096,
7
+ "train_steps_per_second": 53.295
8
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:59bd0190032b8576fa67d50e424c4c48cd36048fc82231e1db1117e678eb5423
3
  size 5432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:03155aa7943de7470f4e3e03615d815edc1696005426d2517ae93b9b94930139
3
  size 5432