ninagroot commited on
Commit
ad43c57
·
verified ·
1 Parent(s): 8c75cbb

ninagroot/GPT2-705Mtest

Browse files
README.md ADDED
@@ -0,0 +1,81 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - generated_from_trainer
4
+ model-index:
5
+ - name: GPT2-705M
6
+ results: []
7
+ ---
8
+
9
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
10
+ should probably proofread and complete it, then remove this comment. -->
11
+
12
+ # GPT2-705M
13
+
14
+ This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset.
15
+ It achieves the following results on the evaluation set:
16
+ - Loss: 5.4628
17
+
18
+ ## Model description
19
+
20
+ More information needed
21
+
22
+ ## Intended uses & limitations
23
+
24
+ More information needed
25
+
26
+ ## Training and evaluation data
27
+
28
+ More information needed
29
+
30
+ ## Training procedure
31
+
32
+ ### Training hyperparameters
33
+
34
+ The following hyperparameters were used during training:
35
+ - learning_rate: 0.00025
36
+ - train_batch_size: 16
37
+ - eval_batch_size: 8
38
+ - seed: 42
39
+ - gradient_accumulation_steps: 8
40
+ - total_train_batch_size: 128
41
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
+ - lr_scheduler_type: cosine
43
+ - lr_scheduler_warmup_steps: 50
44
+ - num_epochs: 40
45
+ - mixed_precision_training: Native AMP
46
+
47
+ ### Training results
48
+
49
+ | Training Loss | Epoch | Step | Validation Loss |
50
+ |:-------------:|:-----:|:----:|:---------------:|
51
+ | 9.7135 | 0.57 | 1 | 9.7272 |
52
+ | 8.0222 | 1.71 | 3 | 9.3213 |
53
+ | 7.6063 | 2.86 | 5 | 8.5841 |
54
+ | 7.5596 | 4.0 | 7 | 7.9271 |
55
+ | 7.4194 | 4.57 | 8 | 8.0942 |
56
+ | 7.1644 | 5.71 | 10 | 7.5409 |
57
+ | 6.8531 | 6.86 | 12 | 7.3028 |
58
+ | 6.3614 | 8.0 | 14 | 9.3796 |
59
+ | 8.5129 | 8.57 | 15 | 7.6361 |
60
+ | 6.1325 | 9.71 | 17 | 6.7577 |
61
+ | 5.8526 | 10.86 | 19 | 6.5249 |
62
+ | 5.5941 | 12.0 | 21 | 6.2490 |
63
+ | 5.4307 | 12.57 | 22 | 6.2442 |
64
+ | 5.1381 | 13.71 | 24 | 5.9595 |
65
+ | 4.8705 | 14.86 | 26 | 5.8944 |
66
+ | 4.7083 | 16.0 | 28 | 5.7005 |
67
+ | 4.5355 | 16.57 | 29 | 5.7459 |
68
+ | 4.4187 | 17.71 | 31 | 5.5387 |
69
+ | 4.3123 | 18.86 | 33 | 5.4863 |
70
+ | 4.0269 | 20.0 | 35 | 5.3277 |
71
+ | 3.942 | 20.57 | 36 | 5.3274 |
72
+ | 3.784 | 21.71 | 38 | 5.3998 |
73
+ | 3.4991 | 22.86 | 40 | 5.4628 |
74
+
75
+
76
+ ### Framework versions
77
+
78
+ - Transformers 4.39.1
79
+ - Pytorch 2.1.2+cu121
80
+ - Datasets 2.16.1
81
+ - Tokenizers 0.15.0
added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "<|endoftext|>": 12198
3
+ }
config.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation_function": "gelu_new",
3
+ "architectures": [
4
+ "GPT2LMHeadModel"
5
+ ],
6
+ "attn_pdrop": 0.0,
7
+ "bos_token_id": 50256,
8
+ "embd_pdrop": 0.0,
9
+ "eos_token_id": 50256,
10
+ "initializer_range": 0.02,
11
+ "layer_norm_epsilon": 1e-05,
12
+ "model_type": "gpt2",
13
+ "n_embd": 1536,
14
+ "n_head": 8,
15
+ "n_inner": null,
16
+ "n_layer": 24,
17
+ "n_positions": 256,
18
+ "pad_token_id": 0,
19
+ "reorder_and_upcast_attn": false,
20
+ "resid_pdrop": 0.0,
21
+ "scale_attn_by_inverse_layer_idx": false,
22
+ "scale_attn_weights": true,
23
+ "summary_activation": null,
24
+ "summary_first_dropout": 0.1,
25
+ "summary_proj_to_labels": true,
26
+ "summary_type": "cls_index",
27
+ "summary_use_proj": true,
28
+ "torch_dtype": "float32",
29
+ "transformers_version": "4.39.1",
30
+ "use_cache": true,
31
+ "vocab_size": 12198
32
+ }
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 50256,
4
+ "eos_token_id": 50256,
5
+ "pad_token_id": 0,
6
+ "transformers_version": "4.39.1"
7
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88489323ca20dc864aa1bbbb0229c3ee79409a0e79fb1746145003251f77a5ac
3
+ size 2796386080
runs/Apr17_12-24-48_gcn61.local.snellius.surf.nl/events.out.tfevents.1713349497.gcn61.local.snellius.surf.nl.3984517.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6f3d980b23d33ec2f66c7aa4389002abb105341cef143cc8b2622ea84f13474
3
+ size 41155
runs/Apr17_14-09-48_gcn20.local.snellius.surf.nl/events.out.tfevents.1713355797.gcn20.local.snellius.surf.nl.1872804.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a6700eac66ebd9f905898d40f4578bc648ede5bf54c07cccb52f61c44369a1d
3
+ size 19424
runs/Apr17_14-30-23_gcn10.local.snellius.surf.nl/events.out.tfevents.1713357035.gcn10.local.snellius.surf.nl.3796627.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df9ca5185347da1ec47e04458b55269d305bede1c4357f58f9c8bdea17c11933
3
+ size 19424
runs/Apr17_14-39-08_gcn53.local.snellius.surf.nl/events.out.tfevents.1713357559.gcn53.local.snellius.surf.nl.4133403.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e820e6647d03c6ab72ff3c33a7673f486461751c2af756c36d13f8666cdaa193
3
+ size 23090
runs/Apr17_14-54-37_gcn38.local.snellius.surf.nl/events.out.tfevents.1713358487.gcn38.local.snellius.surf.nl.1402374.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37322800818c3695e4ce8ec27b3e2f30c509e94edfcd592f0832da014a5999b5
3
+ size 19424
special_tokens_map.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<s>",
3
+ "eos_token": "</s>",
4
+ "pad_token": "<pad>",
5
+ "unk_token": "<|endoftext|>"
6
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "<pad>",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "<s>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "2": {
21
+ "content": "</s>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "12198": {
29
+ "content": "<|endoftext|>",
30
+ "lstrip": false,
31
+ "normalized": false,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": true
35
+ }
36
+ },
37
+ "bos_token": "<s>",
38
+ "clean_up_tokenization_spaces": true,
39
+ "eos_token": "</s>",
40
+ "model_max_length": 128,
41
+ "pad_token": "<pad>",
42
+ "tokenizer_class": "GPT2Tokenizer",
43
+ "unk_token": "<|endoftext|>"
44
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8771d2cfcef1d1ffe22114e70ebead0ecc08c4ee4084fd54fa207a9793c64510
3
+ size 4984
vocab.json ADDED
The diff for this file is too large to render. See raw diff