Ocean82 qiny17 commited on
Commit
05c3f01
·
0 Parent(s):

Duplicate from ECE1786-AG/lyrics-generator

Browse files

Co-authored-by: Yiqian Qin <qiny17@users.noreply.huggingface.co>

.gitattributes ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tflite filter=lfs diff=lfs merge=lfs -text
29
+ *.tgz filter=lfs diff=lfs merge=lfs -text
30
+ *.wasm filter=lfs diff=lfs merge=lfs -text
31
+ *.xz filter=lfs diff=lfs merge=lfs -text
32
+ *.zip filter=lfs diff=lfs merge=lfs -text
33
+ *.zst filter=lfs diff=lfs merge=lfs -text
34
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
.gitignore ADDED
@@ -0,0 +1 @@
 
 
1
+ checkpoint-*/
README.md ADDED
@@ -0,0 +1,83 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ tags:
4
+ - generated_from_trainer
5
+ model-index:
6
+ - name: lyrics-generator-v2
7
+ results: []
8
+ ---
9
+
10
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
11
+ should probably proofread and complete it, then remove this comment. -->
12
+
13
+ # lyrics-generator-v2
14
+
15
+ This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on the None dataset.
16
+ It achieves the following results on the evaluation set:
17
+ - Loss: 2.1114
18
+
19
+ ## Model description
20
+
21
+ More information needed
22
+
23
+ ## Intended uses & limitations
24
+
25
+ More information needed
26
+
27
+ ## Training and evaluation data
28
+
29
+ More information needed
30
+
31
+ ## Training procedure
32
+
33
+ ### Training hyperparameters
34
+
35
+ The following hyperparameters were used during training:
36
+ - learning_rate: 5e-05
37
+ - train_batch_size: 8
38
+ - eval_batch_size: 8
39
+ - seed: 42
40
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
41
+ - lr_scheduler_type: linear
42
+ - num_epochs: 10
43
+
44
+ ### Training results
45
+
46
+ | Training Loss | Epoch | Step | Validation Loss |
47
+ |:-------------:|:-----:|:----:|:---------------:|
48
+ | 3.4325 | 0.36 | 200 | 2.3267 |
49
+ | 2.4121 | 0.71 | 400 | 2.2693 |
50
+ | 2.3582 | 1.07 | 600 | 2.2401 |
51
+ | 2.2903 | 1.42 | 800 | 2.2182 |
52
+ | 2.2738 | 1.78 | 1000 | 2.2046 |
53
+ | 2.2322 | 2.14 | 1200 | 2.1922 |
54
+ | 2.1933 | 2.49 | 1400 | 2.1832 |
55
+ | 2.1944 | 2.85 | 1600 | 2.1736 |
56
+ | 2.1632 | 3.2 | 1800 | 2.1648 |
57
+ | 2.1366 | 3.56 | 2000 | 2.1554 |
58
+ | 2.1492 | 3.91 | 2200 | 2.1491 |
59
+ | 2.1108 | 4.27 | 2400 | 2.1472 |
60
+ | 2.0882 | 4.63 | 2600 | 2.1422 |
61
+ | 2.0971 | 4.98 | 2800 | 2.1343 |
62
+ | 2.0829 | 5.34 | 3000 | 2.1318 |
63
+ | 2.042 | 5.69 | 3200 | 2.1280 |
64
+ | 2.0375 | 6.05 | 3400 | 2.1261 |
65
+ | 2.0146 | 6.41 | 3600 | 2.1245 |
66
+ | 2.0551 | 6.76 | 3800 | 2.1217 |
67
+ | 1.992 | 7.12 | 4000 | 2.1182 |
68
+ | 1.9994 | 7.47 | 4200 | 2.1170 |
69
+ | 2.0189 | 7.83 | 4400 | 2.1156 |
70
+ | 1.9795 | 8.19 | 4600 | 2.1133 |
71
+ | 2.0101 | 8.54 | 4800 | 2.1143 |
72
+ | 1.9864 | 8.9 | 5000 | 2.1111 |
73
+ | 1.9602 | 9.25 | 5200 | 2.1120 |
74
+ | 1.9899 | 9.61 | 5400 | 2.1117 |
75
+ | 1.9928 | 9.96 | 5600 | 2.1114 |
76
+
77
+
78
+ ### Framework versions
79
+
80
+ - Transformers 4.25.1
81
+ - Pytorch 1.12.1+cu113
82
+ - Datasets 2.7.1
83
+ - Tokenizers 0.13.2
added_tokens.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "<BOS>": 50257,
3
+ "<EOS>": 50258,
4
+ "<PAD>": 50260,
5
+ "<UNK>": 50259,
6
+ "<country>": 50263,
7
+ "<generic>": 50266,
8
+ "<pop>": 50261,
9
+ "<r&b>": 50265,
10
+ "<rap>": 50262,
11
+ "<rock>": 50264
12
+ }
config.json ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "gpt2",
3
+ "activation_function": "gelu_new",
4
+ "architectures": [
5
+ "GPT2LMHeadModel"
6
+ ],
7
+ "attn_pdrop": 0.1,
8
+ "bos_token_id": 50257,
9
+ "embd_pdrop": 0.1,
10
+ "eos_token_id": 50258,
11
+ "initializer_range": 0.02,
12
+ "layer_norm_epsilon": 1e-05,
13
+ "model_type": "gpt2",
14
+ "n_ctx": 1024,
15
+ "n_embd": 768,
16
+ "n_head": 12,
17
+ "n_inner": null,
18
+ "n_layer": 12,
19
+ "n_positions": 1024,
20
+ "pad_token_id": 50260,
21
+ "reorder_and_upcast_attn": false,
22
+ "resid_pdrop": 0.1,
23
+ "scale_attn_by_inverse_layer_idx": false,
24
+ "scale_attn_weights": true,
25
+ "summary_activation": null,
26
+ "summary_first_dropout": 0.1,
27
+ "summary_proj_to_labels": true,
28
+ "summary_type": "cls_index",
29
+ "summary_use_proj": true,
30
+ "task_specific_params": {
31
+ "text-generation": {
32
+ "do_sample": true,
33
+ "max_length": 50
34
+ }
35
+ },
36
+ "torch_dtype": "float32",
37
+ "transformers_version": "4.25.1",
38
+ "unk_token_id": 50259,
39
+ "use_cache": true,
40
+ "vocab_size": 50267
41
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ad41da6864a100200d58dd01927753e62c2ca896cf983573c3bda12081ef5838
3
+ size 510427241
runs/Dec05_02-10-54_c6ed9cf582c8/1670224274.1706448/events.out.tfevents.1670224274.c6ed9cf582c8.116.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0aae94c858bc3bd780168715d6e877636c48c82f479b940345e70a33df93ae89
3
+ size 5600
runs/Dec05_02-10-54_c6ed9cf582c8/events.out.tfevents.1670224274.c6ed9cf582c8.116.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43143c6fa886a3330bf2e55cb0d3b2d37659bc2423c462c81443b5cdf0b24fe4
3
+ size 16442
special_tokens_map.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<pop>",
4
+ "<rap>",
5
+ "<country>",
6
+ "<rock>",
7
+ "<r&b>",
8
+ "<generic>"
9
+ ],
10
+ "bos_token": "<BOS>",
11
+ "eos_token": "<EOS>",
12
+ "pad_token": "<PAD>",
13
+ "unk_token": "<UNK>"
14
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "bos_token": "<|endoftext|>",
4
+ "eos_token": "<|endoftext|>",
5
+ "model_max_length": 1024,
6
+ "name_or_path": "gpt2",
7
+ "special_tokens_map_file": null,
8
+ "tokenizer_class": "GPT2Tokenizer",
9
+ "unk_token": "<|endoftext|>"
10
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ee597ed88a2711f849d8301c55edea6741fe02ef0a6a5dfaf18ec376ef01df2
3
+ size 3439
vocab.json ADDED
The diff for this file is too large to render. See raw diff