Commit
·
3b4951b
1
Parent(s):
ffe42be
huggingartists
Browse files- README.md +3 -3
- config.json +1 -1
- evaluation.txt +1 -1
- flax_model.msgpack +1 -1
- optimizer.pt +1 -1
- pytorch_model.bin +2 -2
- rng_state.pth +1 -1
- scheduler.pt +1 -1
- special_tokens_map.json +5 -1
- tokenizer.json +8 -5
- tokenizer_config.json +10 -1
- trainer_state.json +113 -7
- training_args.bin +2 -2
README.md
CHANGED
|
@@ -45,15 +45,15 @@ from datasets import load_dataset
|
|
| 45 |
dataset = load_dataset("huggingartists/queen")
|
| 46 |
```
|
| 47 |
|
| 48 |
-
[Explore the data](https://wandb.ai/huggingartists/huggingartists/runs/
|
| 49 |
|
| 50 |
## Training procedure
|
| 51 |
|
| 52 |
The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on Queen's lyrics.
|
| 53 |
|
| 54 |
-
Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/huggingartists/huggingartists/runs/
|
| 55 |
|
| 56 |
-
At the end of training, [the final model](https://wandb.ai/huggingartists/huggingartists/runs/
|
| 57 |
|
| 58 |
## How to use
|
| 59 |
|
|
|
|
| 45 |
dataset = load_dataset("huggingartists/queen")
|
| 46 |
```
|
| 47 |
|
| 48 |
+
[Explore the data](https://wandb.ai/huggingartists/huggingartists/runs/33biqcks/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline.
|
| 49 |
|
| 50 |
## Training procedure
|
| 51 |
|
| 52 |
The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on Queen's lyrics.
|
| 53 |
|
| 54 |
+
Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/huggingartists/huggingartists/runs/1k3htq7p) for full transparency and reproducibility.
|
| 55 |
|
| 56 |
+
At the end of training, [the final model](https://wandb.ai/huggingartists/huggingartists/runs/1k3htq7p/artifacts) is logged and versioned.
|
| 57 |
|
| 58 |
## How to use
|
| 59 |
|
config.json
CHANGED
|
@@ -37,7 +37,7 @@
|
|
| 37 |
}
|
| 38 |
},
|
| 39 |
"torch_dtype": "float32",
|
| 40 |
-
"transformers_version": "4.
|
| 41 |
"use_cache": true,
|
| 42 |
"vocab_size": 50257
|
| 43 |
}
|
|
|
|
| 37 |
}
|
| 38 |
},
|
| 39 |
"torch_dtype": "float32",
|
| 40 |
+
"transformers_version": "4.20.0",
|
| 41 |
"use_cache": true,
|
| 42 |
"vocab_size": 50257
|
| 43 |
}
|
evaluation.txt
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
{"eval_loss":
|
|
|
|
| 1 |
+
{"eval_loss": 0.9721857309341431, "eval_runtime": 1.902, "eval_samples_per_second": 44.689, "eval_steps_per_second": 5.783, "epoch": 11.0}
|
flax_model.msgpack
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 497764120
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:800bffdd4c52f619bf17474c0d9f571b39795cdad6155abd76ab3d0e828edbee
|
| 3 |
size 497764120
|
optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 995604017
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d44e3f3b76fe08f524c40d07198c2501597e28b7eaa180a40a79abc32258dba5
|
| 3 |
size 995604017
|
pytorch_model.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:56180aae9f0dfc24b3d1cdbb07dc2eb5acaab5a5b5e60892d48dca7feaf72ef9
|
| 3 |
+
size 510396521
|
rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14503
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f441152599d4ecfcd40f2f229153cf77289fd381959e35ff164226f90d545b2b
|
| 3 |
size 14503
|
scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ae81c5f3c8df017bd46e14a116a6befbfbd6eda859addceb4a11946640558415
|
| 3 |
size 623
|
special_tokens_map.json
CHANGED
|
@@ -1 +1,5 @@
|
|
| 1 |
-
{
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"bos_token": "<|endoftext|>",
|
| 3 |
+
"eos_token": "<|endoftext|>",
|
| 4 |
+
"unk_token": "<|endoftext|>"
|
| 5 |
+
}
|
tokenizer.json
CHANGED
|
@@ -5,29 +5,32 @@
|
|
| 5 |
"added_tokens": [
|
| 6 |
{
|
| 7 |
"id": 50256,
|
| 8 |
-
"special": true,
|
| 9 |
"content": "<|endoftext|>",
|
| 10 |
"single_word": false,
|
| 11 |
"lstrip": false,
|
| 12 |
"rstrip": false,
|
| 13 |
-
"normalized": false
|
|
|
|
| 14 |
}
|
| 15 |
],
|
| 16 |
"normalizer": null,
|
| 17 |
"pre_tokenizer": {
|
| 18 |
"type": "ByteLevel",
|
| 19 |
"add_prefix_space": false,
|
| 20 |
-
"trim_offsets": true
|
|
|
|
| 21 |
},
|
| 22 |
"post_processor": {
|
| 23 |
"type": "ByteLevel",
|
| 24 |
"add_prefix_space": true,
|
| 25 |
-
"trim_offsets": false
|
|
|
|
| 26 |
},
|
| 27 |
"decoder": {
|
| 28 |
"type": "ByteLevel",
|
| 29 |
"add_prefix_space": true,
|
| 30 |
-
"trim_offsets": true
|
|
|
|
| 31 |
},
|
| 32 |
"model": {
|
| 33 |
"type": "BPE",
|
|
|
|
| 5 |
"added_tokens": [
|
| 6 |
{
|
| 7 |
"id": 50256,
|
|
|
|
| 8 |
"content": "<|endoftext|>",
|
| 9 |
"single_word": false,
|
| 10 |
"lstrip": false,
|
| 11 |
"rstrip": false,
|
| 12 |
+
"normalized": false,
|
| 13 |
+
"special": true
|
| 14 |
}
|
| 15 |
],
|
| 16 |
"normalizer": null,
|
| 17 |
"pre_tokenizer": {
|
| 18 |
"type": "ByteLevel",
|
| 19 |
"add_prefix_space": false,
|
| 20 |
+
"trim_offsets": true,
|
| 21 |
+
"use_regex": true
|
| 22 |
},
|
| 23 |
"post_processor": {
|
| 24 |
"type": "ByteLevel",
|
| 25 |
"add_prefix_space": true,
|
| 26 |
+
"trim_offsets": false,
|
| 27 |
+
"use_regex": true
|
| 28 |
},
|
| 29 |
"decoder": {
|
| 30 |
"type": "ByteLevel",
|
| 31 |
"add_prefix_space": true,
|
| 32 |
+
"trim_offsets": true,
|
| 33 |
+
"use_regex": true
|
| 34 |
},
|
| 35 |
"model": {
|
| 36 |
"type": "BPE",
|
tokenizer_config.json
CHANGED
|
@@ -1 +1,10 @@
|
|
| 1 |
-
{
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"add_prefix_space": false,
|
| 3 |
+
"bos_token": "<|endoftext|>",
|
| 4 |
+
"eos_token": "<|endoftext|>",
|
| 5 |
+
"model_max_length": 1024,
|
| 6 |
+
"name_or_path": "huggingartists/queen",
|
| 7 |
+
"special_tokens_map_file": null,
|
| 8 |
+
"tokenizer_class": "GPT2Tokenizer",
|
| 9 |
+
"unk_token": "<|endoftext|>"
|
| 10 |
+
}
|
trainer_state.json
CHANGED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
-
"best_metric":
|
| 3 |
-
"best_model_checkpoint": "output/queen/checkpoint-
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -910,11 +910,117 @@
|
|
| 910 |
"eval_samples_per_second": 75.649,
|
| 911 |
"eval_steps_per_second": 9.867,
|
| 912 |
"step": 680
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 913 |
}
|
| 914 |
],
|
| 915 |
-
"max_steps":
|
| 916 |
-
"num_train_epochs":
|
| 917 |
-
"total_flos":
|
| 918 |
"trial_name": null,
|
| 919 |
"trial_params": null
|
| 920 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"best_metric": 0.9721857309341431,
|
| 3 |
+
"best_model_checkpoint": "output/queen/checkpoint-759",
|
| 4 |
+
"epoch": 11.0,
|
| 5 |
+
"global_step": 759,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 910 |
"eval_samples_per_second": 75.649,
|
| 911 |
"eval_steps_per_second": 9.867,
|
| 912 |
"step": 680
|
| 913 |
+
},
|
| 914 |
+
{
|
| 915 |
+
"epoch": 9.93,
|
| 916 |
+
"learning_rate": 0.00013543005707140072,
|
| 917 |
+
"loss": 1.2301,
|
| 918 |
+
"step": 685
|
| 919 |
+
},
|
| 920 |
+
{
|
| 921 |
+
"epoch": 10.0,
|
| 922 |
+
"learning_rate": 0.0001372,
|
| 923 |
+
"loss": 1.2039,
|
| 924 |
+
"step": 690
|
| 925 |
+
},
|
| 926 |
+
{
|
| 927 |
+
"epoch": 10.0,
|
| 928 |
+
"eval_loss": 0.9929937720298767,
|
| 929 |
+
"eval_runtime": 1.7672,
|
| 930 |
+
"eval_samples_per_second": 48.098,
|
| 931 |
+
"eval_steps_per_second": 6.224,
|
| 932 |
+
"step": 690
|
| 933 |
+
},
|
| 934 |
+
{
|
| 935 |
+
"epoch": 10.07,
|
| 936 |
+
"learning_rate": 0.00013543005707140074,
|
| 937 |
+
"loss": 1.1894,
|
| 938 |
+
"step": 695
|
| 939 |
+
},
|
| 940 |
+
{
|
| 941 |
+
"epoch": 10.14,
|
| 942 |
+
"learning_rate": 0.0001302115605879495,
|
| 943 |
+
"loss": 1.2106,
|
| 944 |
+
"step": 700
|
| 945 |
+
},
|
| 946 |
+
{
|
| 947 |
+
"epoch": 10.22,
|
| 948 |
+
"learning_rate": 0.00012181379454232323,
|
| 949 |
+
"loss": 1.144,
|
| 950 |
+
"step": 705
|
| 951 |
+
},
|
| 952 |
+
{
|
| 953 |
+
"epoch": 10.29,
|
| 954 |
+
"learning_rate": 0.00011067009906946301,
|
| 955 |
+
"loss": 1.3006,
|
| 956 |
+
"step": 710
|
| 957 |
+
},
|
| 958 |
+
{
|
| 959 |
+
"epoch": 10.36,
|
| 960 |
+
"learning_rate": 9.735550930058323e-05,
|
| 961 |
+
"loss": 1.2147,
|
| 962 |
+
"step": 715
|
| 963 |
+
},
|
| 964 |
+
{
|
| 965 |
+
"epoch": 10.43,
|
| 966 |
+
"learning_rate": 8.255708249541075e-05,
|
| 967 |
+
"loss": 1.1931,
|
| 968 |
+
"step": 720
|
| 969 |
+
},
|
| 970 |
+
{
|
| 971 |
+
"epoch": 10.51,
|
| 972 |
+
"learning_rate": 6.703844462707122e-05,
|
| 973 |
+
"loss": 1.2735,
|
| 974 |
+
"step": 725
|
| 975 |
+
},
|
| 976 |
+
{
|
| 977 |
+
"epoch": 10.58,
|
| 978 |
+
"learning_rate": 5.160038588088712e-05,
|
| 979 |
+
"loss": 1.1695,
|
| 980 |
+
"step": 730
|
| 981 |
+
},
|
| 982 |
+
{
|
| 983 |
+
"epoch": 10.65,
|
| 984 |
+
"learning_rate": 3.703953841164302e-05,
|
| 985 |
+
"loss": 1.2219,
|
| 986 |
+
"step": 735
|
| 987 |
+
},
|
| 988 |
+
{
|
| 989 |
+
"epoch": 10.72,
|
| 990 |
+
"learning_rate": 2.4107268662854985e-05,
|
| 991 |
+
"loss": 1.1271,
|
| 992 |
+
"step": 740
|
| 993 |
+
},
|
| 994 |
+
{
|
| 995 |
+
"epoch": 10.8,
|
| 996 |
+
"learning_rate": 1.3470905479788523e-05,
|
| 997 |
+
"loss": 1.2963,
|
| 998 |
+
"step": 745
|
| 999 |
+
},
|
| 1000 |
+
{
|
| 1001 |
+
"epoch": 10.87,
|
| 1002 |
+
"learning_rate": 5.679304716725944e-06,
|
| 1003 |
+
"loss": 1.0029,
|
| 1004 |
+
"step": 750
|
| 1005 |
+
},
|
| 1006 |
+
{
|
| 1007 |
+
"epoch": 10.94,
|
| 1008 |
+
"learning_rate": 1.1345272679252971e-06,
|
| 1009 |
+
"loss": 1.2789,
|
| 1010 |
+
"step": 755
|
| 1011 |
+
},
|
| 1012 |
+
{
|
| 1013 |
+
"epoch": 11.0,
|
| 1014 |
+
"eval_loss": 0.9721857309341431,
|
| 1015 |
+
"eval_runtime": 1.8495,
|
| 1016 |
+
"eval_samples_per_second": 45.957,
|
| 1017 |
+
"eval_steps_per_second": 5.947,
|
| 1018 |
+
"step": 759
|
| 1019 |
}
|
| 1020 |
],
|
| 1021 |
+
"max_steps": 759,
|
| 1022 |
+
"num_train_epochs": 11,
|
| 1023 |
+
"total_flos": 785966432256000.0,
|
| 1024 |
"trial_name": null,
|
| 1025 |
"trial_params": null
|
| 1026 |
}
|
training_args.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:04747b0b5177cd34b6949ed0755471acbad69a771c0daf9bdbba80e2464114ac
|
| 3 |
+
size 3311
|