Commit ·
63123e4
1
Parent(s): 2c66833
Training in progress, step 1000
Browse files- .gitignore +1 -0
- config.json +42 -0
- last-checkpoint/config.json +42 -0
- last-checkpoint/generation_config.json +6 -0
- last-checkpoint/optimizer.pt +3 -0
- last-checkpoint/pytorch_model.bin +3 -0
- last-checkpoint/rng_state.pth +3 -0
- last-checkpoint/scheduler.pt +3 -0
- last-checkpoint/trainer_state.json +36 -0
- last-checkpoint/training_args.bin +3 -0
- pytorch_model.bin +3 -0
- runs/Mar21_09-06-28_da74753029b6/1679389598.649497/events.out.tfevents.1679389598.da74753029b6.214.1 +3 -0
- runs/Mar21_09-06-28_da74753029b6/events.out.tfevents.1679389598.da74753029b6.214.0 +3 -0
- training_args.bin +3 -0
.gitignore
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
checkpoint-*/
|
config.json
ADDED
|
@@ -0,0 +1,42 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_name_or_path": "/content/Socrat_4",
|
| 3 |
+
"activation_function": "gelu_new",
|
| 4 |
+
"architectures": [
|
| 5 |
+
"GPT2LMHeadModel"
|
| 6 |
+
],
|
| 7 |
+
"attn_pdrop": 0.1,
|
| 8 |
+
"bos_token_id": 50256,
|
| 9 |
+
"embd_pdrop": 0.1,
|
| 10 |
+
"eos_token_id": 50256,
|
| 11 |
+
"id2label": {
|
| 12 |
+
"0": "LABEL_0"
|
| 13 |
+
},
|
| 14 |
+
"initializer_range": 0.02,
|
| 15 |
+
"label2id": {
|
| 16 |
+
"LABEL_0": 0
|
| 17 |
+
},
|
| 18 |
+
"layer_norm_epsilon": 1e-05,
|
| 19 |
+
"model_type": "gpt2",
|
| 20 |
+
"n_ctx": 2048,
|
| 21 |
+
"n_embd": 1024,
|
| 22 |
+
"n_head": 16,
|
| 23 |
+
"n_inner": null,
|
| 24 |
+
"n_layer": 24,
|
| 25 |
+
"n_positions": 2048,
|
| 26 |
+
"n_special": 0,
|
| 27 |
+
"output_past": true,
|
| 28 |
+
"predict_special_tokens": true,
|
| 29 |
+
"reorder_and_upcast_attn": false,
|
| 30 |
+
"resid_pdrop": 0.1,
|
| 31 |
+
"scale_attn_by_inverse_layer_idx": false,
|
| 32 |
+
"scale_attn_weights": true,
|
| 33 |
+
"summary_activation": null,
|
| 34 |
+
"summary_first_dropout": 0.1,
|
| 35 |
+
"summary_proj_to_labels": true,
|
| 36 |
+
"summary_type": "cls_index",
|
| 37 |
+
"summary_use_proj": true,
|
| 38 |
+
"torch_dtype": "float32",
|
| 39 |
+
"transformers_version": "4.27.2",
|
| 40 |
+
"use_cache": true,
|
| 41 |
+
"vocab_size": 50257
|
| 42 |
+
}
|
last-checkpoint/config.json
ADDED
|
@@ -0,0 +1,42 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_name_or_path": "/content/Socrat_4",
|
| 3 |
+
"activation_function": "gelu_new",
|
| 4 |
+
"architectures": [
|
| 5 |
+
"GPT2LMHeadModel"
|
| 6 |
+
],
|
| 7 |
+
"attn_pdrop": 0.1,
|
| 8 |
+
"bos_token_id": 50256,
|
| 9 |
+
"embd_pdrop": 0.1,
|
| 10 |
+
"eos_token_id": 50256,
|
| 11 |
+
"id2label": {
|
| 12 |
+
"0": "LABEL_0"
|
| 13 |
+
},
|
| 14 |
+
"initializer_range": 0.02,
|
| 15 |
+
"label2id": {
|
| 16 |
+
"LABEL_0": 0
|
| 17 |
+
},
|
| 18 |
+
"layer_norm_epsilon": 1e-05,
|
| 19 |
+
"model_type": "gpt2",
|
| 20 |
+
"n_ctx": 2048,
|
| 21 |
+
"n_embd": 1024,
|
| 22 |
+
"n_head": 16,
|
| 23 |
+
"n_inner": null,
|
| 24 |
+
"n_layer": 24,
|
| 25 |
+
"n_positions": 2048,
|
| 26 |
+
"n_special": 0,
|
| 27 |
+
"output_past": true,
|
| 28 |
+
"predict_special_tokens": true,
|
| 29 |
+
"reorder_and_upcast_attn": false,
|
| 30 |
+
"resid_pdrop": 0.1,
|
| 31 |
+
"scale_attn_by_inverse_layer_idx": false,
|
| 32 |
+
"scale_attn_weights": true,
|
| 33 |
+
"summary_activation": null,
|
| 34 |
+
"summary_first_dropout": 0.1,
|
| 35 |
+
"summary_proj_to_labels": true,
|
| 36 |
+
"summary_type": "cls_index",
|
| 37 |
+
"summary_use_proj": true,
|
| 38 |
+
"torch_dtype": "float32",
|
| 39 |
+
"transformers_version": "4.27.2",
|
| 40 |
+
"use_cache": true,
|
| 41 |
+
"vocab_size": 50257
|
| 42 |
+
}
|
last-checkpoint/generation_config.json
ADDED
|
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_from_model_config": true,
|
| 3 |
+
"bos_token_id": 50256,
|
| 4 |
+
"eos_token_id": 50256,
|
| 5 |
+
"transformers_version": "4.27.2"
|
| 6 |
+
}
|
last-checkpoint/optimizer.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d951dcccb3138caeed1ab4351056967ab2bf699172236aa45edc1328eeb5dfcb
|
| 3 |
+
size 2847145157
|
last-checkpoint/pytorch_model.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a75bb415d99c0da358739989a91937fe22b77d82618cb1946977a2b4e860ef3b
|
| 3 |
+
size 1524261149
|
last-checkpoint/rng_state.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:478d2a12e14dda206ad13c93b28e9383719e71464f75e206b9b3b5a4f9e3ffb4
|
| 3 |
+
size 14575
|
last-checkpoint/scheduler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6fd63048cac0dc24c318a6baa96db9ac8e3f1ebf138acdbe8249be7a05bddc22
|
| 3 |
+
size 627
|
last-checkpoint/trainer_state.json
ADDED
|
@@ -0,0 +1,36 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"best_metric": null,
|
| 3 |
+
"best_model_checkpoint": null,
|
| 4 |
+
"epoch": 0.25012506253126565,
|
| 5 |
+
"global_step": 1000,
|
| 6 |
+
"is_hyper_param_search": false,
|
| 7 |
+
"is_local_process_zero": true,
|
| 8 |
+
"is_world_process_zero": true,
|
| 9 |
+
"log_history": [
|
| 10 |
+
{
|
| 11 |
+
"epoch": 0.13,
|
| 12 |
+
"learning_rate": 1.949974987493747e-05,
|
| 13 |
+
"loss": 2.8007,
|
| 14 |
+
"step": 500
|
| 15 |
+
},
|
| 16 |
+
{
|
| 17 |
+
"epoch": 0.25,
|
| 18 |
+
"learning_rate": 1.8999499749874938e-05,
|
| 19 |
+
"loss": 2.7961,
|
| 20 |
+
"step": 1000
|
| 21 |
+
},
|
| 22 |
+
{
|
| 23 |
+
"epoch": 0.25,
|
| 24 |
+
"eval_loss": 3.1374855041503906,
|
| 25 |
+
"eval_runtime": 136.8005,
|
| 26 |
+
"eval_samples_per_second": 15.475,
|
| 27 |
+
"eval_steps_per_second": 5.161,
|
| 28 |
+
"step": 1000
|
| 29 |
+
}
|
| 30 |
+
],
|
| 31 |
+
"max_steps": 19990,
|
| 32 |
+
"num_train_epochs": 5,
|
| 33 |
+
"total_flos": 1414817464320000.0,
|
| 34 |
+
"trial_name": null,
|
| 35 |
+
"trial_params": null
|
| 36 |
+
}
|
last-checkpoint/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a27a26c548ae69b704f6121b2840b0eae566d932411df704e8d8478fc6c30e2e
|
| 3 |
+
size 3579
|
pytorch_model.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a75bb415d99c0da358739989a91937fe22b77d82618cb1946977a2b4e860ef3b
|
| 3 |
+
size 1524261149
|
runs/Mar21_09-06-28_da74753029b6/1679389598.649497/events.out.tfevents.1679389598.da74753029b6.214.1
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b828ef7e07101367a9bc7e2aaeda6d7cda41c543dfed94de02b8c9187534d920
|
| 3 |
+
size 5742
|
runs/Mar21_09-06-28_da74753029b6/events.out.tfevents.1679389598.da74753029b6.214.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6112b973c976203b085883ed9cf44eb7898dc00e0d719b56ed9eac6c7cac2a96
|
| 3 |
+
size 4783
|
training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a27a26c548ae69b704f6121b2840b0eae566d932411df704e8d8478fc6c30e2e
|
| 3 |
+
size 3579
|