Training in progress, step 80, checkpoint
Browse files
last-checkpoint/adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 140815952
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4404b24076cc63d331e67d7ba3a7fe9520f7a0e8f6d4b52270bfbfafc9664489
|
| 3 |
size 140815952
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 281829907
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d619ba6794210f6129f2e991bb4cca6fb744cc687fb23c48cd2119fe42dacaae
|
| 3 |
size 281829907
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14645
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:417227e46f0a581722968d781d8ea2ba8688c7892e0d69749745f185230daf6a
|
| 3 |
size 14645
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1465
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d2f9c1a66b0f67830f363d7aa55a784bfca1310ea07876b83a0f2f54a5722f96
|
| 3 |
size 1465
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -2,9 +2,9 @@
|
|
| 2 |
"best_global_step": null,
|
| 3 |
"best_metric": null,
|
| 4 |
"best_model_checkpoint": null,
|
| 5 |
-
"epoch":
|
| 6 |
"eval_steps": 100,
|
| 7 |
-
"global_step":
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
@@ -64,6 +64,62 @@
|
|
| 64 |
"learning_rate": 0.0004922997497999166,
|
| 65 |
"loss": 0.5324,
|
| 66 |
"step": 40
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 67 |
}
|
| 68 |
],
|
| 69 |
"logging_steps": 5,
|
|
@@ -83,7 +139,7 @@
|
|
| 83 |
"attributes": {}
|
| 84 |
}
|
| 85 |
},
|
| 86 |
-
"total_flos":
|
| 87 |
"train_batch_size": 1,
|
| 88 |
"trial_name": null,
|
| 89 |
"trial_params": null
|
|
|
|
| 2 |
"best_global_step": null,
|
| 3 |
"best_metric": null,
|
| 4 |
"best_model_checkpoint": null,
|
| 5 |
+
"epoch": 1.0666666666666667,
|
| 6 |
"eval_steps": 100,
|
| 7 |
+
"global_step": 80,
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
|
|
| 64 |
"learning_rate": 0.0004922997497999166,
|
| 65 |
"loss": 0.5324,
|
| 66 |
"step": 40
|
| 67 |
+
},
|
| 68 |
+
{
|
| 69 |
+
"epoch": 0.6,
|
| 70 |
+
"grad_norm": 0.1061202809214592,
|
| 71 |
+
"learning_rate": 0.00048678455732775007,
|
| 72 |
+
"loss": 0.554,
|
| 73 |
+
"step": 45
|
| 74 |
+
},
|
| 75 |
+
{
|
| 76 |
+
"epoch": 0.6666666666666666,
|
| 77 |
+
"grad_norm": 0.10958874970674515,
|
| 78 |
+
"learning_rate": 0.00047983825974555906,
|
| 79 |
+
"loss": 0.6044,
|
| 80 |
+
"step": 50
|
| 81 |
+
},
|
| 82 |
+
{
|
| 83 |
+
"epoch": 0.7333333333333333,
|
| 84 |
+
"grad_norm": 0.09531711041927338,
|
| 85 |
+
"learning_rate": 0.0004715028398670787,
|
| 86 |
+
"loss": 0.5653,
|
| 87 |
+
"step": 55
|
| 88 |
+
},
|
| 89 |
+
{
|
| 90 |
+
"epoch": 0.8,
|
| 91 |
+
"grad_norm": 0.11301957815885544,
|
| 92 |
+
"learning_rate": 0.00046182867623944434,
|
| 93 |
+
"loss": 0.5539,
|
| 94 |
+
"step": 60
|
| 95 |
+
},
|
| 96 |
+
{
|
| 97 |
+
"epoch": 0.8666666666666667,
|
| 98 |
+
"grad_norm": 0.1083487942814827,
|
| 99 |
+
"learning_rate": 0.00045087423865966894,
|
| 100 |
+
"loss": 0.5267,
|
| 101 |
+
"step": 65
|
| 102 |
+
},
|
| 103 |
+
{
|
| 104 |
+
"epoch": 0.9333333333333333,
|
| 105 |
+
"grad_norm": 0.10189539939165115,
|
| 106 |
+
"learning_rate": 0.0004387057347883143,
|
| 107 |
+
"loss": 0.5533,
|
| 108 |
+
"step": 70
|
| 109 |
+
},
|
| 110 |
+
{
|
| 111 |
+
"epoch": 1.0,
|
| 112 |
+
"grad_norm": 0.09216652810573578,
|
| 113 |
+
"learning_rate": 0.0004253967099961942,
|
| 114 |
+
"loss": 0.5016,
|
| 115 |
+
"step": 75
|
| 116 |
+
},
|
| 117 |
+
{
|
| 118 |
+
"epoch": 1.0666666666666667,
|
| 119 |
+
"grad_norm": 0.10134831815958023,
|
| 120 |
+
"learning_rate": 0.0004110276028625994,
|
| 121 |
+
"loss": 0.4856,
|
| 122 |
+
"step": 80
|
| 123 |
}
|
| 124 |
],
|
| 125 |
"logging_steps": 5,
|
|
|
|
| 139 |
"attributes": {}
|
| 140 |
}
|
| 141 |
},
|
| 142 |
+
"total_flos": 2644590082129920.0,
|
| 143 |
"train_batch_size": 1,
|
| 144 |
"trial_name": null,
|
| 145 |
"trial_params": null
|