Training in progress, step 120, checkpoint
Browse files
last-checkpoint/adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 100697728
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3deef287322f16442c1af0f3d4759b9b94c49d764bad6a816bfce60d94c71c48
|
| 3 |
size 100697728
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 201541754
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9a85f3add41e11299940edf2dafaca2ba6ad73c89a200523cd8908a1bc1c8968
|
| 3 |
size 201541754
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14244
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1a1443ffb31b9a8cbe553da1fd18b9036abf1404b7fdd8944cbae23f4aedcc22
|
| 3 |
size 14244
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b7ef7e01c7f7e481304676211a797020a963cacd6c94a5e1c27d96b74a321887
|
| 3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": null,
|
| 3 |
"best_model_checkpoint": null,
|
| 4 |
-
"epoch": 0.
|
| 5 |
"eval_steps": 500,
|
| 6 |
-
"global_step":
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
@@ -87,14 +87,30 @@
|
|
| 87 |
"loss": 0.4824,
|
| 88 |
"num_input_tokens_seen": 66414,
|
| 89 |
"step": 100
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 90 |
}
|
| 91 |
],
|
| 92 |
"logging_steps": 10,
|
| 93 |
"max_steps": 2795,
|
| 94 |
-
"num_input_tokens_seen":
|
| 95 |
"num_train_epochs": 1,
|
| 96 |
"save_steps": 20,
|
| 97 |
-
"total_flos":
|
| 98 |
"train_batch_size": 1,
|
| 99 |
"trial_name": null,
|
| 100 |
"trial_params": null
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": null,
|
| 3 |
"best_model_checkpoint": null,
|
| 4 |
+
"epoch": 0.04293381037567084,
|
| 5 |
"eval_steps": 500,
|
| 6 |
+
"global_step": 120,
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
|
|
| 87 |
"loss": 0.4824,
|
| 88 |
"num_input_tokens_seen": 66414,
|
| 89 |
"step": 100
|
| 90 |
+
},
|
| 91 |
+
{
|
| 92 |
+
"epoch": 0.03935599284436494,
|
| 93 |
+
"grad_norm": 0.35169002413749695,
|
| 94 |
+
"learning_rate": 0.00019212880143112702,
|
| 95 |
+
"loss": 0.4863,
|
| 96 |
+
"num_input_tokens_seen": 73451,
|
| 97 |
+
"step": 110
|
| 98 |
+
},
|
| 99 |
+
{
|
| 100 |
+
"epoch": 0.04293381037567084,
|
| 101 |
+
"grad_norm": 0.4088020324707031,
|
| 102 |
+
"learning_rate": 0.00019141323792486585,
|
| 103 |
+
"loss": 0.4792,
|
| 104 |
+
"num_input_tokens_seen": 81934,
|
| 105 |
+
"step": 120
|
| 106 |
}
|
| 107 |
],
|
| 108 |
"logging_steps": 10,
|
| 109 |
"max_steps": 2795,
|
| 110 |
+
"num_input_tokens_seen": 81934,
|
| 111 |
"num_train_epochs": 1,
|
| 112 |
"save_steps": 20,
|
| 113 |
+
"total_flos": 1842406318927872.0,
|
| 114 |
"train_batch_size": 1,
|
| 115 |
"trial_name": null,
|
| 116 |
"trial_params": null
|