Training in progress, step 2140, checkpoint
Browse files
last-checkpoint/adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 100697728
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2b31cd1bacf359ded93b1a975a832f43cf8275d8876e0472c643b229fff2eb00
|
| 3 |
size 100697728
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 201541754
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e963850275f272c22ef0a3706fe2ae328531160ce5f9a82b5cc306c0bdd72573
|
| 3 |
size 201541754
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14244
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0002b99f842540feaa883ef0561d8d2638b1009ac9f8f241ca4c7a7d22a87a41
|
| 3 |
size 14244
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3b36f7fc3da668a866fd41d662c3dd39a222899bdb3fe635785c3fffd4983ac1
|
| 3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": null,
|
| 3 |
"best_model_checkpoint": null,
|
| 4 |
-
"epoch": 0.
|
| 5 |
"eval_steps": 500,
|
| 6 |
-
"global_step":
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
@@ -1703,14 +1703,30 @@
|
|
| 1703 |
"loss": 0.3879,
|
| 1704 |
"num_input_tokens_seen": 1437151,
|
| 1705 |
"step": 2120
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1706 |
}
|
| 1707 |
],
|
| 1708 |
"logging_steps": 10,
|
| 1709 |
"max_steps": 2795,
|
| 1710 |
-
"num_input_tokens_seen":
|
| 1711 |
"num_train_epochs": 1,
|
| 1712 |
"save_steps": 20,
|
| 1713 |
-
"total_flos": 3.
|
| 1714 |
"train_batch_size": 1,
|
| 1715 |
"trial_name": null,
|
| 1716 |
"trial_params": null
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": null,
|
| 3 |
"best_model_checkpoint": null,
|
| 4 |
+
"epoch": 0.7656529516994633,
|
| 5 |
"eval_steps": 500,
|
| 6 |
+
"global_step": 2140,
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
|
|
| 1703 |
"loss": 0.3879,
|
| 1704 |
"num_input_tokens_seen": 1437151,
|
| 1705 |
"step": 2120
|
| 1706 |
+
},
|
| 1707 |
+
{
|
| 1708 |
+
"epoch": 0.7620751341681574,
|
| 1709 |
+
"grad_norm": 0.2736769914627075,
|
| 1710 |
+
"learning_rate": 4.758497316636852e-05,
|
| 1711 |
+
"loss": 0.3994,
|
| 1712 |
+
"num_input_tokens_seen": 1443666,
|
| 1713 |
+
"step": 2130
|
| 1714 |
+
},
|
| 1715 |
+
{
|
| 1716 |
+
"epoch": 0.7656529516994633,
|
| 1717 |
+
"grad_norm": 0.46681609749794006,
|
| 1718 |
+
"learning_rate": 4.686940966010734e-05,
|
| 1719 |
+
"loss": 0.3754,
|
| 1720 |
+
"num_input_tokens_seen": 1449724,
|
| 1721 |
+
"step": 2140
|
| 1722 |
}
|
| 1723 |
],
|
| 1724 |
"logging_steps": 10,
|
| 1725 |
"max_steps": 2795,
|
| 1726 |
+
"num_input_tokens_seen": 1449724,
|
| 1727 |
"num_train_epochs": 1,
|
| 1728 |
"save_steps": 20,
|
| 1729 |
+
"total_flos": 3.259917321626419e+16,
|
| 1730 |
"train_batch_size": 1,
|
| 1731 |
"trial_name": null,
|
| 1732 |
"trial_params": null
|