Upload folder using huggingface_hub
Browse filesThis view is limited to 50 files because it contains too many changes. See raw diff
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/0_client_model_round10.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/0_client_model_round12.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/0_client_model_round15.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/0_client_model_round17.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/0_client_model_round2.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/0_client_model_round20.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/0_client_model_round5.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/0_client_model_round7.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/0_trainer_state.json +392 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/1_client_model_round10.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/1_client_model_round12.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/1_client_model_round15.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/1_client_model_round17.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/1_client_model_round2.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/1_client_model_round20.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/1_client_model_round5.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/1_client_model_round7.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/1_trainer_state.json +392 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/2_client_model_round10.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/2_client_model_round12.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/2_client_model_round15.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/2_client_model_round17.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/2_client_model_round2.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/2_client_model_round20.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/2_client_model_round5.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/2_client_model_round7.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/2_trainer_state.json +392 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/3_client_model_round10.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/3_client_model_round12.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/3_client_model_round15.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/3_client_model_round17.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/3_client_model_round2.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/3_client_model_round20.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/3_client_model_round5.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/3_client_model_round7.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/3_trainer_state.json +392 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/4_client_model_round10.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/4_client_model_round12.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/4_client_model_round15.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/4_client_model_round17.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/4_client_model_round2.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/4_client_model_round20.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/4_client_model_round5.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/4_client_model_round7.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/4_trainer_state.json +392 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/5_client_model_round10.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/5_client_model_round12.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/5_client_model_round15.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/5_client_model_round17.pth +3 -0
- client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/5_client_model_round2.pth +3 -0
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/0_client_model_round10.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1dac7b6bce352f6c975f0f5656beef509afdab382bc722b932595b99bf50090c
|
| 3 |
+
size 180440142
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/0_client_model_round12.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5dcab448066a13c1f03938d764d26279e454333c8dc6923a707bf1cf0dc2f833
|
| 3 |
+
size 180440142
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/0_client_model_round15.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3c7d3453a23565008a42c4dccd1749cabfbf9ec23ea912014a4d4209943520d8
|
| 3 |
+
size 180440142
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/0_client_model_round17.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:56c05502646862715dc27c5584e6a3e60191712121cfc7f969a9461170eadbe9
|
| 3 |
+
size 180440142
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/0_client_model_round2.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:513aba8649167d639d7183da473d0eb85bdf76970b10f7d12d4d47d0bfa8f4f6
|
| 3 |
+
size 180439850
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/0_client_model_round20.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:269ac2ed7bfedd0bec7453eff915764e747247d28030f63f1ac848a7ed018c9f
|
| 3 |
+
size 180440142
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/0_client_model_round5.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:027f4e35e64efb5d494d33d46953933ce8bdb4a518cb33cecc238cfdbe8c1b84
|
| 3 |
+
size 180439850
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/0_client_model_round7.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:72abb5953c3d401fe841ebfb21c15bdd99db0975371868aa0aff2121d5510e50
|
| 3 |
+
size 180439850
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/0_trainer_state.json
ADDED
|
@@ -0,0 +1,392 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"best_metric": null,
|
| 3 |
+
"best_model_checkpoint": null,
|
| 4 |
+
"epoch": 1.0,
|
| 5 |
+
"eval_steps": 500,
|
| 6 |
+
"global_step": 100,
|
| 7 |
+
"is_hyper_param_search": false,
|
| 8 |
+
"is_local_process_zero": true,
|
| 9 |
+
"is_world_process_zero": true,
|
| 10 |
+
"log_history": [
|
| 11 |
+
{
|
| 12 |
+
"epoch": 0.02,
|
| 13 |
+
"grad_norm": 3.185389757156372,
|
| 14 |
+
"learning_rate": 2e-05,
|
| 15 |
+
"loss": 0.5313,
|
| 16 |
+
"step": 2
|
| 17 |
+
},
|
| 18 |
+
{
|
| 19 |
+
"epoch": 0.04,
|
| 20 |
+
"grad_norm": 1.462697148323059,
|
| 21 |
+
"learning_rate": 2e-05,
|
| 22 |
+
"loss": 2.1023,
|
| 23 |
+
"step": 4
|
| 24 |
+
},
|
| 25 |
+
{
|
| 26 |
+
"epoch": 0.06,
|
| 27 |
+
"grad_norm": 0.06314225494861603,
|
| 28 |
+
"learning_rate": 2e-05,
|
| 29 |
+
"loss": 0.0065,
|
| 30 |
+
"step": 6
|
| 31 |
+
},
|
| 32 |
+
{
|
| 33 |
+
"epoch": 0.08,
|
| 34 |
+
"grad_norm": 8.08350658416748,
|
| 35 |
+
"learning_rate": 2e-05,
|
| 36 |
+
"loss": 0.7299,
|
| 37 |
+
"step": 8
|
| 38 |
+
},
|
| 39 |
+
{
|
| 40 |
+
"epoch": 0.1,
|
| 41 |
+
"grad_norm": 1.8942979574203491,
|
| 42 |
+
"learning_rate": 2e-05,
|
| 43 |
+
"loss": 0.4338,
|
| 44 |
+
"step": 10
|
| 45 |
+
},
|
| 46 |
+
{
|
| 47 |
+
"epoch": 0.12,
|
| 48 |
+
"grad_norm": 7.397281646728516,
|
| 49 |
+
"learning_rate": 2e-05,
|
| 50 |
+
"loss": 0.9851,
|
| 51 |
+
"step": 12
|
| 52 |
+
},
|
| 53 |
+
{
|
| 54 |
+
"epoch": 0.14,
|
| 55 |
+
"grad_norm": 3.0156924724578857,
|
| 56 |
+
"learning_rate": 2e-05,
|
| 57 |
+
"loss": 0.1971,
|
| 58 |
+
"step": 14
|
| 59 |
+
},
|
| 60 |
+
{
|
| 61 |
+
"epoch": 0.16,
|
| 62 |
+
"grad_norm": 9.138948440551758,
|
| 63 |
+
"learning_rate": 2e-05,
|
| 64 |
+
"loss": 0.6213,
|
| 65 |
+
"step": 16
|
| 66 |
+
},
|
| 67 |
+
{
|
| 68 |
+
"epoch": 0.18,
|
| 69 |
+
"grad_norm": 8.548439979553223,
|
| 70 |
+
"learning_rate": 2e-05,
|
| 71 |
+
"loss": 0.8961,
|
| 72 |
+
"step": 18
|
| 73 |
+
},
|
| 74 |
+
{
|
| 75 |
+
"epoch": 0.2,
|
| 76 |
+
"grad_norm": 2.227810859680176,
|
| 77 |
+
"learning_rate": 2e-05,
|
| 78 |
+
"loss": 1.4456,
|
| 79 |
+
"step": 20
|
| 80 |
+
},
|
| 81 |
+
{
|
| 82 |
+
"epoch": 0.22,
|
| 83 |
+
"grad_norm": 4.551537990570068,
|
| 84 |
+
"learning_rate": 2e-05,
|
| 85 |
+
"loss": 1.636,
|
| 86 |
+
"step": 22
|
| 87 |
+
},
|
| 88 |
+
{
|
| 89 |
+
"epoch": 0.24,
|
| 90 |
+
"grad_norm": 6.477214336395264,
|
| 91 |
+
"learning_rate": 2e-05,
|
| 92 |
+
"loss": 1.0221,
|
| 93 |
+
"step": 24
|
| 94 |
+
},
|
| 95 |
+
{
|
| 96 |
+
"epoch": 0.26,
|
| 97 |
+
"grad_norm": 0.2676788866519928,
|
| 98 |
+
"learning_rate": 2e-05,
|
| 99 |
+
"loss": 0.2403,
|
| 100 |
+
"step": 26
|
| 101 |
+
},
|
| 102 |
+
{
|
| 103 |
+
"epoch": 0.28,
|
| 104 |
+
"grad_norm": 14.42641830444336,
|
| 105 |
+
"learning_rate": 2e-05,
|
| 106 |
+
"loss": 1.108,
|
| 107 |
+
"step": 28
|
| 108 |
+
},
|
| 109 |
+
{
|
| 110 |
+
"epoch": 0.3,
|
| 111 |
+
"grad_norm": 0.8010652661323547,
|
| 112 |
+
"learning_rate": 2e-05,
|
| 113 |
+
"loss": 0.3901,
|
| 114 |
+
"step": 30
|
| 115 |
+
},
|
| 116 |
+
{
|
| 117 |
+
"epoch": 0.32,
|
| 118 |
+
"grad_norm": 9.867847442626953,
|
| 119 |
+
"learning_rate": 2e-05,
|
| 120 |
+
"loss": 1.073,
|
| 121 |
+
"step": 32
|
| 122 |
+
},
|
| 123 |
+
{
|
| 124 |
+
"epoch": 0.34,
|
| 125 |
+
"grad_norm": 0.5489710569381714,
|
| 126 |
+
"learning_rate": 2e-05,
|
| 127 |
+
"loss": 0.1765,
|
| 128 |
+
"step": 34
|
| 129 |
+
},
|
| 130 |
+
{
|
| 131 |
+
"epoch": 0.36,
|
| 132 |
+
"grad_norm": 2.8706305027008057,
|
| 133 |
+
"learning_rate": 2e-05,
|
| 134 |
+
"loss": 0.8621,
|
| 135 |
+
"step": 36
|
| 136 |
+
},
|
| 137 |
+
{
|
| 138 |
+
"epoch": 0.38,
|
| 139 |
+
"grad_norm": 3.2662458419799805,
|
| 140 |
+
"learning_rate": 2e-05,
|
| 141 |
+
"loss": 0.1863,
|
| 142 |
+
"step": 38
|
| 143 |
+
},
|
| 144 |
+
{
|
| 145 |
+
"epoch": 0.4,
|
| 146 |
+
"grad_norm": 10.825136184692383,
|
| 147 |
+
"learning_rate": 2e-05,
|
| 148 |
+
"loss": 1.1348,
|
| 149 |
+
"step": 40
|
| 150 |
+
},
|
| 151 |
+
{
|
| 152 |
+
"epoch": 0.42,
|
| 153 |
+
"grad_norm": 0.9670252799987793,
|
| 154 |
+
"learning_rate": 2e-05,
|
| 155 |
+
"loss": 0.6283,
|
| 156 |
+
"step": 42
|
| 157 |
+
},
|
| 158 |
+
{
|
| 159 |
+
"epoch": 0.44,
|
| 160 |
+
"grad_norm": 2.5218558311462402,
|
| 161 |
+
"learning_rate": 2e-05,
|
| 162 |
+
"loss": 0.8367,
|
| 163 |
+
"step": 44
|
| 164 |
+
},
|
| 165 |
+
{
|
| 166 |
+
"epoch": 0.46,
|
| 167 |
+
"grad_norm": 1.4162901639938354,
|
| 168 |
+
"learning_rate": 2e-05,
|
| 169 |
+
"loss": 0.4749,
|
| 170 |
+
"step": 46
|
| 171 |
+
},
|
| 172 |
+
{
|
| 173 |
+
"epoch": 0.48,
|
| 174 |
+
"grad_norm": 5.961083889007568,
|
| 175 |
+
"learning_rate": 2e-05,
|
| 176 |
+
"loss": 0.3601,
|
| 177 |
+
"step": 48
|
| 178 |
+
},
|
| 179 |
+
{
|
| 180 |
+
"epoch": 0.5,
|
| 181 |
+
"grad_norm": 7.117353439331055,
|
| 182 |
+
"learning_rate": 2e-05,
|
| 183 |
+
"loss": 0.6846,
|
| 184 |
+
"step": 50
|
| 185 |
+
},
|
| 186 |
+
{
|
| 187 |
+
"epoch": 0.52,
|
| 188 |
+
"grad_norm": 6.346409797668457,
|
| 189 |
+
"learning_rate": 2e-05,
|
| 190 |
+
"loss": 1.0359,
|
| 191 |
+
"step": 52
|
| 192 |
+
},
|
| 193 |
+
{
|
| 194 |
+
"epoch": 0.54,
|
| 195 |
+
"grad_norm": 3.3846776485443115,
|
| 196 |
+
"learning_rate": 2e-05,
|
| 197 |
+
"loss": 0.2492,
|
| 198 |
+
"step": 54
|
| 199 |
+
},
|
| 200 |
+
{
|
| 201 |
+
"epoch": 0.56,
|
| 202 |
+
"grad_norm": 0.2677325904369354,
|
| 203 |
+
"learning_rate": 2e-05,
|
| 204 |
+
"loss": 1.0758,
|
| 205 |
+
"step": 56
|
| 206 |
+
},
|
| 207 |
+
{
|
| 208 |
+
"epoch": 0.58,
|
| 209 |
+
"grad_norm": 3.839362621307373,
|
| 210 |
+
"learning_rate": 2e-05,
|
| 211 |
+
"loss": 0.7528,
|
| 212 |
+
"step": 58
|
| 213 |
+
},
|
| 214 |
+
{
|
| 215 |
+
"epoch": 0.6,
|
| 216 |
+
"grad_norm": 15.682966232299805,
|
| 217 |
+
"learning_rate": 2e-05,
|
| 218 |
+
"loss": 0.9388,
|
| 219 |
+
"step": 60
|
| 220 |
+
},
|
| 221 |
+
{
|
| 222 |
+
"epoch": 0.62,
|
| 223 |
+
"grad_norm": 9.203821182250977,
|
| 224 |
+
"learning_rate": 2e-05,
|
| 225 |
+
"loss": 1.9389,
|
| 226 |
+
"step": 62
|
| 227 |
+
},
|
| 228 |
+
{
|
| 229 |
+
"epoch": 0.64,
|
| 230 |
+
"grad_norm": 10.753218650817871,
|
| 231 |
+
"learning_rate": 2e-05,
|
| 232 |
+
"loss": 1.8481,
|
| 233 |
+
"step": 64
|
| 234 |
+
},
|
| 235 |
+
{
|
| 236 |
+
"epoch": 0.66,
|
| 237 |
+
"grad_norm": 3.960542678833008,
|
| 238 |
+
"learning_rate": 2e-05,
|
| 239 |
+
"loss": 0.9183,
|
| 240 |
+
"step": 66
|
| 241 |
+
},
|
| 242 |
+
{
|
| 243 |
+
"epoch": 0.68,
|
| 244 |
+
"grad_norm": 0.8610289096832275,
|
| 245 |
+
"learning_rate": 2e-05,
|
| 246 |
+
"loss": 0.4862,
|
| 247 |
+
"step": 68
|
| 248 |
+
},
|
| 249 |
+
{
|
| 250 |
+
"epoch": 0.7,
|
| 251 |
+
"grad_norm": 0.5907980799674988,
|
| 252 |
+
"learning_rate": 2e-05,
|
| 253 |
+
"loss": 0.0323,
|
| 254 |
+
"step": 70
|
| 255 |
+
},
|
| 256 |
+
{
|
| 257 |
+
"epoch": 0.72,
|
| 258 |
+
"grad_norm": 5.849497318267822,
|
| 259 |
+
"learning_rate": 2e-05,
|
| 260 |
+
"loss": 0.5046,
|
| 261 |
+
"step": 72
|
| 262 |
+
},
|
| 263 |
+
{
|
| 264 |
+
"epoch": 0.74,
|
| 265 |
+
"grad_norm": 2.7647833824157715,
|
| 266 |
+
"learning_rate": 2e-05,
|
| 267 |
+
"loss": 0.9464,
|
| 268 |
+
"step": 74
|
| 269 |
+
},
|
| 270 |
+
{
|
| 271 |
+
"epoch": 0.76,
|
| 272 |
+
"grad_norm": 2.2242050170898438,
|
| 273 |
+
"learning_rate": 2e-05,
|
| 274 |
+
"loss": 1.0414,
|
| 275 |
+
"step": 76
|
| 276 |
+
},
|
| 277 |
+
{
|
| 278 |
+
"epoch": 0.78,
|
| 279 |
+
"grad_norm": 0.43988850712776184,
|
| 280 |
+
"learning_rate": 2e-05,
|
| 281 |
+
"loss": 0.0393,
|
| 282 |
+
"step": 78
|
| 283 |
+
},
|
| 284 |
+
{
|
| 285 |
+
"epoch": 0.8,
|
| 286 |
+
"grad_norm": 7.581626892089844,
|
| 287 |
+
"learning_rate": 2e-05,
|
| 288 |
+
"loss": 0.914,
|
| 289 |
+
"step": 80
|
| 290 |
+
},
|
| 291 |
+
{
|
| 292 |
+
"epoch": 0.82,
|
| 293 |
+
"grad_norm": 8.071428298950195,
|
| 294 |
+
"learning_rate": 2e-05,
|
| 295 |
+
"loss": 0.5456,
|
| 296 |
+
"step": 82
|
| 297 |
+
},
|
| 298 |
+
{
|
| 299 |
+
"epoch": 0.84,
|
| 300 |
+
"grad_norm": 9.775481224060059,
|
| 301 |
+
"learning_rate": 2e-05,
|
| 302 |
+
"loss": 1.1202,
|
| 303 |
+
"step": 84
|
| 304 |
+
},
|
| 305 |
+
{
|
| 306 |
+
"epoch": 0.86,
|
| 307 |
+
"grad_norm": 4.938329219818115,
|
| 308 |
+
"learning_rate": 2e-05,
|
| 309 |
+
"loss": 0.4349,
|
| 310 |
+
"step": 86
|
| 311 |
+
},
|
| 312 |
+
{
|
| 313 |
+
"epoch": 0.88,
|
| 314 |
+
"grad_norm": 2.7860143184661865,
|
| 315 |
+
"learning_rate": 2e-05,
|
| 316 |
+
"loss": 0.3314,
|
| 317 |
+
"step": 88
|
| 318 |
+
},
|
| 319 |
+
{
|
| 320 |
+
"epoch": 0.9,
|
| 321 |
+
"grad_norm": 1.644363522529602,
|
| 322 |
+
"learning_rate": 2e-05,
|
| 323 |
+
"loss": 0.2175,
|
| 324 |
+
"step": 90
|
| 325 |
+
},
|
| 326 |
+
{
|
| 327 |
+
"epoch": 0.92,
|
| 328 |
+
"grad_norm": 3.6504173278808594,
|
| 329 |
+
"learning_rate": 2e-05,
|
| 330 |
+
"loss": 0.2242,
|
| 331 |
+
"step": 92
|
| 332 |
+
},
|
| 333 |
+
{
|
| 334 |
+
"epoch": 0.94,
|
| 335 |
+
"grad_norm": 3.1692447662353516,
|
| 336 |
+
"learning_rate": 2e-05,
|
| 337 |
+
"loss": 0.3253,
|
| 338 |
+
"step": 94
|
| 339 |
+
},
|
| 340 |
+
{
|
| 341 |
+
"epoch": 0.96,
|
| 342 |
+
"grad_norm": 7.509362697601318,
|
| 343 |
+
"learning_rate": 2e-05,
|
| 344 |
+
"loss": 0.9171,
|
| 345 |
+
"step": 96
|
| 346 |
+
},
|
| 347 |
+
{
|
| 348 |
+
"epoch": 0.98,
|
| 349 |
+
"grad_norm": 0.7557180523872375,
|
| 350 |
+
"learning_rate": 2e-05,
|
| 351 |
+
"loss": 0.0612,
|
| 352 |
+
"step": 98
|
| 353 |
+
},
|
| 354 |
+
{
|
| 355 |
+
"epoch": 1.0,
|
| 356 |
+
"grad_norm": 0.4876616597175598,
|
| 357 |
+
"learning_rate": 2e-05,
|
| 358 |
+
"loss": 0.054,
|
| 359 |
+
"step": 100
|
| 360 |
+
},
|
| 361 |
+
{
|
| 362 |
+
"epoch": 1.0,
|
| 363 |
+
"step": 100,
|
| 364 |
+
"total_flos": 2050008522162176.0,
|
| 365 |
+
"train_loss": 0.714327917098999,
|
| 366 |
+
"train_runtime": 101.3941,
|
| 367 |
+
"train_samples_per_second": 3.945,
|
| 368 |
+
"train_steps_per_second": 0.986
|
| 369 |
+
}
|
| 370 |
+
],
|
| 371 |
+
"logging_steps": 2,
|
| 372 |
+
"max_steps": 100,
|
| 373 |
+
"num_input_tokens_seen": 0,
|
| 374 |
+
"num_train_epochs": 1,
|
| 375 |
+
"save_steps": 500,
|
| 376 |
+
"stateful_callbacks": {
|
| 377 |
+
"TrainerControl": {
|
| 378 |
+
"args": {
|
| 379 |
+
"should_epoch_stop": false,
|
| 380 |
+
"should_evaluate": false,
|
| 381 |
+
"should_log": false,
|
| 382 |
+
"should_save": false,
|
| 383 |
+
"should_training_stop": false
|
| 384 |
+
},
|
| 385 |
+
"attributes": {}
|
| 386 |
+
}
|
| 387 |
+
},
|
| 388 |
+
"total_flos": 2050008522162176.0,
|
| 389 |
+
"train_batch_size": 1,
|
| 390 |
+
"trial_name": null,
|
| 391 |
+
"trial_params": null
|
| 392 |
+
}
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/1_client_model_round10.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d3d4c4d89bdb3f350e2518ed150ec2a0b791fe39b81f2e0f62a2f5cc4ba7666e
|
| 3 |
+
size 180440142
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/1_client_model_round12.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:459f79aed03d2ed8311a89bfc1543de0307289da1e03e5b61b17859497bd7b94
|
| 3 |
+
size 180440142
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/1_client_model_round15.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d025fa1d64446f3b8d8480477c7c2366e1dc17c670e582008dc6b3d3ebdae8eb
|
| 3 |
+
size 180440142
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/1_client_model_round17.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:88310f262c4da1aa92e63114b5417a63139a526b4adffa1eb7585320a682b4cf
|
| 3 |
+
size 180440142
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/1_client_model_round2.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e6940d88337600557f8ef5224d94bee18273ca815983d5753451f7393f3ae082
|
| 3 |
+
size 180439850
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/1_client_model_round20.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8f2718fac6eeee5d8a92ba8245e122b5832a37a9c8d03f18654502259c52e93e
|
| 3 |
+
size 180440142
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/1_client_model_round5.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0c5ed776c37af82da7e7c990f70a018ca7465e1c03abbf9f8e50c2450e301235
|
| 3 |
+
size 180439850
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/1_client_model_round7.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e310e436e59b58a8b51ebb8bfeeadcd4c23607edc7a36a702c0ef8a1d8775b51
|
| 3 |
+
size 180439850
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/1_trainer_state.json
ADDED
|
@@ -0,0 +1,392 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"best_metric": null,
|
| 3 |
+
"best_model_checkpoint": null,
|
| 4 |
+
"epoch": 1.0,
|
| 5 |
+
"eval_steps": 500,
|
| 6 |
+
"global_step": 100,
|
| 7 |
+
"is_hyper_param_search": false,
|
| 8 |
+
"is_local_process_zero": true,
|
| 9 |
+
"is_world_process_zero": true,
|
| 10 |
+
"log_history": [
|
| 11 |
+
{
|
| 12 |
+
"epoch": 0.02,
|
| 13 |
+
"grad_norm": 0.0033269652631133795,
|
| 14 |
+
"learning_rate": 2e-05,
|
| 15 |
+
"loss": 0.0085,
|
| 16 |
+
"step": 2
|
| 17 |
+
},
|
| 18 |
+
{
|
| 19 |
+
"epoch": 0.04,
|
| 20 |
+
"grad_norm": 0.041157741099596024,
|
| 21 |
+
"learning_rate": 2e-05,
|
| 22 |
+
"loss": 0.0008,
|
| 23 |
+
"step": 4
|
| 24 |
+
},
|
| 25 |
+
{
|
| 26 |
+
"epoch": 0.06,
|
| 27 |
+
"grad_norm": 0.9592494368553162,
|
| 28 |
+
"learning_rate": 2e-05,
|
| 29 |
+
"loss": 0.0262,
|
| 30 |
+
"step": 6
|
| 31 |
+
},
|
| 32 |
+
{
|
| 33 |
+
"epoch": 0.08,
|
| 34 |
+
"grad_norm": 0.6419932246208191,
|
| 35 |
+
"learning_rate": 2e-05,
|
| 36 |
+
"loss": 0.0282,
|
| 37 |
+
"step": 8
|
| 38 |
+
},
|
| 39 |
+
{
|
| 40 |
+
"epoch": 0.1,
|
| 41 |
+
"grad_norm": 5.6463303565979,
|
| 42 |
+
"learning_rate": 2e-05,
|
| 43 |
+
"loss": 0.6913,
|
| 44 |
+
"step": 10
|
| 45 |
+
},
|
| 46 |
+
{
|
| 47 |
+
"epoch": 0.12,
|
| 48 |
+
"grad_norm": 0.000999588519334793,
|
| 49 |
+
"learning_rate": 2e-05,
|
| 50 |
+
"loss": 0.0004,
|
| 51 |
+
"step": 12
|
| 52 |
+
},
|
| 53 |
+
{
|
| 54 |
+
"epoch": 0.14,
|
| 55 |
+
"grad_norm": 0.2222857028245926,
|
| 56 |
+
"learning_rate": 2e-05,
|
| 57 |
+
"loss": 0.0397,
|
| 58 |
+
"step": 14
|
| 59 |
+
},
|
| 60 |
+
{
|
| 61 |
+
"epoch": 0.16,
|
| 62 |
+
"grad_norm": 0.16892485320568085,
|
| 63 |
+
"learning_rate": 2e-05,
|
| 64 |
+
"loss": 0.0769,
|
| 65 |
+
"step": 16
|
| 66 |
+
},
|
| 67 |
+
{
|
| 68 |
+
"epoch": 0.18,
|
| 69 |
+
"grad_norm": 0.009978498332202435,
|
| 70 |
+
"learning_rate": 2e-05,
|
| 71 |
+
"loss": 0.0012,
|
| 72 |
+
"step": 18
|
| 73 |
+
},
|
| 74 |
+
{
|
| 75 |
+
"epoch": 0.2,
|
| 76 |
+
"grad_norm": 2.3538615703582764,
|
| 77 |
+
"learning_rate": 2e-05,
|
| 78 |
+
"loss": 0.4112,
|
| 79 |
+
"step": 20
|
| 80 |
+
},
|
| 81 |
+
{
|
| 82 |
+
"epoch": 0.22,
|
| 83 |
+
"grad_norm": 0.0049792262725532055,
|
| 84 |
+
"learning_rate": 2e-05,
|
| 85 |
+
"loss": 0.0009,
|
| 86 |
+
"step": 22
|
| 87 |
+
},
|
| 88 |
+
{
|
| 89 |
+
"epoch": 0.24,
|
| 90 |
+
"grad_norm": 0.9717739820480347,
|
| 91 |
+
"learning_rate": 2e-05,
|
| 92 |
+
"loss": 0.0344,
|
| 93 |
+
"step": 24
|
| 94 |
+
},
|
| 95 |
+
{
|
| 96 |
+
"epoch": 0.26,
|
| 97 |
+
"grad_norm": 0.007954389788210392,
|
| 98 |
+
"learning_rate": 2e-05,
|
| 99 |
+
"loss": 0.0886,
|
| 100 |
+
"step": 26
|
| 101 |
+
},
|
| 102 |
+
{
|
| 103 |
+
"epoch": 0.28,
|
| 104 |
+
"grad_norm": 0.15055210888385773,
|
| 105 |
+
"learning_rate": 2e-05,
|
| 106 |
+
"loss": 0.982,
|
| 107 |
+
"step": 28
|
| 108 |
+
},
|
| 109 |
+
{
|
| 110 |
+
"epoch": 0.3,
|
| 111 |
+
"grad_norm": 0.05759048089385033,
|
| 112 |
+
"learning_rate": 2e-05,
|
| 113 |
+
"loss": 0.5327,
|
| 114 |
+
"step": 30
|
| 115 |
+
},
|
| 116 |
+
{
|
| 117 |
+
"epoch": 0.32,
|
| 118 |
+
"grad_norm": 0.051013048738241196,
|
| 119 |
+
"learning_rate": 2e-05,
|
| 120 |
+
"loss": 0.0034,
|
| 121 |
+
"step": 32
|
| 122 |
+
},
|
| 123 |
+
{
|
| 124 |
+
"epoch": 0.34,
|
| 125 |
+
"grad_norm": 0.027761617675423622,
|
| 126 |
+
"learning_rate": 2e-05,
|
| 127 |
+
"loss": 0.0036,
|
| 128 |
+
"step": 34
|
| 129 |
+
},
|
| 130 |
+
{
|
| 131 |
+
"epoch": 0.36,
|
| 132 |
+
"grad_norm": 0.010388568975031376,
|
| 133 |
+
"learning_rate": 2e-05,
|
| 134 |
+
"loss": 0.0009,
|
| 135 |
+
"step": 36
|
| 136 |
+
},
|
| 137 |
+
{
|
| 138 |
+
"epoch": 0.38,
|
| 139 |
+
"grad_norm": 0.01740710809826851,
|
| 140 |
+
"learning_rate": 2e-05,
|
| 141 |
+
"loss": 0.2172,
|
| 142 |
+
"step": 38
|
| 143 |
+
},
|
| 144 |
+
{
|
| 145 |
+
"epoch": 0.4,
|
| 146 |
+
"grad_norm": 0.017691656947135925,
|
| 147 |
+
"learning_rate": 2e-05,
|
| 148 |
+
"loss": 0.0017,
|
| 149 |
+
"step": 40
|
| 150 |
+
},
|
| 151 |
+
{
|
| 152 |
+
"epoch": 0.42,
|
| 153 |
+
"grad_norm": 0.31259411573410034,
|
| 154 |
+
"learning_rate": 2e-05,
|
| 155 |
+
"loss": 0.0131,
|
| 156 |
+
"step": 42
|
| 157 |
+
},
|
| 158 |
+
{
|
| 159 |
+
"epoch": 0.44,
|
| 160 |
+
"grad_norm": 0.02895217575132847,
|
| 161 |
+
"learning_rate": 2e-05,
|
| 162 |
+
"loss": 0.0063,
|
| 163 |
+
"step": 44
|
| 164 |
+
},
|
| 165 |
+
{
|
| 166 |
+
"epoch": 0.46,
|
| 167 |
+
"grad_norm": 0.009466350078582764,
|
| 168 |
+
"learning_rate": 2e-05,
|
| 169 |
+
"loss": 0.0024,
|
| 170 |
+
"step": 46
|
| 171 |
+
},
|
| 172 |
+
{
|
| 173 |
+
"epoch": 0.48,
|
| 174 |
+
"grad_norm": 0.41093146800994873,
|
| 175 |
+
"learning_rate": 2e-05,
|
| 176 |
+
"loss": 0.018,
|
| 177 |
+
"step": 48
|
| 178 |
+
},
|
| 179 |
+
{
|
| 180 |
+
"epoch": 0.5,
|
| 181 |
+
"grad_norm": 0.006547864992171526,
|
| 182 |
+
"learning_rate": 2e-05,
|
| 183 |
+
"loss": 0.0008,
|
| 184 |
+
"step": 50
|
| 185 |
+
},
|
| 186 |
+
{
|
| 187 |
+
"epoch": 0.52,
|
| 188 |
+
"grad_norm": 0.009852895513176918,
|
| 189 |
+
"learning_rate": 2e-05,
|
| 190 |
+
"loss": 0.0273,
|
| 191 |
+
"step": 52
|
| 192 |
+
},
|
| 193 |
+
{
|
| 194 |
+
"epoch": 0.54,
|
| 195 |
+
"grad_norm": 0.0796719342470169,
|
| 196 |
+
"learning_rate": 2e-05,
|
| 197 |
+
"loss": 0.0066,
|
| 198 |
+
"step": 54
|
| 199 |
+
},
|
| 200 |
+
{
|
| 201 |
+
"epoch": 0.56,
|
| 202 |
+
"grad_norm": 0.17664939165115356,
|
| 203 |
+
"learning_rate": 2e-05,
|
| 204 |
+
"loss": 0.0089,
|
| 205 |
+
"step": 56
|
| 206 |
+
},
|
| 207 |
+
{
|
| 208 |
+
"epoch": 0.58,
|
| 209 |
+
"grad_norm": 0.03962412849068642,
|
| 210 |
+
"learning_rate": 2e-05,
|
| 211 |
+
"loss": 0.0023,
|
| 212 |
+
"step": 58
|
| 213 |
+
},
|
| 214 |
+
{
|
| 215 |
+
"epoch": 0.6,
|
| 216 |
+
"grad_norm": 0.23390346765518188,
|
| 217 |
+
"learning_rate": 2e-05,
|
| 218 |
+
"loss": 0.0101,
|
| 219 |
+
"step": 60
|
| 220 |
+
},
|
| 221 |
+
{
|
| 222 |
+
"epoch": 0.62,
|
| 223 |
+
"grad_norm": 0.035884492099285126,
|
| 224 |
+
"learning_rate": 2e-05,
|
| 225 |
+
"loss": 0.0022,
|
| 226 |
+
"step": 62
|
| 227 |
+
},
|
| 228 |
+
{
|
| 229 |
+
"epoch": 0.64,
|
| 230 |
+
"grad_norm": 0.006772290449589491,
|
| 231 |
+
"learning_rate": 2e-05,
|
| 232 |
+
"loss": 0.0014,
|
| 233 |
+
"step": 64
|
| 234 |
+
},
|
| 235 |
+
{
|
| 236 |
+
"epoch": 0.66,
|
| 237 |
+
"grad_norm": 0.13574518263339996,
|
| 238 |
+
"learning_rate": 2e-05,
|
| 239 |
+
"loss": 0.0098,
|
| 240 |
+
"step": 66
|
| 241 |
+
},
|
| 242 |
+
{
|
| 243 |
+
"epoch": 0.68,
|
| 244 |
+
"grad_norm": 17.729595184326172,
|
| 245 |
+
"learning_rate": 2e-05,
|
| 246 |
+
"loss": 1.8484,
|
| 247 |
+
"step": 68
|
| 248 |
+
},
|
| 249 |
+
{
|
| 250 |
+
"epoch": 0.7,
|
| 251 |
+
"grad_norm": 0.05980616807937622,
|
| 252 |
+
"learning_rate": 2e-05,
|
| 253 |
+
"loss": 0.0085,
|
| 254 |
+
"step": 70
|
| 255 |
+
},
|
| 256 |
+
{
|
| 257 |
+
"epoch": 0.72,
|
| 258 |
+
"grad_norm": 1.1025233268737793,
|
| 259 |
+
"learning_rate": 2e-05,
|
| 260 |
+
"loss": 1.0268,
|
| 261 |
+
"step": 72
|
| 262 |
+
},
|
| 263 |
+
{
|
| 264 |
+
"epoch": 0.74,
|
| 265 |
+
"grad_norm": 0.00824617501348257,
|
| 266 |
+
"learning_rate": 2e-05,
|
| 267 |
+
"loss": 0.0033,
|
| 268 |
+
"step": 74
|
| 269 |
+
},
|
| 270 |
+
{
|
| 271 |
+
"epoch": 0.76,
|
| 272 |
+
"grad_norm": 4.188517093658447,
|
| 273 |
+
"learning_rate": 2e-05,
|
| 274 |
+
"loss": 0.38,
|
| 275 |
+
"step": 76
|
| 276 |
+
},
|
| 277 |
+
{
|
| 278 |
+
"epoch": 0.78,
|
| 279 |
+
"grad_norm": 0.014042158611118793,
|
| 280 |
+
"learning_rate": 2e-05,
|
| 281 |
+
"loss": 0.0017,
|
| 282 |
+
"step": 78
|
| 283 |
+
},
|
| 284 |
+
{
|
| 285 |
+
"epoch": 0.8,
|
| 286 |
+
"grad_norm": 0.04890606924891472,
|
| 287 |
+
"learning_rate": 2e-05,
|
| 288 |
+
"loss": 0.1103,
|
| 289 |
+
"step": 80
|
| 290 |
+
},
|
| 291 |
+
{
|
| 292 |
+
"epoch": 0.82,
|
| 293 |
+
"grad_norm": 0.023772185668349266,
|
| 294 |
+
"learning_rate": 2e-05,
|
| 295 |
+
"loss": 0.0018,
|
| 296 |
+
"step": 82
|
| 297 |
+
},
|
| 298 |
+
{
|
| 299 |
+
"epoch": 0.84,
|
| 300 |
+
"grad_norm": 3.432652473449707,
|
| 301 |
+
"learning_rate": 2e-05,
|
| 302 |
+
"loss": 0.3927,
|
| 303 |
+
"step": 84
|
| 304 |
+
},
|
| 305 |
+
{
|
| 306 |
+
"epoch": 0.86,
|
| 307 |
+
"grad_norm": 0.6805059909820557,
|
| 308 |
+
"learning_rate": 2e-05,
|
| 309 |
+
"loss": 0.0469,
|
| 310 |
+
"step": 86
|
| 311 |
+
},
|
| 312 |
+
{
|
| 313 |
+
"epoch": 0.88,
|
| 314 |
+
"grad_norm": 1.975512146949768,
|
| 315 |
+
"learning_rate": 2e-05,
|
| 316 |
+
"loss": 0.2212,
|
| 317 |
+
"step": 88
|
| 318 |
+
},
|
| 319 |
+
{
|
| 320 |
+
"epoch": 0.9,
|
| 321 |
+
"grad_norm": 0.9377481937408447,
|
| 322 |
+
"learning_rate": 2e-05,
|
| 323 |
+
"loss": 0.4155,
|
| 324 |
+
"step": 90
|
| 325 |
+
},
|
| 326 |
+
{
|
| 327 |
+
"epoch": 0.92,
|
| 328 |
+
"grad_norm": 0.2486044466495514,
|
| 329 |
+
"learning_rate": 2e-05,
|
| 330 |
+
"loss": 0.0152,
|
| 331 |
+
"step": 92
|
| 332 |
+
},
|
| 333 |
+
{
|
| 334 |
+
"epoch": 0.94,
|
| 335 |
+
"grad_norm": 0.015230941586196423,
|
| 336 |
+
"learning_rate": 2e-05,
|
| 337 |
+
"loss": 0.0133,
|
| 338 |
+
"step": 94
|
| 339 |
+
},
|
| 340 |
+
{
|
| 341 |
+
"epoch": 0.96,
|
| 342 |
+
"grad_norm": 0.1053871437907219,
|
| 343 |
+
"learning_rate": 2e-05,
|
| 344 |
+
"loss": 0.0693,
|
| 345 |
+
"step": 96
|
| 346 |
+
},
|
| 347 |
+
{
|
| 348 |
+
"epoch": 0.98,
|
| 349 |
+
"grad_norm": 0.023564128205180168,
|
| 350 |
+
"learning_rate": 2e-05,
|
| 351 |
+
"loss": 0.0031,
|
| 352 |
+
"step": 98
|
| 353 |
+
},
|
| 354 |
+
{
|
| 355 |
+
"epoch": 1.0,
|
| 356 |
+
"grad_norm": 0.015611840412020683,
|
| 357 |
+
"learning_rate": 2e-05,
|
| 358 |
+
"loss": 0.025,
|
| 359 |
+
"step": 100
|
| 360 |
+
},
|
| 361 |
+
{
|
| 362 |
+
"epoch": 1.0,
|
| 363 |
+
"step": 100,
|
| 364 |
+
"total_flos": 2066355427737600.0,
|
| 365 |
+
"train_loss": 0.1568439483642578,
|
| 366 |
+
"train_runtime": 99.9893,
|
| 367 |
+
"train_samples_per_second": 4.0,
|
| 368 |
+
"train_steps_per_second": 1.0
|
| 369 |
+
}
|
| 370 |
+
],
|
| 371 |
+
"logging_steps": 2,
|
| 372 |
+
"max_steps": 100,
|
| 373 |
+
"num_input_tokens_seen": 0,
|
| 374 |
+
"num_train_epochs": 1,
|
| 375 |
+
"save_steps": 500,
|
| 376 |
+
"stateful_callbacks": {
|
| 377 |
+
"TrainerControl": {
|
| 378 |
+
"args": {
|
| 379 |
+
"should_epoch_stop": false,
|
| 380 |
+
"should_evaluate": false,
|
| 381 |
+
"should_log": false,
|
| 382 |
+
"should_save": false,
|
| 383 |
+
"should_training_stop": false
|
| 384 |
+
},
|
| 385 |
+
"attributes": {}
|
| 386 |
+
}
|
| 387 |
+
},
|
| 388 |
+
"total_flos": 2066355427737600.0,
|
| 389 |
+
"train_batch_size": 1,
|
| 390 |
+
"trial_name": null,
|
| 391 |
+
"trial_params": null
|
| 392 |
+
}
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/2_client_model_round10.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:de2a14f0034bf1a514488a2a3c3dd156f6cd9486176464301efb43f273eeaa69
|
| 3 |
+
size 389170582
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/2_client_model_round12.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ea2995ff2139b05e73dd687bf519c8a8de7d3e2ada63a7ff7c04c34ae1853fd2
|
| 3 |
+
size 389170582
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/2_client_model_round15.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9b71a7bf9a66fe1dfe1482aa1d0d500b7ebfd0950d854e6159f8c7f469e5a0c2
|
| 3 |
+
size 389170582
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/2_client_model_round17.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e7db80190dcd29b88d459a63955e96ddfe0d6783a855674d557761b674e3e07e
|
| 3 |
+
size 389170582
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/2_client_model_round2.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3b28230a69be1080fce36c0aa88635abdbef0bee9461b7e7f17386ee963856aa
|
| 3 |
+
size 389170122
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/2_client_model_round20.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8fc05025bf8c8f9ceebe415e6840941b0c064ff4afe2125e74ae593b3c6d5b48
|
| 3 |
+
size 389170582
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/2_client_model_round5.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:49fc6d328f225665d42987b8a4bba2cd04fdf3ee8634fe9e9699f73ffb9ea04e
|
| 3 |
+
size 389170122
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/2_client_model_round7.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3acfe893b0c739503151ea961300cc6c25f8f270719d005c1737ae4d31e60eca
|
| 3 |
+
size 389170122
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/2_trainer_state.json
ADDED
|
@@ -0,0 +1,392 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"best_metric": null,
|
| 3 |
+
"best_model_checkpoint": null,
|
| 4 |
+
"epoch": 1.0,
|
| 5 |
+
"eval_steps": 500,
|
| 6 |
+
"global_step": 100,
|
| 7 |
+
"is_hyper_param_search": false,
|
| 8 |
+
"is_local_process_zero": true,
|
| 9 |
+
"is_world_process_zero": true,
|
| 10 |
+
"log_history": [
|
| 11 |
+
{
|
| 12 |
+
"epoch": 0.02,
|
| 13 |
+
"grad_norm": 1.806666374206543,
|
| 14 |
+
"learning_rate": 2e-05,
|
| 15 |
+
"loss": 0.3127,
|
| 16 |
+
"step": 2
|
| 17 |
+
},
|
| 18 |
+
{
|
| 19 |
+
"epoch": 0.04,
|
| 20 |
+
"grad_norm": 2.772969961166382,
|
| 21 |
+
"learning_rate": 2e-05,
|
| 22 |
+
"loss": 0.6935,
|
| 23 |
+
"step": 4
|
| 24 |
+
},
|
| 25 |
+
{
|
| 26 |
+
"epoch": 0.06,
|
| 27 |
+
"grad_norm": 5.492923736572266,
|
| 28 |
+
"learning_rate": 2e-05,
|
| 29 |
+
"loss": 1.7379,
|
| 30 |
+
"step": 6
|
| 31 |
+
},
|
| 32 |
+
{
|
| 33 |
+
"epoch": 0.08,
|
| 34 |
+
"grad_norm": 6.6056976318359375,
|
| 35 |
+
"learning_rate": 2e-05,
|
| 36 |
+
"loss": 0.8991,
|
| 37 |
+
"step": 8
|
| 38 |
+
},
|
| 39 |
+
{
|
| 40 |
+
"epoch": 0.1,
|
| 41 |
+
"grad_norm": 2.7061705589294434,
|
| 42 |
+
"learning_rate": 2e-05,
|
| 43 |
+
"loss": 0.5627,
|
| 44 |
+
"step": 10
|
| 45 |
+
},
|
| 46 |
+
{
|
| 47 |
+
"epoch": 0.12,
|
| 48 |
+
"grad_norm": 1.2979965209960938,
|
| 49 |
+
"learning_rate": 2e-05,
|
| 50 |
+
"loss": 0.3267,
|
| 51 |
+
"step": 12
|
| 52 |
+
},
|
| 53 |
+
{
|
| 54 |
+
"epoch": 0.14,
|
| 55 |
+
"grad_norm": 1.3533953428268433,
|
| 56 |
+
"learning_rate": 2e-05,
|
| 57 |
+
"loss": 0.3217,
|
| 58 |
+
"step": 14
|
| 59 |
+
},
|
| 60 |
+
{
|
| 61 |
+
"epoch": 0.16,
|
| 62 |
+
"grad_norm": 1.877424716949463,
|
| 63 |
+
"learning_rate": 2e-05,
|
| 64 |
+
"loss": 0.2856,
|
| 65 |
+
"step": 16
|
| 66 |
+
},
|
| 67 |
+
{
|
| 68 |
+
"epoch": 0.18,
|
| 69 |
+
"grad_norm": 1.0798438787460327,
|
| 70 |
+
"learning_rate": 2e-05,
|
| 71 |
+
"loss": 0.4367,
|
| 72 |
+
"step": 18
|
| 73 |
+
},
|
| 74 |
+
{
|
| 75 |
+
"epoch": 0.2,
|
| 76 |
+
"grad_norm": 5.245316982269287,
|
| 77 |
+
"learning_rate": 2e-05,
|
| 78 |
+
"loss": 0.5249,
|
| 79 |
+
"step": 20
|
| 80 |
+
},
|
| 81 |
+
{
|
| 82 |
+
"epoch": 0.22,
|
| 83 |
+
"grad_norm": 1.5087791681289673,
|
| 84 |
+
"learning_rate": 2e-05,
|
| 85 |
+
"loss": 0.3091,
|
| 86 |
+
"step": 22
|
| 87 |
+
},
|
| 88 |
+
{
|
| 89 |
+
"epoch": 0.24,
|
| 90 |
+
"grad_norm": 3.263746500015259,
|
| 91 |
+
"learning_rate": 2e-05,
|
| 92 |
+
"loss": 1.0143,
|
| 93 |
+
"step": 24
|
| 94 |
+
},
|
| 95 |
+
{
|
| 96 |
+
"epoch": 0.26,
|
| 97 |
+
"grad_norm": 0.557353138923645,
|
| 98 |
+
"learning_rate": 2e-05,
|
| 99 |
+
"loss": 0.1558,
|
| 100 |
+
"step": 26
|
| 101 |
+
},
|
| 102 |
+
{
|
| 103 |
+
"epoch": 0.28,
|
| 104 |
+
"grad_norm": 1.3562333583831787,
|
| 105 |
+
"learning_rate": 2e-05,
|
| 106 |
+
"loss": 0.55,
|
| 107 |
+
"step": 28
|
| 108 |
+
},
|
| 109 |
+
{
|
| 110 |
+
"epoch": 0.3,
|
| 111 |
+
"grad_norm": 2.261547088623047,
|
| 112 |
+
"learning_rate": 2e-05,
|
| 113 |
+
"loss": 0.8929,
|
| 114 |
+
"step": 30
|
| 115 |
+
},
|
| 116 |
+
{
|
| 117 |
+
"epoch": 0.32,
|
| 118 |
+
"grad_norm": 3.1925289630889893,
|
| 119 |
+
"learning_rate": 2e-05,
|
| 120 |
+
"loss": 0.4653,
|
| 121 |
+
"step": 32
|
| 122 |
+
},
|
| 123 |
+
{
|
| 124 |
+
"epoch": 0.34,
|
| 125 |
+
"grad_norm": 1.4827537536621094,
|
| 126 |
+
"learning_rate": 2e-05,
|
| 127 |
+
"loss": 0.4744,
|
| 128 |
+
"step": 34
|
| 129 |
+
},
|
| 130 |
+
{
|
| 131 |
+
"epoch": 0.36,
|
| 132 |
+
"grad_norm": 0.6158387660980225,
|
| 133 |
+
"learning_rate": 2e-05,
|
| 134 |
+
"loss": 0.1999,
|
| 135 |
+
"step": 36
|
| 136 |
+
},
|
| 137 |
+
{
|
| 138 |
+
"epoch": 0.38,
|
| 139 |
+
"grad_norm": 3.348236560821533,
|
| 140 |
+
"learning_rate": 2e-05,
|
| 141 |
+
"loss": 0.5648,
|
| 142 |
+
"step": 38
|
| 143 |
+
},
|
| 144 |
+
{
|
| 145 |
+
"epoch": 0.4,
|
| 146 |
+
"grad_norm": 0.5702676177024841,
|
| 147 |
+
"learning_rate": 2e-05,
|
| 148 |
+
"loss": 0.5624,
|
| 149 |
+
"step": 40
|
| 150 |
+
},
|
| 151 |
+
{
|
| 152 |
+
"epoch": 0.42,
|
| 153 |
+
"grad_norm": 3.970848321914673,
|
| 154 |
+
"learning_rate": 2e-05,
|
| 155 |
+
"loss": 1.4146,
|
| 156 |
+
"step": 42
|
| 157 |
+
},
|
| 158 |
+
{
|
| 159 |
+
"epoch": 0.44,
|
| 160 |
+
"grad_norm": 5.26162052154541,
|
| 161 |
+
"learning_rate": 2e-05,
|
| 162 |
+
"loss": 1.5653,
|
| 163 |
+
"step": 44
|
| 164 |
+
},
|
| 165 |
+
{
|
| 166 |
+
"epoch": 0.46,
|
| 167 |
+
"grad_norm": 5.289398193359375,
|
| 168 |
+
"learning_rate": 2e-05,
|
| 169 |
+
"loss": 1.6653,
|
| 170 |
+
"step": 46
|
| 171 |
+
},
|
| 172 |
+
{
|
| 173 |
+
"epoch": 0.48,
|
| 174 |
+
"grad_norm": 4.84953498840332,
|
| 175 |
+
"learning_rate": 2e-05,
|
| 176 |
+
"loss": 0.9846,
|
| 177 |
+
"step": 48
|
| 178 |
+
},
|
| 179 |
+
{
|
| 180 |
+
"epoch": 0.5,
|
| 181 |
+
"grad_norm": 4.963735103607178,
|
| 182 |
+
"learning_rate": 2e-05,
|
| 183 |
+
"loss": 1.5258,
|
| 184 |
+
"step": 50
|
| 185 |
+
},
|
| 186 |
+
{
|
| 187 |
+
"epoch": 0.52,
|
| 188 |
+
"grad_norm": 2.125373601913452,
|
| 189 |
+
"learning_rate": 2e-05,
|
| 190 |
+
"loss": 0.5031,
|
| 191 |
+
"step": 52
|
| 192 |
+
},
|
| 193 |
+
{
|
| 194 |
+
"epoch": 0.54,
|
| 195 |
+
"grad_norm": 1.5217068195343018,
|
| 196 |
+
"learning_rate": 2e-05,
|
| 197 |
+
"loss": 0.3672,
|
| 198 |
+
"step": 54
|
| 199 |
+
},
|
| 200 |
+
{
|
| 201 |
+
"epoch": 0.56,
|
| 202 |
+
"grad_norm": 1.5323041677474976,
|
| 203 |
+
"learning_rate": 2e-05,
|
| 204 |
+
"loss": 0.6326,
|
| 205 |
+
"step": 56
|
| 206 |
+
},
|
| 207 |
+
{
|
| 208 |
+
"epoch": 0.58,
|
| 209 |
+
"grad_norm": 1.9375790357589722,
|
| 210 |
+
"learning_rate": 2e-05,
|
| 211 |
+
"loss": 0.236,
|
| 212 |
+
"step": 58
|
| 213 |
+
},
|
| 214 |
+
{
|
| 215 |
+
"epoch": 0.6,
|
| 216 |
+
"grad_norm": 2.5400359630584717,
|
| 217 |
+
"learning_rate": 2e-05,
|
| 218 |
+
"loss": 1.02,
|
| 219 |
+
"step": 60
|
| 220 |
+
},
|
| 221 |
+
{
|
| 222 |
+
"epoch": 0.62,
|
| 223 |
+
"grad_norm": 1.9115395545959473,
|
| 224 |
+
"learning_rate": 2e-05,
|
| 225 |
+
"loss": 0.3253,
|
| 226 |
+
"step": 62
|
| 227 |
+
},
|
| 228 |
+
{
|
| 229 |
+
"epoch": 0.64,
|
| 230 |
+
"grad_norm": 4.119795322418213,
|
| 231 |
+
"learning_rate": 2e-05,
|
| 232 |
+
"loss": 0.7132,
|
| 233 |
+
"step": 64
|
| 234 |
+
},
|
| 235 |
+
{
|
| 236 |
+
"epoch": 0.66,
|
| 237 |
+
"grad_norm": 2.27884840965271,
|
| 238 |
+
"learning_rate": 2e-05,
|
| 239 |
+
"loss": 1.2836,
|
| 240 |
+
"step": 66
|
| 241 |
+
},
|
| 242 |
+
{
|
| 243 |
+
"epoch": 0.68,
|
| 244 |
+
"grad_norm": 2.7405014038085938,
|
| 245 |
+
"learning_rate": 2e-05,
|
| 246 |
+
"loss": 0.3597,
|
| 247 |
+
"step": 68
|
| 248 |
+
},
|
| 249 |
+
{
|
| 250 |
+
"epoch": 0.7,
|
| 251 |
+
"grad_norm": 1.0723375082015991,
|
| 252 |
+
"learning_rate": 2e-05,
|
| 253 |
+
"loss": 1.1792,
|
| 254 |
+
"step": 70
|
| 255 |
+
},
|
| 256 |
+
{
|
| 257 |
+
"epoch": 0.72,
|
| 258 |
+
"grad_norm": 1.9426579475402832,
|
| 259 |
+
"learning_rate": 2e-05,
|
| 260 |
+
"loss": 0.2904,
|
| 261 |
+
"step": 72
|
| 262 |
+
},
|
| 263 |
+
{
|
| 264 |
+
"epoch": 0.74,
|
| 265 |
+
"grad_norm": 3.871702194213867,
|
| 266 |
+
"learning_rate": 2e-05,
|
| 267 |
+
"loss": 1.4492,
|
| 268 |
+
"step": 74
|
| 269 |
+
},
|
| 270 |
+
{
|
| 271 |
+
"epoch": 0.76,
|
| 272 |
+
"grad_norm": 0.5469585657119751,
|
| 273 |
+
"learning_rate": 2e-05,
|
| 274 |
+
"loss": 0.1543,
|
| 275 |
+
"step": 76
|
| 276 |
+
},
|
| 277 |
+
{
|
| 278 |
+
"epoch": 0.78,
|
| 279 |
+
"grad_norm": 5.215576648712158,
|
| 280 |
+
"learning_rate": 2e-05,
|
| 281 |
+
"loss": 1.2247,
|
| 282 |
+
"step": 78
|
| 283 |
+
},
|
| 284 |
+
{
|
| 285 |
+
"epoch": 0.8,
|
| 286 |
+
"grad_norm": 5.197208881378174,
|
| 287 |
+
"learning_rate": 2e-05,
|
| 288 |
+
"loss": 0.7877,
|
| 289 |
+
"step": 80
|
| 290 |
+
},
|
| 291 |
+
{
|
| 292 |
+
"epoch": 0.82,
|
| 293 |
+
"grad_norm": 0.622382402420044,
|
| 294 |
+
"learning_rate": 2e-05,
|
| 295 |
+
"loss": 0.6431,
|
| 296 |
+
"step": 82
|
| 297 |
+
},
|
| 298 |
+
{
|
| 299 |
+
"epoch": 0.84,
|
| 300 |
+
"grad_norm": 1.4372755289077759,
|
| 301 |
+
"learning_rate": 2e-05,
|
| 302 |
+
"loss": 0.543,
|
| 303 |
+
"step": 84
|
| 304 |
+
},
|
| 305 |
+
{
|
| 306 |
+
"epoch": 0.86,
|
| 307 |
+
"grad_norm": 1.7911620140075684,
|
| 308 |
+
"learning_rate": 2e-05,
|
| 309 |
+
"loss": 0.3779,
|
| 310 |
+
"step": 86
|
| 311 |
+
},
|
| 312 |
+
{
|
| 313 |
+
"epoch": 0.88,
|
| 314 |
+
"grad_norm": 1.9083788394927979,
|
| 315 |
+
"learning_rate": 2e-05,
|
| 316 |
+
"loss": 0.315,
|
| 317 |
+
"step": 88
|
| 318 |
+
},
|
| 319 |
+
{
|
| 320 |
+
"epoch": 0.9,
|
| 321 |
+
"grad_norm": 0.12868903577327728,
|
| 322 |
+
"learning_rate": 2e-05,
|
| 323 |
+
"loss": 0.1472,
|
| 324 |
+
"step": 90
|
| 325 |
+
},
|
| 326 |
+
{
|
| 327 |
+
"epoch": 0.92,
|
| 328 |
+
"grad_norm": 0.37657198309898376,
|
| 329 |
+
"learning_rate": 2e-05,
|
| 330 |
+
"loss": 0.344,
|
| 331 |
+
"step": 92
|
| 332 |
+
},
|
| 333 |
+
{
|
| 334 |
+
"epoch": 0.94,
|
| 335 |
+
"grad_norm": 2.369081735610962,
|
| 336 |
+
"learning_rate": 2e-05,
|
| 337 |
+
"loss": 0.9474,
|
| 338 |
+
"step": 94
|
| 339 |
+
},
|
| 340 |
+
{
|
| 341 |
+
"epoch": 0.96,
|
| 342 |
+
"grad_norm": 0.9848467707633972,
|
| 343 |
+
"learning_rate": 2e-05,
|
| 344 |
+
"loss": 0.1981,
|
| 345 |
+
"step": 96
|
| 346 |
+
},
|
| 347 |
+
{
|
| 348 |
+
"epoch": 0.98,
|
| 349 |
+
"grad_norm": 0.4741188883781433,
|
| 350 |
+
"learning_rate": 2e-05,
|
| 351 |
+
"loss": 0.1882,
|
| 352 |
+
"step": 98
|
| 353 |
+
},
|
| 354 |
+
{
|
| 355 |
+
"epoch": 1.0,
|
| 356 |
+
"grad_norm": 4.421746730804443,
|
| 357 |
+
"learning_rate": 2e-05,
|
| 358 |
+
"loss": 1.1065,
|
| 359 |
+
"step": 100
|
| 360 |
+
},
|
| 361 |
+
{
|
| 362 |
+
"epoch": 1.0,
|
| 363 |
+
"step": 100,
|
| 364 |
+
"total_flos": 4911038537400320.0,
|
| 365 |
+
"train_loss": 0.6756564712524414,
|
| 366 |
+
"train_runtime": 161.515,
|
| 367 |
+
"train_samples_per_second": 2.477,
|
| 368 |
+
"train_steps_per_second": 0.619
|
| 369 |
+
}
|
| 370 |
+
],
|
| 371 |
+
"logging_steps": 2,
|
| 372 |
+
"max_steps": 100,
|
| 373 |
+
"num_input_tokens_seen": 0,
|
| 374 |
+
"num_train_epochs": 1,
|
| 375 |
+
"save_steps": 500,
|
| 376 |
+
"stateful_callbacks": {
|
| 377 |
+
"TrainerControl": {
|
| 378 |
+
"args": {
|
| 379 |
+
"should_epoch_stop": false,
|
| 380 |
+
"should_evaluate": false,
|
| 381 |
+
"should_log": false,
|
| 382 |
+
"should_save": false,
|
| 383 |
+
"should_training_stop": false
|
| 384 |
+
},
|
| 385 |
+
"attributes": {}
|
| 386 |
+
}
|
| 387 |
+
},
|
| 388 |
+
"total_flos": 4911038537400320.0,
|
| 389 |
+
"train_batch_size": 1,
|
| 390 |
+
"trial_name": null,
|
| 391 |
+
"trial_params": null
|
| 392 |
+
}
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/3_client_model_round10.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:925e02bab65a8c21b13961b7c1c2504de2854c66f54cb1db5c5db095445d9dfb
|
| 3 |
+
size 180440142
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/3_client_model_round12.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7dfe2459c0ed9a99de2b5da1cc8c1a3d564a7665992fe018017fad9a0eea2b55
|
| 3 |
+
size 180440142
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/3_client_model_round15.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9a21077d8411c069fc73ccf68c83a8476cf01f5314a1c6eba65d534185182233
|
| 3 |
+
size 180440142
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/3_client_model_round17.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f92d68232d50118aa91f6cd58d30ddcbc66f1a3de7a10fb77f8a0c9b3c5bc168
|
| 3 |
+
size 180440142
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/3_client_model_round2.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5f59030679d64456e72c0ef422af79b028f254ee3a6b11793905566e90314704
|
| 3 |
+
size 180439850
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/3_client_model_round20.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:133dc402d125ddaafc27f06bc1e6410077a9202486323d7f05f062dee2fd8afe
|
| 3 |
+
size 180440142
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/3_client_model_round5.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1eb1c71ea609345555a1aa2ed7e222ce23f61a50458542ffa87292b16afd3cda
|
| 3 |
+
size 180439850
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/3_client_model_round7.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a62cd085080e075ce63299e262c2c23ff7573f2d1ac1354c18f23e91751e8647
|
| 3 |
+
size 180439850
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/3_trainer_state.json
ADDED
|
@@ -0,0 +1,392 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"best_metric": null,
|
| 3 |
+
"best_model_checkpoint": null,
|
| 4 |
+
"epoch": 1.0,
|
| 5 |
+
"eval_steps": 500,
|
| 6 |
+
"global_step": 100,
|
| 7 |
+
"is_hyper_param_search": false,
|
| 8 |
+
"is_local_process_zero": true,
|
| 9 |
+
"is_world_process_zero": true,
|
| 10 |
+
"log_history": [
|
| 11 |
+
{
|
| 12 |
+
"epoch": 0.02,
|
| 13 |
+
"grad_norm": 5.710101127624512,
|
| 14 |
+
"learning_rate": 2e-05,
|
| 15 |
+
"loss": 1.4172,
|
| 16 |
+
"step": 2
|
| 17 |
+
},
|
| 18 |
+
{
|
| 19 |
+
"epoch": 0.04,
|
| 20 |
+
"grad_norm": 3.9397523403167725,
|
| 21 |
+
"learning_rate": 2e-05,
|
| 22 |
+
"loss": 0.8797,
|
| 23 |
+
"step": 4
|
| 24 |
+
},
|
| 25 |
+
{
|
| 26 |
+
"epoch": 0.06,
|
| 27 |
+
"grad_norm": 3.1615402698516846,
|
| 28 |
+
"learning_rate": 2e-05,
|
| 29 |
+
"loss": 1.0671,
|
| 30 |
+
"step": 6
|
| 31 |
+
},
|
| 32 |
+
{
|
| 33 |
+
"epoch": 0.08,
|
| 34 |
+
"grad_norm": 13.820159912109375,
|
| 35 |
+
"learning_rate": 2e-05,
|
| 36 |
+
"loss": 3.6283,
|
| 37 |
+
"step": 8
|
| 38 |
+
},
|
| 39 |
+
{
|
| 40 |
+
"epoch": 0.1,
|
| 41 |
+
"grad_norm": 11.3993501663208,
|
| 42 |
+
"learning_rate": 2e-05,
|
| 43 |
+
"loss": 2.2113,
|
| 44 |
+
"step": 10
|
| 45 |
+
},
|
| 46 |
+
{
|
| 47 |
+
"epoch": 0.12,
|
| 48 |
+
"grad_norm": 13.748973846435547,
|
| 49 |
+
"learning_rate": 2e-05,
|
| 50 |
+
"loss": 2.2509,
|
| 51 |
+
"step": 12
|
| 52 |
+
},
|
| 53 |
+
{
|
| 54 |
+
"epoch": 0.14,
|
| 55 |
+
"grad_norm": 10.254955291748047,
|
| 56 |
+
"learning_rate": 2e-05,
|
| 57 |
+
"loss": 1.3235,
|
| 58 |
+
"step": 14
|
| 59 |
+
},
|
| 60 |
+
{
|
| 61 |
+
"epoch": 0.16,
|
| 62 |
+
"grad_norm": 3.2683303356170654,
|
| 63 |
+
"learning_rate": 2e-05,
|
| 64 |
+
"loss": 0.3648,
|
| 65 |
+
"step": 16
|
| 66 |
+
},
|
| 67 |
+
{
|
| 68 |
+
"epoch": 0.18,
|
| 69 |
+
"grad_norm": 7.850968837738037,
|
| 70 |
+
"learning_rate": 2e-05,
|
| 71 |
+
"loss": 1.1944,
|
| 72 |
+
"step": 18
|
| 73 |
+
},
|
| 74 |
+
{
|
| 75 |
+
"epoch": 0.2,
|
| 76 |
+
"grad_norm": 5.817758560180664,
|
| 77 |
+
"learning_rate": 2e-05,
|
| 78 |
+
"loss": 1.5678,
|
| 79 |
+
"step": 20
|
| 80 |
+
},
|
| 81 |
+
{
|
| 82 |
+
"epoch": 0.22,
|
| 83 |
+
"grad_norm": 3.871739625930786,
|
| 84 |
+
"learning_rate": 2e-05,
|
| 85 |
+
"loss": 1.3073,
|
| 86 |
+
"step": 22
|
| 87 |
+
},
|
| 88 |
+
{
|
| 89 |
+
"epoch": 0.24,
|
| 90 |
+
"grad_norm": 8.361922264099121,
|
| 91 |
+
"learning_rate": 2e-05,
|
| 92 |
+
"loss": 0.9962,
|
| 93 |
+
"step": 24
|
| 94 |
+
},
|
| 95 |
+
{
|
| 96 |
+
"epoch": 0.26,
|
| 97 |
+
"grad_norm": 4.391210079193115,
|
| 98 |
+
"learning_rate": 2e-05,
|
| 99 |
+
"loss": 1.3205,
|
| 100 |
+
"step": 26
|
| 101 |
+
},
|
| 102 |
+
{
|
| 103 |
+
"epoch": 0.28,
|
| 104 |
+
"grad_norm": 6.039274215698242,
|
| 105 |
+
"learning_rate": 2e-05,
|
| 106 |
+
"loss": 1.9638,
|
| 107 |
+
"step": 28
|
| 108 |
+
},
|
| 109 |
+
{
|
| 110 |
+
"epoch": 0.3,
|
| 111 |
+
"grad_norm": 2.7958240509033203,
|
| 112 |
+
"learning_rate": 2e-05,
|
| 113 |
+
"loss": 0.46,
|
| 114 |
+
"step": 30
|
| 115 |
+
},
|
| 116 |
+
{
|
| 117 |
+
"epoch": 0.32,
|
| 118 |
+
"grad_norm": 7.391857147216797,
|
| 119 |
+
"learning_rate": 2e-05,
|
| 120 |
+
"loss": 1.1388,
|
| 121 |
+
"step": 32
|
| 122 |
+
},
|
| 123 |
+
{
|
| 124 |
+
"epoch": 0.34,
|
| 125 |
+
"grad_norm": 8.654966354370117,
|
| 126 |
+
"learning_rate": 2e-05,
|
| 127 |
+
"loss": 1.5172,
|
| 128 |
+
"step": 34
|
| 129 |
+
},
|
| 130 |
+
{
|
| 131 |
+
"epoch": 0.36,
|
| 132 |
+
"grad_norm": 10.035390853881836,
|
| 133 |
+
"learning_rate": 2e-05,
|
| 134 |
+
"loss": 1.7336,
|
| 135 |
+
"step": 36
|
| 136 |
+
},
|
| 137 |
+
{
|
| 138 |
+
"epoch": 0.38,
|
| 139 |
+
"grad_norm": 4.6513237953186035,
|
| 140 |
+
"learning_rate": 2e-05,
|
| 141 |
+
"loss": 1.2339,
|
| 142 |
+
"step": 38
|
| 143 |
+
},
|
| 144 |
+
{
|
| 145 |
+
"epoch": 0.4,
|
| 146 |
+
"grad_norm": 2.69884991645813,
|
| 147 |
+
"learning_rate": 2e-05,
|
| 148 |
+
"loss": 0.4367,
|
| 149 |
+
"step": 40
|
| 150 |
+
},
|
| 151 |
+
{
|
| 152 |
+
"epoch": 0.42,
|
| 153 |
+
"grad_norm": 3.5067331790924072,
|
| 154 |
+
"learning_rate": 2e-05,
|
| 155 |
+
"loss": 0.7892,
|
| 156 |
+
"step": 42
|
| 157 |
+
},
|
| 158 |
+
{
|
| 159 |
+
"epoch": 0.44,
|
| 160 |
+
"grad_norm": 8.914396286010742,
|
| 161 |
+
"learning_rate": 2e-05,
|
| 162 |
+
"loss": 1.5515,
|
| 163 |
+
"step": 44
|
| 164 |
+
},
|
| 165 |
+
{
|
| 166 |
+
"epoch": 0.46,
|
| 167 |
+
"grad_norm": 6.865166187286377,
|
| 168 |
+
"learning_rate": 2e-05,
|
| 169 |
+
"loss": 1.689,
|
| 170 |
+
"step": 46
|
| 171 |
+
},
|
| 172 |
+
{
|
| 173 |
+
"epoch": 0.48,
|
| 174 |
+
"grad_norm": 2.9471137523651123,
|
| 175 |
+
"learning_rate": 2e-05,
|
| 176 |
+
"loss": 0.9443,
|
| 177 |
+
"step": 48
|
| 178 |
+
},
|
| 179 |
+
{
|
| 180 |
+
"epoch": 0.5,
|
| 181 |
+
"grad_norm": 2.263576030731201,
|
| 182 |
+
"learning_rate": 2e-05,
|
| 183 |
+
"loss": 0.3431,
|
| 184 |
+
"step": 50
|
| 185 |
+
},
|
| 186 |
+
{
|
| 187 |
+
"epoch": 0.52,
|
| 188 |
+
"grad_norm": 3.667490243911743,
|
| 189 |
+
"learning_rate": 2e-05,
|
| 190 |
+
"loss": 1.6956,
|
| 191 |
+
"step": 52
|
| 192 |
+
},
|
| 193 |
+
{
|
| 194 |
+
"epoch": 0.54,
|
| 195 |
+
"grad_norm": 10.96878433227539,
|
| 196 |
+
"learning_rate": 2e-05,
|
| 197 |
+
"loss": 1.092,
|
| 198 |
+
"step": 54
|
| 199 |
+
},
|
| 200 |
+
{
|
| 201 |
+
"epoch": 0.56,
|
| 202 |
+
"grad_norm": 4.746389865875244,
|
| 203 |
+
"learning_rate": 2e-05,
|
| 204 |
+
"loss": 0.7459,
|
| 205 |
+
"step": 56
|
| 206 |
+
},
|
| 207 |
+
{
|
| 208 |
+
"epoch": 0.58,
|
| 209 |
+
"grad_norm": 6.84367561340332,
|
| 210 |
+
"learning_rate": 2e-05,
|
| 211 |
+
"loss": 0.8944,
|
| 212 |
+
"step": 58
|
| 213 |
+
},
|
| 214 |
+
{
|
| 215 |
+
"epoch": 0.6,
|
| 216 |
+
"grad_norm": 1.7791221141815186,
|
| 217 |
+
"learning_rate": 2e-05,
|
| 218 |
+
"loss": 1.3167,
|
| 219 |
+
"step": 60
|
| 220 |
+
},
|
| 221 |
+
{
|
| 222 |
+
"epoch": 0.62,
|
| 223 |
+
"grad_norm": 4.828078746795654,
|
| 224 |
+
"learning_rate": 2e-05,
|
| 225 |
+
"loss": 1.8402,
|
| 226 |
+
"step": 62
|
| 227 |
+
},
|
| 228 |
+
{
|
| 229 |
+
"epoch": 0.64,
|
| 230 |
+
"grad_norm": 13.462433815002441,
|
| 231 |
+
"learning_rate": 2e-05,
|
| 232 |
+
"loss": 1.1954,
|
| 233 |
+
"step": 64
|
| 234 |
+
},
|
| 235 |
+
{
|
| 236 |
+
"epoch": 0.66,
|
| 237 |
+
"grad_norm": 9.061487197875977,
|
| 238 |
+
"learning_rate": 2e-05,
|
| 239 |
+
"loss": 1.31,
|
| 240 |
+
"step": 66
|
| 241 |
+
},
|
| 242 |
+
{
|
| 243 |
+
"epoch": 0.68,
|
| 244 |
+
"grad_norm": 1.7035795450210571,
|
| 245 |
+
"learning_rate": 2e-05,
|
| 246 |
+
"loss": 0.6877,
|
| 247 |
+
"step": 68
|
| 248 |
+
},
|
| 249 |
+
{
|
| 250 |
+
"epoch": 0.7,
|
| 251 |
+
"grad_norm": 7.510168075561523,
|
| 252 |
+
"learning_rate": 2e-05,
|
| 253 |
+
"loss": 1.5226,
|
| 254 |
+
"step": 70
|
| 255 |
+
},
|
| 256 |
+
{
|
| 257 |
+
"epoch": 0.72,
|
| 258 |
+
"grad_norm": 5.2935638427734375,
|
| 259 |
+
"learning_rate": 2e-05,
|
| 260 |
+
"loss": 1.5661,
|
| 261 |
+
"step": 72
|
| 262 |
+
},
|
| 263 |
+
{
|
| 264 |
+
"epoch": 0.74,
|
| 265 |
+
"grad_norm": 4.344234466552734,
|
| 266 |
+
"learning_rate": 2e-05,
|
| 267 |
+
"loss": 1.6382,
|
| 268 |
+
"step": 74
|
| 269 |
+
},
|
| 270 |
+
{
|
| 271 |
+
"epoch": 0.76,
|
| 272 |
+
"grad_norm": 3.743943691253662,
|
| 273 |
+
"learning_rate": 2e-05,
|
| 274 |
+
"loss": 0.713,
|
| 275 |
+
"step": 76
|
| 276 |
+
},
|
| 277 |
+
{
|
| 278 |
+
"epoch": 0.78,
|
| 279 |
+
"grad_norm": 8.068772315979004,
|
| 280 |
+
"learning_rate": 2e-05,
|
| 281 |
+
"loss": 1.1778,
|
| 282 |
+
"step": 78
|
| 283 |
+
},
|
| 284 |
+
{
|
| 285 |
+
"epoch": 0.8,
|
| 286 |
+
"grad_norm": 3.1857481002807617,
|
| 287 |
+
"learning_rate": 2e-05,
|
| 288 |
+
"loss": 0.9007,
|
| 289 |
+
"step": 80
|
| 290 |
+
},
|
| 291 |
+
{
|
| 292 |
+
"epoch": 0.82,
|
| 293 |
+
"grad_norm": 5.7237138748168945,
|
| 294 |
+
"learning_rate": 2e-05,
|
| 295 |
+
"loss": 2.0638,
|
| 296 |
+
"step": 82
|
| 297 |
+
},
|
| 298 |
+
{
|
| 299 |
+
"epoch": 0.84,
|
| 300 |
+
"grad_norm": 6.149637699127197,
|
| 301 |
+
"learning_rate": 2e-05,
|
| 302 |
+
"loss": 0.9916,
|
| 303 |
+
"step": 84
|
| 304 |
+
},
|
| 305 |
+
{
|
| 306 |
+
"epoch": 0.86,
|
| 307 |
+
"grad_norm": 2.0919957160949707,
|
| 308 |
+
"learning_rate": 2e-05,
|
| 309 |
+
"loss": 0.6464,
|
| 310 |
+
"step": 86
|
| 311 |
+
},
|
| 312 |
+
{
|
| 313 |
+
"epoch": 0.88,
|
| 314 |
+
"grad_norm": 3.3683154582977295,
|
| 315 |
+
"learning_rate": 2e-05,
|
| 316 |
+
"loss": 0.6144,
|
| 317 |
+
"step": 88
|
| 318 |
+
},
|
| 319 |
+
{
|
| 320 |
+
"epoch": 0.9,
|
| 321 |
+
"grad_norm": 4.998460292816162,
|
| 322 |
+
"learning_rate": 2e-05,
|
| 323 |
+
"loss": 0.8671,
|
| 324 |
+
"step": 90
|
| 325 |
+
},
|
| 326 |
+
{
|
| 327 |
+
"epoch": 0.92,
|
| 328 |
+
"grad_norm": 2.805094003677368,
|
| 329 |
+
"learning_rate": 2e-05,
|
| 330 |
+
"loss": 0.6894,
|
| 331 |
+
"step": 92
|
| 332 |
+
},
|
| 333 |
+
{
|
| 334 |
+
"epoch": 0.94,
|
| 335 |
+
"grad_norm": 2.4422361850738525,
|
| 336 |
+
"learning_rate": 2e-05,
|
| 337 |
+
"loss": 0.3568,
|
| 338 |
+
"step": 94
|
| 339 |
+
},
|
| 340 |
+
{
|
| 341 |
+
"epoch": 0.96,
|
| 342 |
+
"grad_norm": 3.961075782775879,
|
| 343 |
+
"learning_rate": 2e-05,
|
| 344 |
+
"loss": 0.8177,
|
| 345 |
+
"step": 96
|
| 346 |
+
},
|
| 347 |
+
{
|
| 348 |
+
"epoch": 0.98,
|
| 349 |
+
"grad_norm": 9.63961124420166,
|
| 350 |
+
"learning_rate": 2e-05,
|
| 351 |
+
"loss": 2.5556,
|
| 352 |
+
"step": 98
|
| 353 |
+
},
|
| 354 |
+
{
|
| 355 |
+
"epoch": 1.0,
|
| 356 |
+
"grad_norm": 2.2371578216552734,
|
| 357 |
+
"learning_rate": 2e-05,
|
| 358 |
+
"loss": 0.4797,
|
| 359 |
+
"step": 100
|
| 360 |
+
},
|
| 361 |
+
{
|
| 362 |
+
"epoch": 1.0,
|
| 363 |
+
"step": 100,
|
| 364 |
+
"total_flos": 2094816317931520.0,
|
| 365 |
+
"train_loss": 1.2341706466674804,
|
| 366 |
+
"train_runtime": 100.7881,
|
| 367 |
+
"train_samples_per_second": 3.969,
|
| 368 |
+
"train_steps_per_second": 0.992
|
| 369 |
+
}
|
| 370 |
+
],
|
| 371 |
+
"logging_steps": 2,
|
| 372 |
+
"max_steps": 100,
|
| 373 |
+
"num_input_tokens_seen": 0,
|
| 374 |
+
"num_train_epochs": 1,
|
| 375 |
+
"save_steps": 500,
|
| 376 |
+
"stateful_callbacks": {
|
| 377 |
+
"TrainerControl": {
|
| 378 |
+
"args": {
|
| 379 |
+
"should_epoch_stop": false,
|
| 380 |
+
"should_evaluate": false,
|
| 381 |
+
"should_log": false,
|
| 382 |
+
"should_save": false,
|
| 383 |
+
"should_training_stop": false
|
| 384 |
+
},
|
| 385 |
+
"attributes": {}
|
| 386 |
+
}
|
| 387 |
+
},
|
| 388 |
+
"total_flos": 2094816317931520.0,
|
| 389 |
+
"train_batch_size": 1,
|
| 390 |
+
"trial_name": null,
|
| 391 |
+
"trial_params": null
|
| 392 |
+
}
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/4_client_model_round10.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:20b7d2e033a4b3d1dd6203f1de7275223b49ea702986d1bcc4bc023f61efac5c
|
| 3 |
+
size 389170582
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/4_client_model_round12.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:448423352706cc1f24e4a54a1813344e0787aa887633b802cf0e89e39d7b71b4
|
| 3 |
+
size 389170582
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/4_client_model_round15.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fb4840eadbd7ef1db67e9ceca295cc87f0fe1a605516db0cb6646e53aada077a
|
| 3 |
+
size 389170582
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/4_client_model_round17.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:22187ed7cdb78ce527269fc7896018eb44dd8bd197e63cb13b6dd630c05aac58
|
| 3 |
+
size 389170582
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/4_client_model_round2.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c5a3056867ad1c4ca79975c0c9bace7ffb0bbb72c2de8042a61fdba76ee6983f
|
| 3 |
+
size 389170122
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/4_client_model_round20.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:491ed5be2cd16c68c92a753673103ac5a7d180f35c90e218b096fcd9390e8090
|
| 3 |
+
size 389170582
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/4_client_model_round5.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5a8b283879be4d8ee9c79c7127666354510b854e5227bb598eb78c61238b9f4b
|
| 3 |
+
size 389170122
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/4_client_model_round7.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:000d4f0e9c4b4d86561759eb13ba6397ea829bb5be2c694c51bacc84bfcb746f
|
| 3 |
+
size 389170122
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/4_trainer_state.json
ADDED
|
@@ -0,0 +1,392 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"best_metric": null,
|
| 3 |
+
"best_model_checkpoint": null,
|
| 4 |
+
"epoch": 1.0,
|
| 5 |
+
"eval_steps": 500,
|
| 6 |
+
"global_step": 100,
|
| 7 |
+
"is_hyper_param_search": false,
|
| 8 |
+
"is_local_process_zero": true,
|
| 9 |
+
"is_world_process_zero": true,
|
| 10 |
+
"log_history": [
|
| 11 |
+
{
|
| 12 |
+
"epoch": 0.02,
|
| 13 |
+
"grad_norm": 2.1811344623565674,
|
| 14 |
+
"learning_rate": 2e-05,
|
| 15 |
+
"loss": 0.9004,
|
| 16 |
+
"step": 2
|
| 17 |
+
},
|
| 18 |
+
{
|
| 19 |
+
"epoch": 0.04,
|
| 20 |
+
"grad_norm": 4.057499885559082,
|
| 21 |
+
"learning_rate": 2e-05,
|
| 22 |
+
"loss": 0.4477,
|
| 23 |
+
"step": 4
|
| 24 |
+
},
|
| 25 |
+
{
|
| 26 |
+
"epoch": 0.06,
|
| 27 |
+
"grad_norm": 0.4118136167526245,
|
| 28 |
+
"learning_rate": 2e-05,
|
| 29 |
+
"loss": 0.4494,
|
| 30 |
+
"step": 6
|
| 31 |
+
},
|
| 32 |
+
{
|
| 33 |
+
"epoch": 0.08,
|
| 34 |
+
"grad_norm": 1.3802590370178223,
|
| 35 |
+
"learning_rate": 2e-05,
|
| 36 |
+
"loss": 0.5171,
|
| 37 |
+
"step": 8
|
| 38 |
+
},
|
| 39 |
+
{
|
| 40 |
+
"epoch": 0.1,
|
| 41 |
+
"grad_norm": 3.8723020553588867,
|
| 42 |
+
"learning_rate": 2e-05,
|
| 43 |
+
"loss": 0.2846,
|
| 44 |
+
"step": 10
|
| 45 |
+
},
|
| 46 |
+
{
|
| 47 |
+
"epoch": 0.12,
|
| 48 |
+
"grad_norm": 1.808058738708496,
|
| 49 |
+
"learning_rate": 2e-05,
|
| 50 |
+
"loss": 0.7729,
|
| 51 |
+
"step": 12
|
| 52 |
+
},
|
| 53 |
+
{
|
| 54 |
+
"epoch": 0.14,
|
| 55 |
+
"grad_norm": 0.9369915127754211,
|
| 56 |
+
"learning_rate": 2e-05,
|
| 57 |
+
"loss": 0.4847,
|
| 58 |
+
"step": 14
|
| 59 |
+
},
|
| 60 |
+
{
|
| 61 |
+
"epoch": 0.16,
|
| 62 |
+
"grad_norm": 1.7796673774719238,
|
| 63 |
+
"learning_rate": 2e-05,
|
| 64 |
+
"loss": 0.7306,
|
| 65 |
+
"step": 16
|
| 66 |
+
},
|
| 67 |
+
{
|
| 68 |
+
"epoch": 0.18,
|
| 69 |
+
"grad_norm": 3.8881521224975586,
|
| 70 |
+
"learning_rate": 2e-05,
|
| 71 |
+
"loss": 0.7441,
|
| 72 |
+
"step": 18
|
| 73 |
+
},
|
| 74 |
+
{
|
| 75 |
+
"epoch": 0.2,
|
| 76 |
+
"grad_norm": 1.4085545539855957,
|
| 77 |
+
"learning_rate": 2e-05,
|
| 78 |
+
"loss": 0.6283,
|
| 79 |
+
"step": 20
|
| 80 |
+
},
|
| 81 |
+
{
|
| 82 |
+
"epoch": 0.22,
|
| 83 |
+
"grad_norm": 3.587538719177246,
|
| 84 |
+
"learning_rate": 2e-05,
|
| 85 |
+
"loss": 1.1968,
|
| 86 |
+
"step": 22
|
| 87 |
+
},
|
| 88 |
+
{
|
| 89 |
+
"epoch": 0.24,
|
| 90 |
+
"grad_norm": 1.6639175415039062,
|
| 91 |
+
"learning_rate": 2e-05,
|
| 92 |
+
"loss": 0.7579,
|
| 93 |
+
"step": 24
|
| 94 |
+
},
|
| 95 |
+
{
|
| 96 |
+
"epoch": 0.26,
|
| 97 |
+
"grad_norm": 2.090292453765869,
|
| 98 |
+
"learning_rate": 2e-05,
|
| 99 |
+
"loss": 1.3047,
|
| 100 |
+
"step": 26
|
| 101 |
+
},
|
| 102 |
+
{
|
| 103 |
+
"epoch": 0.28,
|
| 104 |
+
"grad_norm": 2.301203966140747,
|
| 105 |
+
"learning_rate": 2e-05,
|
| 106 |
+
"loss": 0.2738,
|
| 107 |
+
"step": 28
|
| 108 |
+
},
|
| 109 |
+
{
|
| 110 |
+
"epoch": 0.3,
|
| 111 |
+
"grad_norm": 3.2021005153656006,
|
| 112 |
+
"learning_rate": 2e-05,
|
| 113 |
+
"loss": 0.3979,
|
| 114 |
+
"step": 30
|
| 115 |
+
},
|
| 116 |
+
{
|
| 117 |
+
"epoch": 0.32,
|
| 118 |
+
"grad_norm": 6.288852691650391,
|
| 119 |
+
"learning_rate": 2e-05,
|
| 120 |
+
"loss": 1.585,
|
| 121 |
+
"step": 32
|
| 122 |
+
},
|
| 123 |
+
{
|
| 124 |
+
"epoch": 0.34,
|
| 125 |
+
"grad_norm": 4.023402214050293,
|
| 126 |
+
"learning_rate": 2e-05,
|
| 127 |
+
"loss": 1.1474,
|
| 128 |
+
"step": 34
|
| 129 |
+
},
|
| 130 |
+
{
|
| 131 |
+
"epoch": 0.36,
|
| 132 |
+
"grad_norm": 2.4434189796447754,
|
| 133 |
+
"learning_rate": 2e-05,
|
| 134 |
+
"loss": 0.2991,
|
| 135 |
+
"step": 36
|
| 136 |
+
},
|
| 137 |
+
{
|
| 138 |
+
"epoch": 0.38,
|
| 139 |
+
"grad_norm": 1.773408055305481,
|
| 140 |
+
"learning_rate": 2e-05,
|
| 141 |
+
"loss": 0.6277,
|
| 142 |
+
"step": 38
|
| 143 |
+
},
|
| 144 |
+
{
|
| 145 |
+
"epoch": 0.4,
|
| 146 |
+
"grad_norm": 2.1902174949645996,
|
| 147 |
+
"learning_rate": 2e-05,
|
| 148 |
+
"loss": 0.5112,
|
| 149 |
+
"step": 40
|
| 150 |
+
},
|
| 151 |
+
{
|
| 152 |
+
"epoch": 0.42,
|
| 153 |
+
"grad_norm": 0.60333251953125,
|
| 154 |
+
"learning_rate": 2e-05,
|
| 155 |
+
"loss": 0.2128,
|
| 156 |
+
"step": 42
|
| 157 |
+
},
|
| 158 |
+
{
|
| 159 |
+
"epoch": 0.44,
|
| 160 |
+
"grad_norm": 4.51239013671875,
|
| 161 |
+
"learning_rate": 2e-05,
|
| 162 |
+
"loss": 1.3767,
|
| 163 |
+
"step": 44
|
| 164 |
+
},
|
| 165 |
+
{
|
| 166 |
+
"epoch": 0.46,
|
| 167 |
+
"grad_norm": 0.47873491048812866,
|
| 168 |
+
"learning_rate": 2e-05,
|
| 169 |
+
"loss": 0.1478,
|
| 170 |
+
"step": 46
|
| 171 |
+
},
|
| 172 |
+
{
|
| 173 |
+
"epoch": 0.48,
|
| 174 |
+
"grad_norm": 2.3050646781921387,
|
| 175 |
+
"learning_rate": 2e-05,
|
| 176 |
+
"loss": 0.4301,
|
| 177 |
+
"step": 48
|
| 178 |
+
},
|
| 179 |
+
{
|
| 180 |
+
"epoch": 0.5,
|
| 181 |
+
"grad_norm": 1.4680174589157104,
|
| 182 |
+
"learning_rate": 2e-05,
|
| 183 |
+
"loss": 0.4868,
|
| 184 |
+
"step": 50
|
| 185 |
+
},
|
| 186 |
+
{
|
| 187 |
+
"epoch": 0.52,
|
| 188 |
+
"grad_norm": 5.060284614562988,
|
| 189 |
+
"learning_rate": 2e-05,
|
| 190 |
+
"loss": 1.4766,
|
| 191 |
+
"step": 52
|
| 192 |
+
},
|
| 193 |
+
{
|
| 194 |
+
"epoch": 0.54,
|
| 195 |
+
"grad_norm": 2.9290950298309326,
|
| 196 |
+
"learning_rate": 2e-05,
|
| 197 |
+
"loss": 0.521,
|
| 198 |
+
"step": 54
|
| 199 |
+
},
|
| 200 |
+
{
|
| 201 |
+
"epoch": 0.56,
|
| 202 |
+
"grad_norm": 3.033177614212036,
|
| 203 |
+
"learning_rate": 2e-05,
|
| 204 |
+
"loss": 0.8623,
|
| 205 |
+
"step": 56
|
| 206 |
+
},
|
| 207 |
+
{
|
| 208 |
+
"epoch": 0.58,
|
| 209 |
+
"grad_norm": 1.6711596250534058,
|
| 210 |
+
"learning_rate": 2e-05,
|
| 211 |
+
"loss": 0.6449,
|
| 212 |
+
"step": 58
|
| 213 |
+
},
|
| 214 |
+
{
|
| 215 |
+
"epoch": 0.6,
|
| 216 |
+
"grad_norm": 2.6369800567626953,
|
| 217 |
+
"learning_rate": 2e-05,
|
| 218 |
+
"loss": 0.7327,
|
| 219 |
+
"step": 60
|
| 220 |
+
},
|
| 221 |
+
{
|
| 222 |
+
"epoch": 0.62,
|
| 223 |
+
"grad_norm": 3.0031280517578125,
|
| 224 |
+
"learning_rate": 2e-05,
|
| 225 |
+
"loss": 0.5193,
|
| 226 |
+
"step": 62
|
| 227 |
+
},
|
| 228 |
+
{
|
| 229 |
+
"epoch": 0.64,
|
| 230 |
+
"grad_norm": 3.3499245643615723,
|
| 231 |
+
"learning_rate": 2e-05,
|
| 232 |
+
"loss": 0.4866,
|
| 233 |
+
"step": 64
|
| 234 |
+
},
|
| 235 |
+
{
|
| 236 |
+
"epoch": 0.66,
|
| 237 |
+
"grad_norm": 3.0874881744384766,
|
| 238 |
+
"learning_rate": 2e-05,
|
| 239 |
+
"loss": 0.8739,
|
| 240 |
+
"step": 66
|
| 241 |
+
},
|
| 242 |
+
{
|
| 243 |
+
"epoch": 0.68,
|
| 244 |
+
"grad_norm": 2.215085506439209,
|
| 245 |
+
"learning_rate": 2e-05,
|
| 246 |
+
"loss": 1.1589,
|
| 247 |
+
"step": 68
|
| 248 |
+
},
|
| 249 |
+
{
|
| 250 |
+
"epoch": 0.7,
|
| 251 |
+
"grad_norm": 1.7898951768875122,
|
| 252 |
+
"learning_rate": 2e-05,
|
| 253 |
+
"loss": 0.2845,
|
| 254 |
+
"step": 70
|
| 255 |
+
},
|
| 256 |
+
{
|
| 257 |
+
"epoch": 0.72,
|
| 258 |
+
"grad_norm": 2.463358163833618,
|
| 259 |
+
"learning_rate": 2e-05,
|
| 260 |
+
"loss": 0.3765,
|
| 261 |
+
"step": 72
|
| 262 |
+
},
|
| 263 |
+
{
|
| 264 |
+
"epoch": 0.74,
|
| 265 |
+
"grad_norm": 0.7925405502319336,
|
| 266 |
+
"learning_rate": 2e-05,
|
| 267 |
+
"loss": 0.3446,
|
| 268 |
+
"step": 74
|
| 269 |
+
},
|
| 270 |
+
{
|
| 271 |
+
"epoch": 0.76,
|
| 272 |
+
"grad_norm": 1.6757383346557617,
|
| 273 |
+
"learning_rate": 2e-05,
|
| 274 |
+
"loss": 0.7917,
|
| 275 |
+
"step": 76
|
| 276 |
+
},
|
| 277 |
+
{
|
| 278 |
+
"epoch": 0.78,
|
| 279 |
+
"grad_norm": 1.5198004245758057,
|
| 280 |
+
"learning_rate": 2e-05,
|
| 281 |
+
"loss": 0.3082,
|
| 282 |
+
"step": 78
|
| 283 |
+
},
|
| 284 |
+
{
|
| 285 |
+
"epoch": 0.8,
|
| 286 |
+
"grad_norm": 2.186577081680298,
|
| 287 |
+
"learning_rate": 2e-05,
|
| 288 |
+
"loss": 0.6401,
|
| 289 |
+
"step": 80
|
| 290 |
+
},
|
| 291 |
+
{
|
| 292 |
+
"epoch": 0.82,
|
| 293 |
+
"grad_norm": 0.8060414791107178,
|
| 294 |
+
"learning_rate": 2e-05,
|
| 295 |
+
"loss": 0.0603,
|
| 296 |
+
"step": 82
|
| 297 |
+
},
|
| 298 |
+
{
|
| 299 |
+
"epoch": 0.84,
|
| 300 |
+
"grad_norm": 2.0299384593963623,
|
| 301 |
+
"learning_rate": 2e-05,
|
| 302 |
+
"loss": 0.4283,
|
| 303 |
+
"step": 84
|
| 304 |
+
},
|
| 305 |
+
{
|
| 306 |
+
"epoch": 0.86,
|
| 307 |
+
"grad_norm": 1.539688229560852,
|
| 308 |
+
"learning_rate": 2e-05,
|
| 309 |
+
"loss": 0.2565,
|
| 310 |
+
"step": 86
|
| 311 |
+
},
|
| 312 |
+
{
|
| 313 |
+
"epoch": 0.88,
|
| 314 |
+
"grad_norm": 1.5283581018447876,
|
| 315 |
+
"learning_rate": 2e-05,
|
| 316 |
+
"loss": 0.1788,
|
| 317 |
+
"step": 88
|
| 318 |
+
},
|
| 319 |
+
{
|
| 320 |
+
"epoch": 0.9,
|
| 321 |
+
"grad_norm": 0.6798237562179565,
|
| 322 |
+
"learning_rate": 2e-05,
|
| 323 |
+
"loss": 0.7287,
|
| 324 |
+
"step": 90
|
| 325 |
+
},
|
| 326 |
+
{
|
| 327 |
+
"epoch": 0.92,
|
| 328 |
+
"grad_norm": 3.418360471725464,
|
| 329 |
+
"learning_rate": 2e-05,
|
| 330 |
+
"loss": 0.3788,
|
| 331 |
+
"step": 92
|
| 332 |
+
},
|
| 333 |
+
{
|
| 334 |
+
"epoch": 0.94,
|
| 335 |
+
"grad_norm": 0.48986631631851196,
|
| 336 |
+
"learning_rate": 2e-05,
|
| 337 |
+
"loss": 2.0909,
|
| 338 |
+
"step": 94
|
| 339 |
+
},
|
| 340 |
+
{
|
| 341 |
+
"epoch": 0.96,
|
| 342 |
+
"grad_norm": 0.2811102271080017,
|
| 343 |
+
"learning_rate": 2e-05,
|
| 344 |
+
"loss": 0.0683,
|
| 345 |
+
"step": 96
|
| 346 |
+
},
|
| 347 |
+
{
|
| 348 |
+
"epoch": 0.98,
|
| 349 |
+
"grad_norm": 1.9367895126342773,
|
| 350 |
+
"learning_rate": 2e-05,
|
| 351 |
+
"loss": 0.4261,
|
| 352 |
+
"step": 98
|
| 353 |
+
},
|
| 354 |
+
{
|
| 355 |
+
"epoch": 1.0,
|
| 356 |
+
"grad_norm": 2.887495279312134,
|
| 357 |
+
"learning_rate": 2e-05,
|
| 358 |
+
"loss": 0.4067,
|
| 359 |
+
"step": 100
|
| 360 |
+
},
|
| 361 |
+
{
|
| 362 |
+
"epoch": 1.0,
|
| 363 |
+
"step": 100,
|
| 364 |
+
"total_flos": 5689210367377408.0,
|
| 365 |
+
"train_loss": 0.6352152729034424,
|
| 366 |
+
"train_runtime": 162.5562,
|
| 367 |
+
"train_samples_per_second": 2.461,
|
| 368 |
+
"train_steps_per_second": 0.615
|
| 369 |
+
}
|
| 370 |
+
],
|
| 371 |
+
"logging_steps": 2,
|
| 372 |
+
"max_steps": 100,
|
| 373 |
+
"num_input_tokens_seen": 0,
|
| 374 |
+
"num_train_epochs": 1,
|
| 375 |
+
"save_steps": 500,
|
| 376 |
+
"stateful_callbacks": {
|
| 377 |
+
"TrainerControl": {
|
| 378 |
+
"args": {
|
| 379 |
+
"should_epoch_stop": false,
|
| 380 |
+
"should_evaluate": false,
|
| 381 |
+
"should_log": false,
|
| 382 |
+
"should_save": false,
|
| 383 |
+
"should_training_stop": false
|
| 384 |
+
},
|
| 385 |
+
"attributes": {}
|
| 386 |
+
}
|
| 387 |
+
},
|
| 388 |
+
"total_flos": 5689210367377408.0,
|
| 389 |
+
"train_batch_size": 1,
|
| 390 |
+
"trial_name": null,
|
| 391 |
+
"trial_params": null
|
| 392 |
+
}
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/5_client_model_round10.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:07f03b0ba909c845d5b5b49c2c076935f837f2786c99faff543722d4798bbed1
|
| 3 |
+
size 389170582
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/5_client_model_round12.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:71e00c2c325bfc7ee21add5d449c752326a25edbf3d18251963cb643297bb9ae
|
| 3 |
+
size 389170582
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/5_client_model_round15.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5fbba7266a66e0096e2065265be474ac545ee8600fc54ed14af585a8031e9726
|
| 3 |
+
size 389170582
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/5_client_model_round17.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6756fbbfec598b27becfb99f39fdd78b7398d2e636031e0b73a8e62aa614ad2b
|
| 3 |
+
size 389170582
|
client_states_fedavg_hetero_NOCONT_bs4_saveoptim_lr2e-5_5e-5_sc315_4tasks_5rounds_fixitr100_T0125_decay099/5_client_model_round2.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:cdc2aa4f97bd4e09072b338e21a46703e6b339f7d3a0baa7ac9633d09088b73b
|
| 3 |
+
size 389170122
|