Training Step 945
Browse files- model.safetensors +1 -1
- trainer_state.pt +1 -1
- training_config.json +1 -1
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 68876952
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8ae60708378fb8c0aac28a98aa9a483b6b52cc9a1fae88be1e8ba9dd2be77248
|
| 3 |
size 68876952
|
trainer_state.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 129403747
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:38b6a24480759e2dc0c5204bdb0565079a141a499c54587c34c33d2b50e0efe9
|
| 3 |
size 129403747
|
training_config.json
CHANGED
|
@@ -66,7 +66,7 @@
|
|
| 66 |
"hf_api": "<huggingface_hub.hf_api.HfApi object at 0x14e5868dd3a0>",
|
| 67 |
"wandb_writer": "<wandb.sdk.wandb_run.Run object at 0x14e57f3612e0>",
|
| 68 |
"wandb_table": null,
|
| 69 |
-
"optimizer": "AdamW (\nParameter Group 0\n amsgrad: False\n betas: (0.9, 0.999)\n capturable: False\n decoupled_weight_decay: True\n differentiable: False\n eps: 1e-08\n foreach: None\n fused: None\n initial_lr: 0.000125\n lr: 2.
|
| 70 |
"lr_scheduler": "<torch.optim.lr_scheduler.SequentialLR object at 0x14e586a1dca0>",
|
| 71 |
"wd_scheduler": "<trainers.schedulers.weight_decay.WeightDecayScheduler object at 0x14e57e241820>",
|
| 72 |
"momentum_scheduler": "<trainers.schedulers.momentum.MomentumScheduler object at 0x14e57e241850>",
|
|
|
|
| 66 |
"hf_api": "<huggingface_hub.hf_api.HfApi object at 0x14e5868dd3a0>",
|
| 67 |
"wandb_writer": "<wandb.sdk.wandb_run.Run object at 0x14e57f3612e0>",
|
| 68 |
"wandb_table": null,
|
| 69 |
+
"optimizer": "AdamW (\nParameter Group 0\n amsgrad: False\n betas: (0.9, 0.999)\n capturable: False\n decoupled_weight_decay: True\n differentiable: False\n eps: 1e-08\n foreach: None\n fused: None\n initial_lr: 0.000125\n lr: 2.7469758064516153e-06\n maximize: False\n weight_decay: 0.040001297063542884\n)",
|
| 70 |
"lr_scheduler": "<torch.optim.lr_scheduler.SequentialLR object at 0x14e586a1dca0>",
|
| 71 |
"wd_scheduler": "<trainers.schedulers.weight_decay.WeightDecayScheduler object at 0x14e57e241820>",
|
| 72 |
"momentum_scheduler": "<trainers.schedulers.momentum.MomentumScheduler object at 0x14e57e241850>",
|