Training Step 1890
Browse files- model.safetensors +1 -1
- trainer_state.pt +1 -1
- training_config.json +1 -1
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 124660456
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:122d00eaf24ff70417426bb0511ba131fca6ec7347a2d0acdfca9ba2279a5a78
|
| 3 |
size 124660456
|
trainer_state.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 215673699
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ecfd3e8045f9392f4cc785105af3e84e0c2ae3b7a0e95cc8d783e54ff51d0e89
|
| 3 |
size 215673699
|
training_config.json
CHANGED
|
@@ -61,7 +61,7 @@
|
|
| 61 |
"hf_api": "<huggingface_hub.hf_api.HfApi object at 0x14e641e1e600>",
|
| 62 |
"wandb_writer": "<wandb.sdk.wandb_run.Run object at 0x14e642ade690>",
|
| 63 |
"wandb_table": null,
|
| 64 |
-
"optimizer": "AdamW (\nParameter Group 0\n amsgrad: False\n betas: (0.9, 0.999)\n capturable: False\n decoupled_weight_decay: True\n differentiable: False\n eps: 1e-08\n foreach: None\n fused: None\n initial_lr: 0.0002\n lr:
|
| 65 |
"lr_scheduler": "<torch.optim.lr_scheduler.SequentialLR object at 0x14e642129e50>",
|
| 66 |
"wd_scheduler": "<trainers.schedulers.weight_decay.WeightDecayScheduler object at 0x14e642129f40>",
|
| 67 |
"momentum_scheduler": "<trainers.schedulers.momentum.MomentumScheduler object at 0x14e642129e80>",
|
|
|
|
| 61 |
"hf_api": "<huggingface_hub.hf_api.HfApi object at 0x14e641e1e600>",
|
| 62 |
"wandb_writer": "<wandb.sdk.wandb_run.Run object at 0x14e642ade690>",
|
| 63 |
"wandb_table": null,
|
| 64 |
+
"optimizer": "AdamW (\nParameter Group 0\n amsgrad: False\n betas: (0.9, 0.999)\n capturable: False\n decoupled_weight_decay: True\n differentiable: False\n eps: 1e-08\n foreach: None\n fused: None\n initial_lr: 0.0002\n lr: 2.116129032258056e-05\n maximize: False\n weight_decay: 0.04008309373244129\n)",
|
| 65 |
"lr_scheduler": "<torch.optim.lr_scheduler.SequentialLR object at 0x14e642129e50>",
|
| 66 |
"wd_scheduler": "<trainers.schedulers.weight_decay.WeightDecayScheduler object at 0x14e642129f40>",
|
| 67 |
"momentum_scheduler": "<trainers.schedulers.momentum.MomentumScheduler object at 0x14e642129e80>",
|