Upload folder using huggingface_hub
Browse files- pretrained_model/config.json +1 -1
- pretrained_model/model.safetensors +1 -1
- pretrained_model/train_config.json +4 -4
- training_state/optimizer_param_groups.json +1 -1
- training_state/optimizer_state.safetensors +1 -1
- training_state/rng_state.safetensors +1 -1
- training_state/scheduler_state.json +3 -3
- training_state/training_step.json +1 -1
pretrained_model/config.json
CHANGED
|
@@ -80,7 +80,7 @@
|
|
| 80 |
],
|
| 81 |
"optimizer_eps": 1e-08,
|
| 82 |
"optimizer_weight_decay": 1e-10,
|
| 83 |
-
"optimizer_grad_clip_norm": 10,
|
| 84 |
"scheduler_warmup_steps": 1000,
|
| 85 |
"scheduler_decay_steps": 30000,
|
| 86 |
"scheduler_decay_lr": 2.5e-06,
|
|
|
|
| 80 |
],
|
| 81 |
"optimizer_eps": 1e-08,
|
| 82 |
"optimizer_weight_decay": 1e-10,
|
| 83 |
+
"optimizer_grad_clip_norm": 10.0,
|
| 84 |
"scheduler_warmup_steps": 1000,
|
| 85 |
"scheduler_decay_steps": 30000,
|
| 86 |
"scheduler_decay_lr": 2.5e-06,
|
pretrained_model/model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1197790216
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8d1f59797019d099af5f774fcf28738bb433a838e310d22dc378f61121d76f4b
|
| 3 |
size 1197790216
|
pretrained_model/train_config.json
CHANGED
|
@@ -147,7 +147,7 @@
|
|
| 147 |
],
|
| 148 |
"optimizer_eps": 1e-08,
|
| 149 |
"optimizer_weight_decay": 1e-10,
|
| 150 |
-
"optimizer_grad_clip_norm": 10,
|
| 151 |
"scheduler_warmup_steps": 1000,
|
| 152 |
"scheduler_decay_steps": 30000,
|
| 153 |
"scheduler_decay_lr": 2.5e-06,
|
|
@@ -166,7 +166,7 @@
|
|
| 166 |
},
|
| 167 |
"output_dir": "outputs/train/smolvla_fold_cloth",
|
| 168 |
"job_name": "act_so101_test",
|
| 169 |
-
"resume":
|
| 170 |
"seed": 1000,
|
| 171 |
"num_workers": 4,
|
| 172 |
"batch_size": 8,
|
|
@@ -174,13 +174,13 @@
|
|
| 174 |
"eval_freq": 20000,
|
| 175 |
"log_freq": 200,
|
| 176 |
"save_checkpoint": true,
|
| 177 |
-
"save_freq":
|
| 178 |
"use_policy_training_preset": true,
|
| 179 |
"optimizer": {
|
| 180 |
"type": "adamw",
|
| 181 |
"lr": 0.0001,
|
| 182 |
"weight_decay": 1e-10,
|
| 183 |
-
"grad_clip_norm": 10,
|
| 184 |
"betas": [
|
| 185 |
0.9,
|
| 186 |
0.95
|
|
|
|
| 147 |
],
|
| 148 |
"optimizer_eps": 1e-08,
|
| 149 |
"optimizer_weight_decay": 1e-10,
|
| 150 |
+
"optimizer_grad_clip_norm": 10.0,
|
| 151 |
"scheduler_warmup_steps": 1000,
|
| 152 |
"scheduler_decay_steps": 30000,
|
| 153 |
"scheduler_decay_lr": 2.5e-06,
|
|
|
|
| 166 |
},
|
| 167 |
"output_dir": "outputs/train/smolvla_fold_cloth",
|
| 168 |
"job_name": "act_so101_test",
|
| 169 |
+
"resume": true,
|
| 170 |
"seed": 1000,
|
| 171 |
"num_workers": 4,
|
| 172 |
"batch_size": 8,
|
|
|
|
| 174 |
"eval_freq": 20000,
|
| 175 |
"log_freq": 200,
|
| 176 |
"save_checkpoint": true,
|
| 177 |
+
"save_freq": 50000,
|
| 178 |
"use_policy_training_preset": true,
|
| 179 |
"optimizer": {
|
| 180 |
"type": "adamw",
|
| 181 |
"lr": 0.0001,
|
| 182 |
"weight_decay": 1e-10,
|
| 183 |
+
"grad_clip_norm": 10.0,
|
| 184 |
"betas": [
|
| 185 |
0.9,
|
| 186 |
0.95
|
training_state/optimizer_param_groups.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
| 1 |
[
|
| 2 |
{
|
| 3 |
-
"lr":
|
| 4 |
"betas": [
|
| 5 |
0.9,
|
| 6 |
0.95
|
|
|
|
| 1 |
[
|
| 2 |
{
|
| 3 |
+
"lr": 2.5e-06,
|
| 4 |
"betas": [
|
| 5 |
0.9,
|
| 6 |
0.95
|
training_state/optimizer_state.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 412659164
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5fcb672bbd44c271104c97ae9270686ba2ed249477d1825466f3a29f0a3cb056
|
| 3 |
size 412659164
|
training_state/rng_state.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 15708
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dccdadddc3d5ccb003e120c4e84836d0efcfbb753c1a80375cf0cc4494252618
|
| 3 |
size 15708
|
training_state/scheduler_state.json
CHANGED
|
@@ -2,11 +2,11 @@
|
|
| 2 |
"base_lrs": [
|
| 3 |
0.0001
|
| 4 |
],
|
| 5 |
-
"last_epoch":
|
| 6 |
-
"_step_count":
|
| 7 |
"_get_lr_called_within_step": false,
|
| 8 |
"_last_lr": [
|
| 9 |
-
|
| 10 |
],
|
| 11 |
"lr_lambdas": [
|
| 12 |
null
|
|
|
|
| 2 |
"base_lrs": [
|
| 3 |
0.0001
|
| 4 |
],
|
| 5 |
+
"last_epoch": 100000,
|
| 6 |
+
"_step_count": 100001,
|
| 7 |
"_get_lr_called_within_step": false,
|
| 8 |
"_last_lr": [
|
| 9 |
+
2.5e-06
|
| 10 |
],
|
| 11 |
"lr_lambdas": [
|
| 12 |
null
|
training_state/training_step.json
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
{
|
| 2 |
-
"step":
|
| 3 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"step": 100000
|
| 3 |
}
|