Upload folder using huggingface_hub
Browse files- mld_miko_finetune_v5/args.yaml +24 -0
- mld_miko_finetune_v5/args_read.yaml +74 -0
- mld_miko_finetune_v5/checkpoint_305000.pt +3 -0
- mld_miko_finetune_v5/checkpoint_310000.pt +3 -0
- mld_miko_finetune_v5/checkpoint_315000.pt +3 -0
- mld_miko_finetune_v5/checkpoint_320000.pt +3 -0
- mld_miko_finetune_v5/checkpoint_325000.pt +3 -0
- mld_miko_finetune_v5/checkpoint_330000.pt +3 -0
- mld_miko_finetune_v5/checkpoint_335000.pt +3 -0
- mld_miko_finetune_v5/checkpoint_340000.pt +3 -0
mld_miko_finetune_v5/args.yaml
ADDED
|
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"# tyro YAML.\n!dataclass:MLDArgs\ndata_args: !dataclass:DataArgs\n body_type: smplx\n\
|
| 2 |
+
\ cfg_path: ./config_files/config_hydra/motion_primitive/mp_h2_f8_r4.yaml\n data_dir:\
|
| 3 |
+
\ ./data/mixed_filtered_seq_data_zero_male\n dataset: mp_seq_v2\n enforce_gender:\
|
| 4 |
+
\ male\n enforce_zero_beta: 1\n feature_dim: 276\n future_length: 8\n history_length:\
|
| 5 |
+
\ 2\n num_primitive: 4\n prob_static: 0.0\n text_tolerance: 0.0\n weight_scheme:\
|
| 6 |
+
\ text_samp:0.\ndenoiser_args: !dataclass:DenoiserArgs\n diffusion_args: !dataclass:DiffusionArgs\n\
|
| 7 |
+
\ diffusion_steps: 10\n noise_schedule: cosine\n respacing: ''\n sigma_small:\
|
| 8 |
+
\ true\n model_args: !dataclass:DenoiserTransformerArgs\n activation: gelu\n\
|
| 9 |
+
\ clip_dim: 512\n cond_mask_prob: 0.1\n dropout: 0.1\n ff_size: 1024\n\
|
| 10 |
+
\ h_dim: 512\n history_shape: !!python/tuple\n - 2\n - 276\n noise_shape:\
|
| 11 |
+
\ !!python/tuple\n - 1\n - 256\n num_heads: 4\n num_layers: 8\n model_type:\
|
| 12 |
+
\ transformer\n mvae_path: ./mvae/mvae_miko_finetuned_v5/checkpoint_230000.pt\n\
|
| 13 |
+
\ rescale_latent: 1\n train_rollout_history: rollout\n train_rollout_type: full\n\
|
| 14 |
+
device: cuda\nexp_name: mld_miko_finetune_v5\nsave_dir: !!python/object/apply:pathlib.PosixPath\n\
|
| 15 |
+
- mld_denoiser\n- mld_miko_finetune_v5\nseed: 0\ntorch_deterministic: true\ntrack:\
|
| 16 |
+
\ 1\ntrain_args: !dataclass:TrainArgs\n anneal_lr: 1\n batch_size: 1024\n ema_decay:\
|
| 17 |
+
\ 0.999\n grad_clip: 1.0\n learning_rate: 0.0001\n log_interval: 1000\n resume_checkpoint:\
|
| 18 |
+
\ ./mld_denoiser/mld_miko_finetune_v5/checkpoint_310000.pt\n save_interval: 5000\n\
|
| 19 |
+
\ stage1_steps: 120000\n stage2_steps: 120000\n stage3_steps: 120000\n use_amp:\
|
| 20 |
+
\ 1\n use_predicted_joints: 0\n val_interval: 10000\n weight_feature_rec: 1.0\n\
|
| 21 |
+
\ weight_joints_consistency: 0.0\n weight_joints_delta: 10000.0\n weight_kl:\
|
| 22 |
+
\ 0.0001\n weight_latent_rec: 1.0\n weight_orient_delta: 10000.0\n weight_rec:\
|
| 23 |
+
\ 1.0\n weight_smpl_joints_rec: 0.0\n weight_transl_delta: 10000.0\nwandb_entity:\
|
| 24 |
+
\ interaction\nwandb_project_name: mld_denoiser\n"
|
mld_miko_finetune_v5/args_read.yaml
ADDED
|
@@ -0,0 +1,74 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
data_args:
|
| 2 |
+
body_type: smplx
|
| 3 |
+
cfg_path: ./config_files/config_hydra/motion_primitive/mp_h2_f8_r4.yaml
|
| 4 |
+
data_dir: ./data/mixed_filtered_seq_data_zero_male
|
| 5 |
+
dataset: mp_seq_v2
|
| 6 |
+
enforce_gender: male
|
| 7 |
+
enforce_zero_beta: 1
|
| 8 |
+
feature_dim: 276
|
| 9 |
+
future_length: 8
|
| 10 |
+
history_length: 2
|
| 11 |
+
num_primitive: 4
|
| 12 |
+
prob_static: 0.0
|
| 13 |
+
text_tolerance: 0.0
|
| 14 |
+
weight_scheme: text_samp:0.
|
| 15 |
+
denoiser_args:
|
| 16 |
+
diffusion_args:
|
| 17 |
+
diffusion_steps: 10
|
| 18 |
+
noise_schedule: cosine
|
| 19 |
+
respacing: ''
|
| 20 |
+
sigma_small: true
|
| 21 |
+
model_args:
|
| 22 |
+
activation: gelu
|
| 23 |
+
clip_dim: 512
|
| 24 |
+
cond_mask_prob: 0.1
|
| 25 |
+
dropout: 0.1
|
| 26 |
+
ff_size: 1024
|
| 27 |
+
h_dim: 512
|
| 28 |
+
history_shape: !!python/tuple
|
| 29 |
+
- 2
|
| 30 |
+
- 276
|
| 31 |
+
noise_shape: !!python/tuple
|
| 32 |
+
- 1
|
| 33 |
+
- 256
|
| 34 |
+
num_heads: 4
|
| 35 |
+
num_layers: 8
|
| 36 |
+
model_type: transformer
|
| 37 |
+
mvae_path: ./mvae/mvae_miko_finetuned_v5/checkpoint_230000.pt
|
| 38 |
+
rescale_latent: 1
|
| 39 |
+
train_rollout_history: rollout
|
| 40 |
+
train_rollout_type: full
|
| 41 |
+
device: cuda
|
| 42 |
+
exp_name: mld_miko_finetune_v5
|
| 43 |
+
save_dir: !!python/object/apply:pathlib.PosixPath
|
| 44 |
+
- mld_denoiser
|
| 45 |
+
- mld_miko_finetune_v5
|
| 46 |
+
seed: 0
|
| 47 |
+
torch_deterministic: true
|
| 48 |
+
track: 1
|
| 49 |
+
train_args:
|
| 50 |
+
anneal_lr: 1
|
| 51 |
+
batch_size: 1024
|
| 52 |
+
ema_decay: 0.999
|
| 53 |
+
grad_clip: 1.0
|
| 54 |
+
learning_rate: 0.0001
|
| 55 |
+
log_interval: 1000
|
| 56 |
+
resume_checkpoint: ./mld_denoiser/mld_miko_finetune_v5/checkpoint_310000.pt
|
| 57 |
+
save_interval: 5000
|
| 58 |
+
stage1_steps: 120000
|
| 59 |
+
stage2_steps: 120000
|
| 60 |
+
stage3_steps: 120000
|
| 61 |
+
use_amp: 1
|
| 62 |
+
use_predicted_joints: 0
|
| 63 |
+
val_interval: 10000
|
| 64 |
+
weight_feature_rec: 1.0
|
| 65 |
+
weight_joints_consistency: 0.0
|
| 66 |
+
weight_joints_delta: 10000.0
|
| 67 |
+
weight_kl: 0.0001
|
| 68 |
+
weight_latent_rec: 1.0
|
| 69 |
+
weight_orient_delta: 10000.0
|
| 70 |
+
weight_rec: 1.0
|
| 71 |
+
weight_smpl_joints_rec: 0.0
|
| 72 |
+
weight_transl_delta: 10000.0
|
| 73 |
+
wandb_entity: interaction
|
| 74 |
+
wandb_project_name: mld_denoiser
|
mld_miko_finetune_v5/checkpoint_305000.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:53e8a85b2600ff01724633cf6a53bd962c88391bc4d07713729b09bf19cc15c9
|
| 3 |
+
size 82342402
|
mld_miko_finetune_v5/checkpoint_310000.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:17e25755a8452d80ea300ce0a28b447fbf5447b6b64d2f1b876d405af1e0e22d
|
| 3 |
+
size 82342402
|
mld_miko_finetune_v5/checkpoint_315000.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:da0995b083474fd5c5fd6a43dc67b78d6ce71ee870b36e29eb2293a3fe24109b
|
| 3 |
+
size 82342402
|
mld_miko_finetune_v5/checkpoint_320000.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:904e481606bbddf0c094af4d21f87817d9aff874fddee2a368b4508fe386a00c
|
| 3 |
+
size 82342402
|
mld_miko_finetune_v5/checkpoint_325000.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8e219542304d40fe5d76c39883b636992735ea0b9722404039c2fadb006b3a4d
|
| 3 |
+
size 82342402
|
mld_miko_finetune_v5/checkpoint_330000.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2e6c9405b40c11dd7dce67c3a87ef21f1730bcd0b53133b6766b9fc02c7c0a5a
|
| 3 |
+
size 82342402
|
mld_miko_finetune_v5/checkpoint_335000.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:71adde7810ae0cfe0f7a834f900f7d6aec23c8ab7b6ab509a570ecdf2be6a390
|
| 3 |
+
size 82342402
|
mld_miko_finetune_v5/checkpoint_340000.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1efea50a685a46ad70ffb23d03a34b126a4a4148ae7e632a36f8a4d2bfc4cdf3
|
| 3 |
+
size 82342402
|