zirobtc commited on
Commit
7ee7e96
·
verified ·
1 Parent(s): 827294e

Upload folder using huggingface_hub

Browse files
mld_miko_finetune_v5/args.yaml ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ "# tyro YAML.\n!dataclass:MLDArgs\ndata_args: !dataclass:DataArgs\n body_type: smplx\n\
2
+ \ cfg_path: ./config_files/config_hydra/motion_primitive/mp_h2_f8_r4.yaml\n data_dir:\
3
+ \ ./data/mixed_filtered_seq_data_zero_male\n dataset: mp_seq_v2\n enforce_gender:\
4
+ \ male\n enforce_zero_beta: 1\n feature_dim: 276\n future_length: 8\n history_length:\
5
+ \ 2\n num_primitive: 4\n prob_static: 0.0\n text_tolerance: 0.0\n weight_scheme:\
6
+ \ text_samp:0.\ndenoiser_args: !dataclass:DenoiserArgs\n diffusion_args: !dataclass:DiffusionArgs\n\
7
+ \ diffusion_steps: 10\n noise_schedule: cosine\n respacing: ''\n sigma_small:\
8
+ \ true\n model_args: !dataclass:DenoiserTransformerArgs\n activation: gelu\n\
9
+ \ clip_dim: 512\n cond_mask_prob: 0.1\n dropout: 0.1\n ff_size: 1024\n\
10
+ \ h_dim: 512\n history_shape: !!python/tuple\n - 2\n - 276\n noise_shape:\
11
+ \ !!python/tuple\n - 1\n - 256\n num_heads: 4\n num_layers: 8\n model_type:\
12
+ \ transformer\n mvae_path: ./mvae/mvae_miko_finetuned_v5/checkpoint_230000.pt\n\
13
+ \ rescale_latent: 1\n train_rollout_history: rollout\n train_rollout_type: full\n\
14
+ device: cuda\nexp_name: mld_miko_finetune_v5\nsave_dir: !!python/object/apply:pathlib.PosixPath\n\
15
+ - mld_denoiser\n- mld_miko_finetune_v5\nseed: 0\ntorch_deterministic: true\ntrack:\
16
+ \ 1\ntrain_args: !dataclass:TrainArgs\n anneal_lr: 1\n batch_size: 1024\n ema_decay:\
17
+ \ 0.999\n grad_clip: 1.0\n learning_rate: 0.0001\n log_interval: 1000\n resume_checkpoint:\
18
+ \ ./mld_denoiser/mld_miko_finetune_v5/checkpoint_310000.pt\n save_interval: 5000\n\
19
+ \ stage1_steps: 120000\n stage2_steps: 120000\n stage3_steps: 120000\n use_amp:\
20
+ \ 1\n use_predicted_joints: 0\n val_interval: 10000\n weight_feature_rec: 1.0\n\
21
+ \ weight_joints_consistency: 0.0\n weight_joints_delta: 10000.0\n weight_kl:\
22
+ \ 0.0001\n weight_latent_rec: 1.0\n weight_orient_delta: 10000.0\n weight_rec:\
23
+ \ 1.0\n weight_smpl_joints_rec: 0.0\n weight_transl_delta: 10000.0\nwandb_entity:\
24
+ \ interaction\nwandb_project_name: mld_denoiser\n"
mld_miko_finetune_v5/args_read.yaml ADDED
@@ -0,0 +1,74 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ data_args:
2
+ body_type: smplx
3
+ cfg_path: ./config_files/config_hydra/motion_primitive/mp_h2_f8_r4.yaml
4
+ data_dir: ./data/mixed_filtered_seq_data_zero_male
5
+ dataset: mp_seq_v2
6
+ enforce_gender: male
7
+ enforce_zero_beta: 1
8
+ feature_dim: 276
9
+ future_length: 8
10
+ history_length: 2
11
+ num_primitive: 4
12
+ prob_static: 0.0
13
+ text_tolerance: 0.0
14
+ weight_scheme: text_samp:0.
15
+ denoiser_args:
16
+ diffusion_args:
17
+ diffusion_steps: 10
18
+ noise_schedule: cosine
19
+ respacing: ''
20
+ sigma_small: true
21
+ model_args:
22
+ activation: gelu
23
+ clip_dim: 512
24
+ cond_mask_prob: 0.1
25
+ dropout: 0.1
26
+ ff_size: 1024
27
+ h_dim: 512
28
+ history_shape: !!python/tuple
29
+ - 2
30
+ - 276
31
+ noise_shape: !!python/tuple
32
+ - 1
33
+ - 256
34
+ num_heads: 4
35
+ num_layers: 8
36
+ model_type: transformer
37
+ mvae_path: ./mvae/mvae_miko_finetuned_v5/checkpoint_230000.pt
38
+ rescale_latent: 1
39
+ train_rollout_history: rollout
40
+ train_rollout_type: full
41
+ device: cuda
42
+ exp_name: mld_miko_finetune_v5
43
+ save_dir: !!python/object/apply:pathlib.PosixPath
44
+ - mld_denoiser
45
+ - mld_miko_finetune_v5
46
+ seed: 0
47
+ torch_deterministic: true
48
+ track: 1
49
+ train_args:
50
+ anneal_lr: 1
51
+ batch_size: 1024
52
+ ema_decay: 0.999
53
+ grad_clip: 1.0
54
+ learning_rate: 0.0001
55
+ log_interval: 1000
56
+ resume_checkpoint: ./mld_denoiser/mld_miko_finetune_v5/checkpoint_310000.pt
57
+ save_interval: 5000
58
+ stage1_steps: 120000
59
+ stage2_steps: 120000
60
+ stage3_steps: 120000
61
+ use_amp: 1
62
+ use_predicted_joints: 0
63
+ val_interval: 10000
64
+ weight_feature_rec: 1.0
65
+ weight_joints_consistency: 0.0
66
+ weight_joints_delta: 10000.0
67
+ weight_kl: 0.0001
68
+ weight_latent_rec: 1.0
69
+ weight_orient_delta: 10000.0
70
+ weight_rec: 1.0
71
+ weight_smpl_joints_rec: 0.0
72
+ weight_transl_delta: 10000.0
73
+ wandb_entity: interaction
74
+ wandb_project_name: mld_denoiser
mld_miko_finetune_v5/checkpoint_305000.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:53e8a85b2600ff01724633cf6a53bd962c88391bc4d07713729b09bf19cc15c9
3
+ size 82342402
mld_miko_finetune_v5/checkpoint_310000.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17e25755a8452d80ea300ce0a28b447fbf5447b6b64d2f1b876d405af1e0e22d
3
+ size 82342402
mld_miko_finetune_v5/checkpoint_315000.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da0995b083474fd5c5fd6a43dc67b78d6ce71ee870b36e29eb2293a3fe24109b
3
+ size 82342402
mld_miko_finetune_v5/checkpoint_320000.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:904e481606bbddf0c094af4d21f87817d9aff874fddee2a368b4508fe386a00c
3
+ size 82342402
mld_miko_finetune_v5/checkpoint_325000.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e219542304d40fe5d76c39883b636992735ea0b9722404039c2fadb006b3a4d
3
+ size 82342402
mld_miko_finetune_v5/checkpoint_330000.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e6c9405b40c11dd7dce67c3a87ef21f1730bcd0b53133b6766b9fc02c7c0a5a
3
+ size 82342402
mld_miko_finetune_v5/checkpoint_335000.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71adde7810ae0cfe0f7a834f900f7d6aec23c8ab7b6ab509a570ecdf2be6a390
3
+ size 82342402
mld_miko_finetune_v5/checkpoint_340000.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1efea50a685a46ad70ffb23d03a34b126a4a4148ae7e632a36f8a4d2bfc4cdf3
3
+ size 82342402