zirobtc commited on
Commit
ad403bd
·
verified ·
1 Parent(s): 482257f

Upload folder using huggingface_hub

Browse files
checkpoints/mvae_miko_finetuned_v2/args.yaml ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ "# tyro YAML.\n!dataclass:Args\ndata_args: !dataclass:DataArgs\n body_type: smplx\n\
2
+ \ cfg_path: ./config_files/config_hydra/motion_primitive/mp_h2_f8_r8.yaml\n data_dir:\
3
+ \ ./data/miko_seq_data_zero_male\n dataset: mp_seq_v2\n enforce_gender: male\n\
4
+ \ enforce_zero_beta: 1\n feature_dim: 276\n future_length: 8\n history_length:\
5
+ \ 2\n num_primitive: 8\n prob_static: 0.0\n text_tolerance: 0.0\n weight_scheme:\
6
+ \ text_samp:0.\ndevice: cuda\nexp_name: mvae_miko_finetuned_v2\nmodel_args: !dataclass:VAEArgs\n\
7
+ \ activation: gelu\n arch: all_encoder\n dropout: 0.1\n ff_size: 1024\n h_dim:\
8
+ \ 256\n latent_dim: !!python/tuple\n - 1\n - 256\n nfeats: 276\n normalize_before:\
9
+ \ false\n num_heads: 4\n num_layers: 7\n position_embedding: learned\nsave_dir:\
10
+ \ !!python/object/apply:pathlib.PosixPath\n- mvae\n- mvae_miko_finetuned_v2\nseed:\
11
+ \ 0\ntorch_deterministic: true\ntrack: 1\ntrain_args: !dataclass:TrainArgs\n anneal_lr:\
12
+ \ 1\n batch_size: 128\n ema_decay: 0.999\n grad_clip: 1.0\n learning_rate: 1.0e-06\n\
13
+ \ log_interval: 1000\n resume_checkpoint: ./mvae/mvae_fps_clip/checkpoint_200000.pt\n\
14
+ \ save_interval: 18000\n stage1_steps: 110000\n stage2_steps: 60000\n stage3_steps:\
15
+ \ 60000\n use_amp: 0\n use_predicted_joints: 0\n val_interval: 15000\n weight_feature_rec:\
16
+ \ 0.0\n weight_joints_consistency: 10.0\n weight_joints_delta: 100.0\n weight_kl:\
17
+ \ 1.0e-06\n weight_latent_rec: 1.0\n weight_orient_delta: 100.0\n weight_rec:\
18
+ \ 1.0\n weight_smpl_joints_rec: 10.0\n weight_transl_delta: 100.0\nwandb_entity:\
19
+ \ interaction\nwandb_project_name: mld_vae\n"
checkpoints/mvae_miko_finetuned_v2/args_read.yaml ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ data_args:
2
+ body_type: smplx
3
+ cfg_path: ./config_files/config_hydra/motion_primitive/mp_h2_f8_r8.yaml
4
+ data_dir: ./data/miko_seq_data_zero_male
5
+ dataset: mp_seq_v2
6
+ enforce_gender: male
7
+ enforce_zero_beta: 1
8
+ feature_dim: 276
9
+ future_length: 8
10
+ history_length: 2
11
+ num_primitive: 8
12
+ prob_static: 0.0
13
+ text_tolerance: 0.0
14
+ weight_scheme: text_samp:0.
15
+ device: cuda
16
+ exp_name: mvae_miko_finetuned_v2
17
+ model_args:
18
+ activation: gelu
19
+ arch: all_encoder
20
+ dropout: 0.1
21
+ ff_size: 1024
22
+ h_dim: 256
23
+ latent_dim: !!python/tuple
24
+ - 1
25
+ - 256
26
+ nfeats: 276
27
+ normalize_before: false
28
+ num_heads: 4
29
+ num_layers: 7
30
+ position_embedding: learned
31
+ save_dir: !!python/object/apply:pathlib.PosixPath
32
+ - mvae
33
+ - mvae_miko_finetuned_v2
34
+ seed: 0
35
+ torch_deterministic: true
36
+ track: 1
37
+ train_args:
38
+ anneal_lr: 1
39
+ batch_size: 128
40
+ ema_decay: 0.999
41
+ grad_clip: 1.0
42
+ learning_rate: 1.0e-06
43
+ log_interval: 1000
44
+ resume_checkpoint: ./mvae/mvae_fps_clip/checkpoint_200000.pt
45
+ save_interval: 18000
46
+ stage1_steps: 110000
47
+ stage2_steps: 60000
48
+ stage3_steps: 60000
49
+ use_amp: 0
50
+ use_predicted_joints: 0
51
+ val_interval: 15000
52
+ weight_feature_rec: 0.0
53
+ weight_joints_consistency: 10.0
54
+ weight_joints_delta: 100.0
55
+ weight_kl: 1.0e-06
56
+ weight_latent_rec: 1.0
57
+ weight_orient_delta: 100.0
58
+ weight_rec: 1.0
59
+ weight_smpl_joints_rec: 10.0
60
+ weight_transl_delta: 100.0
61
+ wandb_entity: interaction
62
+ wandb_project_name: mld_vae
checkpoints/mvae_miko_finetuned_v2/checkpoint_216000.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db16a24b35d9a78050867bfadd71601ec91ae520d0b6c300b97eead47f8edeb1
3
+ size 49577410
checkpoints/mvae_miko_finetuned_v2/checkpoint_230000.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45aa726cda767b75591908b0af6e333df9a5f7c83e47b4e91001594c01743f28
3
+ size 49577410