Hanbo-Cheng commited on
Commit
331a927
·
1 Parent(s): 625baac

seperate inference of blink and pose

Browse files
pbnet_seperate/blink/checkpoint_95000.pth.tar ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0b23c8d2f39acdf11f868aa5e8a2c227eb6dc8922ecdbf1cd3fa90c9733d8ba
3
+ size 35021478
pbnet_seperate/blink/opt.yaml ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ activation: gelu
2
+ archiname: transformerreemb5
3
+ audio_latent_dim: 256
4
+ batch_size: 82
5
+ cuda: true
6
+ dataset: hdtf
7
+ expname: exps
8
+ eye: 'True'
9
+ ff_size: 64
10
+ first3: 'False'
11
+ folder: exps_delta_eye_rope/HDTF_200_onlyeye_fast
12
+ gpu: '0'
13
+ lambda_kl: 0.001
14
+ lambda_rc: 1.0
15
+ lambda_rcw: 1.0
16
+ lambda_ssim: 1.0
17
+ lambdas:
18
+ kl: 0.001
19
+ rc: 1.0
20
+ latent_dim: 256
21
+ losses:
22
+ - rc
23
+ - kl
24
+ lr: 0.0001
25
+ max_distance: 128
26
+ max_len: -1
27
+ min_len: -1
28
+ modelname: cvae_transformerreemb5_rc_kl
29
+ modeltype: cvae
30
+ num_buckets: 128
31
+ num_epochs: 100000
32
+ num_frames: 200
33
+ num_layers: 2
34
+ num_seq_max: -1
35
+ pose_latent_dim: 32
36
+ snapshot: 5000
pbnet_seperate/pose/checkpoint_40000.pth.tar ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:730b1c6cfd219df82b7d40d71e022d8022474d836bedb1bc6538944725afbe0d
3
+ size 45262502
pbnet_seperate/pose/opt.yaml ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ activation: gelu
2
+ archiname: transformerreemb6
3
+ audio_latent_dim: 256
4
+ batch_size: 82
5
+ ckpt: ''
6
+ cuda: true
7
+ dataset: hdtf
8
+ expname: exps
9
+ ff_size: 128
10
+ folder: exps_delta_pose_rope/HDTF_l2_nf200_rope_3drop_noattnmask_no_init_pe
11
+ gpu: '0'
12
+ lambda_freq: 1.0
13
+ lambda_kl: 0.001
14
+ lambda_rc: 1.0
15
+ lambda_rcw: 1.0
16
+ lambda_ssim: 1.0
17
+ lambdas:
18
+ kl: 0.001
19
+ rc: 1.0
20
+ latent_dim: 256
21
+ losses:
22
+ - rc
23
+ - kl
24
+ lr: 0.0004
25
+ max_distance: 128
26
+ max_len: -1
27
+ min_len: -1
28
+ modelname: cvae_transformerreemb6_rc_kl
29
+ modeltype: cvae
30
+ num_buckets: 128
31
+ num_epochs: 100000
32
+ num_frames: 200
33
+ num_layers: 2
34
+ num_seq_max: -1
35
+ snapshot: 10000