JarrentWu commited on
Commit
01dffea
·
verified ·
1 Parent(s): 01cf306

Upload folder using huggingface_hub

Browse files
dvae/dvae_f.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af78499ca80c8761086cbc6cd48e6dcc735e09238b64784621ac2b776db2c710
3
+ size 146248555
dvae/model_config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "enc_depth": 8,
3
+ "dec_depth": 8,
4
+ "dim": 256,
5
+ "output_dim": 48,
6
+ "latent_dim": 32,
7
+ "T": 16,
8
+ "num_traj": 512,
9
+ "n_layers": 1
10
+ }
dvae_factors/dvae_f.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "f0_mean": 0.205227330327034,
3
+ "f0_std": 9.064417839050293,
4
+ "ft_mean": 0.95684415102005,
5
+ "ft_std": 6.397393703460693
6
+ }
rf_model/rf_epoch_f.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5f1ad53e6616ef909f2234a0b80289c26538806cb397679c28279be4e122506
3
+ size 805255167
rf_model/training_config.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "vae_config": {
3
+ "enc_depth": 8,
4
+ "dec_depth": 8,
5
+ "dim": 256,
6
+ "output_dim": 48,
7
+ "latent_dim": 32,
8
+ "T": 16,
9
+ "num_traj": 512,
10
+ "n_layers": 1
11
+ },
12
+ "rf_config": {
13
+ "width": 512,
14
+ "layers": 12,
15
+ "heads": 8,
16
+ "cond_drop_prob": 0.1,
17
+ "input_channels": 64,
18
+ "output_channels": 64,
19
+ "use_flash2": true
20
+ },
21
+ "training_args": {
22
+ "exp_name": "rf_model",
23
+ "vae_exp_dependency": "dvae",
24
+ "vae_epoch_dependency": "f",
25
+ "learning_rate": 0.0002,
26
+ "batch_size_per_gpu": 64,
27
+ "total_epochs": 2000,
28
+ "seed": 666,
29
+ "rescale_stats": true
30
+ }
31
+ }