harryhyzhao commited on
Commit
4183cec
·
1 Parent(s): c2649d6
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ diffusion_pytorch_model.safetensors filter=lfs diff=lfs merge=lfs -text
diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9f87f5bddadd135d541d3e5a30e1e44fdb08bbf286c8b50558bbd9ed4f1371f
3
+ size 11540979448
transformer/config.json DELETED
@@ -1,33 +0,0 @@
1
- {
2
- "_class_name": "DynamiCtrlTransformerModel",
3
- "_diffusers_version": "0.32.0.dev0",
4
- "_name_or_path": "dynamictrl_mask_b01",
5
- "activation_fn": "gelu-approximate",
6
- "attention_bias": true,
7
- "attention_head_dim": 64,
8
- "dropout": 0.0,
9
- "flip_sin_to_cos": true,
10
- "freq_shift": 0,
11
- "in_channels": 32,
12
- "max_text_seq_length": 226,
13
- "norm_elementwise_affine": true,
14
- "norm_eps": 1e-05,
15
- "num_attention_heads": 48,
16
- "num_layers": 42,
17
- "ofs_embed_dim": 512,
18
- "out_channels": 16,
19
- "patch_bias": false,
20
- "patch_size": 2,
21
- "patch_size_t": 2,
22
- "sample_frames": 81,
23
- "sample_height": 300,
24
- "sample_width": 300,
25
- "spatial_interpolation_scale": 1.875,
26
- "temporal_compression_ratio": 4,
27
- "temporal_interpolation_scale": 1.0,
28
- "text_embed_dim": 4096,
29
- "time_embed_dim": 512,
30
- "timestep_activation_fn": "silu",
31
- "use_learned_positional_embeddings": false,
32
- "use_rotary_positional_embeddings": true
33
- }