gulucaptain commited on
Commit
10b0a90
·
verified ·
1 Parent(s): ae6dad7

Delete config.json

Browse files
Files changed (1) hide show
  1. config.json +0 -33
config.json DELETED
@@ -1,33 +0,0 @@
1
- {
2
- "_class_name": "DynamiCtrlTransformerModel",
3
- "_diffusers_version": "0.32.0",
4
- "_name_or_path": "dynamictrl",
5
- "activation_fn": "gelu-approximate",
6
- "attention_bias": true,
7
- "attention_head_dim": 64,
8
- "dropout": 0.0,
9
- "flip_sin_to_cos": true,
10
- "freq_shift": 0,
11
- "in_channels": 32,
12
- "max_text_seq_length": 226,
13
- "norm_elementwise_affine": true,
14
- "norm_eps": 1e-05,
15
- "num_attention_heads": 48,
16
- "num_layers": 42,
17
- "ofs_embed_dim": 512,
18
- "out_channels": 16,
19
- "patch_bias": false,
20
- "patch_size": 2,
21
- "patch_size_t": 2,
22
- "sample_frames": 81,
23
- "sample_height": 300,
24
- "sample_width": 300,
25
- "spatial_interpolation_scale": 1.875,
26
- "temporal_compression_ratio": 4,
27
- "temporal_interpolation_scale": 1.0,
28
- "text_embed_dim": 4096,
29
- "time_embed_dim": 512,
30
- "timestep_activation_fn": "silu",
31
- "use_learned_positional_embeddings": false,
32
- "use_rotary_positional_embeddings": true
33
- }