ramu0e commited on
Commit
e45c299
·
verified ·
1 Parent(s): 106c5ef

Upload folder using huggingface_hub

Browse files
lam/config.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "action_depth": 5,
3
+ "action_dropout": 0.0,
4
+ "action_hidden_dim": 96,
5
+ "action_obs_dim": 0,
6
+ "action_prev_dim": 10,
7
+ "action_state_dim": 5,
8
+ "action_target_dim": 10,
9
+ "action_wide_dim": 512,
10
+ "architectures": [
11
+ "LAMModel"
12
+ ],
13
+ "decoder_attention_head_dim": 64,
14
+ "decoder_attn_implementation": "flash_attention_2",
15
+ "decoder_encoder_hidden_dim": 1,
16
+ "decoder_eps": 1e-06,
17
+ "decoder_ffn_dim": 768,
18
+ "decoder_freq_dim": 64,
19
+ "decoder_in_channels": 3,
20
+ "decoder_num_attention_heads": 3,
21
+ "decoder_num_layers": 12,
22
+ "decoder_out_channels": 3,
23
+ "decoder_patch_size": [
24
+ 4,
25
+ 4
26
+ ],
27
+ "decoder_pos_embed_seq_len": null,
28
+ "decoder_rope_max_seq_len": 1024,
29
+ "dtype": "bfloat16",
30
+ "encoder_height": 64,
31
+ "encoder_width": 64,
32
+ "fsq_levels": [
33
+ 8
34
+ ],
35
+ "initializer_range": 0.02,
36
+ "is_diffusion": false,
37
+ "latent_channels": 1,
38
+ "max_tokens": 4,
39
+ "min_tokens": 1,
40
+ "model_type": "lam",
41
+ "null_latent": 0,
42
+ "transformers_version": "4.57.1",
43
+ "use_tail_drop": false,
44
+ "videomae_config": {
45
+ "attn_drop_rate": 0.0,
46
+ "cos_attn": false,
47
+ "depth": 8,
48
+ "drop_path_rate": 0.0,
49
+ "drop_rate": 0.0,
50
+ "embed_dim": 192,
51
+ "img_size": [
52
+ 64,
53
+ 64
54
+ ],
55
+ "in_chans": 3,
56
+ "init_values": 0.0,
57
+ "layer_norm_eps": 1e-06,
58
+ "mlp_ratio": 4,
59
+ "norm_layer": "nn.LayerNorm",
60
+ "num_classes": 0,
61
+ "num_frames": 2,
62
+ "num_heads": 3,
63
+ "patch_size": 4,
64
+ "qk_scale": null,
65
+ "qkv_bias": true,
66
+ "tubelet_size": 2,
67
+ "use_learnable_pos_emb": false,
68
+ "use_mean_pooling": false,
69
+ "with_cp": false
70
+ },
71
+ "videomae_from_pretrained": null,
72
+ "vocab_size": 8
73
+ }
lam/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:09f3b9f1444be39d4638355b1ad6b5ebd29eaedc5aa8c1433f2818964041fda0
3
+ size 23028000
model_index.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "LAMPipeline",
3
+ "_diffusers_version": "0.35.2",
4
+ "lam": [
5
+ "flexlam_mini.models.lam.modeling_lam",
6
+ "LAMModel"
7
+ ],
8
+ "processor": [
9
+ "flexlam_mini.models.lam.processing_lam",
10
+ "LAMProcessorFast"
11
+ ],
12
+ "scheduler": [
13
+ "diffusers",
14
+ "FlowMatchEulerDiscreteScheduler"
15
+ ]
16
+ }
processor/processor_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "encoder_height": 64,
3
+ "encoder_width": 64,
4
+ "height": 64,
5
+ "processor_class": "LAMProcessorFast",
6
+ "width": 64
7
+ }
scheduler/scheduler_config.json ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "FlowMatchEulerDiscreteScheduler",
3
+ "_diffusers_version": "0.35.2",
4
+ "base_image_seq_len": 256,
5
+ "base_shift": 0.5,
6
+ "invert_sigmas": false,
7
+ "max_image_seq_len": 4096,
8
+ "max_shift": 1.15,
9
+ "num_train_timesteps": 1000,
10
+ "shift": 1.0,
11
+ "shift_terminal": null,
12
+ "stochastic_sampling": false,
13
+ "time_shift_type": "exponential",
14
+ "use_beta_sigmas": false,
15
+ "use_dynamic_shifting": false,
16
+ "use_exponential_sigmas": false,
17
+ "use_karras_sigmas": false
18
+ }