Alexhu1999 commited on
Commit
a042e38
·
verified ·
1 Parent(s): 2c8b6a1

Upload folder using huggingface_hub

Browse files
config.json ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Gemma3nForConditionalGeneration"
4
+ ],
5
+ "audio_config": {
6
+ "conf_attention_chunk_size": 12,
7
+ "conf_attention_context_left": 13,
8
+ "conf_attention_context_right": 0,
9
+ "conf_attention_logit_cap": 50.0,
10
+ "conf_conv_kernel_size": 5,
11
+ "conf_num_attention_heads": 8,
12
+ "conf_num_hidden_layers": 12,
13
+ "conf_reduction_factor": 4,
14
+ "conf_residual_weight": 0.5,
15
+ "gradient_clipping": 10000000000.0,
16
+ "hidden_size": 1536,
17
+ "input_feat_size": 128,
18
+ "model_type": "gemma3n_audio",
19
+ "rms_norm_eps": 1e-06,
20
+ "sscp_conv_channel_size": [
21
+ 128,
22
+ 32
23
+ ],
24
+ "sscp_conv_group_norm_eps": 0.001,
25
+ "sscp_conv_kernel_size": [
26
+ [
27
+ 3,
28
+ 3
29
+ ],
30
+ [
31
+ 3,
32
+ 3
33
+ ]
34
+ ],
35
+ "sscp_conv_stride_size": [
36
+ [
37
+ 2,
38
+ 2
39
+ ],
40
+ [
41
+ 2,
42
+ 2
43
+ ]
44
+ ],
45
+ "torch_dtype": "float32",
46
+ "vocab_offset": 262272,
47
+ "vocab_size": 128
48
+ },
49
+ "audio_soft_tokens_per_image": 188,
50
+ "audio_token_id": 262273,
51
+ "boa_token_id": 256000,
52
+ "boi_token": "<start_of_image>",
53
+ "boi_token_id": 64401,
54
+ "bos_token_id": 1,
55
+ "eoa_token_id": 262272,
56
+ "eoi_token": "<end_of_image>",
57
+ "eoi_token_id": 64402,
58
+ "eos_token_id": 7,
59
+ "image_token": "<image_soft_token>",
60
+ "image_token_id": 64400,
61
+ "initializer_range": 0.02,
62
+ "model_type": "gemma3n",
63
+ "pad_token_id": 0,
64
+ "text_config": {
65
+ "architectures": [
66
+ "Lfm2ForCausalLM"
67
+ ],
68
+ "block_auto_adjust_ff_dim": true,
69
+ "block_dim": 2048,
70
+ "block_ff_dim": 12288,
71
+ "block_ffn_dim_multiplier": 1.0,
72
+ "block_mlp_init_scale": 1.0,
73
+ "block_multiple_of": 256,
74
+ "block_norm_eps": 1e-05,
75
+ "block_out_init_scale": 1.0,
76
+ "block_use_swiglu": true,
77
+ "block_use_xavier_init": true,
78
+ "bos_token_id": 2,
79
+ "conv_L_cache": 3,
80
+ "conv_bias": false,
81
+ "conv_dim": 2048,
82
+ "conv_dim_out": 2048,
83
+ "conv_use_xavier_init": true,
84
+ "eos_token_id": 1,
85
+ "hidden_size": 2048,
86
+ "initializer_range": 0.02,
87
+ "intermediate_size": 12288,
88
+ "layer_types": [
89
+ "conv",
90
+ "conv",
91
+ "full_attention",
92
+ "conv",
93
+ "conv",
94
+ "full_attention",
95
+ "conv",
96
+ "conv",
97
+ "full_attention",
98
+ "conv",
99
+ "full_attention",
100
+ "conv",
101
+ "full_attention",
102
+ "conv",
103
+ "full_attention",
104
+ "conv"
105
+ ],
106
+ "max_position_embeddings": 128000,
107
+ "model_type": "lfm2",
108
+ "norm_eps": 1e-05,
109
+ "num_attention_heads": 32,
110
+ "num_heads": 32,
111
+ "num_hidden_layers": 16,
112
+ "num_key_value_heads": 8,
113
+ "rope_theta": 1000000.0,
114
+ "torch_dtype": "float32",
115
+ "use_cache": true,
116
+ "use_pos_enc": true,
117
+ "vocab_size": 64403
118
+ },
119
+ "torch_dtype": "float32",
120
+ "transformers_version": "4.54.1",
121
+ "use_cache": true,
122
+ "vision_config": {
123
+ "architecture": "mobilenetv5_300m_enc",
124
+ "do_pooling": false,
125
+ "hidden_size": 2048,
126
+ "initializer_range": 0.02,
127
+ "label_names": [
128
+ "LABEL_0",
129
+ "LABEL_1"
130
+ ],
131
+ "model_args": null,
132
+ "model_type": "gemma3n_vision",
133
+ "num_classes": 2,
134
+ "rms_norm_eps": 1e-06,
135
+ "torch_dtype": "float32",
136
+ "vocab_offset": 262144,
137
+ "vocab_size": 128
138
+ },
139
+ "vision_soft_tokens_per_image": 256,
140
+ "vocab_size": 64403
141
+ }
generation_config.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 2,
3
+ "cache_implementation": "hybrid",
4
+ "do_sample": true,
5
+ "eos_token_id": [
6
+ 1,
7
+ 106
8
+ ],
9
+ "pad_token_id": 0,
10
+ "top_k": 64,
11
+ "top_p": 0.95,
12
+ "transformers_version": "4.54.1"
13
+ }
model-00001-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4d0f810cd65290ffeed92619b2afcd17396e320c899726ab9e727394f814618
3
+ size 4934196592
model-00002-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef3626fa9c02e6b3f36f3e9a58a651245e049847d1fac42cd22413e7c31b3df3
3
+ size 3671141560
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff