pjajal commited on
Commit
9e466c9
·
verified ·
1 Parent(s): e68d3bd

Push model using huggingface_hub.

Browse files
Files changed (3) hide show
  1. README.md +10 -0
  2. config.json +41 -0
  3. model.safetensors +3 -0
README.md ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - model_hub_mixin
4
+ - pytorch_model_hub_mixin
5
+ ---
6
+
7
+ This model has been pushed to the Hub using the [PytorchModelHubMixin](https://huggingface.co/docs/huggingface_hub/package_reference/mixins#huggingface_hub.PyTorchModelHubMixin) integration:
8
+ - Code: https://github.com/pjjajal/adaperceiver-public
9
+ - Paper: https://arxiv.org/abs/2511.18105
10
+ - Docs: [More Information Needed]
config.json ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "act_layer": "gelu",
3
+ "attn_drop": 0.0,
4
+ "attn_layer": "flex",
5
+ "block_mask": "block",
6
+ "depth": 21,
7
+ "drop_path": 0.0,
8
+ "embed_dim": 832,
9
+ "ffn_layer": "mlp",
10
+ "ffn_ratio": 2.57,
11
+ "head_drop": 0.0,
12
+ "img_size": 224,
13
+ "in_channels": 3,
14
+ "ls_init_values": 1e-05,
15
+ "mask_token_grans": [
16
+ 32,
17
+ 64,
18
+ 96,
19
+ 128,
20
+ 192,
21
+ 256
22
+ ],
23
+ "mat_dims": [
24
+ 416,
25
+ 624,
26
+ 832
27
+ ],
28
+ "max_latent_tokens": 256,
29
+ "norm_layer": "layernorm",
30
+ "num_classes": 11821,
31
+ "num_heads": 13,
32
+ "output_feat_dim": 1280,
33
+ "patch_size": 14,
34
+ "process_token_init": "learned",
35
+ "proj_bias": true,
36
+ "proj_drop": 0.0,
37
+ "qkv_bias": true,
38
+ "rope_theta": 10000,
39
+ "use_embed_ffn": true,
40
+ "use_output_ffn": false
41
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf37c09acab15354b579ffc6e73a8ea12f43d34876a693891e474af9db9e293c
3
+ size 643033772