euhidaman commited on
Commit
8514404
·
verified ·
1 Parent(s): a1428dc

Add model configuration

Browse files
Files changed (1) hide show
  1. config.json +110 -0
config.json ADDED
@@ -0,0 +1,110 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "qwen2_5_config": {
3
+ "hidden_size": 896,
4
+ "num_hidden_layers": 24,
5
+ "num_attention_heads": 14,
6
+ "num_key_value_heads": 2,
7
+ "intermediate_size": 4864,
8
+ "vocab_size": 151936,
9
+ "max_position_embeddings": 32768,
10
+ "rope_theta": 1000000.0,
11
+ "rms_norm_eps": 1e-06,
12
+ "tie_word_embeddings": true,
13
+ "use_sliding_window": false,
14
+ "sliding_window": null,
15
+ "max_window_layers": 21,
16
+ "head_dim": 64,
17
+ "kv_head_dim": 128,
18
+ "model_type": "qwen2"
19
+ },
20
+ "deit_tiny_config": {
21
+ "image_size": 224,
22
+ "patch_size": 16,
23
+ "num_channels": 3,
24
+ "hidden_size": 192,
25
+ "num_hidden_layers": 12,
26
+ "num_attention_heads": 3,
27
+ "intermediate_size": 768,
28
+ "hidden_dropout_prob": 0.0,
29
+ "attention_probs_dropout_prob": 0.0,
30
+ "layer_norm_eps": 1e-06,
31
+ "num_patches": 196,
32
+ "encoder_stride": 16
33
+ },
34
+ "quantization_config": {
35
+ "memory_quantization": {
36
+ "bits": 1.58,
37
+ "method": "bitnet_158",
38
+ "description": "1.58-bit quantization for episodic memory and final model"
39
+ },
40
+ "training_quantization": {
41
+ "weight_bits": 4,
42
+ "activation_bits": 4,
43
+ "method": "symmetric",
44
+ "description": "4-bit quantization for weights and activations during training"
45
+ },
46
+ "inference_quantization": {
47
+ "model_bits": 1.58,
48
+ "method": "bitnet_158",
49
+ "description": "1.58-bit quantization for final deployed model"
50
+ }
51
+ },
52
+ "model_dimensions": {
53
+ "qwen_hidden_dim": 896,
54
+ "deit_embed_dim": 192,
55
+ "vision_hidden_size": 192,
56
+ "language_hidden_size": 896,
57
+ "num_patches": 196,
58
+ "k_prefix": 25,
59
+ "adapter_projection_dim": 896,
60
+ "alignment_dim": 128,
61
+ "memory_size": 64,
62
+ "memory_dim": 896,
63
+ "memory_target_layers": 6,
64
+ "memory_num_heads": 4,
65
+ "memory_num_layers": 6,
66
+ "head_dim": 64,
67
+ "scope_hidden_dim": 256,
68
+ "itm_hidden_dim": 256,
69
+ "fusion_layers": [
70
+ 9,
71
+ 11
72
+ ],
73
+ "num_fusion_heads": 2,
74
+ "fusion_dim": 384,
75
+ "w_logvar_setting": 1,
76
+ "deterministic_memory": false
77
+ },
78
+ "itc_config": {
79
+ "use_itc_queue": true,
80
+ "itc_queue_size": 256,
81
+ "itc_embed_dim": 128,
82
+ "description": "FIBER-style ITC queue for better negative sampling (lightweight version)"
83
+ },
84
+ "estimated_sizes": {
85
+ "qwen_original_mb": 1976,
86
+ "deit_original_mb": 22.8,
87
+ "qwen_4bit_mb": 247.0,
88
+ "deit_4bit_mb": 2.85,
89
+ "qwen_158bit_mb": 98.80000000000001,
90
+ "adapter_mb": 0.777728,
91
+ "memory_original_mb": 0.229376,
92
+ "memory_158bit_mb": 0.011468800000000001,
93
+ "wm_projection_mb": 11.010048,
94
+ "scopenet_mb": 1.180672,
95
+ "fusion_mb": 0.009216,
96
+ "alignment_mb": 0.557056,
97
+ "itm_head_mb": 1.11616,
98
+ "overhead_total_mb": 14.880256,
99
+ "total_fp32_mb": 2013.6802559999999,
100
+ "total_4bit_qwen_mb": 273.28025599999995,
101
+ "total_fully_quantized_mb": 116.53025600000001
102
+ },
103
+ "metadata": {
104
+ "qwen_model": "Qwen/Qwen2.5-0.5B",
105
+ "deit_model": "facebook/deit-tiny-patch16-224",
106
+ "target_total_size_mb": 500,
107
+ "description": "MicroVLM-V compact model configuration with 4-bit quantization for <1GB deployment",
108
+ "quantization_default": "4-bit Qwen, FP16 DeiT, FP32 overhead"
109
+ }
110
+ }