itriedcoding's picture
Add config.json
00617eb verified
{
"name": "HFMT-8 (High-Fidelity Multimodal Transformer)",
"layers": [
{
"type": "Conv2d",
"params": {
"in_channels": 3,
"out_channels": 1152,
"kernel_size": 14,
"stride": 14,
"note": "SigLIP-style Patch Embedding for high-resolution input"
}
},
{
"type": "TransformerBlock",
"params": {
"hidden_size": 1152,
"num_heads": 16,
"mlp_ratio": 4,
"activation": "GELU",
"note": "SigLIP SO400M Vision Encoder Backbone"
}
},
{
"type": "Conv2d",
"params": {
"in_channels": 1152,
"out_channels": 1152,
"kernel_size": 2,
"stride": 2,
"note": "Adaptive Patch-Merging for 50% Visual Token Reduction"
}
},
{
"type": "Linear",
"params": {
"in_features": 1152,
"out_features": 4096,
"note": "Cross-Modal Projection Bridge to LLM Latent Space"
}
},
{
"type": "TransformerBlock",
"params": {
"hidden_size": 4096,
"num_attention_heads": 32,
"num_key_value_groups": 8,
"attention_type": "Grouped-Query Attention (GQA)",
"positional_encoding": "RoPE (Rotary)",
"note": "Llama-3 Decoder Block with 4-bit NF4 Quantization Support"
}
},
{
"type": "Linear",
"params": {
"in_features": 4096,
"out_features": 14336,
"activation": "SwiGLU",
"note": "Gated Linear Unit for Enhanced Representational Capacity"
}
},
{
"type": "RMSNorm",
"params": {
"normalized_shape": 4096,
"eps": 0.00001,
"note": "Pre-block Normalization for Numerical Stability"
}
},
{
"type": "Linear",
"params": {
"in_features": 4096,
"out_features": 128256,
"bias": false,
"note": "Language Modeling Head (Uncensored Configuration)"
}
}
],
"explanation": "The HFMT-8 architecture is designed to balance multimodal reasoning with extreme memory efficiency for 8GB VRAM environments. By utilizing SigLIP for vision, we achieve better zero-shot alignment than CLIP with fewer parameters. The 'C-Abstractor' via patch-merging reduces visual tokens significantly, preventing KV-cache explosion during multimodal tasks. The LLM backbone utilizes Grouped-Query Attention (GQA) to minimize the memory footprint of the attention mechanism by a factor of 4, and the transition to 4-bit NF4 quantization ensures the 8B parameter model fits comfortably within 4.5GB, leaving ample room for the visual buffer and context window."
}