SAID-LAM-v1 / config.json
SAIDHOME's picture
Upload folder using huggingface_hub
346e942 verified
{
"model_type": "lam",
"model_category": "LAM (Linear Attention Models) — SAID-LAM-v1: Linear Attention Memory",
"architectures": [
"LAM"
],
"attention_mechanism": "Hierarchical DeltaNet (Linear Attention with Dual-State Memory)",
"recall_engine": "SCA (Said Crystalline Attention)",
"vocab_size": 30522,
"hidden_size": 384,
"num_hidden_layers": 6,
"num_attention_heads": 12,
"intermediate_size": 1536,
"max_position_embeddings": 32768,
"embedding_dim": 384,
"n_parameters": 23848788,
"memory_usage_mb": 90,
"framework": "Candle",
"runtime": "Pure Rust (no PyTorch dependency)",
"distilled_from": "all-MiniLM-L6-v2",
"distillation": true,
"linear_attention": true,
"hierarchical_deltanet": true,
"dual_state_memory": true,
"complexity": "O(n)",
"license": "apache-2.0",
"license_note": "Apache 2.0 for model weights; Proprietary for compiled software",
"reference": "https://saidhome.ai"
}