Lighttiny-4M-H / config.json
Levelfive's picture
Upload LIGHTBRAIN model
add86cc verified
{
"model_type": "lighttiny",
"architectures": [
"LightTinyForCausalLM"
],
"name": "lighttiny-4m",
"vocab_size": 8000,
"hidden_dim": 128,
"num_regions": 32,
"field_size": 256,
"field_depth": 488,
"total_configs": 3997696,
"total_params": 7102464,
"layer3_resonance": {
"max_active_ratio": 0.05,
"activation_threshold": 0.08,
"avg_active_ratio": 0.0,
"total_activations": 0
},
"layer4_dynamics": {
"coupling_strength": 0.1,
"damping": 0.01,
"dt": 0.01,
"max_evolution_steps": 40,
"convergence_threshold": 0.008,
"min_coherence": 0.2
},
"bos_token_id": 2,
"eos_token_id": 3,
"pad_token_id": 0
}