{ "engine": "qora", "model_type": "qora_llm", "base_model": "SmolLM3-3B", "parameters": "3.07B", "quantization": "Q4", "hidden_size": 2048, "num_hidden_layers": 36, "num_attention_heads": 16, "num_key_value_heads": 4, "intermediate_size": 11008, "vocab_size": 128256, "max_position_embeddings": 65536, "rope_theta": 5000000.0, "rms_norm_eps": 1e-6, "tie_word_embeddings": true, "no_rope_layer_interval": 4, "bos_token_id": 128000, "eos_token_id": 128012, "hidden_act": "silu", "no_rope_layers": [] }