{ "module": "keras_hub.src.models.smollm3.smollm3_causal_lm", "class_name": "SmolLM3CausalLM", "config": { "backbone": { "module": "keras_hub.src.models.smollm3.smollm3_backbone", "class_name": "SmolLM3Backbone", "config": { "name": "smol_lm3_backbone", "trainable": true, "dtype": { "module": "keras", "class_name": "DTypePolicy", "config": { "name": "float32" }, "registered_name": null }, "vocabulary_size": 128256, "hidden_dim": 2048, "intermediate_dim": 11008, "num_layers": 36, "num_attention_heads": 16, "num_key_value_heads": 4, "attention_bias": false, "attention_dropout": 0.0, "rope_layer_enabled_list": [ true, true, true, false, true, true, true, false, true, true, true, false, true, true, true, false, true, true, true, false, true, true, true, false, true, true, true, false, true, true, true, false, true, true, true, false ], "layer_types": [ "full_attention", "full_attention", "full_attention", "full_attention", "full_attention", "full_attention", "full_attention", "full_attention", "full_attention", "full_attention", "full_attention", "full_attention", "full_attention", "full_attention", "full_attention", "full_attention", "full_attention", "full_attention", "full_attention", "full_attention", "full_attention", "full_attention", "full_attention", "full_attention", "full_attention", "full_attention", "full_attention", "full_attention", "full_attention", "full_attention", "full_attention", "full_attention", "full_attention", "full_attention", "full_attention", "full_attention" ], "mlp_bias": false, "layer_norm_epsilon": 1e-06, "max_position_embeddings": 65536, "rope_theta": 5000000.0, "partial_rotary_factor": 1.0 }, "registered_name": "keras_hub>SmolLM3Backbone" }, "preprocessor": { "module": "keras_hub.src.models.smollm3.smollm3_causal_lm_preprocessor", "class_name": "SmolLM3CausalLMPreprocessor", "config": { "name": "smol_lm3_causal_lm_preprocessor_1", "trainable": true, "dtype": { "module": "keras", "class_name": "DTypePolicy", "config": { "name": "float32" }, "registered_name": null }, "tokenizer": { "module": "keras_hub.src.models.smollm3.smollm3_tokenizer", "class_name": "SmolLM3Tokenizer", "config": { "name": "smol_lm3_tokenizer", "trainable": true, "dtype": { "module": "keras", "class_name": "DTypePolicy", "config": { "name": "int32" }, "registered_name": null }, "config_file": "tokenizer.json", "sequence_length": null, "add_prefix_space": false, "unsplittable_tokens": [ "", "", "<|end_header_id|>", "<|eot_id|>", "<|im_end|>", "<|end_of_text|>", "", "", "<|im_start|>", "", "", "<|start_header_id|>", "<|begin_of_text|>", "<|eom_id|>", "", "", "<|python_tag|>", "<|finetune_right_pad_id|>" ] }, "registered_name": "keras_hub>SmolLM3Tokenizer" }, "config_file": "preprocessor.json", "sequence_length": 1024, "add_start_token": true, "add_end_token": true }, "registered_name": "keras_hub>SmolLM3CausalLMPreprocessor" }, "name": "smol_lm3_causal_lm" }, "registered_name": "keras_hub>SmolLM3CausalLM" }