{
"module": "keras_hub.src.models.smollm3.smollm3_causal_lm",
"class_name": "SmolLM3CausalLM",
"config": {
"backbone": {
"module": "keras_hub.src.models.smollm3.smollm3_backbone",
"class_name": "SmolLM3Backbone",
"config": {
"name": "smol_lm3_backbone",
"trainable": true,
"dtype": {
"module": "keras",
"class_name": "DTypePolicy",
"config": {
"name": "float32"
},
"registered_name": null
},
"vocabulary_size": 128256,
"hidden_dim": 2048,
"intermediate_dim": 11008,
"num_layers": 36,
"num_attention_heads": 16,
"num_key_value_heads": 4,
"attention_bias": false,
"attention_dropout": 0.0,
"rope_layer_enabled_list": [
true,
true,
true,
false,
true,
true,
true,
false,
true,
true,
true,
false,
true,
true,
true,
false,
true,
true,
true,
false,
true,
true,
true,
false,
true,
true,
true,
false,
true,
true,
true,
false,
true,
true,
true,
false
],
"layer_types": [
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention",
"full_attention"
],
"mlp_bias": false,
"layer_norm_epsilon": 1e-06,
"max_position_embeddings": 65536,
"rope_theta": 5000000.0,
"partial_rotary_factor": 1.0
},
"registered_name": "keras_hub>SmolLM3Backbone"
},
"preprocessor": {
"module": "keras_hub.src.models.smollm3.smollm3_causal_lm_preprocessor",
"class_name": "SmolLM3CausalLMPreprocessor",
"config": {
"name": "smol_lm3_causal_lm_preprocessor_1",
"trainable": true,
"dtype": {
"module": "keras",
"class_name": "DTypePolicy",
"config": {
"name": "float32"
},
"registered_name": null
},
"tokenizer": {
"module": "keras_hub.src.models.smollm3.smollm3_tokenizer",
"class_name": "SmolLM3Tokenizer",
"config": {
"name": "smol_lm3_tokenizer",
"trainable": true,
"dtype": {
"module": "keras",
"class_name": "DTypePolicy",
"config": {
"name": "int32"
},
"registered_name": null
},
"config_file": "tokenizer.json",
"sequence_length": null,
"add_prefix_space": false,
"unsplittable_tokens": [
"",
"",
"<|end_header_id|>",
"<|eot_id|>",
"<|im_end|>",
"<|end_of_text|>",
"",
"",
"<|im_start|>",
"",
"",
"<|start_header_id|>",
"<|begin_of_text|>",
"<|eom_id|>",
"",
"",
"<|python_tag|>",
"<|finetune_right_pad_id|>"
]
},
"registered_name": "keras_hub>SmolLM3Tokenizer"
},
"config_file": "preprocessor.json",
"sequence_length": 1024,
"add_start_token": true,
"add_end_token": true
},
"registered_name": "keras_hub>SmolLM3CausalLMPreprocessor"
},
"name": "smol_lm3_causal_lm"
},
"registered_name": "keras_hub>SmolLM3CausalLM"
}