suchirsalhan's picture
Fix: vocab_size=32000 (BPE base from model.vocab); top-level weights; all compat fixes
878a2a7 verified
raw
history blame contribute delete
538 Bytes
{
"architectures": [
"PicoDecoderHF"
],
"model_type": "pico_decoder",
"auto_map": {
"AutoConfig": "pico_decoder.PicoDecoderHFConfig",
"AutoModelForCausalLM": "pico_decoder.PicoDecoderHF"
},
"n_layers": 14,
"d_model": 768,
"attention_n_heads": 12,
"attention_n_kv_heads": 1,
"max_seq_len": 512,
"batch_size": 64,
"position_emb_theta": 10000.0,
"activation_hidden_dim": 3072,
"norm_eps": 1e-05,
"dropout": 0.1,
"torch_dtype": "float32",
"transformers_version": "4.48.3",
"vocab_size": 32000
}