{ "vocab_size": 128256, "d_model": 384, "n_layers": 8, "n_heads": 6, "d_ff": 1536, "seq_len": 1024, "architectures": [ "Model" ], "model_type": "llama", "torch_dtype": "bfloat16" }