{ "architectures": [ "VSBForCausalLM" ], "d_e": 768, "hidden_size": 1536, "max_position_embeddings": 131072, "model_type": "vsbssm", "num_hidden_layers": 16, "r": 384, "torch_dtype": "float32", "transformers_version": "4.55.4", "vocab_size": 50257 }