{ "architectures": [ "HFWrappedLLM" ], "hidden_size": 64, "max_position_embeddings": 128, "num_heads": 1, "num_layers": 1, "torch_dtype": "float32", "transformers_version": "4.48.1", "vocab_size": 500 }