{ "architectures": [ "DuchifatCore" ], "auto_map": { "AutoConfig": "configuration_duchifat_v2.DuchifatConfig", "AutoModelForCausalLM": "modeling_duchifat_v2.DuchifatCore" }, "dtype": "bfloat16", "hidden_size": 768, "max_seq": 1024, "model_type": "duchifat_v2", "nhead": 12, "num_layers": 12, "transformers_version": "5.2.0", "vocab_size": 33152 }