{ "architectures": [ "HFByteETM" ], "block_size": 512, "dtype": "float32", "model_type": "byteetm", "n_embd": 512, "n_head": 16, "n_layer": 4, "transformers_version": "4.57.1", "vocab_size": 258, "auto_map": { "AutoModelForCausalLM": "modeling_byteetm.HFByteETM", "AutoConfig": "modeling_byteetm.ByteETMConfig" } }