{ "architectures": [ "ChessForCausalLM" ], "auto_map": { "AutoConfig": "model.ChessConfig", "AutoModelForCausalLM": "model.ChessForCausalLM" }, "dropout": 0.1, "dtype": "float32", "model_type": "chess_gpt", "n_embd": 120, "n_head": 8, "n_layer": 5, "n_positions": 256, "transformers_version": "4.57.6", "vocab_size": 255 }