File size: 421 Bytes
ebccc4e | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 | {
"architectures": [
"RandyGPTForCausalLM"
],
"model_type": "randygpt",
"auto_map": {
"AutoConfig": "modeling_randygpt.RandyGPTConfig",
"AutoModelForCausalLM": "modeling_randygpt.RandyGPTForCausalLM"
},
"vocab_size": 1500,
"n_embd": 128,
"n_head": 4,
"n_layer": 8,
"block_size": 256,
"bos_token_id": 0,
"eos_token_id": 1,
"trained_iters": 1800,
"best_val_loss": 4.479381561279297
} |