{ "architectures": [ "NepaliGPTForCausalLM" ], "auto_map": { "AutoConfig": "configuration_nepali_gpt.NepaliGPTConfig", "AutoModelForCausalLM": "modeling_nepali_gpt.NepaliGPTForCausalLM" }, "context_length": 1024, "drop_rate": 0.1, "emb_dim": 768, "model_type": "nep_gptv1", "n_heads": 12, "n_layers": 9, "qkv_bias": false, "torch_dtype": "float32", "transformers_version": "4.51.0.dev0", "vocab_size": 51728 }