| { | |
| "architectures": [ | |
| "DeepseekForCausalLM" | |
| ], | |
| "model_type": "deepseek", | |
| "base_model": "deepseek-ai/deepseek-coder-1.3b-base", | |
| "fine_tuning": { | |
| "type": "LoRA", | |
| "r": 16, | |
| "alpha": 32, | |
| "dropout": 0.05 | |
| }, | |
| "generation_config": { | |
| "temperature": 0.7, | |
| "top_p": 0.95, | |
| "max_length": 512 | |
| } | |
| } |