| { | |
| "model_type": "causal_lm", | |
| "base_model": "Open-Orca/Mistral-7B-OpenOrca", | |
| "quantization": { | |
| "load_in_8bit": true, | |
| "quantization_type": "standard_8bit", | |
| "compute_dtype": "bfloat16" | |
| }, | |
| "fine_tuning": { | |
| "method": "PEFT", | |
| "technique": "LoRA", | |
| "lora_rank": 16, | |
| "lora_alpha": 16, | |
| "target_modules": [ | |
| "qproj", | |
| "kproj", | |
| "vproj", | |
| "oproj", | |
| "gateproj", | |
| "upproj", | |
| "downproj" | |
| ], | |
| "dropout": 0.05 | |
| }, | |
| "device_map": "auto", | |
| "trust_remote_code": true, | |
| "tokenizer_options": { | |
| "use_fast": true | |
| } | |
| } | |