| # Fine-tuned Model: Legal-gemma3-12b-it-lora-thinking | |
| ## 📚 Training Configuration | |
| - **data_path**: `QomSSLab/Legal_DS_SFT` | |
| - **output_dir**: `gemma312b_lora_chckpnts` | |
| - **new_model_name**: `Legal-gemma3-12b-it-lora-thinking` | |
| - **data_ratio**: `1.0` | |
| - **model_name**: `QomSSLab/Legal-gemma3-12b-pt-v2` | |
| - **use_4bit**: `False` | |
| - **use_lora**: `True` | |
| - **max_seq_length**: `8500` | |
| - **batch_size**: `1` | |
| - **gradient_accu**: `8` | |
| - **epochs**: `2` | |
| - **learning_rate**: `5e-05` | |
| - **lora_alpha**: `64` | |
| - **lora_drop**: `0.05` | |
| - **lora_r**: `64` | |
| - **tune_embedding_layer**: `False` | |
| - **hf_token**: `********` | |
| - **resume_from_checkpoint**: `False` | |
| - **use_8bit_optimizer**: `True` | |
| - **push_to_hub**: `True` | |
| --- | |
| Auto-generated after training. | |