Update README.md
Browse files
README.md
CHANGED
|
@@ -155,10 +155,10 @@ The model was trained on **~157,000 examples** from multiple high-quality Kubern
|
|
| 155 |
- **Base Model**: unsloth/gemma-3-12b-it-qat-bnb-4bit
|
| 156 |
- **Method**: LoRA (Low-Rank Adaptation)
|
| 157 |
- **Framework**: Unsloth
|
| 158 |
-
- **LoRA Rank**:
|
| 159 |
- **Target Modules**: q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj
|
| 160 |
- **Training Checkpoint**: checkpoint-8175
|
| 161 |
-
- **Max Sequence Length**:
|
| 162 |
- **Training Time**: 28 hours
|
| 163 |
- **Hardware**: NVIDIA GeForce RTX 5070 12GB
|
| 164 |
|
|
|
|
| 155 |
- **Base Model**: unsloth/gemma-3-12b-it-qat-bnb-4bit
|
| 156 |
- **Method**: LoRA (Low-Rank Adaptation)
|
| 157 |
- **Framework**: Unsloth
|
| 158 |
+
- **LoRA Rank**: 8
|
| 159 |
- **Target Modules**: q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj
|
| 160 |
- **Training Checkpoint**: checkpoint-8175
|
| 161 |
+
- **Max Sequence Length**: 1024 tokens
|
| 162 |
- **Training Time**: 28 hours
|
| 163 |
- **Hardware**: NVIDIA GeForce RTX 5070 12GB
|
| 164 |
|