--- library_name: peft tags: - axolotl - lora - transformers - gemma pipeline_tag: text-generation base_model: google/gemma-3-27b-it --- # EmoN1 Fine-tuned from [google/gemma-3-27b-it](https://huggingface.co/google/gemma-3-27b-it) using QLoRA. ## Training Details - **Base Model:** google/gemma-3-27b-it - **Method:** QLoRA (4-bit quantization + LoRA) - **LoRA Rank:** 32 - **LoRA Alpha:** 64 - **Sequence Length:** 8192 - **Epochs:** 3 - **Learning Rate:** 2e-4 ### Training Results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 0.9058 | 1.0 | 63 | 0.8959 | | 0.8279 | 2.0 | 126 | 0.8607 | ### Framework Versions - PEFT 0.17.1 - Transformers 4.55.4 - Pytorch 2.7.1+cu126