| --- |
| library_name: transformers |
| pipeline_tag: text-generation |
| license: gemma |
| datasets: |
| - Crownelius/Opus-4.6-Reasoning-2100x-formatted |
| base_model: |
| - google/gemma-4-E4B-it |
| base_model_relation: adapter |
| tags: |
| - gemma |
| - peft |
| - lora |
| - reasoning |
| - puzzle-solving |
| - code |
| --- |
| |
| # Gemma 4 E4B Opus4.6 Reasoning |
|
|
| A PEFT LoRA adapter fine-tuned on top of `google/gemma-4-e4b-it` using the `Crownelius/Opus-4.6-Reasoning-2100x-formatted` dataset. |
|
|
| This adapter is optimized for: |
|
|
| - structured step-by-step reasoning |
| - logic puzzles |
| - planning and decomposition |
| - algorithm explanations |
| - conceptual problem solving |
| - code reasoning workflows |
|
|
| The strongest improvements are visible on: |
| - multi-step logic puzzles |
| - algorithm design explanations |
| - state-tracking tasks |
| - proof-style conceptual reasoning |
|
|
| The adapter shows strongest gains on deliberate decomposition, planning, and educational reasoning prompts. |
| --- |
|
|
| ## Base Model |
| - `google/gemma-4-e4b-it` |
|
|
| ## Dataset |
| - `Crownelius/Opus-4.6-Reasoning-2100x-formatted` |
|
|
| ## Training Setup |
| - PEFT LoRA fine-tuning |
| - 4-bit QLoRA loading |
| - 2 training epochs |
| - training max sequence length: 512 tokens |
| - gradient accumulation: 16 |
| - trained on Google Colab T4 |
|
|
| --- |
|
|
| ## Training Metrics |
| - training loss: 192.38 |
| - validation loss: 11.95 |
| - entropy: 3.91 |
| - mean token accuracy: 0.0462 |
| - train runtime: 5783 seconds |
| - train rows: 2010 |
| - validation rows: 106 |
|
|
| ## Example Use |
|
|
| ```python |
| from transformers import AutoModelForCausalLM, AutoTokenizer |
| from peft import PeftModel |
| |
| base_model = AutoModelForCausalLM.from_pretrained( |
| "google/gemma-4-e4b-it", |
| device_map="auto" |
| ) |
| |
| model = PeftModel.from_pretrained( |
| base_model, |
| "krishnamraja13/gemma-4-e4b-opus46-reasoning" |
| ) |
| |
| tokenizer = AutoTokenizer.from_pretrained( |
| "krishnamraja13/gemma-4-e4b-opus46-reasoning" |
| ) |
| ``` |
|
|
| ## Requirements |
|
|
| Use a recent version of `transformers` with Gemma 4 support. |
|
|
| ```bash |
| pip install -U transformers peft accelerate bitsandbytes |
| ``` |
|
|
| --- |
|
|
| ## Known Strengths |
| This adapter performs best on: |
|
|
| - logic riddles |
| - switch / state puzzles |
| - recursive explanation prompts |
| - dynamic programming intuition |
| - binary search reasoning |
| - linked list cycle detection explanations |
| - proof-style educational prompts |
| - intermediate reasoning scaffolds and invariant-based explanations |
| --- |
|
|
| ## Known Limitations |
| The adapter is stronger at: |
|
|
| - structured reasoning |
| - decomposition |
| - planning |
| - conceptual explanation |
|
|
| than strict symbolic algebra fidelity. |
|
|
| For exact equation solving, outputs may sometimes over-interpret terse symbolic prompts. |
|
|
| --- |
|
|
| ## License |
| This adapter is a derivative of Gemma 4 and follows the Gemma license terms. |