Gemma 4 E4B Opus4.6 Reasoning

A PEFT LoRA adapter fine-tuned on top of google/gemma-4-e4b-it using the Crownelius/Opus-4.6-Reasoning-2100x-formatted dataset.

This adapter is optimized for:

  • structured step-by-step reasoning
  • logic puzzles
  • planning and decomposition
  • algorithm explanations
  • conceptual problem solving
  • code reasoning workflows

The strongest improvements are visible on:

  • multi-step logic puzzles
  • algorithm design explanations
  • state-tracking tasks
  • proof-style conceptual reasoning

The adapter shows strongest gains on deliberate decomposition, planning, and educational reasoning prompts.

Base Model

  • google/gemma-4-e4b-it

Dataset

  • Crownelius/Opus-4.6-Reasoning-2100x-formatted

Training Setup

  • PEFT LoRA fine-tuning
  • 4-bit QLoRA loading
  • 2 training epochs
  • training max sequence length: 512 tokens
  • gradient accumulation: 16
  • trained on Google Colab T4

Training Metrics

  • training loss: 192.38
  • validation loss: 11.95
  • entropy: 3.91
  • mean token accuracy: 0.0462
  • train runtime: 5783 seconds
  • train rows: 2010
  • validation rows: 106

Example Use

from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel

base_model = AutoModelForCausalLM.from_pretrained(
    "google/gemma-4-e4b-it",
    device_map="auto"
)

model = PeftModel.from_pretrained(
    base_model,
    "krishnamraja13/gemma-4-e4b-opus46-reasoning"
)

tokenizer = AutoTokenizer.from_pretrained(
    "krishnamraja13/gemma-4-e4b-opus46-reasoning"
)

Requirements

Use a recent version of transformers with Gemma 4 support.

pip install -U transformers peft accelerate bitsandbytes

Known Strengths

This adapter performs best on:

  • logic riddles
  • switch / state puzzles
  • recursive explanation prompts
  • dynamic programming intuition
  • binary search reasoning
  • linked list cycle detection explanations
  • proof-style educational prompts
  • intermediate reasoning scaffolds and invariant-based explanations

Known Limitations

The adapter is stronger at:

  • structured reasoning
  • decomposition
  • planning
  • conceptual explanation

than strict symbolic algebra fidelity.

For exact equation solving, outputs may sometimes over-interpret terse symbolic prompts.


License

This adapter is a derivative of Gemma 4 and follows the Gemma license terms.

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for krishnamraja13/gemma-4-e4b-opus46-reasoning

Adapter
(8)
this model

Dataset used to train krishnamraja13/gemma-4-e4b-opus46-reasoning