Uploaded Model

  • Developed by: Harsha901
  • License: Apache-2.0
  • Finetuned from model: unsloth/Qwen2.5-7B-Instruct

This Qwen2.5-7B model was fine-tuned using Unsloth for faster and more memory-efficient training, together with Hugging Faceโ€™s TRL library for supervised fine-tuning.


Model Overview

This is an instruction-tuned causal language model based on Qwen2.5-7B, designed to follow user prompts accurately and generate coherent, high-quality responses.

The model preserves the general-purpose strengths of Qwen2.5 while benefiting from domain-focused supervised fine-tuning.


Training Details

  • Base model: Qwen2.5-7B-Instruct (Unsloth variant)
  • Fine-tuning method: Supervised Fine-Tuning (SFT)
  • Frameworks: Hugging Face Transformers + TRL
  • Acceleration: Unsloth (2ร— faster training, reduced VRAM usage)
  • Precision: FP16 / BF16 (hardware dependent)

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model_id = "Harsha901/<YOUR-MODEL-NAME>"

tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
    model_id,
    device_map="auto",
    torch_dtype="auto"
)

Limitations

  • Outputs may contain factual or reasoning errors
  • Not intended for high-stakes or safety-critical applications
  • Performance depends on prompt quality and context length

License

Released under the Apache 2.0 License, consistent with the base Qwen2.5 model.


Acknowledgements

  • Qwen Team for the Qwen2.5 base model
  • Unsloth for efficient fine-tuning optimizations
  • Hugging Face for the training and hosting ecosystem

Downloads last month
20
Safetensors
Model size
8B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for Harsha901/Qwen2.5-7B-Inst-Math-Reasoning-SFT

Base model

Qwen/Qwen2.5-7B
Finetuned
(1290)
this model