Model Card for Model ID

This repository hosts a supervised fine-tuned (SFT) version of the Qwen/Qwen3-0.6B-Base language model, trained on the Tandogan/sft_dataset_big dataset.


Model Details

  • Model name: Qwen/Qwen3-0.6B-Base
  • Fine-tuned on: Tandogan/sft_dataset_big

Intended Uses

  • Primary use case: Tasks requiring generation of human-like text in domains covered by the fine-tuning dataset.
  • Examples: Question answering, text summarization, code completion, conversational agents.
  • Not suitable for: Safety-critical applications, generating legal or medical advice without human oversight.

How to Use

You can use the model with the transformers and trl libraries for inference or evaluation:

from transformers import AutoTokenizer, AutoModelForCausalLM

model = AutoModelForCausalLM.from_pretrained("Tandogan/sft_finetuned_big").to("cuda")
tokenizer = AutoTokenizer.from_pretrained("Tandogan/sft_finetuned_big")

prompt = "Explain recursion in simple terms."
inputs = tokenizer(prompt, return_tensors="pt").to("cuda")
outputs = model.generate(**inputs, max_new_tokens=256)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
Downloads last month
9
Safetensors
Model size
0.6B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support