(Dataset based on Pinkstack/syngen-reasoning-0.6b-dataset)

This is a 1.7B parameter LLM designed for synthetic grounded reasoning generation between the user prompt and the final model output, specifically for dataset modifications, but can be used for multiple use cases which require reasoning.

For example, this model allows you to turn any chat dataset into a reasoning dataset as if it was generated by DeepSeek R1 or OpenAI's GPT OSS!

Sampler Settings: Pretty standard, temp = 0.7, top_p = 0.95.


Prompt Format

System Message

<reasoning_style>deepseek_r1</reasoning_style> # Can replace deepseek_r1 with gpt_oss
<system_prompt>Original System Prompt</system_prompt>

Prompt Message

<user>User Message Here</user>
<assistant>Assistant Final Response Here (without reasoning)</assistant>

Output Format

<think>Generated Reasoning</think>

Training Details

  • Base Model: Qwen/Qwen3-1.7B
  • Training Epochs: 0.4
  • Learning Rate: 2e-5
  • Batch Size: 32
  • Training Method: Full Fine-Tune (FFT)
  • Training Hardware: 2 x A100-40GB
  • Training Platform: Burla
  • Total Cost: $0.00 USD
  • Seed: 42
Downloads last month
15
Safetensors
Model size
2B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for qingy2024/SynGen-Nano-1.7B

Finetuned
Qwen/Qwen3-1.7B
Finetuned
(411)
this model
Quantizations
2 models

Dataset used to train qingy2024/SynGen-Nano-1.7B

Collection including qingy2024/SynGen-Nano-1.7B