midorin-Linux's picture
Update README.md
21fc28a verified
metadata
license: apache-2.0
datasets:
  - TeichAI/gpt-5.2-high-reasoning-250x
  - TeichAI/gpt-5.1-codex-max-1000x
  - TeichAI/claude-4.5-opus-high-reasoning-250x
  - TeichAI/claude-sonnet-4.5-high-reasoning-250x
base_model:
  - unsloth/gpt-oss-20b
tags:
  - gpt_oss
  - openai
  - unsloth
  - conversational
  - code
pipeline_tag: text-generation
library_name: transformers

gpt-oss-20b-Coding-Distill

This project uses Unsloth for fine-tuning. All training data is converted to OpenAI Harmony format before training, but there may be cases where the output format doesn't conform to the OpenAI Harmony specification.

Do you want to use pre-trained model?

You can download pre-trained data from HuggingFace.

Safetensors repo: midorin-Linux/gpt-oss-20b-Coding-Distill
GGUF repo: midorin-Linux/gpt-oss-20b-Coding-Distill-GGUF

Overview

This project implements a sophisticated multi-phase fine-tuning pipeline for the GPT-OSS-20B model, leveraging conversation data from multiple state-of-the-art AI models to create a balanced, high-performance language model optimized for:

  • Advanced Coding (via GPT-5.2-codex-max)
  • Complex Reasoning (via Claude 4.5 Opus and GPT-5.2 high reasoning)
  • Balanced General Intelligence (via Claude 4.5 Sonnet)

Why This Approach? Traditional fine-tuning often suffers from:

  • Catastrophic forgetting when training on sequential datasets
  • Imbalanced capabilities from single-source training
  • Style inconsistencies across different task types