File size: 1,645 Bytes
2d4deab
 
 
 
 
cebe538
 
2d4deab
 
f023f25
 
 
 
8ede691
 
 
 
dfbd05c
 
 
 
e6eb4e7
 
 
e7b389a
 
e6eb4e7
dfbd05c
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
---
license: apache-2.0
datasets:
- TeichAI/gpt-5.2-high-reasoning-250x
- TeichAI/gpt-5.1-codex-max-1000x
- TeichAI/claude-4.5-opus-high-reasoning-250x
- TeichAI/claude-sonnet-4.5-high-reasoning-250x
base_model:
- unsloth/gpt-oss-20b
tags:
- gpt_oss
- openai
- unsloth
- conversational
- code
pipeline_tag: text-generation
library_name: transformers
---
# gpt-oss-20b-Coding-Distill
This project uses Unsloth for fine-tuning. All training data is converted to OpenAI Harmony format before training, but there may be cases where the output format doesn't conform to the OpenAI Harmony specification.

## Do you want to use pre-trained model?
You can download pre-trained data from HuggingFace.  
  
**Safetensors repo**: [midorin-Linux/gpt-oss-20b-Coding-Distill](https://huggingface.co/midorin-Linux/gpt-oss-20b-Coding-Distill)  
**GGUF repo**: [midorin-Linux/gpt-oss-20b-Coding-Distill-GGUF](https://huggingface.co/midorin-Linux/gpt-oss-20b-Coding-Distill-GGUF)

## Overview
This project implements a sophisticated multi-phase fine-tuning pipeline for the GPT-OSS-20B model, leveraging conversation data from multiple state-of-the-art AI models to create a balanced, high-performance language model optimized for:

- **Advanced Coding** (via GPT-5.2-codex-max)
- **Complex Reasoning** (via Claude 4.5 Opus and GPT-5.2 high reasoning)
- **Balanced General Intelligence** (via Claude 4.5 Sonnet)

**Why This Approach?**
Traditional fine-tuning often suffers from:

- **Catastrophic forgetting** when training on sequential datasets
- **Imbalanced capabilities** from single-source training
- **Style inconsistencies** across different task types