midorin-Linux's picture
Update README.md
21fc28a verified
---
license: apache-2.0
datasets:
- TeichAI/gpt-5.2-high-reasoning-250x
- TeichAI/gpt-5.1-codex-max-1000x
- TeichAI/claude-4.5-opus-high-reasoning-250x
- TeichAI/claude-sonnet-4.5-high-reasoning-250x
base_model:
- unsloth/gpt-oss-20b
tags:
- gpt_oss
- openai
- unsloth
- conversational
- code
pipeline_tag: text-generation
library_name: transformers
---
# gpt-oss-20b-Coding-Distill
This project uses Unsloth for fine-tuning. All training data is converted to OpenAI Harmony format before training, but there may be cases where the output format doesn't conform to the OpenAI Harmony specification.
## Do you want to use pre-trained model?
You can download pre-trained data from HuggingFace.
**Safetensors repo**: [midorin-Linux/gpt-oss-20b-Coding-Distill](https://huggingface.co/midorin-Linux/gpt-oss-20b-Coding-Distill)
**GGUF repo**: [midorin-Linux/gpt-oss-20b-Coding-Distill-GGUF](https://huggingface.co/midorin-Linux/gpt-oss-20b-Coding-Distill-GGUF)
## Overview
This project implements a sophisticated multi-phase fine-tuning pipeline for the GPT-OSS-20B model, leveraging conversation data from multiple state-of-the-art AI models to create a balanced, high-performance language model optimized for:
- **Advanced Coding** (via GPT-5.2-codex-max)
- **Complex Reasoning** (via Claude 4.5 Opus and GPT-5.2 high reasoning)
- **Balanced General Intelligence** (via Claude 4.5 Sonnet)
**Why This Approach?**
Traditional fine-tuning often suffers from:
- **Catastrophic forgetting** when training on sequential datasets
- **Imbalanced capabilities** from single-source training
- **Style inconsistencies** across different task types