|
|
--- |
|
|
license: apache-2.0 |
|
|
datasets: |
|
|
- TeichAI/gpt-5.2-high-reasoning-250x |
|
|
- TeichAI/gpt-5.1-codex-max-1000x |
|
|
- TeichAI/claude-4.5-opus-high-reasoning-250x |
|
|
- TeichAI/claude-sonnet-4.5-high-reasoning-250x |
|
|
base_model: |
|
|
- unsloth/gpt-oss-20b |
|
|
tags: |
|
|
- gpt_oss |
|
|
- openai |
|
|
- unsloth |
|
|
- conversational |
|
|
- code |
|
|
pipeline_tag: text-generation |
|
|
library_name: transformers |
|
|
--- |
|
|
# gpt-oss-20b-Coding-Distill |
|
|
This project uses Unsloth for fine-tuning. All training data is converted to OpenAI Harmony format before training, but there may be cases where the output format doesn't conform to the OpenAI Harmony specification. |
|
|
|
|
|
## Do you want to use pre-trained model? |
|
|
You can download pre-trained data from HuggingFace. |
|
|
|
|
|
**Safetensors repo**: [midorin-Linux/gpt-oss-20b-Coding-Distill](https://huggingface.co/midorin-Linux/gpt-oss-20b-Coding-Distill) |
|
|
**GGUF repo**: [midorin-Linux/gpt-oss-20b-Coding-Distill-GGUF](https://huggingface.co/midorin-Linux/gpt-oss-20b-Coding-Distill-GGUF) |
|
|
|
|
|
## Overview |
|
|
This project implements a sophisticated multi-phase fine-tuning pipeline for the GPT-OSS-20B model, leveraging conversation data from multiple state-of-the-art AI models to create a balanced, high-performance language model optimized for: |
|
|
|
|
|
- **Advanced Coding** (via GPT-5.2-codex-max) |
|
|
- **Complex Reasoning** (via Claude 4.5 Opus and GPT-5.2 high reasoning) |
|
|
- **Balanced General Intelligence** (via Claude 4.5 Sonnet) |
|
|
|
|
|
**Why This Approach?** |
|
|
Traditional fine-tuning often suffers from: |
|
|
|
|
|
- **Catastrophic forgetting** when training on sequential datasets |
|
|
- **Imbalanced capabilities** from single-source training |
|
|
- **Style inconsistencies** across different task types |
|
|
|