File size: 4,395 Bytes
b9f1fe3
 
 
 
 
 
 
 
 
 
 
 
5913fdf
 
7736f66
 
 
2c424d6
7736f66
10102c1
7736f66
 
e3162a5
7736f66
ca2be1d
a6b1174
 
7736f66
512a07b
 
ca2be1d
512a07b
7736f66
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0f3a934
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
729977f
 
a03a01f
01240ea
 
c59fc21
 
 
 
4cd94e4
 
 
 
b04b70c
 
 
1803f93
 
 
 
 
 
 
b9f1fe3
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
---
base_model: unsloth/gpt-oss-20b-unsloth-bnb-4bit
tags:
- text-generation-inference
- transformers
- unsloth
- gpt_oss
license: apache-2.0
language:
- en
---

## Model Card
### We release open-weight metatune-gpt20b, fine tuned version of OpenAI's gpt-oss-20b model,  this is one of the first public release recursive self improving AI.
- Generates new data for itself,
- Evaluates its performance, and
- Adjusts its own hyperparameters based on improvement metrics.
- Fine tune automaticlaly using unsloth [SFT tuning techniques](https://docs.unsloth.ai/get-started/fine-tuning-llms-guide)


## Use cases: 
- genuinely demonstrate scientific and mathematical understanding at a postdoctoral level.
- coding
- - Topics: Euler–Lagrange equation, vector calculus, statistical mechanics
 
## additional information
Due to recursive self improvement method, there is no final model, but improved model, this is a 5th metacycle(generation) improved checkpoint model.

## Guardrails:
- generally, please set reasoning = "high", it will usually prevent jailbreaking and prompt injection
- use safety gpt oss 20b for guardrails before this model:  [openai/gpt-oss-safeguard-20b](https://huggingface.co/openai/gpt-oss-safeguard-20b)

# Inference examples

## Transformers

You can use `gpt-oss-120b` and `gpt-oss-20b` with Transformers. If you use the Transformers chat template, it will automatically apply the [harmony response format](https://github.com/openai/harmony). If you use `model.generate` directly, you need to apply the harmony format manually using the chat template or use our [openai-harmony](https://github.com/openai/harmony) package.

To get started, install the necessary dependencies to setup your environment:

```
pip install -U transformers kernels torch 
```

For Google Colab (free/Pro)
```
!pip install -q --upgrade torch

!pip install -q transformers triton==3.4 kernels

!pip uninstall -q torchvision torchaudio -y
```

Once, setup you can proceed to run the model by running the snippet below:

```py
from transformers import pipeline
import torch
model_id = "EpistemeAI/metatune-gpt20b"
pipe = pipeline(
    "text-generation",
    model=model_id,
    torch_dtype="auto",
    device_map="auto",
)
messages = [
    {"role": "user", "content": "Derive the Euler–Lagrange equation from the principle of stationary action.""},
]
outputs = pipe(
    messages,
    max_new_tokens=3000,
)
print(outputs[0]["generated_text"][-1])
```

# Reasoning levels

You can adjust the reasoning level that suits your task across three levels:

* **Low:** Fast responses for general dialogue.  
* **Medium:** Balanced speed and detail.  
* **High:** Deep and detailed analysis.

The reasoning level can be set in the system prompts, e.g., "Reasoning: high".

# Tool use

The gpt-oss models are excellent for:
* Web browsing (using built-in browsing tools)
* Function calling with defined schemas
* Agentic operations like browser tasks

# Fine-tuning

Both gpt-oss models can be fine-tuned for a variety of specialized use cases.

This smaller model `gpt-oss-20b` can be fine-tuned on consumer hardware, whereas the larger [`gpt-oss-120b`](https://huggingface.co/openai/gpt-oss-120b) can be fine-tuned on a single H100 node.



## Benchmark
These benchmark are current benchmark and not final benchmark, due to recursive fine tuning techniques self improves over time:

hf (pretrained=EpistemeAI/metatune-gpt20b-R0,parallelize=True,dtype=bfloat16), gen_kwargs: (temperature=1,top_p=1,max_new_tokens=1000), limit: 30.0, num_fewshot: 5, batch_size: 1

|             Tasks           |metatune|MiniMax M1 80k|Llama 4 Maverick|
|:----------------------------|:-----|:-----|:----- |
|gsm8k_cot                    |0.91  |   -  |   -  |
|gpqa_diamond_cot_n_shot      |0.722 |0.70  |0.67|
|winigrande     |0.785| -  |-|
|hellaswag      |0.421| -  |-|
|arc_challenge  |0.349| -  |-|




## Thank you
- OpenAI
- Unsloth
- Google Colab
- Nvidia for A100


# Uploaded finetuned  model

- **Developed by:** EpistemeAI
- **License:** apache-2.0
- **Finetuned from model :** unsloth/gpt-oss-20b-unsloth-bnb-4bit

This gpt_oss model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.

[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)