AIGym/free-gpt-oss
Viewer • Updated • 5.54k • 4
How to use AIGym/oss-adapter with Transformers:
# Load model directly
from transformers import AutoModel
model = AutoModel.from_pretrained("AIGym/oss-adapter", dtype="auto")AIGym/oss-adapter
openai/gpt-oss-20b using supervised fine-tuning (SFT) on the AIGym/free-gpt-oss dataset ([Hugging Face][1]).Fine-tuning method: Supervised fine-tuning (SFT) using the TRL library ([Hugging Face][1]).
Tooling and versions:
Dataset: AIGym/free-gpt-oss, which presumably includes examples crafted to expose harmful behaviors in the base GPT-OSS-20B model (specific content should be described here if available).
from transformers import pipeline
generator = pipeline(
"text-generation",
model="AIGym/oss-multi-lingual", # Or "AIGym/oss-adapter" depending on naming
device="cuda"
)
question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
output = generator(
[{"role": "user", "content": question}],
max_new_tokens=128,
return_full_text=False
)[0]
print(output["generated_text"])
This snippet demonstrates how to query the model in an interactive pipeline, useful for both red-teaming experiments and exploratory analysis ([Hugging Face][1]).
| Section | Highlights |
|---|---|
| Overview | Fine-tuned GPT-OSS-20B adapter for red-teaming, using AIGym dataset |
| Motivation | Built for the Kaggle Red-Teaming Challenge targeting safety analysis |
| Tools & Versions | TRL 0.21.0, Transformers 4.55.2, PyTorch dev build, Datasets 4.0.0 etc. |
| Usage Example | Provided pipeline snippet for quick start |
| Caveats | Generates potentially harmful outputs; meant only for controlled eval |
| Citation | TRL GitHub repository |
Base model
openai/gpt-oss-20b
# Load model directly from transformers import AutoModel model = AutoModel.from_pretrained("AIGym/oss-adapter", dtype="auto")