| tags: | |
| - generative | |
| - text-generation | |
| - openai | |
| - gpt-oss | |
| - reap | |
| - bf16 | |
| metrics: | |
| - perplexity | |
| - accuracy | |
| model-index: | |
| - name: gpt-oss-20b-reap-0.4-bf16 | |
| results: [] | |
| # gpt-oss-20b-reap-0.4-bf16 | |
| This repository contains a bfloat16 version of the `sandeshrajx/gpt-oss-20b-reap-0.4-mxfp4` model. | |
| ## Model Description | |
| This model is a bfloat16 version of the MXFP4 quantized `openai/gpt-oss-20b` model. | |
| - **Original Model:** `openai/gpt-oss-20b` | |
| - **Pruning Method:** `reap` with a compression ratio of `0.4` | |
| - **Original Quantization Method:** MXFP4 weight-only quantization | |
| - **Current Format:** bfloat16 | |
| - **Dataset used for pruning/quantization (if applicable):** `theblackcat102/evol-codealpaca-v1` | |
| ## Usage | |
| You can load this model using the `transformers` library: | |
| ```python | |
| from transformers import AutoModelForCausalLM, AutoTokenizer | |
| model_name = "sandeshrajx/gpt-oss-20b-reap-0.4-bf16" | |
| tokenizer = AutoTokenizer.from_pretrained(model_name) | |
| model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.bfloat16) | |
| # Use the model for inference | |
| # ... | |
| ``` | |
| ## License | |
| (Please specify the license of the original model and any modifications) | |