WithIn US AI (((GGUF MODELS))
Collection
LLM MODELS TRAINED, FINE-TUNED, MERGED BY (WITHIN US AI) • 6 items • Updated
• 1
YAML Metadata Warning: empty or missing yaml metadata in repo card
Check out the documentation for more information.
language:
pipeline_tag: text-generation
tags:
license: other license_name: withinusai-custom-license license_link: LICENSE
base_model: WithinUsAI/GPT2.5.2-high-reasoning-codex-0.4B base_model_relation: quantized
metrics:
GGUF quantizations of the GPT-2 Medium → “GPT-5.2 twin target” finetune.
Pick your quant, run local, move fast. ⚡🧠
This repository provides GGUF quantizations for local inference (llama.cpp ecosystem) of:
WithinUsAI/GPT2.5.2-high-reasoning-codex-0.4B (source Transformers model)WithinUsAI/GPT2.5.2-high-reasoning-codex-0.4Bopenai-community/gpt2-medium| Quant | Bits | Size |
|---|---|---|
| Q4_K_M | 4-bit | 242 MB |
| Q5_K_M | 5-bit | 274 MB |
| F16 | 16-bit | 714 MB |
Replace MODEL.gguf with the quant file you downloaded:
./llama-cli -m MODEL.gguf \
-p "You are a senior engineer. List edge cases, then write the code.\nTask: Implement an LRU cache in Python.\n\nAnswer:\n" \
-n 256
4-bit
5-bit
16-bit