Vicens commited on
Commit
036f667
·
verified ·
1 Parent(s): e951e23

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -23,7 +23,7 @@ base_model: openai/gpt-oss-120b
23
  Engineered by [SCALAI](https://scalai.es), this model was surgically distilled from OpenAI's dense 117B parameter MoE (`gpt-oss-120b`) down to a 60B active parameter footprint. Quantized in MXFP4, **ScaLite-60B-Coder requires only ~30GB of VRAM, making it fully deployable on a single NVIDIA L40S (48GB) GPU** with ample room for large KV-caches in production environments.
24
 
25
  ## 🧠 Model Details
26
- * **Developer:** SCALAI (Vicens Gaitan)
27
  * **Model Type:** Pruned Mixture-of-Experts (MoE) Causal Language Model
28
  * **Base Model:** `openai/gpt-oss-120b` (128 experts)
29
  * **Pruned Architecture:** 60B active parameters (64 experts)
 
23
  Engineered by [SCALAI](https://scalai.es), this model was surgically distilled from OpenAI's dense 117B parameter MoE (`gpt-oss-120b`) down to a 60B active parameter footprint. Quantized in MXFP4, **ScaLite-60B-Coder requires only ~30GB of VRAM, making it fully deployable on a single NVIDIA L40S (48GB) GPU** with ample room for large KV-caches in production environments.
24
 
25
  ## 🧠 Model Details
26
+ * **Developer:** SCALAI
27
  * **Model Type:** Pruned Mixture-of-Experts (MoE) Causal Language Model
28
  * **Base Model:** `openai/gpt-oss-120b` (128 experts)
29
  * **Pruned Architecture:** 60B active parameters (64 experts)