Update README.md
Browse files
README.md
CHANGED
|
@@ -23,7 +23,7 @@ base_model: openai/gpt-oss-120b
|
|
| 23 |
Engineered by [SCALAI](https://scalai.es), this model was surgically distilled from OpenAI's dense 117B parameter MoE (`gpt-oss-120b`) down to a 60B active parameter footprint. Quantized in MXFP4, **ScaLite-60B-Coder requires only ~30GB of VRAM, making it fully deployable on a single NVIDIA L40S (48GB) GPU** with ample room for large KV-caches in production environments.
|
| 24 |
|
| 25 |
## 🧠 Model Details
|
| 26 |
-
* **Developer:** SCALAI
|
| 27 |
* **Model Type:** Pruned Mixture-of-Experts (MoE) Causal Language Model
|
| 28 |
* **Base Model:** `openai/gpt-oss-120b` (128 experts)
|
| 29 |
* **Pruned Architecture:** 60B active parameters (64 experts)
|
|
|
|
| 23 |
Engineered by [SCALAI](https://scalai.es), this model was surgically distilled from OpenAI's dense 117B parameter MoE (`gpt-oss-120b`) down to a 60B active parameter footprint. Quantized in MXFP4, **ScaLite-60B-Coder requires only ~30GB of VRAM, making it fully deployable on a single NVIDIA L40S (48GB) GPU** with ample room for large KV-caches in production environments.
|
| 24 |
|
| 25 |
## 🧠 Model Details
|
| 26 |
+
* **Developer:** SCALAI
|
| 27 |
* **Model Type:** Pruned Mixture-of-Experts (MoE) Causal Language Model
|
| 28 |
* **Base Model:** `openai/gpt-oss-120b` (128 experts)
|
| 29 |
* **Pruned Architecture:** 60B active parameters (64 experts)
|