Upload README.md with huggingface_hub
Browse files
README.md
ADDED
|
@@ -0,0 +1,49 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
title: JANGQ-AI
|
| 3 |
+
---
|
| 4 |
+
|
| 5 |
+
<p align="center">
|
| 6 |
+
<a href="https://mlx.studio"><img src="https://raw.githubusercontent.com/jjang-ai/jangq/main/assets/mlx-studio-light.png" alt="MLX Studio" width="500"></a>
|
| 7 |
+
</p>
|
| 8 |
+
|
| 9 |
+
<p align="center">
|
| 10 |
+
<a href="https://mlx.studio"><img src="https://mlx.studio/assets/screenshots/mlx-studio-featured.png?v=1" alt="MLX Studio App" width="600"></a>
|
| 11 |
+
</p>
|
| 12 |
+
|
| 13 |
+
<h3 align="center"><a href="https://mlx.studio">MLX Studio</a> — the only app that natively supports JANG models</h3>
|
| 14 |
+
|
| 15 |
+
---
|
| 16 |
+
|
| 17 |
+
> **LM Studio, Ollama, oMLX, Inferencer** and other MLX apps do **not** support JANG yet. Use [MLX Studio](https://mlx.studio) for native JANG support, or pip install jang for Python inference. **Ask your favorite app's creators to add JANG support!**
|
| 18 |
+
|
| 19 |
+
---
|
| 20 |
+
|
| 21 |
+
<p align="center">
|
| 22 |
+
<img src="https://raw.githubusercontent.com/jjang-ai/jangq/main/assets/jangq-logo-dark.png" alt="JANG" width="300">
|
| 23 |
+
</p>
|
| 24 |
+
|
| 25 |
+
# JANGQ-AI — JANG Quantized Models for Apple Silicon
|
| 26 |
+
|
| 27 |
+
**JANG** (**J**ang **A**daptive **N**-bit **G**rading) — the GGUF equivalent for MLX.
|
| 28 |
+
|
| 29 |
+
Same size as MLX, smarter bit allocation. Models stay quantized in GPU memory at full Metal speed.
|
| 30 |
+
|
| 31 |
+
## Install
|
| 32 |
+
|
| 33 |
+
```
|
| 34 |
+
pip install "jang[mlx]"
|
| 35 |
+
```
|
| 36 |
+
|
| 37 |
+
## Models
|
| 38 |
+
|
| 39 |
+
| Model | Profile | MMLU | HumanEval | Size |
|
| 40 |
+
|-------|---------|------|-----------|------|
|
| 41 |
+
| [Qwen3.5-122B-A10B-JANG_2S](https://huggingface.co/JANGQ-AI/Qwen3.5-122B-A10B-JANG_2S) | 2-bit | **84%** | **90%** | 38 GB |
|
| 42 |
+
| [Qwen3.5-35B-A3B-JANG_4K](https://huggingface.co/JANGQ-AI/Qwen3.5-35B-A3B-JANG_4K) | 4-bit K-quant | **84%** | 90% | 16.7 GB |
|
| 43 |
+
| [Qwen3.5-35B-A3B-JANG_2S](https://huggingface.co/JANGQ-AI/Qwen3.5-35B-A3B-JANG_2S) | 2-bit | 62% | — | 12 GB |
|
| 44 |
+
|
| 45 |
+
## Links
|
| 46 |
+
|
| 47 |
+
[GitHub](https://github.com/jjang-ai/jangq) · [PyPI](https://pypi.org/project/jang/) · [MLX Studio](https://mlx.studio)
|
| 48 |
+
|
| 49 |
+
Created by Jinho Jang — [jangq.ai](https://jangq.ai)
|