jangq commited on
Commit
570d7d4
·
verified ·
1 Parent(s): aa8ab40

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +49 -0
README.md ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ title: JANGQ-AI
3
+ ---
4
+
5
+ <p align="center">
6
+ <a href="https://mlx.studio"><img src="https://raw.githubusercontent.com/jjang-ai/jangq/main/assets/mlx-studio-light.png" alt="MLX Studio" width="500"></a>
7
+ </p>
8
+
9
+ <p align="center">
10
+ <a href="https://mlx.studio"><img src="https://mlx.studio/assets/screenshots/mlx-studio-featured.png?v=1" alt="MLX Studio App" width="600"></a>
11
+ </p>
12
+
13
+ <h3 align="center"><a href="https://mlx.studio">MLX Studio</a> — the only app that natively supports JANG models</h3>
14
+
15
+ ---
16
+
17
+ > **LM Studio, Ollama, oMLX, Inferencer** and other MLX apps do **not** support JANG yet. Use [MLX Studio](https://mlx.studio) for native JANG support, or pip install jang for Python inference. **Ask your favorite app's creators to add JANG support!**
18
+
19
+ ---
20
+
21
+ <p align="center">
22
+ <img src="https://raw.githubusercontent.com/jjang-ai/jangq/main/assets/jangq-logo-dark.png" alt="JANG" width="300">
23
+ </p>
24
+
25
+ # JANGQ-AI — JANG Quantized Models for Apple Silicon
26
+
27
+ **JANG** (**J**ang **A**daptive **N**-bit **G**rading) — the GGUF equivalent for MLX.
28
+
29
+ Same size as MLX, smarter bit allocation. Models stay quantized in GPU memory at full Metal speed.
30
+
31
+ ## Install
32
+
33
+ ```
34
+ pip install "jang[mlx]"
35
+ ```
36
+
37
+ ## Models
38
+
39
+ | Model | Profile | MMLU | HumanEval | Size |
40
+ |-------|---------|------|-----------|------|
41
+ | [Qwen3.5-122B-A10B-JANG_2S](https://huggingface.co/JANGQ-AI/Qwen3.5-122B-A10B-JANG_2S) | 2-bit | **84%** | **90%** | 38 GB |
42
+ | [Qwen3.5-35B-A3B-JANG_4K](https://huggingface.co/JANGQ-AI/Qwen3.5-35B-A3B-JANG_4K) | 4-bit K-quant | **84%** | 90% | 16.7 GB |
43
+ | [Qwen3.5-35B-A3B-JANG_2S](https://huggingface.co/JANGQ-AI/Qwen3.5-35B-A3B-JANG_2S) | 2-bit | 62% | — | 12 GB |
44
+
45
+ ## Links
46
+
47
+ [GitHub](https://github.com/jjang-ai/jangq) · [PyPI](https://pypi.org/project/jang/) · [MLX Studio](https://mlx.studio)
48
+
49
+ Created by Jinho Jang — [jangq.ai](https://jangq.ai)