kernelpool commited on
Commit
dbe0aaf
·
verified ·
1 Parent(s): 98e1d84

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +32 -0
README.md CHANGED
@@ -7,3 +7,35 @@ tags:
7
  - mlx
8
  base_model: meituan-longcat/LongCat-Flash-Thinking-ZigZag
9
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
  - mlx
8
  base_model: meituan-longcat/LongCat-Flash-Thinking-ZigZag
9
  ---
10
+
11
+ # mlx-community/LongCat-Flash-Thinking-ZigZag-6bit
12
+
13
+ This model [mlx-community/LongCat-Flash-Thinking-ZigZag-6bit](https://huggingface.co/mlx-community/LongCat-Flash-Thinking-ZigZag-6bit) was
14
+ converted to MLX format from [meituan-longcat/LongCat-Flash-Thinking-ZigZag](https://huggingface.co/meituan-longcat/LongCat-Flash-Thinking-ZigZag)
15
+ using mlx-lm version **0.30.5**.
16
+
17
+ ## Warning
18
+
19
+ This model is not yet compatible with MLX-LM, but you can track its progress [here](https://github.com/ml-explore/mlx-lm/pull/802) or use [this branch](https://github.com/kernelpool/mlx-lm/tree/feature/longcat-zigzag-attn) to test it.
20
+
21
+ ## Use with mlx
22
+
23
+ ```bash
24
+ pip install mlx-lm
25
+ ```
26
+
27
+ ```python
28
+ from mlx_lm import load, generate
29
+
30
+ model, tokenizer = load("mlx-community/LongCat-Flash-Thinking-ZigZag-6bit")
31
+
32
+ prompt = "hello"
33
+
34
+ if tokenizer.chat_template is not None:
35
+ messages = [{"role": "user", "content": prompt}]
36
+ prompt = tokenizer.apply_chat_template(
37
+ messages, add_generation_prompt=True, return_dict=False,
38
+ )
39
+
40
+ response = generate(model, tokenizer, prompt=prompt, verbose=True)
41
+ ```