bibproj commited on
Commit
982f99b
·
verified ·
1 Parent(s): ab3e276

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +37 -3
README.md CHANGED
@@ -1,7 +1,41 @@
1
  ---
2
- language: en
 
 
3
  tags:
 
4
  - mlx
5
- pipeline_tag: text-generation
6
- library_name: mlx
 
 
 
 
 
 
 
 
 
7
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ license: mit
3
+ library_name: mlx
4
+ pipeline_tag: text-generation
5
  tags:
6
+ - transformers
7
  - mlx
8
+ base_model:
9
+ - XiaomiMiMo/MiMo-V2-Flash
10
+ ---
11
+
12
+ # mlx-community/MiMo-V2-Flash-mlx-8bit
13
+
14
+ This model [mlx-community/MiMo-V2-Flash-mlx-8bit-](https://huggingface.co/mlx-community/MiMo-V2-Flash-mlx-8bit) was converted to MLX format from [XiaomiMiMo/MiMo-V2-Flash](https://huggingface.co/XiaomiMiMo/MiMo-V2-Flash)
15
+ using mlx-lm version **0.30.0**.
16
+
17
+ You can find more similar MLX model quants for a single Apple Mac Studio M3 Ultra with 512 GB at https://huggingface.co/bibproj
18
+
19
  ---
20
+
21
+ ## Use with mlx
22
+
23
+ ```bash
24
+ pip install mlx-lm
25
+ ```
26
+
27
+ ```python
28
+ from mlx_lm import load, generate
29
+
30
+ model, tokenizer = load("mlx-community/MiMo-V2-Flash-mlx-8bit")
31
+
32
+ prompt = "hello"
33
+
34
+ if tokenizer.chat_template is not None:
35
+ messages = [{"role": "user", "content": prompt}]
36
+ prompt = tokenizer.apply_chat_template(
37
+ messages, add_generation_prompt=True
38
+ )
39
+
40
+ response = generate(model, tokenizer, prompt=prompt, verbose=True)
41
+ ```