kernelpool commited on
Commit
426b031
·
verified ·
1 Parent(s): b2921ca

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +33 -1
README.md CHANGED
@@ -3,7 +3,39 @@ license: apache-2.0
3
  pipeline_tag: text-generation
4
  library_name: mlx
5
  tags:
6
- - vllm
7
  - mlx
8
  base_model: ByteDance-Seed/Seed-OSS-36B-Instruct
 
 
 
9
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  pipeline_tag: text-generation
4
  library_name: mlx
5
  tags:
 
6
  - mlx
7
  base_model: ByteDance-Seed/Seed-OSS-36B-Instruct
8
+ language:
9
+ - en
10
+ - zh
11
  ---
12
+
13
+ # catalystsec/Seed-OSS-36B-Instruct-4bit-DWQ
14
+
15
+ This model was quantized to 4-bit and optimized using DWQ (Dynamic Weight Quantization) with mlx-lm version **0.27.1**.
16
+
17
+ | Learning Rate | Total Loss | KL Loss | Activation Loss | Improvement |
18
+ |---------------|------------|---------|-----------------|-------------|
19
+ | 2e-7 | 0.415 | 0.025 | 0.390 | 15.8% |
20
+
21
+ ## Use with mlx
22
+
23
+ ```bash
24
+ pip install mlx-lm
25
+ ```
26
+
27
+ ```python
28
+ from mlx_lm import load, generate
29
+
30
+ model, tokenizer = load("catalystsec/Seed-OSS-36B-Instruct-4bit-DWQ")
31
+
32
+ prompt = "hello"
33
+
34
+ if tokenizer.chat_template is not None:
35
+ messages = [{"role": "user", "content": prompt}]
36
+ prompt = tokenizer.apply_chat_template(
37
+ messages, add_generation_prompt=True
38
+ )
39
+
40
+ response = generate(model, tokenizer, prompt=prompt, verbose=True)
41
+ ```