Update model card for v0.8
Browse files
README.md
CHANGED
|
@@ -1,3 +1,4 @@
|
|
|
|
|
| 1 |
---
|
| 2 |
base_model: unsloth/Qwen2.5-Coder-32B-Instruct-bnb-4bit
|
| 3 |
tags:
|
|
@@ -19,25 +20,25 @@ Pre-fine-tuned LoRA adapters for unsloth/Qwen2.5-Coder-32B-Instruct.
|
|
| 19 |
- https://huggingface.co/datasets/Jake5/wmx-doc-user
|
| 20 |
- https://huggingface.co/datasets/Jake5/wmx-doc-robot
|
| 21 |
|
| 22 |
-
## Version v0.
|
| 23 |
- Source: lora_model
|
| 24 |
- Base model: unsloth/Qwen2.5-Coder-32B-Instruct
|
| 25 |
-
- Uploaded on: 2025-09-
|
| 26 |
|
| 27 |
## Usage
|
| 28 |
```python
|
| 29 |
from peft import PeftModel
|
| 30 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 31 |
base_model = AutoModelForCausalLM.from_pretrained("unsloth/Qwen2.5-Coder-32B-Instruct")
|
| 32 |
-
model = PeftModel.from_pretrained(base_model, "Jake5/Qwen2.5-Coder-32B-Instruct-WMX", subfolder="adapters_v0.
|
| 33 |
-
tokenizer = AutoTokenizer.from_pretrained("Jake5/Qwen2.5-Coder-32B-Instruct-WMX", subfolder="adapters_v0.
|
| 34 |
```
|
| 35 |
|
| 36 |
## vLLM Serving
|
| 37 |
```bash
|
| 38 |
python -m vllm.entrypoints.openai.api_server \
|
| 39 |
--model unsloth/Qwen2.5-Coder-32B-Instruct \
|
| 40 |
-
--lora-modules my-lora=Jake5/Qwen2.5-Coder-32B-Instruct-WMX/adapters_v0.
|
| 41 |
--dtype bfloat16 \
|
| 42 |
--port 8000
|
| 43 |
```
|
|
|
|
| 1 |
+
|
| 2 |
---
|
| 3 |
base_model: unsloth/Qwen2.5-Coder-32B-Instruct-bnb-4bit
|
| 4 |
tags:
|
|
|
|
| 20 |
- https://huggingface.co/datasets/Jake5/wmx-doc-user
|
| 21 |
- https://huggingface.co/datasets/Jake5/wmx-doc-robot
|
| 22 |
|
| 23 |
+
## Version v0.8
|
| 24 |
- Source: lora_model
|
| 25 |
- Base model: unsloth/Qwen2.5-Coder-32B-Instruct
|
| 26 |
+
- Uploaded on: 2025-09-11
|
| 27 |
|
| 28 |
## Usage
|
| 29 |
```python
|
| 30 |
from peft import PeftModel
|
| 31 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 32 |
base_model = AutoModelForCausalLM.from_pretrained("unsloth/Qwen2.5-Coder-32B-Instruct")
|
| 33 |
+
model = PeftModel.from_pretrained(base_model, "Jake5/Qwen2.5-Coder-32B-Instruct-WMX", subfolder="adapters_v0.8")
|
| 34 |
+
tokenizer = AutoTokenizer.from_pretrained("Jake5/Qwen2.5-Coder-32B-Instruct-WMX", subfolder="adapters_v0.8")
|
| 35 |
```
|
| 36 |
|
| 37 |
## vLLM Serving
|
| 38 |
```bash
|
| 39 |
python -m vllm.entrypoints.openai.api_server \
|
| 40 |
--model unsloth/Qwen2.5-Coder-32B-Instruct \
|
| 41 |
+
--lora-modules my-lora=Jake5/Qwen2.5-Coder-32B-Instruct-WMX/adapters_v0.8 \
|
| 42 |
--dtype bfloat16 \
|
| 43 |
--port 8000
|
| 44 |
```
|