GGUF
conversational
catpp commited on
Commit
ffcd0ee
·
verified ·
1 Parent(s): 5efff8c

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +36 -0
README.md ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ ---
4
+ # 🧠 Skywork-R1V3-38B - GGUF Quantized
5
+
6
+ This repository provides a `GGUF` quantized version of the [Skywork-R1V3-38B](https://huggingface.co/Skywork/Skywork-R1V3-38B) model, converted using the latest `master` branch of [llama.cpp](https://github.com/ggerganov/llama.cpp). This version is optimized for **fast and memory-efficient local inference** on CPU or GPU.
7
+
8
+ ## 💻 How to Use
9
+
10
+ You can run this model with [`llama.cpp`](https://github.com/ggerganov/llama.cpp):
11
+
12
+ ```bash
13
+ ./llama-server -m /path/to/Skywork-R1V3-38B-Q8_0.gguf --mmproj /path/to/mmproj-Skywork-R1V3-38B-f16.gguf --port 8080
14
+ ```
15
+ You can now use OpenAI-compatible tools (like curl) to query the model:
16
+
17
+ ```bash
18
+ BASE64_IMAGE=$(base64 -w 0 /path/to/image)
19
+ curl -X POST http://localhost:8080/v1/chat/completions \
20
+ -H "Content-Type: application/json" \
21
+ -d '{
22
+ "model": "Skywork-R1V3",
23
+ "messages": [
24
+ {
25
+ "role": "user",
26
+ "content": [
27
+ {"type": "text", "text": "Please describe this image."},
28
+ {"type": "image_url", "image_url": {"url": "data:image/jpeg;base64,'"${BASE64_IMAGE}"'" }}
29
+ ]
30
+ }
31
+ ],
32
+ "temperature": 0.7,
33
+ "max_tokens": 512
34
+ }'
35
+
36
+ ```