Update README.md
Browse files
README.md
CHANGED
|
@@ -50,8 +50,6 @@ tags:
|
|
| 50 |
|
| 51 |
**Jan-v3-4B-base-instruct** is a 4B-parameter model obtained via post-training distillation from a larger teacher, transferring capabilities while preserving general-purpose performance on standard benchmarks. The result is a compact, ownable base that is straightforward to fine-tune, broadly applicable and minimizing the usual capacity–capability trade-offs.
|
| 52 |
|
| 53 |
-
Building on this base, **Jan-Code**, a code-tuned variant, **will be released soon.**
|
| 54 |
-
|
| 55 |
## Model Overview
|
| 56 |
|
| 57 |
> **Note:** Jan-v3-4B-base-instruct inherits its core architecture from **Qwen/Qwen3-4B-Instruct-2507**.
|
|
@@ -81,7 +79,7 @@ Jan-v3 demo is hosted on **Jan Browser** at **[chat.jan.ai](https://chat.jan.ai/
|
|
| 81 |
|
| 82 |
**Using vLLM:**
|
| 83 |
```bash
|
| 84 |
-
vllm serve
|
| 85 |
--host 0.0.0.0 \
|
| 86 |
--port 1234 \
|
| 87 |
--enable-auto-tool-choice \
|
|
@@ -89,9 +87,9 @@ vllm serve janhq/Jan-v3-4B-base-instruct \
|
|
| 89 |
|
| 90 |
```
|
| 91 |
|
| 92 |
-
**Using llama.cpp:**
|
| 93 |
```bash
|
| 94 |
-
llama-server --model Jan-v3-4B-base-instruct-Q8_0.gguf \
|
| 95 |
--host 0.0.0.0 \
|
| 96 |
--port 1234 \
|
| 97 |
--jinja \
|
|
|
|
| 50 |
|
| 51 |
**Jan-v3-4B-base-instruct** is a 4B-parameter model obtained via post-training distillation from a larger teacher, transferring capabilities while preserving general-purpose performance on standard benchmarks. The result is a compact, ownable base that is straightforward to fine-tune, broadly applicable and minimizing the usual capacity–capability trade-offs.
|
| 52 |
|
|
|
|
|
|
|
| 53 |
## Model Overview
|
| 54 |
|
| 55 |
> **Note:** Jan-v3-4B-base-instruct inherits its core architecture from **Qwen/Qwen3-4B-Instruct-2507**.
|
|
|
|
| 79 |
|
| 80 |
**Using vLLM:**
|
| 81 |
```bash
|
| 82 |
+
vllm serve megabytes/Jan-v3-4B-base-instruct-heretic \
|
| 83 |
--host 0.0.0.0 \
|
| 84 |
--port 1234 \
|
| 85 |
--enable-auto-tool-choice \
|
|
|
|
| 87 |
|
| 88 |
```
|
| 89 |
|
| 90 |
+
**Using llama.cpp:** (Use a quantized version of this)
|
| 91 |
```bash
|
| 92 |
+
llama-server --model Jan-v3-4B-base-instruct-heretic-Q8_0.gguf \
|
| 93 |
--host 0.0.0.0 \
|
| 94 |
--port 1234 \
|
| 95 |
--jinja \
|