Upload folder using huggingface_hub
Browse files- .claude/settings.local.json +11 -0
- README.md +366 -0
- chat_template.jinja +1 -0
- config.json +76 -0
- generation_config.json +6 -0
- help.txt +304 -0
- model-00001-of-00015.safetensors +3 -0
- model-00002-of-00015.safetensors +3 -0
- model-00003-of-00015.safetensors +3 -0
- model-00004-of-00015.safetensors +3 -0
- model-00005-of-00015.safetensors +3 -0
- model-00006-of-00015.safetensors +3 -0
- model-00007-of-00015.safetensors +3 -0
- model-00008-of-00015.safetensors +3 -0
- model-00009-of-00015.safetensors +3 -0
- model-00010-of-00015.safetensors +3 -0
- model-00011-of-00015.safetensors +3 -0
- model-00012-of-00015.safetensors +3 -0
- model-00013-of-00015.safetensors +3 -0
- model-00014-of-00015.safetensors +3 -0
- model-00015-of-00015.safetensors +3 -0
- model.safetensors.index.json +0 -0
- recipe.yaml +6 -0
- sglang.sh +19 -0
- special_tokens_map.json +23 -0
- tokenizer.json +0 -0
- tokenizer_config.json +0 -0
.claude/settings.local.json
ADDED
|
@@ -0,0 +1,11 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"permissions": {
|
| 3 |
+
"allow": [
|
| 4 |
+
"WebSearch",
|
| 5 |
+
"Bash(gh issue view:*)",
|
| 6 |
+
"Bash(python3 -c \" import json, sys data = json.load\\(sys.stdin\\) print\\(''TITLE:'', data[''title'']\\) print\\(''BODY:'', data[''body''][:3000]\\) print\\(\\) for c in data.get\\(''comments'', []\\)[:10]: print\\(''---COMMENT---''\\) print\\(c[''body''][:2000]\\) print\\(\\) \")",
|
| 7 |
+
"WebFetch(domain:docs.sglang.io)",
|
| 8 |
+
"Bash(python3 -c \" import json, sys data = json.load\\(sys.stdin\\) print\\(''TITLE:'', data[''title'']\\) print\\(''BODY:'', data[''body''][:3000]\\) print\\(\\) for c in data.get\\(''comments'', []\\)[:8]: print\\(''---COMMENT---''\\) print\\(c[''body''][:1500]\\) print\\(\\) \")"
|
| 9 |
+
]
|
| 10 |
+
}
|
| 11 |
+
}
|
README.md
ADDED
|
@@ -0,0 +1,366 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
base_model:
|
| 3 |
+
- mistralai/Mistral-Large-Instruct-2411
|
| 4 |
+
base_model_relation: quantized
|
| 5 |
+
library_name: vllm
|
| 6 |
+
language:
|
| 7 |
+
- en
|
| 8 |
+
- fr
|
| 9 |
+
- de
|
| 10 |
+
- es
|
| 11 |
+
- it
|
| 12 |
+
- pt
|
| 13 |
+
- zh
|
| 14 |
+
- ja
|
| 15 |
+
- ru
|
| 16 |
+
- ko
|
| 17 |
+
license: other
|
| 18 |
+
license_name: mrl
|
| 19 |
+
inference: false
|
| 20 |
+
license_link: https://mistral.ai/licenses/MRL-0.1.md
|
| 21 |
+
tags:
|
| 22 |
+
- quantization
|
| 23 |
+
- nvfp4
|
| 24 |
+
- fp4
|
| 25 |
+
- mistral-common
|
| 26 |
+
---
|
| 27 |
+
|
| 28 |
+
# Model Card for Mistral-Large-Instruct-2411
|
| 29 |
+
|
| 30 |
+
Mistral-Large-Instruct-2411 is an advanced dense Large Language Model (LLM) of 123B parameters with state-of-the-art reasoning, knowledge and coding capabilities extending [Mistral-Large-Instruct-2407](https://huggingface.co/mistralai/Mistral-Large-Instruct-2407) with better Long Context, Function Calling and System Prompt.
|
| 31 |
+
|
| 32 |
+
## Key features
|
| 33 |
+
- **Multi-lingual by design:** Dozens of languages supported, including English, French, German, Spanish, Italian, Chinese, Japanese, Korean, Portuguese, Dutch and Polish.
|
| 34 |
+
- **Proficient in coding:** Trained on 80+ coding languages such as Python, Java, C, C++, Javacsript, and Bash. Also trained on more specific languages such as Swift and Fortran.
|
| 35 |
+
- **Agent-centric:** Best-in-class agentic capabilities with native function calling and JSON outputting.
|
| 36 |
+
- **Advanced Reasoning:** State-of-the-art mathematical and reasoning capabilities.
|
| 37 |
+
- **Mistral Research License:** Allows usage and modification for non-commercial usages.
|
| 38 |
+
- **Large Context:** A large 128k context window.
|
| 39 |
+
- **Robust Context Adherence:** Ensures strong adherence for RAG and large context applications.
|
| 40 |
+
- **System Prompt:** Maintains strong adherence and support for more reliable system prompts.
|
| 41 |
+
|
| 42 |
+
### System Prompt
|
| 43 |
+
We appreciate the feedback received from our community regarding our system prompt handling.
|
| 44 |
+
In response, we have implemented stronger support for system prompts.
|
| 45 |
+
To achieve optimal results, we recommend always including a system prompt that clearly outlines the bot's purpose, even if it is minimal.
|
| 46 |
+
|
| 47 |
+
### Basic Instruct Template (V7)
|
| 48 |
+
|
| 49 |
+
```
|
| 50 |
+
<s>[SYSTEM_PROMPT] <system prompt>[/SYSTEM_PROMPT][INST] <user message>[/INST] <assistant response></s>[INST] <user message>[/INST]
|
| 51 |
+
```
|
| 52 |
+
|
| 53 |
+
**Be careful with subtle missing or trailing white spaces!**
|
| 54 |
+
|
| 55 |
+
*Please make sure to use [mistral-common](https://github.com/mistralai/mistral-common) as the source of truth*
|
| 56 |
+
|
| 57 |
+
## Usage
|
| 58 |
+
|
| 59 |
+
The model can be used with the following frameworks
|
| 60 |
+
|
| 61 |
+
- [`vllm`](https://github.com/vllm-project/vllm): See [here](#vLLM)
|
| 62 |
+
|
| 63 |
+
### vLLM
|
| 64 |
+
|
| 65 |
+
We recommend using this model with the [vLLM library](https://github.com/vllm-project/vllm)
|
| 66 |
+
to implement production-ready inference pipelines.
|
| 67 |
+
|
| 68 |
+
**_Installation_**
|
| 69 |
+
|
| 70 |
+
Make sure you install [`vLLM >= v0.6.4.post1`](https://github.com/vllm-project/vllm/releases/tag/v0.6.4.post1):
|
| 71 |
+
|
| 72 |
+
```
|
| 73 |
+
pip install --upgrade vllm
|
| 74 |
+
```
|
| 75 |
+
|
| 76 |
+
Also make sure you have [`mistral_common >= 1.5.0`](https://github.com/mistralai/mistral-common/releases/tag/v1.5.0) installed:
|
| 77 |
+
|
| 78 |
+
```
|
| 79 |
+
pip install --upgrade mistral_common
|
| 80 |
+
```
|
| 81 |
+
|
| 82 |
+
You can also make use of a ready-to-go [docker image](https://github.com/vllm-project/vllm/blob/main/Dockerfile) or on the [docker hub](https://hub.docker.com/layers/vllm/vllm-openai/latest/images/sha256-55a88146a4da0b6e193431b5b1d3492dfd7bebdc16919df4d031273e85a6157c?context=explore).
|
| 83 |
+
|
| 84 |
+
### Server
|
| 85 |
+
|
| 86 |
+
We recommand that you use Mistral-Large-Instruct-2411 in a server/client setting.
|
| 87 |
+
|
| 88 |
+
1. Spin up a server:
|
| 89 |
+
|
| 90 |
+
|
| 91 |
+
```
|
| 92 |
+
vllm serve mistralai/Mistral-Large-Instruct-2411 --tokenizer_mode mistral --config_format mistral --load_format mistral --tensor_parallel_size 8
|
| 93 |
+
```
|
| 94 |
+
|
| 95 |
+
**Note:** Running Mistral-Large-Instruct-2411 on GPU requires over 300 GB of GPU RAM.
|
| 96 |
+
|
| 97 |
+
|
| 98 |
+
2. To ping the client you can use a simple Python snippet.
|
| 99 |
+
|
| 100 |
+
```py
|
| 101 |
+
import requests
|
| 102 |
+
import json
|
| 103 |
+
from huggingface_hub import hf_hub_download
|
| 104 |
+
from datetime import datetime, timedelta
|
| 105 |
+
|
| 106 |
+
url = "http://<your-server>:8000/v1/chat/completions"
|
| 107 |
+
headers = {"Content-Type": "application/json", "Authorization": "Bearer token"}
|
| 108 |
+
|
| 109 |
+
model = "mistralai/Mistral-Large-Instruct-2411"
|
| 110 |
+
|
| 111 |
+
|
| 112 |
+
def load_system_prompt(repo_id: str, filename: str) -> str:
|
| 113 |
+
file_path = hf_hub_download(repo_id=repo_id, filename=filename)
|
| 114 |
+
with open(file_path, "r") as file:
|
| 115 |
+
system_prompt = file.read()
|
| 116 |
+
today = datetime.today().strftime("%Y-%m-%d")
|
| 117 |
+
yesterday = (datetime.today() - timedelta(days=1)).strftime("%Y-%m-%d")
|
| 118 |
+
model_name = repo_id.split("/")[-1]
|
| 119 |
+
return system_prompt.format(name=model_name, today=today, yesterday=yesterday)
|
| 120 |
+
|
| 121 |
+
|
| 122 |
+
SYSTEM_PROMPT = load_system_prompt(model, "SYSTEM_PROMPT.txt")
|
| 123 |
+
|
| 124 |
+
|
| 125 |
+
messages = [
|
| 126 |
+
{"role": "system", "content": SYSTEM_PROMPT + "\n\nThink step by step. You're a math genius."},
|
| 127 |
+
{
|
| 128 |
+
"role": "user",
|
| 129 |
+
"content": "Think of four random numbers. Then add, substract or multiply them so that the solution is 10. If it's not possible, say it."
|
| 130 |
+
},
|
| 131 |
+
]
|
| 132 |
+
|
| 133 |
+
data = {"model": model, "messages": messages}
|
| 134 |
+
|
| 135 |
+
response = requests.post(url, headers=headers, data=json.dumps(data))
|
| 136 |
+
print(response.json()["choices"][0]["message"]["content"])
|
| 137 |
+
# Sure, let's start by thinking of four random numbers. For example, let's take 3, 5, 2, and 1.
|
| 138 |
+
#
|
| 139 |
+
# Now, we need to find a combination of addition, subtraction, or multiplication that results in 10.
|
| 140 |
+
|
| 141 |
+
# Let's try:
|
| 142 |
+
|
| 143 |
+
# \[ 3 + 5 + 2 - 1 = 9 \]
|
| 144 |
+
|
| 145 |
+
# This doesn't work. Let's try another combination:
|
| 146 |
+
|
| 147 |
+
# \[ 3 \times 2 + 5 - 1 = 6 + 5 - 1 = 10 \]
|
| 148 |
+
|
| 149 |
+
# This works! So, with the numbers 3, 5, 2, and 1, we can achieve the result 10 by performing the operations \( 3 \times 2 + 5 - 1 \).
|
| 150 |
+
```
|
| 151 |
+
|
| 152 |
+
### Offline
|
| 153 |
+
|
| 154 |
+
```py
|
| 155 |
+
from vllm import LLM
|
| 156 |
+
from vllm.sampling_params import SamplingParams
|
| 157 |
+
from huggingface_hub import hf_hub_download
|
| 158 |
+
from datetime import datetime, timedelta
|
| 159 |
+
|
| 160 |
+
model_name = "mistralai/Mistral-Large-Instruct-2411"
|
| 161 |
+
|
| 162 |
+
def load_system_prompt(repo_id: str, filename: str) -> str:
|
| 163 |
+
file_path = hf_hub_download(repo_id=repo_id, filename=filename)
|
| 164 |
+
with open(file_path, 'r') as file:
|
| 165 |
+
system_prompt = file.read()
|
| 166 |
+
today = datetime.today().strftime('%Y-%m-%d')
|
| 167 |
+
yesterday = (datetime.today() - timedelta(days=1)).strftime('%Y-%m-%d')
|
| 168 |
+
model_name = repo_id.split("/")[-1]
|
| 169 |
+
return system_prompt.format(name=model_name, today=today, yesterday=yesterday)
|
| 170 |
+
|
| 171 |
+
|
| 172 |
+
SYSTEM_PROMPT = load_system_prompt(model_name, "SYSTEM_PROMPT.txt") + "\n\nThink step by step. You're a math genius."
|
| 173 |
+
|
| 174 |
+
user_prompt = "Without browsing the web, how many days ago was Mistral founded?"
|
| 175 |
+
|
| 176 |
+
messages = [
|
| 177 |
+
{
|
| 178 |
+
"role": "system",
|
| 179 |
+
"content": SYSTEM_PROMPT
|
| 180 |
+
},
|
| 181 |
+
{
|
| 182 |
+
"role": "user",
|
| 183 |
+
"content": user_prompt
|
| 184 |
+
},
|
| 185 |
+
]
|
| 186 |
+
|
| 187 |
+
# note that running this model on GPU requires over 300 GB of GPU RAM
|
| 188 |
+
llm = LLM(model=model_name, tokenizer_mode="mistral", tensor_parallel_size=8)
|
| 189 |
+
|
| 190 |
+
sampling_params = SamplingParams(max_tokens=512)
|
| 191 |
+
|
| 192 |
+
outputs = llm.chat(messages, sampling_params=sampling_params)
|
| 193 |
+
|
| 194 |
+
print(outputs[0].outputs[0].text)
|
| 195 |
+
# I don't have real-time web browsing capabilities or access to current data, but I can help you calculate the number of days based on the information I have.
|
| 196 |
+
#
|
| 197 |
+
#Mistral AI was founded in April 2023. To determine how many days ago that was from today's date, November 18, 2024, we need to calculate the total number of days between April 2023 and November 2024.
|
| 198 |
+
#
|
| 199 |
+
#Here's the step-by-step calculation:
|
| 200 |
+
#
|
| 201 |
+
#1. **Days from April 2023 to December 2023:**
|
| 202 |
+
# - April 2023: 30 days (April has 30 days)
|
| 203 |
+
# - May 2023: 31 days
|
| 204 |
+
# - June 2023: 30 days
|
| 205 |
+
# - July 2023: 31 days
|
| 206 |
+
# - August 2023: 31 days
|
| 207 |
+
# - September 2023: 30 days
|
| 208 |
+
# - October 2023: 31 days
|
| 209 |
+
# - November 2023: 30 days
|
| 210 |
+
# - December 2023: 31 days
|
| 211 |
+
#
|
| 212 |
+
# Total days in 2023 from April to December = 30 + 31 + 30 + 31 + 31 + 30 + 31 + 30 + 31 = 275 days
|
| 213 |
+
#
|
| 214 |
+
#2. **Days from January 2024 to November 18, 2024:**
|
| 215 |
+
# - January 2024: 31 days
|
| 216 |
+
# - February 2024: 29 days (2024 is a leap year)
|
| 217 |
+
# - March 2024: 31 days
|
| 218 |
+
# - April 2024: 30 days
|
| 219 |
+
# - May 2024: 31 days
|
| 220 |
+
# - June 2024: 30 days
|
| 221 |
+
# - July 2024: 31 days
|
| 222 |
+
# - August 2024: 31 days
|
| 223 |
+
# - September 2024: 30 days
|
| 224 |
+
# - October 2024: 31 days
|
| 225 |
+
# - November 2024 (up to the 18th): 18 days
|
| 226 |
+
#
|
| 227 |
+
# Total days in 2024 from January to November 18 = 31 + 29 + 31 + 30 + 31 + 30 + 31 + 31 + 30 + 31 + 18 = 323 days
|
| 228 |
+
#
|
| 229 |
+
#3. **Total days from April 2023 to November 18, 2024:**
|
| 230 |
+
# Total days = 275 days (2023) + 323 days (2024) = 598 days
|
| 231 |
+
#
|
| 232 |
+
#Therefore, Mistral AI was founded 598 days ago from today's date, November 18, 2024.
|
| 233 |
+
```
|
| 234 |
+
|
| 235 |
+
### Improved Function Calling
|
| 236 |
+
|
| 237 |
+
Mistral-Large-2411 has much improved function calling capabilities that are fully supported
|
| 238 |
+
using [`mistral_common >= 1.5.0`](https://github.com/mistralai/mistral-common/releases/tag/v1.5.0) and [`vLLM >= v0.6.4.post1`](https://github.com/vllm-project/vllm/releases/tag/v0.6.4.post1).
|
| 239 |
+
|
| 240 |
+
Make sure to serve the model with the following flags in vLLM:
|
| 241 |
+
|
| 242 |
+
```
|
| 243 |
+
vllm serve mistralai/Pixtral-Large-Instruct-2411 --tokenizer_mode mistral --config_format mistral --load_format mistral --tensor-parallel-size 8 --tool-call-parser mistral --enable-auto-tool-choice
|
| 244 |
+
```
|
| 245 |
+
|
| 246 |
+
<details>
|
| 247 |
+
<summary>Example</summary>
|
| 248 |
+
|
| 249 |
+
```py
|
| 250 |
+
import requests
|
| 251 |
+
import json
|
| 252 |
+
from huggingface_hub import hf_hub_download
|
| 253 |
+
from datetime import datetime, timedelta
|
| 254 |
+
|
| 255 |
+
url = "http://<your-server>:8000/v1/chat/completions"
|
| 256 |
+
headers = {"Content-Type": "application/json", "Authorization": "Bearer token"}
|
| 257 |
+
|
| 258 |
+
model = "mistralai/Mistral-Large-Instruct-2411"
|
| 259 |
+
|
| 260 |
+
|
| 261 |
+
def load_system_prompt(repo_id: str, filename: str) -> str:
|
| 262 |
+
file_path = hf_hub_download(repo_id=repo_id, filename=filename)
|
| 263 |
+
with open(file_path, "r") as file:
|
| 264 |
+
system_prompt = file.read()
|
| 265 |
+
today = datetime.today().strftime("%Y-%m-%d")
|
| 266 |
+
yesterday = (datetime.today() - timedelta(days=1)).strftime("%Y-%m-%d")
|
| 267 |
+
model_name = repo_id.split("/")[-1]
|
| 268 |
+
return system_prompt.format(name=model_name, today=today, yesterday=yesterday)
|
| 269 |
+
|
| 270 |
+
|
| 271 |
+
SYSTEM_PROMPT = load_system_prompt(model, "SYSTEM_PROMPT.txt")
|
| 272 |
+
|
| 273 |
+
|
| 274 |
+
tools = [
|
| 275 |
+
{
|
| 276 |
+
"type": "function",
|
| 277 |
+
"function": {
|
| 278 |
+
"name": "get_current_weather",
|
| 279 |
+
"description": "Get the current weather in a given location",
|
| 280 |
+
"parameters": {
|
| 281 |
+
"type": "object",
|
| 282 |
+
"properties": {
|
| 283 |
+
"city": {
|
| 284 |
+
"type": "string",
|
| 285 |
+
"description": "The city to find the weather for, e.g. 'San Francisco'",
|
| 286 |
+
},
|
| 287 |
+
"state": {
|
| 288 |
+
"type": "string",
|
| 289 |
+
"description": "The state abbreviation, e.g. 'CA' for California",
|
| 290 |
+
},
|
| 291 |
+
"unit": {
|
| 292 |
+
"type": "string",
|
| 293 |
+
"description": "The unit for temperature",
|
| 294 |
+
"enum": ["celsius", "fahrenheit"],
|
| 295 |
+
},
|
| 296 |
+
},
|
| 297 |
+
"required": ["city", "state", "unit"],
|
| 298 |
+
},
|
| 299 |
+
},
|
| 300 |
+
},
|
| 301 |
+
{
|
| 302 |
+
"type": "function",
|
| 303 |
+
"function": {
|
| 304 |
+
"name": "rewrite",
|
| 305 |
+
"description": "Rewrite a given text for improved clarity",
|
| 306 |
+
"parameters": {
|
| 307 |
+
"type": "object",
|
| 308 |
+
"properties": {
|
| 309 |
+
"text": {
|
| 310 |
+
"type": "string",
|
| 311 |
+
"description": "The input text to rewrite",
|
| 312 |
+
}
|
| 313 |
+
},
|
| 314 |
+
},
|
| 315 |
+
},
|
| 316 |
+
},
|
| 317 |
+
]
|
| 318 |
+
|
| 319 |
+
messages = [
|
| 320 |
+
{"role": "system", "content": SYSTEM_PROMPT},
|
| 321 |
+
{
|
| 322 |
+
"role": "user",
|
| 323 |
+
"content": "Could you please make the below article more concise?\n\nOpenAI is an artificial intelligence research laboratory consisting of the non-profit OpenAI Incorporated and its for-profit subsidiary corporation OpenAI Limited Partnership.",
|
| 324 |
+
},
|
| 325 |
+
{
|
| 326 |
+
"role": "assistant",
|
| 327 |
+
"content": "",
|
| 328 |
+
"tool_calls": [
|
| 329 |
+
{
|
| 330 |
+
"id": "bbc5b7ede",
|
| 331 |
+
"type": "function",
|
| 332 |
+
"function": {
|
| 333 |
+
"name": "rewrite",
|
| 334 |
+
"arguments": '{"text": "OpenAI is an artificial intelligence research laboratory consisting of the non-profit OpenAI Incorporated and its for-profit subsidiary corporation OpenAI Limited Partnership."}',
|
| 335 |
+
},
|
| 336 |
+
}
|
| 337 |
+
],
|
| 338 |
+
},
|
| 339 |
+
{
|
| 340 |
+
"role": "tool",
|
| 341 |
+
"content": '{"action":"rewrite","outcome":"OpenAI is a FOR-profit company."}',
|
| 342 |
+
"tool_call_id": "bbc5b7ede",
|
| 343 |
+
"name": "rewrite",
|
| 344 |
+
},
|
| 345 |
+
{
|
| 346 |
+
"role": "assistant",
|
| 347 |
+
"content": "---\n\nOpenAI is a FOR-profit company.",
|
| 348 |
+
},
|
| 349 |
+
{
|
| 350 |
+
"role": "user",
|
| 351 |
+
"content": "Can you tell me what the temperature will be in Dallas, in Fahrenheit?",
|
| 352 |
+
},
|
| 353 |
+
]
|
| 354 |
+
|
| 355 |
+
data = {"model": model, "messages": messages, "tools": tools}
|
| 356 |
+
|
| 357 |
+
response = requests.post(url, headers=headers, data=json.dumps(data))
|
| 358 |
+
print(response.json()["choices"][0]["message"]["tool_calls"])
|
| 359 |
+
# [{'id': '8PdihwL6d', 'type': 'function', 'function': {'name': 'get_current_weather', 'arguments': '{"city": "Dallas", "state": "TX", "unit": "fahrenheit"}'}}]
|
| 360 |
+
```
|
| 361 |
+
|
| 362 |
+
</details>
|
| 363 |
+
|
| 364 |
+
## The Mistral AI Team
|
| 365 |
+
|
| 366 |
+
Albert Jiang, Alexandre Sablayrolles, Alexis Tacnet, Alok Kothari, Antoine Roux, Arthur Mensch, Audrey Herblin-Stoop, Augustin Garreau, Austin Birky, Bam4d, Baptiste Bout, Baudouin de Monicault, Blanche Savary, Carole Rambaud, Caroline Feldman, Devendra Singh Chaplot, Diego de las Casas, Diogo Costa, Eleonore Arcelin, Emma Bou Hanna, Etienne Metzger, Gaspard Blanchet, Gianna Lengyel, Guillaume Bour, Guillaume Lample, Harizo Rajaona, Henri Roussez, Hichem Sattouf, Ian Mack, Jean-Malo Delignon, Jessica Chudnovsky, Justus Murke, Kartik Khandelwal, Lawrence Stewart, Louis Martin, Louis Ternon, Lucile Saulnier, Lélio Renard Lavaud, Margaret Jennings, Marie Pellat, Marie Torelli, Marie-Anne Lachaux, Marjorie Janiewicz, Mickaël Seznec, Nicolas Schuhl, Niklas Muhs, Olivier de Garrigues, Patrick von Platen, Paul Jacob, Pauline Buche, Pavan Kumar Reddy, Perry Savas, Pierre Stock, Romain Sauvestre, Sagar Vaze, Sandeep Subramanian, Saurabh Garg, Sophia Yang, Szymon Antoniak, Teven Le Scao, Thibault Schueller, Thibaut Lavril, Thomas Wang, Théophile Gervet, Timothée Lacroix, Valera Nemychnikova, Wendy Shang, William El Sayed, William Marshall
|
chat_template.jinja
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{{ bos_token }}{% for message in messages %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + '[/INST]' }}{% elif message['role'] == 'system' %}{{ '[SYSTEM_PROMPT] ' + message['content'] + '[/SYSTEM_PROMPT]' }}{% elif message['role'] == 'assistant' %}{{ ' ' + message['content'] + eos_token }}{% else %}{{ raise_exception('Only user, system and assistant roles are supported!') }}{% endif %}{% endfor %}
|
config.json
ADDED
|
@@ -0,0 +1,76 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"architectures": [
|
| 3 |
+
"MistralForCausalLM"
|
| 4 |
+
],
|
| 5 |
+
"attention_dropout": 0.0,
|
| 6 |
+
"bos_token_id": 1,
|
| 7 |
+
"dtype": "bfloat16",
|
| 8 |
+
"eos_token_id": 2,
|
| 9 |
+
"head_dim": 128,
|
| 10 |
+
"hidden_act": "silu",
|
| 11 |
+
"hidden_size": 12288,
|
| 12 |
+
"initializer_range": 0.02,
|
| 13 |
+
"intermediate_size": 28672,
|
| 14 |
+
"max_position_embeddings": 131072,
|
| 15 |
+
"model_type": "mistral",
|
| 16 |
+
"num_attention_heads": 96,
|
| 17 |
+
"num_hidden_layers": 88,
|
| 18 |
+
"num_key_value_heads": 8,
|
| 19 |
+
"quantization_config": {
|
| 20 |
+
"config_groups": {
|
| 21 |
+
"group_0": {
|
| 22 |
+
"format": "nvfp4-pack-quantized",
|
| 23 |
+
"input_activations": {
|
| 24 |
+
"actorder": null,
|
| 25 |
+
"block_structure": null,
|
| 26 |
+
"dynamic": "local",
|
| 27 |
+
"group_size": 16,
|
| 28 |
+
"num_bits": 4,
|
| 29 |
+
"observer": "static_minmax",
|
| 30 |
+
"observer_kwargs": {},
|
| 31 |
+
"scale_dtype": "torch.float8_e4m3fn",
|
| 32 |
+
"strategy": "tensor_group",
|
| 33 |
+
"symmetric": true,
|
| 34 |
+
"type": "float",
|
| 35 |
+
"zp_dtype": null
|
| 36 |
+
},
|
| 37 |
+
"output_activations": null,
|
| 38 |
+
"targets": [
|
| 39 |
+
"Linear"
|
| 40 |
+
],
|
| 41 |
+
"weights": {
|
| 42 |
+
"actorder": null,
|
| 43 |
+
"block_structure": null,
|
| 44 |
+
"dynamic": false,
|
| 45 |
+
"group_size": 16,
|
| 46 |
+
"num_bits": 4,
|
| 47 |
+
"observer": "static_minmax",
|
| 48 |
+
"observer_kwargs": {},
|
| 49 |
+
"scale_dtype": "torch.float8_e4m3fn",
|
| 50 |
+
"strategy": "tensor_group",
|
| 51 |
+
"symmetric": true,
|
| 52 |
+
"type": "float",
|
| 53 |
+
"zp_dtype": null
|
| 54 |
+
}
|
| 55 |
+
}
|
| 56 |
+
},
|
| 57 |
+
"format": "nvfp4-pack-quantized",
|
| 58 |
+
"global_compression_ratio": null,
|
| 59 |
+
"ignore": [
|
| 60 |
+
"lm_head"
|
| 61 |
+
],
|
| 62 |
+
"kv_cache_scheme": null,
|
| 63 |
+
"quant_method": "compressed-tensors",
|
| 64 |
+
"quantization_status": "compressed",
|
| 65 |
+
"sparsity_config": {},
|
| 66 |
+
"transform_config": {},
|
| 67 |
+
"version": "0.13.0"
|
| 68 |
+
},
|
| 69 |
+
"rms_norm_eps": 1e-05,
|
| 70 |
+
"rope_theta": 1000000.0,
|
| 71 |
+
"sliding_window": null,
|
| 72 |
+
"tie_word_embeddings": false,
|
| 73 |
+
"transformers_version": "4.57.3",
|
| 74 |
+
"use_cache": true,
|
| 75 |
+
"vocab_size": 32768
|
| 76 |
+
}
|
generation_config.json
ADDED
|
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_from_model_config": true,
|
| 3 |
+
"bos_token_id": 1,
|
| 4 |
+
"eos_token_id": 2,
|
| 5 |
+
"transformers_version": "4.57.3"
|
| 6 |
+
}
|
help.txt
ADDED
|
@@ -0,0 +1,304 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
usage: launch_server.py [-h] --model-path MODEL_PATH
|
| 2 |
+
[--tokenizer-path TOKENIZER_PATH]
|
| 3 |
+
[--tokenizer-mode {auto,slow}]
|
| 4 |
+
[--tokenizer-worker-num TOKENIZER_WORKER_NUM]
|
| 5 |
+
[--skip-tokenizer-init]
|
| 6 |
+
[--load-format {auto,pt,safetensors,npcache,dummy,sharded_state,gguf,bitsandbytes,layered,flash_rl,remote,remote_instance,fastsafetensors,private}]
|
| 7 |
+
[--model-loader-extra-config MODEL_LOADER_EXTRA_CONFIG]
|
| 8 |
+
[--trust-remote-code]
|
| 9 |
+
[--context-length CONTEXT_LENGTH] [--is-embedding]
|
| 10 |
+
[--enable-multimodal] [--revision REVISION]
|
| 11 |
+
[--model-impl MODEL_IMPL] [--host HOST] [--port PORT]
|
| 12 |
+
[--fastapi-root-path FASTAPI_ROOT_PATH] [--grpc-mode]
|
| 13 |
+
[--skip-server-warmup] [--warmups WARMUPS]
|
| 14 |
+
[--nccl-port NCCL_PORT]
|
| 15 |
+
[--checkpoint-engine-wait-weights-before-ready]
|
| 16 |
+
[--dtype {auto,half,float16,bfloat16,float,float32}]
|
| 17 |
+
[--quantization {awq,fp8,gptq,marlin,gptq_marlin,awq_marlin,bitsandbytes,gguf,modelopt,modelopt_fp8,modelopt_fp4,petit_nvfp4,w8a8_int8,w8a8_fp8,moe_wna16,qoq,w4afp8,mxfp4,auto-round,compressed-tensors,modelslim,quark_int4fp8_moe}]
|
| 18 |
+
[--quantization-param-path QUANTIZATION_PARAM_PATH]
|
| 19 |
+
[--kv-cache-dtype {auto,fp8_e5m2,fp8_e4m3,bf16,bfloat16,fp4_e2m1}]
|
| 20 |
+
[--enable-fp32-lm-head]
|
| 21 |
+
[--modelopt-quant MODELOPT_QUANT]
|
| 22 |
+
[--modelopt-checkpoint-restore-path MODELOPT_CHECKPOINT_RESTORE_PATH]
|
| 23 |
+
[--modelopt-checkpoint-save-path MODELOPT_CHECKPOINT_SAVE_PATH]
|
| 24 |
+
[--modelopt-export-path MODELOPT_EXPORT_PATH]
|
| 25 |
+
[--quantize-and-serve]
|
| 26 |
+
[--rl-quant-profile RL_QUANT_PROFILE]
|
| 27 |
+
[--mem-fraction-static MEM_FRACTION_STATIC]
|
| 28 |
+
[--max-running-requests MAX_RUNNING_REQUESTS]
|
| 29 |
+
[--max-queued-requests MAX_QUEUED_REQUESTS]
|
| 30 |
+
[--max-total-tokens MAX_TOTAL_TOKENS]
|
| 31 |
+
[--chunked-prefill-size CHUNKED_PREFILL_SIZE]
|
| 32 |
+
[--prefill-max-requests PREFILL_MAX_REQUESTS]
|
| 33 |
+
[--enable-dynamic-chunking]
|
| 34 |
+
[--max-prefill-tokens MAX_PREFILL_TOKENS]
|
| 35 |
+
[--schedule-policy {lpm,random,fcfs,dfs-weight,lof,priority,routing-key}]
|
| 36 |
+
[--enable-priority-scheduling]
|
| 37 |
+
[--abort-on-priority-when-disabled]
|
| 38 |
+
[--schedule-low-priority-values-first]
|
| 39 |
+
[--priority-scheduling-preemption-threshold PRIORITY_SCHEDULING_PREEMPTION_THRESHOLD]
|
| 40 |
+
[--schedule-conservativeness SCHEDULE_CONSERVATIVENESS]
|
| 41 |
+
[--page-size PAGE_SIZE] [--hybrid-kvcache-ratio]
|
| 42 |
+
[--swa-full-tokens-ratio SWA_FULL_TOKENS_RATIO]
|
| 43 |
+
[--disable-hybrid-swa-memory]
|
| 44 |
+
[--radix-eviction-policy {lru,lfu}]
|
| 45 |
+
[--enable-prefill-delayer]
|
| 46 |
+
[--prefill-delayer-max-delay-passes PREFILL_DELAYER_MAX_DELAY_PASSES]
|
| 47 |
+
[--prefill-delayer-token-usage-low-watermark PREFILL_DELAYER_TOKEN_USAGE_LOW_WATERMARK]
|
| 48 |
+
[--prefill-delayer-forward-passes-buckets PREFILL_DELAYER_FORWARD_PASSES_BUCKETS [PREFILL_DELAYER_FORWARD_PASSES_BUCKETS ...]]
|
| 49 |
+
[--prefill-delayer-wait-seconds-buckets PREFILL_DELAYER_WAIT_SECONDS_BUCKETS [PREFILL_DELAYER_WAIT_SECONDS_BUCKETS ...]]
|
| 50 |
+
[--device DEVICE]
|
| 51 |
+
[--tensor-parallel-size TENSOR_PARALLEL_SIZE]
|
| 52 |
+
[--pipeline-parallel-size PIPELINE_PARALLEL_SIZE]
|
| 53 |
+
[--pp-max-micro-batch-size PP_MAX_MICRO_BATCH_SIZE]
|
| 54 |
+
[--pp-async-batch-depth PP_ASYNC_BATCH_DEPTH]
|
| 55 |
+
[--stream-interval STREAM_INTERVAL] [--stream-output]
|
| 56 |
+
[--random-seed RANDOM_SEED]
|
| 57 |
+
[--constrained-json-whitespace-pattern CONSTRAINED_JSON_WHITESPACE_PATTERN]
|
| 58 |
+
[--constrained-json-disable-any-whitespace]
|
| 59 |
+
[--watchdog-timeout WATCHDOG_TIMEOUT]
|
| 60 |
+
[--soft-watchdog-timeout SOFT_WATCHDOG_TIMEOUT]
|
| 61 |
+
[--dist-timeout DIST_TIMEOUT]
|
| 62 |
+
[--download-dir DOWNLOAD_DIR]
|
| 63 |
+
[--model-checksum [MODEL_CHECKSUM]]
|
| 64 |
+
[--base-gpu-id BASE_GPU_ID]
|
| 65 |
+
[--gpu-id-step GPU_ID_STEP] [--sleep-on-idle]
|
| 66 |
+
[--custom-sigquit-handler CUSTOM_SIGQUIT_HANDLER]
|
| 67 |
+
[--log-level LOG_LEVEL]
|
| 68 |
+
[--log-level-http LOG_LEVEL_HTTP] [--log-requests]
|
| 69 |
+
[--log-requests-level {0,1,2,3}]
|
| 70 |
+
[--log-requests-format {text,json}]
|
| 71 |
+
[--log-requests-target LOG_REQUESTS_TARGET [LOG_REQUESTS_TARGET ...]]
|
| 72 |
+
[--uvicorn-access-log-exclude-prefixes [UVICORN_ACCESS_LOG_EXCLUDE_PREFIXES ...]]
|
| 73 |
+
[--crash-dump-folder CRASH_DUMP_FOLDER]
|
| 74 |
+
[--show-time-cost] [--enable-metrics]
|
| 75 |
+
[--enable-metrics-for-all-schedulers]
|
| 76 |
+
[--tokenizer-metrics-custom-labels-header TOKENIZER_METRICS_CUSTOM_LABELS_HEADER]
|
| 77 |
+
[--tokenizer-metrics-allowed-custom-labels TOKENIZER_METRICS_ALLOWED_CUSTOM_LABELS [TOKENIZER_METRICS_ALLOWED_CUSTOM_LABELS ...]]
|
| 78 |
+
[--bucket-time-to-first-token BUCKET_TIME_TO_FIRST_TOKEN [BUCKET_TIME_TO_FIRST_TOKEN ...]]
|
| 79 |
+
[--bucket-inter-token-latency BUCKET_INTER_TOKEN_LATENCY [BUCKET_INTER_TOKEN_LATENCY ...]]
|
| 80 |
+
[--bucket-e2e-request-latency BUCKET_E2E_REQUEST_LATENCY [BUCKET_E2E_REQUEST_LATENCY ...]]
|
| 81 |
+
[--collect-tokens-histogram]
|
| 82 |
+
[--prompt-tokens-buckets PROMPT_TOKENS_BUCKETS [PROMPT_TOKENS_BUCKETS ...]]
|
| 83 |
+
[--generation-tokens-buckets GENERATION_TOKENS_BUCKETS [GENERATION_TOKENS_BUCKETS ...]]
|
| 84 |
+
[--gc-warning-threshold-secs GC_WARNING_THRESHOLD_SECS]
|
| 85 |
+
[--decode-log-interval DECODE_LOG_INTERVAL]
|
| 86 |
+
[--enable-request-time-stats-logging]
|
| 87 |
+
[--kv-events-config KV_EVENTS_CONFIG] [--enable-trace]
|
| 88 |
+
[--otlp-traces-endpoint OTLP_TRACES_ENDPOINT]
|
| 89 |
+
[--export-metrics-to-file]
|
| 90 |
+
[--export-metrics-to-file-dir EXPORT_METRICS_TO_FILE_DIR]
|
| 91 |
+
[--api-key API_KEY] [--admin-api-key ADMIN_API_KEY]
|
| 92 |
+
[--served-model-name SERVED_MODEL_NAME]
|
| 93 |
+
[--weight-version WEIGHT_VERSION]
|
| 94 |
+
[--chat-template CHAT_TEMPLATE]
|
| 95 |
+
[--hf-chat-template-name HF_CHAT_TEMPLATE_NAME]
|
| 96 |
+
[--completion-template COMPLETION_TEMPLATE]
|
| 97 |
+
[--file-storage-path FILE_STORAGE_PATH]
|
| 98 |
+
[--enable-cache-report]
|
| 99 |
+
[--reasoning-parser {deepseek-r1,deepseek-v3,glm45,gpt-oss,kimi,kimi_k2,qwen3,qwen3-thinking,minimax,minimax-append-think,step3,nano_v3,interns1}]
|
| 100 |
+
[--tool-call-parser {deepseekv3,deepseekv31,deepseekv32,glm,glm45,glm47,gpt-oss,kimi_k2,llama3,mimo,mistral,pythonic,qwen,qwen25,qwen3_coder,step3,minimax-m2,interns1}]
|
| 101 |
+
[--tool-server TOOL_SERVER]
|
| 102 |
+
[--sampling-defaults {openai,model}]
|
| 103 |
+
[--data-parallel-size DATA_PARALLEL_SIZE]
|
| 104 |
+
[--load-balance-method {auto,round_robin,follow_bootstrap_room,total_requests,total_tokens}]
|
| 105 |
+
[--prefill-round-robin-balance]
|
| 106 |
+
[--dist-init-addr DIST_INIT_ADDR] [--nnodes NNODES]
|
| 107 |
+
[--node-rank NODE_RANK]
|
| 108 |
+
[--json-model-override-args JSON_MODEL_OVERRIDE_ARGS]
|
| 109 |
+
[--preferred-sampling-params PREFERRED_SAMPLING_PARAMS]
|
| 110 |
+
[--enable-lora] [--enable-lora-overlap-loading]
|
| 111 |
+
[--max-lora-rank MAX_LORA_RANK]
|
| 112 |
+
[--lora-target-modules [{q_proj,k_proj,v_proj,o_proj,gate_proj,up_proj,down_proj,qkv_proj,gate_up_proj,embed_tokens,lm_head,all} ...]]
|
| 113 |
+
[--lora-paths [LORA_PATHS ...]]
|
| 114 |
+
[--max-loras-per-batch MAX_LORAS_PER_BATCH]
|
| 115 |
+
[--max-loaded-loras MAX_LOADED_LORAS]
|
| 116 |
+
[--lora-eviction-policy {lru,fifo}]
|
| 117 |
+
[--lora-backend {triton,csgmv,ascend,torch_native}]
|
| 118 |
+
[--max-lora-chunk-size {16,32,64,128}]
|
| 119 |
+
[--attention-backend {triton,torch_native,flex_attention,nsa,cutlass_mla,fa3,fa4,flashinfer,flashmla,trtllm_mla,trtllm_mha,dual_chunk_flash_attn,aiter,wave,intel_amx,ascend,intel_xpu}]
|
| 120 |
+
[--prefill-attention-backend {triton,torch_native,flex_attention,nsa,cutlass_mla,fa3,fa4,flashinfer,flashmla,trtllm_mla,trtllm_mha,dual_chunk_flash_attn,aiter,wave,intel_amx,ascend,intel_xpu}]
|
| 121 |
+
[--decode-attention-backend {triton,torch_native,flex_attention,nsa,cutlass_mla,fa3,fa4,flashinfer,flashmla,trtllm_mla,trtllm_mha,dual_chunk_flash_attn,aiter,wave,intel_amx,ascend,intel_xpu}]
|
| 122 |
+
[--sampling-backend {flashinfer,pytorch,ascend}]
|
| 123 |
+
[--grammar-backend {xgrammar,outlines,llguidance,none}]
|
| 124 |
+
[--mm-attention-backend {sdpa,fa3,triton_attn,ascend_attn,aiter_attn}]
|
| 125 |
+
[--nsa-prefill-backend {flashmla_sparse,flashmla_kv,flashmla_auto,fa3,tilelang,aiter}]
|
| 126 |
+
[--nsa-decode-backend {flashmla_sparse,flashmla_kv,flashmla_auto,fa3,tilelang,aiter}]
|
| 127 |
+
[--fp8-gemm-backend {auto,deep_gemm,flashinfer_trtllm,cutlass,triton,aiter}]
|
| 128 |
+
[--fp4-gemm-backend {auto,flashinfer_cudnn,flashinfer_cutlass,flashinfer_trtllm}]
|
| 129 |
+
[--disable-flashinfer-autotune]
|
| 130 |
+
[--speculative-algorithm {EAGLE,EAGLE3,NEXTN,STANDALONE,NGRAM}]
|
| 131 |
+
[--speculative-draft-model-path SPECULATIVE_DRAFT_MODEL_PATH]
|
| 132 |
+
[--speculative-draft-model-revision SPECULATIVE_DRAFT_MODEL_REVISION]
|
| 133 |
+
[--speculative-draft-load-format {auto,pt,safetensors,npcache,dummy,sharded_state,gguf,bitsandbytes,layered,flash_rl,remote,remote_instance,fastsafetensors,private}]
|
| 134 |
+
[--speculative-num-steps SPECULATIVE_NUM_STEPS]
|
| 135 |
+
[--speculative-eagle-topk SPECULATIVE_EAGLE_TOPK]
|
| 136 |
+
[--speculative-num-draft-tokens SPECULATIVE_NUM_DRAFT_TOKENS]
|
| 137 |
+
[--speculative-accept-threshold-single SPECULATIVE_ACCEPT_THRESHOLD_SINGLE]
|
| 138 |
+
[--speculative-accept-threshold-acc SPECULATIVE_ACCEPT_THRESHOLD_ACC]
|
| 139 |
+
[--speculative-token-map SPECULATIVE_TOKEN_MAP]
|
| 140 |
+
[--speculative-attention-mode {prefill,decode}]
|
| 141 |
+
[--speculative-draft-attention-backend SPECULATIVE_DRAFT_ATTENTION_BACKEND]
|
| 142 |
+
[--speculative-moe-runner-backend {auto,deep_gemm,triton,triton_kernel,flashinfer_trtllm,flashinfer_cutlass,flashinfer_mxfp4,flashinfer_cutedsl,cutlass}]
|
| 143 |
+
[--speculative-moe-a2a-backend {none,deepep,mooncake,ascend_fuseep}]
|
| 144 |
+
[--speculative-draft-model-quantization {awq,fp8,gptq,marlin,gptq_marlin,awq_marlin,bitsandbytes,gguf,modelopt,modelopt_fp8,modelopt_fp4,petit_nvfp4,w8a8_int8,w8a8_fp8,moe_wna16,qoq,w4afp8,mxfp4,auto-round,compressed-tensors,modelslim,quark_int4fp8_moe,unquant}]
|
| 145 |
+
[--speculative-ngram-min-match-window-size SPECULATIVE_NGRAM_MIN_MATCH_WINDOW_SIZE]
|
| 146 |
+
[--speculative-ngram-max-match-window-size SPECULATIVE_NGRAM_MAX_MATCH_WINDOW_SIZE]
|
| 147 |
+
[--speculative-ngram-min-bfs-breadth SPECULATIVE_NGRAM_MIN_BFS_BREADTH]
|
| 148 |
+
[--speculative-ngram-max-bfs-breadth SPECULATIVE_NGRAM_MAX_BFS_BREADTH]
|
| 149 |
+
[--speculative-ngram-match-type {BFS,PROB}]
|
| 150 |
+
[--speculative-ngram-branch-length SPECULATIVE_NGRAM_BRANCH_LENGTH]
|
| 151 |
+
[--speculative-ngram-capacity SPECULATIVE_NGRAM_CAPACITY]
|
| 152 |
+
[--enable-multi-layer-eagle]
|
| 153 |
+
[--expert-parallel-size EXPERT_PARALLEL_SIZE]
|
| 154 |
+
[--moe-a2a-backend {none,deepep,mooncake,ascend_fuseep}]
|
| 155 |
+
[--moe-runner-backend {auto,deep_gemm,triton,triton_kernel,flashinfer_trtllm,flashinfer_cutlass,flashinfer_mxfp4,flashinfer_cutedsl,cutlass}]
|
| 156 |
+
[--flashinfer-mxfp4-moe-precision {default,bf16}]
|
| 157 |
+
[--enable-flashinfer-allreduce-fusion]
|
| 158 |
+
[--deepep-mode {normal,low_latency,auto}]
|
| 159 |
+
[--ep-num-redundant-experts EP_NUM_REDUNDANT_EXPERTS]
|
| 160 |
+
[--ep-dispatch-algorithm EP_DISPATCH_ALGORITHM]
|
| 161 |
+
[--init-expert-location INIT_EXPERT_LOCATION]
|
| 162 |
+
[--enable-eplb] [--eplb-algorithm EPLB_ALGORITHM]
|
| 163 |
+
[--eplb-rebalance-num-iterations EPLB_REBALANCE_NUM_ITERATIONS]
|
| 164 |
+
[--eplb-rebalance-layers-per-chunk EPLB_REBALANCE_LAYERS_PER_CHUNK]
|
| 165 |
+
[--eplb-min-rebalancing-utilization-threshold EPLB_MIN_REBALANCING_UTILIZATION_THRESHOLD]
|
| 166 |
+
[--expert-distribution-recorder-mode EXPERT_DISTRIBUTION_RECORDER_MODE]
|
| 167 |
+
[--expert-distribution-recorder-buffer-size EXPERT_DISTRIBUTION_RECORDER_BUFFER_SIZE]
|
| 168 |
+
[--enable-expert-distribution-metrics]
|
| 169 |
+
[--deepep-config DEEPEP_CONFIG]
|
| 170 |
+
[--moe-dense-tp-size MOE_DENSE_TP_SIZE]
|
| 171 |
+
[--elastic-ep-backend {none,mooncake}]
|
| 172 |
+
[--mooncake-ib-device MOONCAKE_IB_DEVICE]
|
| 173 |
+
[--max-mamba-cache-size MAX_MAMBA_CACHE_SIZE]
|
| 174 |
+
[--mamba-ssm-dtype {float32,bfloat16}]
|
| 175 |
+
[--mamba-full-memory-ratio MAMBA_FULL_MEMORY_RATIO]
|
| 176 |
+
[--mamba-scheduler-strategy {auto,no_buffer,extra_buffer}]
|
| 177 |
+
[--mamba-track-interval MAMBA_TRACK_INTERVAL]
|
| 178 |
+
[--enable-hierarchical-cache]
|
| 179 |
+
[--hicache-ratio HICACHE_RATIO]
|
| 180 |
+
[--hicache-size HICACHE_SIZE]
|
| 181 |
+
[--hicache-write-policy {write_back,write_through,write_through_selective}]
|
| 182 |
+
[--hicache-io-backend {direct,kernel,kernel_ascend}]
|
| 183 |
+
[--hicache-mem-layout {layer_first,page_first,page_first_direct,page_first_kv_split,page_head}]
|
| 184 |
+
[--disable-hicache-numa-detect]
|
| 185 |
+
[--hicache-storage-backend {file,mooncake,hf3fs,nixl,aibrix,dynamic,eic}]
|
| 186 |
+
[--hicache-storage-prefetch-policy {best_effort,wait_complete,timeout}]
|
| 187 |
+
[--hicache-storage-backend-extra-config HICACHE_STORAGE_BACKEND_EXTRA_CONFIG]
|
| 188 |
+
[--hierarchical-sparse-attention-extra-config HIERARCHICAL_SPARSE_ATTENTION_EXTRA_CONFIG]
|
| 189 |
+
[--enable-lmcache] [--kt-weight-path KT_WEIGHT_PATH]
|
| 190 |
+
[--kt-method KT_METHOD] [--kt-cpuinfer KT_CPUINFER]
|
| 191 |
+
[--kt-threadpool-count KT_THREADPOOL_COUNT]
|
| 192 |
+
[--kt-num-gpu-experts KT_NUM_GPU_EXPERTS]
|
| 193 |
+
[--kt-max-deferred-experts-per-token KT_MAX_DEFERRED_EXPERTS_PER_TOKEN]
|
| 194 |
+
[--dllm-algorithm DLLM_ALGORITHM]
|
| 195 |
+
[--dllm-algorithm-config DLLM_ALGORITHM_CONFIG]
|
| 196 |
+
[--enable-double-sparsity]
|
| 197 |
+
[--ds-channel-config-path DS_CHANNEL_CONFIG_PATH]
|
| 198 |
+
[--ds-heavy-channel-num DS_HEAVY_CHANNEL_NUM]
|
| 199 |
+
[--ds-heavy-token-num DS_HEAVY_TOKEN_NUM]
|
| 200 |
+
[--ds-heavy-channel-type DS_HEAVY_CHANNEL_TYPE]
|
| 201 |
+
[--ds-sparse-decode-threshold DS_SPARSE_DECODE_THRESHOLD]
|
| 202 |
+
[--cpu-offload-gb CPU_OFFLOAD_GB]
|
| 203 |
+
[--offload-group-size OFFLOAD_GROUP_SIZE]
|
| 204 |
+
[--offload-num-in-group OFFLOAD_NUM_IN_GROUP]
|
| 205 |
+
[--offload-prefetch-step OFFLOAD_PREFETCH_STEP]
|
| 206 |
+
[--offload-mode OFFLOAD_MODE]
|
| 207 |
+
[--multi-item-scoring-delimiter MULTI_ITEM_SCORING_DELIMITER]
|
| 208 |
+
[--disable-radix-cache]
|
| 209 |
+
[--cuda-graph-max-bs CUDA_GRAPH_MAX_BS]
|
| 210 |
+
[--cuda-graph-bs CUDA_GRAPH_BS [CUDA_GRAPH_BS ...]]
|
| 211 |
+
[--disable-cuda-graph] [--disable-cuda-graph-padding]
|
| 212 |
+
[--enable-profile-cuda-graph] [--enable-cudagraph-gc]
|
| 213 |
+
[--enable-layerwise-nvtx-marker] [--enable-nccl-nvls]
|
| 214 |
+
[--enable-symm-mem]
|
| 215 |
+
[--disable-flashinfer-cutlass-moe-fp4-allgather]
|
| 216 |
+
[--enable-tokenizer-batch-encode]
|
| 217 |
+
[--disable-tokenizer-batch-decode]
|
| 218 |
+
[--disable-outlines-disk-cache]
|
| 219 |
+
[--disable-custom-all-reduce] [--enable-mscclpp]
|
| 220 |
+
[--enable-torch-symm-mem] [--disable-overlap-schedule]
|
| 221 |
+
[--enable-mixed-chunk] [--enable-dp-attention]
|
| 222 |
+
[--enable-dp-lm-head] [--enable-two-batch-overlap]
|
| 223 |
+
[--enable-single-batch-overlap]
|
| 224 |
+
[--tbo-token-distribution-threshold TBO_TOKEN_DISTRIBUTION_THRESHOLD]
|
| 225 |
+
[--enable-torch-compile]
|
| 226 |
+
[--enable-torch-compile-debug-mode]
|
| 227 |
+
[--enable-piecewise-cuda-graph]
|
| 228 |
+
[--piecewise-cuda-graph-tokens PIECEWISE_CUDA_GRAPH_TOKENS [PIECEWISE_CUDA_GRAPH_TOKENS ...]]
|
| 229 |
+
[--piecewise-cuda-graph-compiler {eager,inductor}]
|
| 230 |
+
[--torch-compile-max-bs TORCH_COMPILE_MAX_BS]
|
| 231 |
+
[--piecewise-cuda-graph-max-tokens PIECEWISE_CUDA_GRAPH_MAX_TOKENS]
|
| 232 |
+
[--torchao-config TORCHAO_CONFIG]
|
| 233 |
+
[--enable-nan-detection] [--enable-p2p-check]
|
| 234 |
+
[--triton-attention-reduce-in-fp32]
|
| 235 |
+
[--triton-attention-num-kv-splits TRITON_ATTENTION_NUM_KV_SPLITS]
|
| 236 |
+
[--triton-attention-split-tile-size TRITON_ATTENTION_SPLIT_TILE_SIZE]
|
| 237 |
+
[--num-continuous-decode-steps NUM_CONTINUOUS_DECODE_STEPS]
|
| 238 |
+
[--delete-ckpt-after-loading] [--enable-memory-saver]
|
| 239 |
+
[--enable-weights-cpu-backup]
|
| 240 |
+
[--enable-draft-weights-cpu-backup]
|
| 241 |
+
[--allow-auto-truncate]
|
| 242 |
+
[--enable-custom-logit-processor]
|
| 243 |
+
[--flashinfer-mla-disable-ragged]
|
| 244 |
+
[--disable-shared-experts-fusion]
|
| 245 |
+
[--disable-chunked-prefix-cache]
|
| 246 |
+
[--disable-fast-image-processor]
|
| 247 |
+
[--keep-mm-feature-on-device]
|
| 248 |
+
[--enable-return-hidden-states]
|
| 249 |
+
[--enable-return-routed-experts]
|
| 250 |
+
[--scheduler-recv-interval SCHEDULER_RECV_INTERVAL]
|
| 251 |
+
[--numa-node NUMA_NODE [NUMA_NODE ...]]
|
| 252 |
+
[--enable-deterministic-inference]
|
| 253 |
+
[--rl-on-policy-target {fsdp}]
|
| 254 |
+
[--enable-attn-tp-input-scattered]
|
| 255 |
+
[--enable-nsa-prefill-context-parallel]
|
| 256 |
+
[--nsa-prefill-cp-mode {in-seq-split,round-robin-split}]
|
| 257 |
+
[--enable-fused-qk-norm-rope]
|
| 258 |
+
[--enable-precise-embedding-interpolation]
|
| 259 |
+
[--enable-dynamic-batch-tokenizer]
|
| 260 |
+
[--dynamic-batch-tokenizer-batch-size DYNAMIC_BATCH_TOKENIZER_BATCH_SIZE]
|
| 261 |
+
[--dynamic-batch-tokenizer-batch-timeout DYNAMIC_BATCH_TOKENIZER_BATCH_TIMEOUT]
|
| 262 |
+
[--debug-tensor-dump-output-folder DEBUG_TENSOR_DUMP_OUTPUT_FOLDER]
|
| 263 |
+
[--debug-tensor-dump-layers DEBUG_TENSOR_DUMP_LAYERS [DEBUG_TENSOR_DUMP_LAYERS ...]]
|
| 264 |
+
[--debug-tensor-dump-input-file DEBUG_TENSOR_DUMP_INPUT_FILE]
|
| 265 |
+
[--debug-tensor-dump-inject DEBUG_TENSOR_DUMP_INJECT]
|
| 266 |
+
[--disaggregation-mode {null,prefill,decode}]
|
| 267 |
+
[--disaggregation-transfer-backend {mooncake,nixl,ascend,fake}]
|
| 268 |
+
[--disaggregation-bootstrap-port DISAGGREGATION_BOOTSTRAP_PORT]
|
| 269 |
+
[--disaggregation-decode-tp DISAGGREGATION_DECODE_TP]
|
| 270 |
+
[--disaggregation-decode-dp DISAGGREGATION_DECODE_DP]
|
| 271 |
+
[--disaggregation-prefill-pp DISAGGREGATION_PREFILL_PP]
|
| 272 |
+
[--disaggregation-ib-device DISAGGREGATION_IB_DEVICE]
|
| 273 |
+
[--disaggregation-decode-enable-offload-kvcache]
|
| 274 |
+
[--disaggregation-decode-enable-fake-auto]
|
| 275 |
+
[--num-reserved-decode-tokens NUM_RESERVED_DECODE_TOKENS]
|
| 276 |
+
[--disaggregation-decode-polling-interval DISAGGREGATION_DECODE_POLLING_INTERVAL]
|
| 277 |
+
[--encoder-only] [--language-only]
|
| 278 |
+
[--encoder-transfer-backend {zmq_to_scheduler,zmq_to_tokenizer,mooncake}]
|
| 279 |
+
[--encoder-urls ENCODER_URLS [ENCODER_URLS ...]]
|
| 280 |
+
[--custom-weight-loader [CUSTOM_WEIGHT_LOADER ...]]
|
| 281 |
+
[--weight-loader-disable-mmap]
|
| 282 |
+
[--remote-instance-weight-loader-seed-instance-ip REMOTE_INSTANCE_WEIGHT_LOADER_SEED_INSTANCE_IP]
|
| 283 |
+
[--remote-instance-weight-loader-seed-instance-service-port REMOTE_INSTANCE_WEIGHT_LOADER_SEED_INSTANCE_SERVICE_PORT]
|
| 284 |
+
[--remote-instance-weight-loader-send-weights-group-ports REMOTE_INSTANCE_WEIGHT_LOADER_SEND_WEIGHTS_GROUP_PORTS]
|
| 285 |
+
[--remote-instance-weight-loader-backend {transfer_engine,nccl}]
|
| 286 |
+
[--remote-instance-weight-loader-start-seed-via-transfer-engine]
|
| 287 |
+
[--enable-pdmux]
|
| 288 |
+
[--pdmux-config-path PDMUX_CONFIG_PATH]
|
| 289 |
+
[--sm-group-num SM_GROUP_NUM] [--config CONFIG]
|
| 290 |
+
[--mm-max-concurrent-calls MM_MAX_CONCURRENT_CALLS]
|
| 291 |
+
[--mm-per-request-timeout MM_PER_REQUEST_TIMEOUT]
|
| 292 |
+
[--enable-broadcast-mm-inputs-process]
|
| 293 |
+
[--mm-process-config MM_PROCESS_CONFIG]
|
| 294 |
+
[--mm-enable-dp-encoder]
|
| 295 |
+
[--limit-mm-data-per-request LIMIT_MM_DATA_PER_REQUEST]
|
| 296 |
+
[--decrypted-config-file DECRYPTED_CONFIG_FILE]
|
| 297 |
+
[--decrypted-draft-config-file DECRYPTED_DRAFT_CONFIG_FILE]
|
| 298 |
+
[--enable-prefix-mm-cache]
|
| 299 |
+
[--forward-hooks FORWARD_HOOKS]
|
| 300 |
+
$ nvidia-smi -L
|
| 301 |
+
GPU 0: NVIDIA RTX PRO 6000 Blackwell Workstation Edition (UUID: GPU-9d2a0cb2-d380-60e5-580d-cd14950c67a9)
|
| 302 |
+
$ nvidia-smi --query-gpu=memory.free,memory.used,memory.total --format=csv
|
| 303 |
+
memory.free [MiB], memory.used [MiB], memory.total [MiB]
|
| 304 |
+
97215 MiB, 30 MiB, 97887 MiB
|
model-00001-of-00015.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:447d62c327213921556b02ca0d1ac13c562bd9de6d56c9bf046beefee297dd9d
|
| 3 |
+
size 4882434912
|
model-00002-of-00015.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3abeb410e6c2c7b26029bdb1ceffadfc8676c56524e2549d830528843721a33d
|
| 3 |
+
size 4869903000
|
model-00003-of-00015.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:736f493ed8214d6e07aac0cdf74359b0daa13bbf1dacfd0f052e798dc7b45bc2
|
| 3 |
+
size 4869903136
|
model-00004-of-00015.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c61a87b5cc0be08071aad2705fe28354c44d3465f0c242fe7511f023614fcddb
|
| 3 |
+
size 4969044352
|
model-00005-of-00015.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2471290d5d20d6b78aedb970236dcb796980b7f3edcf7b21464af2f00acaaeca
|
| 3 |
+
size 4954838264
|
model-00006-of-00015.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c52bab97a0bd7113e89828a21c301fbcd5a4e5ac27ae04f1d2aa7318cc4fac63
|
| 3 |
+
size 4869903136
|
model-00007-of-00015.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:44f4497add18953da7d0071b9c722e50521d4ce2a97fc5cba814500388d6e1af
|
| 3 |
+
size 4969044352
|
model-00008-of-00015.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:985384ef888837d4d89060ba0e9b8586b772e232f2d13b7571a6d1dbc38bcc82
|
| 3 |
+
size 4954838264
|
model-00009-of-00015.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2457facc7d24efd840f2a2fa6ce7c416553aeecd5dbb9ce52e4d13f84c731394
|
| 3 |
+
size 4869903136
|
model-00010-of-00015.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:47dade11e77d8e5f48a0e6f202b5da33b90c3fc07112b4c718337cf7c9c3b98a
|
| 3 |
+
size 4969044352
|
model-00011-of-00015.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4d3b421695eca4697020448f10873255625a1334eb002963e7266835e762fa64
|
| 3 |
+
size 4954838264
|
model-00012-of-00015.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ff074803df240f9c5ca64686b459a51e5d32007d4ea7886f91bb7c7e054f8e02
|
| 3 |
+
size 4869903136
|
model-00013-of-00015.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5acbc16127454a327ab2afad1b9e328cfeb8d9fc6f9e4ff9947fa01dd4685fc0
|
| 3 |
+
size 4969044352
|
model-00014-of-00015.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5b5cebc7c7d0aee6588b8424c457c5d527623ebb660992bcd59e45aa3903b913
|
| 3 |
+
size 4954838264
|
model-00015-of-00015.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:56f4a96b87c468ed924a6661edd0f918e7fedbd39eda45711654e57d31d45722
|
| 3 |
+
size 1201743176
|
model.safetensors.index.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
recipe.yaml
ADDED
|
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
default_stage:
|
| 2 |
+
default_modifiers:
|
| 3 |
+
QuantizationModifier:
|
| 4 |
+
targets: [Linear]
|
| 5 |
+
ignore: [lm_head]
|
| 6 |
+
scheme: NVFP4
|
sglang.sh
ADDED
|
@@ -0,0 +1,19 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#!/bin/bash
|
| 2 |
+
|
| 3 |
+
export SGLANG_ENABLE_JIT_DEEPGEMM=0
|
| 4 |
+
|
| 5 |
+
/home/administrator/workspace/sglang/python/.venv/bin/python -m sglang.launch_server \
|
| 6 |
+
--model-path "/home/administrator/workspace2/models/Mistral-Large-Instruct-2411-NVFP4" \
|
| 7 |
+
--served-model-name "Mistral-Large-Instruct" \
|
| 8 |
+
--tp-size 1 \
|
| 9 |
+
--attention-backend triton \
|
| 10 |
+
--fp4-gemm-backend flashinfer_cutlass \
|
| 11 |
+
--fp8-gemm-backend cutlass \
|
| 12 |
+
--kv-cache-dtype fp8_e4m3 \
|
| 13 |
+
--disable-radix-cache \
|
| 14 |
+
--chunked-prefill-size -1 \
|
| 15 |
+
--mem-fraction-static 0.85 \
|
| 16 |
+
--tool-call-parser mistral \
|
| 17 |
+
--max-total-tokens 65536 \
|
| 18 |
+
--host "192.168.0.104" \
|
| 19 |
+
--port 8006
|
special_tokens_map.json
ADDED
|
@@ -0,0 +1,23 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"bos_token": {
|
| 3 |
+
"content": "<s>",
|
| 4 |
+
"lstrip": false,
|
| 5 |
+
"normalized": false,
|
| 6 |
+
"rstrip": false,
|
| 7 |
+
"single_word": false
|
| 8 |
+
},
|
| 9 |
+
"eos_token": {
|
| 10 |
+
"content": "</s>",
|
| 11 |
+
"lstrip": false,
|
| 12 |
+
"normalized": false,
|
| 13 |
+
"rstrip": false,
|
| 14 |
+
"single_word": false
|
| 15 |
+
},
|
| 16 |
+
"unk_token": {
|
| 17 |
+
"content": "<unk>",
|
| 18 |
+
"lstrip": false,
|
| 19 |
+
"normalized": false,
|
| 20 |
+
"rstrip": false,
|
| 21 |
+
"single_word": false
|
| 22 |
+
}
|
| 23 |
+
}
|
tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
tokenizer_config.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|