update tool call

#12
by Navanit-AI - opened
Files changed (4) hide show
  1. README.md +4 -4
  2. SYSTEM_PROMPT.txt +2 -1
  3. params.json +1 -2
  4. tokenizer_config.json +1 -1
README.md CHANGED
@@ -1,5 +1,4 @@
1
  ---
2
- library_name: vllm
3
  language:
4
  - en
5
  - fr
@@ -12,6 +11,7 @@ language:
12
  - ru
13
  - ko
14
  license: apache-2.0
 
15
  inference: false
16
  base_model:
17
  - mistralai/Mistral-Small-24B-Base-2501
@@ -19,7 +19,7 @@ extra_gated_description: >-
19
  If you want to learn more about how we process your personal data, please read
20
  our <a href="https://mistral.ai/terms/">Privacy Policy</a>.
21
  tags:
22
- - vllm
23
  ---
24
 
25
  # Model Card for Mistral-Small-24B-Instruct-2501
@@ -113,8 +113,8 @@ Model developper: Mistral AI Team
113
  ## Usage
114
 
115
  The model can be used with the following frameworks;
116
- - [`vllm`](https://github.com/vllm-project/vllm): See [here](#vllm)
117
- - [`transformers`](https://github.com/huggingface/transformers): See [here](#transformers)
118
 
119
  ### vLLM
120
 
 
1
  ---
 
2
  language:
3
  - en
4
  - fr
 
11
  - ru
12
  - ko
13
  license: apache-2.0
14
+ library_name: vllm
15
  inference: false
16
  base_model:
17
  - mistralai/Mistral-Small-24B-Base-2501
 
19
  If you want to learn more about how we process your personal data, please read
20
  our <a href="https://mistral.ai/terms/">Privacy Policy</a>.
21
  tags:
22
+ - transformers
23
  ---
24
 
25
  # Model Card for Mistral-Small-24B-Instruct-2501
 
113
  ## Usage
114
 
115
  The model can be used with the following frameworks;
116
+ - [`vllm`](https://github.com/vllm-project/vllm): See [here](#vLLM)
117
+ - [`transformers`](https://github.com/huggingface/transformers): See [here](#Transformers)
118
 
119
  ### vLLM
120
 
SYSTEM_PROMPT.txt CHANGED
@@ -6,4 +6,5 @@ The current date is {today}.
6
  When you're not sure about some information, you say that you don't have the information and don't make up anything.
7
  If the user's question is not clear, ambiguous, or does not provide enough context for you to accurately answer the question, you do not try to answer it right away and you rather ask the user to clarify their request (e.g. "What are some good restaurants around me?" => "Where are you?" or "When is the next flight to Tokyo" => "Where do you travel from?").
8
  You are always very attentive to dates, in particular you try to resolve dates (e.g. "yesterday" is {yesterday}) and when asked about information at specific dates, you discard information that is at another date.
9
- You follow these instructions in all languages, and always respond to the user in the language they use or request.
 
 
6
  When you're not sure about some information, you say that you don't have the information and don't make up anything.
7
  If the user's question is not clear, ambiguous, or does not provide enough context for you to accurately answer the question, you do not try to answer it right away and you rather ask the user to clarify their request (e.g. "What are some good restaurants around me?" => "Where are you?" or "When is the next flight to Tokyo" => "Where do you travel from?").
8
  You are always very attentive to dates, in particular you try to resolve dates (e.g. "yesterday" is {yesterday}) and when asked about information at specific dates, you discard information that is at another date.
9
+ You follow these instructions in all languages, and always respond to the user in the language they use or request.
10
+ Next sections describe the capabilities that you have.
params.json CHANGED
@@ -8,6 +8,5 @@
8
  "norm_eps": 1e-05,
9
  "vocab_size": 131072,
10
  "rope_theta": 100000000.0,
11
- "max_seq_len": 32768,
12
- "max_position_embeddings": 32768
13
  }
 
8
  "norm_eps": 1e-05,
9
  "vocab_size": 131072,
10
  "rope_theta": 100000000.0,
11
+ "max_seq_len": 32768
 
12
  }
tokenizer_config.json CHANGED
@@ -9012,7 +9012,7 @@
9012
  "eos_token": "</s>",
9013
  "extra_special_tokens": {},
9014
  "legacy": true,
9015
- "model_max_length": 32768,
9016
  "tokenizer_class": "LlamaTokenizer",
9017
  "unk_token": "<unk>",
9018
  "use_default_system_prompt": false
 
9012
  "eos_token": "</s>",
9013
  "extra_special_tokens": {},
9014
  "legacy": true,
9015
+ "model_max_length": 1000000000000000019884624838656,
9016
  "tokenizer_class": "LlamaTokenizer",
9017
  "unk_token": "<unk>",
9018
  "use_default_system_prompt": false