| { | |
| "model_lib": "mistral", | |
| "local_id": "mistral", | |
| "conv_template": "mistral_default", | |
| "temperature": 0.7, | |
| "repetition_penalty": 1.0, | |
| "conv_config": { | |
| "seps": [ | |
| " " | |
| ], | |
| "stop_tokens": [ | |
| 2 | |
| ], | |
| "offset": 0, | |
| "messages": [ | |
| ], | |
| "stop_str": "</s>", | |
| "roles": [ | |
| "[INST]", | |
| "[/INST]" | |
| ], | |
| "role_msg_sep": ": ", | |
| "role_empty_sep": ": ", | |
| "system": "<s>[INST] A chat between a curious user and an artificial intelligence research data management assistant. The assistant gives helpful, detailed, and polite answers to the user's questions. [/INST]", | |
| "add_bos": true, | |
| "name": "Llama2" | |
| }, | |
| "top_p": 0.95, | |
| "mean_gen_len": 128, | |
| "max_gen_len": 512, | |
| "num_shards": 1, | |
| "shift_fill_factor": 0.3, | |
| "tokenizer_files": [ | |
| "tokenizer.json", | |
| "tokenizer.model" | |
| ], | |
| "model_category": "mistral", | |
| "model_name": "Mistral-7B-Instruct-v0.1", | |
| "vocab_size": 32000, | |
| "sliding_window": 4096, | |
| "prefill_chunk_size": 4096 | |
| } |