| | |
| | id: smollm2 |
| | model: smollm2 |
| | name: smollm2 |
| | version: 1 |
| | |
| |
|
| | |
| | |
| | stop: |
| | - <|im_end|> |
| | |
| |
|
| | |
| | stream: true |
| | top_p: 0.9 |
| | temperature: 0.7 |
| | frequency_penalty: 0 |
| | presence_penalty: 0 |
| | max_tokens: 4096 |
| | seed: -1 |
| | dynatemp_range: 0 |
| | dynatemp_exponent: 1 |
| | top_k: 40 |
| | min_p: 0.05 |
| | tfs_z: 1 |
| | typ_p: 1 |
| | repeat_last_n: 64 |
| | repeat_penalty: 1 |
| | mirostat: false |
| | mirostat_tau: 5 |
| | mirostat_eta: 0.100000001 |
| | penalize_nl: false |
| | ignore_eos: false |
| | n_probs: 0 |
| | min_keep: 0 |
| | |
| | |
| |
|
| | |
| | |
| | engine: llama-cpp |
| | ctx_len: 4096 |
| | ngl: 34 |
| | prompt_template: "<|im_start|>system |
| | {system_message}<|im_end|> |
| | <|im_start|>user |
| | {prompt}<|im_end|> |
| | <|im_start|>assistant" |
| | |
| | |
| |
|