File size: 1,459 Bytes
481e42a | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 | # Modelfile for ThalisAI/DeepSeek-R1-Distill-Llama-70B-heretic
#
# This model is based on Llama 3.1 architecture and requires the Llama 3.1
# chat template instead of the DeepSeek native template. The default template
# from the GGUF metadata uses DeepSeek's fullwidth Unicode special tokens
# which are not correctly handled by the Llama BPE tokenizer in some backends.
#
# Usage:
# ollama create deepseek-r1-70b-heretic -f DeepSeek-R1-Distill-Llama-70B-heretic.Modelfile
# ollama run deepseek-r1-70b-heretic "Hello!"
#
# To use a different quantization, change the FROM line:
# FROM hf.co/ThalisAI/DeepSeek-R1-Distill-Llama-70B-heretic:Q4_K_M
# FROM hf.co/ThalisAI/DeepSeek-R1-Distill-Llama-70B-heretic:Q3_K_M
FROM hf.co/ThalisAI/DeepSeek-R1-Distill-Llama-70B-heretic:Q6_K
TEMPLATE """{{- if .System }}<|start_header_id|>system<|end_header_id|>
{{ .System }}<|eot_id|>{{ end }}
{{- range $i, $_ := .Messages }}
{{- $last := eq (len (slice $.Messages $i)) 1 }}
{{- if eq .Role "user" }}<|start_header_id|>user<|end_header_id|>
{{ .Content }}<|eot_id|>
{{- else if eq .Role "assistant" }}<|start_header_id|>assistant<|end_header_id|>
{{- if .Content }}{{ .Content }}
{{- end }}{{- if not $last }}<|eot_id|>{{- end }}
{{- end }}
{{- if and $last (ne .Role "assistant") }}<|start_header_id|>assistant<|end_header_id|>
{{ end }}
{{- end }}"""
PARAMETER stop <|eot_id|>
PARAMETER stop <|end_of_text|>
PARAMETER temperature 0.6
PARAMETER top_p 0.95
|