# Modelfile for ThalisAI/DeepSeek-R1-Distill-Llama-70B-heretic # # This model is based on Llama 3.1 architecture and requires the Llama 3.1 # chat template instead of the DeepSeek native template. The default template # from the GGUF metadata uses DeepSeek's fullwidth Unicode special tokens # which are not correctly handled by the Llama BPE tokenizer in some backends. # # Usage: # ollama create deepseek-r1-70b-heretic -f DeepSeek-R1-Distill-Llama-70B-heretic.Modelfile # ollama run deepseek-r1-70b-heretic "Hello!" # # To use a different quantization, change the FROM line: # FROM hf.co/ThalisAI/DeepSeek-R1-Distill-Llama-70B-heretic:Q4_K_M # FROM hf.co/ThalisAI/DeepSeek-R1-Distill-Llama-70B-heretic:Q3_K_M FROM hf.co/ThalisAI/DeepSeek-R1-Distill-Llama-70B-heretic:Q6_K TEMPLATE """{{- if .System }}<|start_header_id|>system<|end_header_id|> {{ .System }}<|eot_id|>{{ end }} {{- range $i, $_ := .Messages }} {{- $last := eq (len (slice $.Messages $i)) 1 }} {{- if eq .Role "user" }}<|start_header_id|>user<|end_header_id|> {{ .Content }}<|eot_id|> {{- else if eq .Role "assistant" }}<|start_header_id|>assistant<|end_header_id|> {{- if .Content }}{{ .Content }} {{- end }}{{- if not $last }}<|eot_id|>{{- end }} {{- end }} {{- if and $last (ne .Role "assistant") }}<|start_header_id|>assistant<|end_header_id|> {{ end }} {{- end }}""" PARAMETER stop <|eot_id|> PARAMETER stop <|end_of_text|> PARAMETER temperature 0.6 PARAMETER top_p 0.95