| context_size: 4096 | |
| f16: true | |
| backend: llama-cpp | |
| function: | |
| capture_llm_results: | |
| - (?s)<Thought>(.*?)</Thought> | |
| grammar: | |
| properties_order: name,arguments | |
| json_regex_match: | |
| - (?s)<Output>(.*?)</Output> | |
| replace_llm_results: | |
| - key: (?s)<Thought>(.*?)</Thought> | |
| value: "" | |
| mmap: true | |
| name: gpt-4 | |
| parameters: | |
| model: localai-functioncall-qwen2.5-7b-v0.5-q4_k_m.gguf | |
| stopwords: | |
| - <|im_end|> | |
| - <dummy32000> | |
| - </s> | |
| template: | |
| chat: | | |
| {{.Input -}} | |
| <|im_start|>assistant | |
| chat_message: | | |
| <|im_start|>{{ .RoleName }} | |
| {{ if .FunctionCall -}} | |
| Function call: | |
| {{ else if eq .RoleName "tool" -}} | |
| Function response: | |
| {{ end -}} | |
| {{ if .Content -}} | |
| {{.Content }} | |
| {{ end -}} | |
| {{ if .FunctionCall -}} | |
| {{toJson .FunctionCall}} | |
| {{ end -}}<|im_end|> | |
| completion: | | |
| {{.Input}} | |
| function: | | |
| <|im_start|>system | |
| You are an AI assistant that executes function calls, and these are the tools at your disposal: | |
| {{range .Functions}} | |
| {'type': 'function', 'function': {'name': '{{.Name}}', 'description': '{{.Description}}', 'parameters': {{toJson .Parameters}} }} | |
| {{end}} | |
| <|im_end|> | |
| {{.Input -}} | |
| <|im_start|>assistant | |
| download_files: | |
| - filename: localai-functioncall-phi-4-v0.3-q4_k_m.gguf | |
| sha256: 23fee048ded2a6e2e1a7b6bbefa6cbf83068f194caa9552aecbaa00fec8a16d5 | |
| uri: huggingface://mudler/LocalAI-functioncall-phi-4-v0.3-Q4_K_M-GGUF/localai-functioncall-phi-4-v0.3-q4_k_m.gguf |