# Ollama Modelfile to use this fine-tuned model with Ollama. # TODO (Ryan Parker, 7-May-2026): this Modelfile currently doesn't work. # When using safetensors, the base model must be downloaded first, and # Ollama raises an error stating that MLX is required for quantization. # Note that the base model can be downloaded like this: # uv tool run --from huggingface-hub -- hf download LiquidAI/LFM2.5-1.2B-Instruct # When using GGUF, Ollama successfully creates the model, but then raises # an error that LoRA is not supported. # safetensors version # Uses the adapter_model.safetensors trained in this repository. # Usage: # ollama create lfm-coder --file Modelfile --quantize q4_K_M --experimental # FROM hf.co/LiquidAI/LFM2.5-1.2B-Instruct # ADAPTER adapter_model.safetensors # GGUF version # Usage: # ollama create lfm-coder --file Modelfile FROM hf.co/LiquidAI/LFM2.5-1.2B-Instruct-GGUF:Q4_K_M ADAPTER hf.co/rparkr/LFM2.5-1.2B-Instruct-Coding-F16-GGUF # Recommended settings from: https://huggingface.co/LiquidAI/LFM2.5-1.2B-Instruct#:~:text=Generation%20parameters%3A PARAMETER temperature 0.1 PARAMETER top_k 50 PARAMETER repeat_penalty 1.05 PARAMETER num_ctx 128000