How to use from
llama.cpp
Install from brew
brew install llama.cpp
# Start a local OpenAI-compatible server with a web UI:
llama-server -hf devopsforflops/delia-functiongemma-270m-gguf:F16
# Run inference directly in the terminal:
llama-cli -hf devopsforflops/delia-functiongemma-270m-gguf:F16
Install from WinGet (Windows)
winget install llama.cpp
# Start a local OpenAI-compatible server with a web UI:
llama-server -hf devopsforflops/delia-functiongemma-270m-gguf:F16
# Run inference directly in the terminal:
llama-cli -hf devopsforflops/delia-functiongemma-270m-gguf:F16
Use pre-built binary
# Download pre-built binary from:
# https://github.com/ggerganov/llama.cpp/releases
# Start a local OpenAI-compatible server with a web UI:
./llama-server -hf devopsforflops/delia-functiongemma-270m-gguf:F16
# Run inference directly in the terminal:
./llama-cli -hf devopsforflops/delia-functiongemma-270m-gguf:F16
Build from source code
git clone https://github.com/ggerganov/llama.cpp.git
cd llama.cpp
cmake -B build
cmake --build build -j --target llama-server llama-cli
# Start a local OpenAI-compatible server with a web UI:
./build/bin/llama-server -hf devopsforflops/delia-functiongemma-270m-gguf:F16
# Run inference directly in the terminal:
./build/bin/llama-cli -hf devopsforflops/delia-functiongemma-270m-gguf:F16
Use Docker
docker model run hf.co/devopsforflops/delia-functiongemma-270m-gguf:F16
Quick Links

Delia FunctionGemma 270M - GGUF

This is the GGUF version of delia-functiongemma-270m, fine-tuned for Delia MCP tool orchestration.

Quick Start with Ollama

# Download the GGUF file
wget https://huggingface.co/devopsforflops/delia-functiongemma-270m-gguf/resolve/main/functiongemma-delia-f16.gguf

# Create Modelfile
cat > Modelfile << 'MODELFILE'
FROM ./functiongemma-delia-f16.gguf

TEMPLATE """{{ if .System }}<start_of_turn>developer
{{ .System }}
<end_of_turn>
{{ end }}<start_of_turn>user
{{ .Prompt }}
<end_of_turn>
<start_of_turn>model
"""

PARAMETER stop <end_of_turn>
PARAMETER stop <start_of_turn>
PARAMETER temperature 0.1
PARAMETER num_ctx 2048
MODELFILE

# Import to Ollama
ollama create functiongemma-delia -f Modelfile

# Test it
ollama run functiongemma-delia "Hello!"

Model Details

Property Value
Base Model google/functiongemma-270m-it
Architecture Gemma3
Parameters 268M
Quantization F16 (full precision)
File Size ~518 MB
Context Length 2048 tokens

Training

Fine-tuned using LoRA on Delia MCP tool calling examples:

  • LoRA rank: 16
  • LoRA alpha: 64
  • Epochs: 20
  • Dataset: 27 training examples from Delia test suite

Use with Delia

Add to your Delia settings.json:

{
  "model_dispatcher": {
    "name": "functiongemma-delia",
    "num_ctx": 2048
  }
}

Important: The model name must contain "functiongemma" for Delia to apply the correct prompt formatting.

Related Models

License

Apache 2.0

Downloads last month
4
GGUF
Model size
0.3B params
Architecture
gemma3
Hardware compatibility
Log In to add your hardware

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for devopsforflops/delia-functiongemma-270m-gguf

Quantized
(49)
this model