File size: 2,060 Bytes
32bf06e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
---
license: apache-2.0
base_model: google/functiongemma-270m-it
tags:
  - gemma
  - function-calling
  - tool-use
  - gguf
  - ollama
model-index:
  - name: delia-functiongemma-270m-gguf
    results: []
---

# Delia FunctionGemma 270M - GGUF

This is the GGUF version of [delia-functiongemma-270m](https://huggingface.co/devopsforflops/delia-functiongemma-270m), fine-tuned for Delia MCP tool orchestration.

## Quick Start with Ollama

```bash
# Download the GGUF file
wget https://huggingface.co/devopsforflops/delia-functiongemma-270m-gguf/resolve/main/functiongemma-delia-f16.gguf

# Create Modelfile
cat > Modelfile << 'MODELFILE'
FROM ./functiongemma-delia-f16.gguf

TEMPLATE """{{ if .System }}<start_of_turn>developer
{{ .System }}
<end_of_turn>
{{ end }}<start_of_turn>user
{{ .Prompt }}
<end_of_turn>
<start_of_turn>model
"""

PARAMETER stop <end_of_turn>
PARAMETER stop <start_of_turn>
PARAMETER temperature 0.1
PARAMETER num_ctx 2048
MODELFILE

# Import to Ollama
ollama create functiongemma-delia -f Modelfile

# Test it
ollama run functiongemma-delia "Hello!"
```

## Model Details

| Property | Value |
|----------|-------|
| Base Model | google/functiongemma-270m-it |
| Architecture | Gemma3 |
| Parameters | 268M |
| Quantization | F16 (full precision) |
| File Size | ~518 MB |
| Context Length | 2048 tokens |

## Training

Fine-tuned using LoRA on Delia MCP tool calling examples:
- LoRA rank: 16
- LoRA alpha: 64
- Epochs: 20
- Dataset: 27 training examples from Delia test suite

## Use with Delia

Add to your Delia `settings.json`:

```json
{
  "model_dispatcher": {
    "name": "functiongemma-delia",
    "num_ctx": 2048
  }
}
```

**Important:** The model name must contain "functiongemma" for Delia to apply the correct prompt formatting.

## Related Models

- [delia-functiongemma-270m](https://huggingface.co/devopsforflops/delia-functiongemma-270m) - Full merged HuggingFace model
- [delia-functiongemma-270m-lora](https://huggingface.co/devopsforflops/delia-functiongemma-270m-lora) - LoRA adapter only

## License

Apache 2.0