File size: 1,585 Bytes
1978816
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
---
base_model: google/functiongemma-270m-it
library_name: transformers.js
model_name: functiongemma-tasky-onnx
tags:
- onnx
- transformers.js
- text-generation
- function-calling
- task-management
---

# FunctionGemma Tasky (ONNX)

This repository contains an ONNX export of `functiongemma-tasky`, a fine-tuned variant of `google/functiongemma-270m-it` trained for task/todo function-calling. It targets Transformers.js and includes both full precision and Q4 quantized weights.

## Files

- `onnx/model.onnx`: full-precision weights (fp32)
- `onnx/model_q4.onnx`: 4-bit quantized weights (q4)

## Usage (Transformers.js)

```javascript
import { pipeline } from '@huggingface/transformers';

// Q4 (smaller, faster)
const pipe = await pipeline('text-generation', 'REPLACE_WITH_HF_REPO', {
  dtype: 'q4',
});

const out = await pipe('Add a task to call Alice tomorrow at 9am', {
  max_new_tokens: 128,
});
console.log(out[0].generated_text);
```

To load full precision instead:

```javascript
const pipe = await pipeline('text-generation', 'REPLACE_WITH_HF_REPO', {
  dtype: 'fp32',
});
```

Transformers.js expects ONNX weights under an `onnx/` subfolder, which this repo provides.

## Training summary

- Base model: `google/functiongemma-270m-it`
- Fine-tuning data: synthetic task/todo function-calling prompts, mixed English/Italian, includes user-style typos
- Eval success rate: ~99.5% on a 1500/500 train/eval split

## Notes

- Quantized models trade some accuracy for faster inference and smaller size.
- Outputs may not be strict JSON; validate and post-process if needed.