|
|
--- |
|
|
base_model: google/functiongemma-270m-it |
|
|
library_name: transformers.js |
|
|
model_name: functiongemma-tasky-onnx |
|
|
tags: |
|
|
- onnx |
|
|
- transformers.js |
|
|
- text-generation |
|
|
- function-calling |
|
|
- task-management |
|
|
--- |
|
|
|
|
|
# FunctionGemma Tasky (ONNX) |
|
|
|
|
|
This repository contains an ONNX export of `functiongemma-tasky`, a fine-tuned variant of `google/functiongemma-270m-it` trained for task/todo function-calling. It targets Transformers.js and includes both full precision and Q4 quantized weights. |
|
|
|
|
|
## Files |
|
|
|
|
|
- `onnx/model.onnx`: full-precision weights (fp32) |
|
|
- `onnx/model_q4.onnx`: 4-bit quantized weights (q4) |
|
|
|
|
|
## Usage (Transformers.js) |
|
|
|
|
|
```javascript |
|
|
import { pipeline } from '@huggingface/transformers'; |
|
|
|
|
|
// Q4 (smaller, faster) |
|
|
const pipe = await pipeline('text-generation', 'REPLACE_WITH_HF_REPO', { |
|
|
dtype: 'q4', |
|
|
}); |
|
|
|
|
|
const out = await pipe('Add a task to call Alice tomorrow at 9am', { |
|
|
max_new_tokens: 128, |
|
|
}); |
|
|
console.log(out[0].generated_text); |
|
|
``` |
|
|
|
|
|
To load full precision instead: |
|
|
|
|
|
```javascript |
|
|
const pipe = await pipeline('text-generation', 'REPLACE_WITH_HF_REPO', { |
|
|
dtype: 'fp32', |
|
|
}); |
|
|
``` |
|
|
|
|
|
Transformers.js expects ONNX weights under an `onnx/` subfolder, which this repo provides. |
|
|
|
|
|
## Training summary |
|
|
|
|
|
- Base model: `google/functiongemma-270m-it` |
|
|
- Fine-tuning data: synthetic task/todo function-calling prompts, mixed English/Italian, includes user-style typos |
|
|
- Eval success rate: ~99.5% on a 1500/500 train/eval split |
|
|
|
|
|
## Notes |
|
|
|
|
|
- Quantized models trade some accuracy for faster inference and smaller size. |
|
|
- Outputs may not be strict JSON; validate and post-process if needed. |
|
|
|