--- base_model: google/functiongemma-270m-it library_name: transformers.js model_name: functiongemma-tasky-onnx tags: - onnx - transformers.js - text-generation - function-calling - task-management --- # FunctionGemma Tasky (ONNX) This repository contains an ONNX export of `functiongemma-tasky`, a fine-tuned variant of `google/functiongemma-270m-it` trained for task/todo function-calling. It targets Transformers.js and includes both full precision and Q4 quantized weights. ## Files - `onnx/model.onnx`: full-precision weights (fp32) - `onnx/model_q4.onnx`: 4-bit quantized weights (q4) ## Usage (Transformers.js) ```javascript import { pipeline } from '@huggingface/transformers'; // Q4 (smaller, faster) const pipe = await pipeline('text-generation', 'REPLACE_WITH_HF_REPO', { dtype: 'q4', }); const out = await pipe('Add a task to call Alice tomorrow at 9am', { max_new_tokens: 128, }); console.log(out[0].generated_text); ``` To load full precision instead: ```javascript const pipe = await pipeline('text-generation', 'REPLACE_WITH_HF_REPO', { dtype: 'fp32', }); ``` Transformers.js expects ONNX weights under an `onnx/` subfolder, which this repo provides. ## Training summary - Base model: `google/functiongemma-270m-it` - Fine-tuning data: synthetic task/todo function-calling prompts, mixed English/Italian, includes user-style typos - Eval success rate: ~99.5% on a 1500/500 train/eval split ## Notes - Quantized models trade some accuracy for faster inference and smaller size. - Outputs may not be strict JSON; validate and post-process if needed.