File size: 2,247 Bytes
b2343db ef3ae22 244fe7e ef3ae22 b2343db 244fe7e ef3ae22 244fe7e ef3ae22 244fe7e ef3ae22 244fe7e ef3ae22 244fe7e ef3ae22 244fe7e ef3ae22 244fe7e ef3ae22 244fe7e ef3ae22 244fe7e ef3ae22 244fe7e ef3ae22 244fe7e ef3ae22 244fe7e ef3ae22 244fe7e b2343db |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 |
---
datasets:
- independently-platform/tasky
language:
- en
- it
base_model:
- google/functiongemma-270m-it
library_name: transformers
---
# Tasky
## About the model
This model is a fine-tuned **function-calling assistant** for a todo/task application. It maps user requests to one of four tools and produces valid tool
arguments according to the schema in `AI-TRAINING-TOOLS.md`.
- **Base model:** `google/functiongemma-270m-it`
- **Primary languages:** English and Italian (with light spelling errors/typos to mimic real users)
- **Task:** Structured tool selection + argument generation
## Intended Use
Use this model to translate natural language task requests into tool calls for:
- `create_tasks`
- `search_tasks`
- `update_tasks`
- `delete_tasks`
It is designed for **task/todo management** workflows and should be paired with strict validation of tool arguments before execution.
### Example
**Input (user):**
Aggiungi un task per pagare la bolletta della luce domani mattina
**Expected output (model):**
```json
{
"tool_name": "create_tasks",
"tool_arguments": "{\"tasks\":[{\"content\":\"pagare la bolletta della luce\",\"dueDate\":\"2026-01-13T09:00:00.000Z\"}]}"
}
## Training Data
Synthetic, bilingual tool-calling data built from the tool schema, including:
- Multiple phrasings and paraphrases
- Mixed English/Italian prompts
- Light typos and user mistakes in user_content
- Broad coverage of optional parameters
Splits:
- Train: 1,500 examples
- Eval: 500 examples
## Training Procedure
- Fine-tuning on synthetic tool-calling samples
- Deduplicated examples
- Balanced coverage of all tools and key parameters
## Evaluation
Reported success rate: 99.5% on the 500‑example eval split vs 0% base model.
Success was measured as exact match on the predicted tool name and the JSON arguments after normalization.
## Limitations
- Trained for a specific tool schema; not a general-purpose assistant.
- Outputs may include incorrect or incomplete tool arguments; validate before execution.
- Language coverage is strongest in English and Italian.
- Synthetic data may not capture all real-world user phrasing or ambiguity. |