independently-platform commited on
Commit
244fe7e
·
verified ·
1 Parent(s): ef3ae22

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +64 -43
README.md CHANGED
@@ -1,58 +1,79 @@
1
- ---
2
- base_model: google/functiongemma-270m-it
3
- library_name: transformers
4
- model_name: functiongemma-tasky
5
- tags:
6
- - generated_from_trainer
7
- - sft
8
- - trl
9
- licence: license
10
- ---
 
 
 
 
 
11
 
12
- # Model Card for functiongemma-tasky
13
 
14
- This model is a fine-tuned version of [google/functiongemma-270m-it](https://huggingface.co/google/functiongemma-270m-it).
15
- It has been trained using [TRL](https://github.com/huggingface/trl).
 
16
 
17
- ## Quick start
 
 
18
 
19
- ```python
20
- from transformers import pipeline
 
 
 
 
21
 
22
- question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
23
- generator = pipeline("text-generation", model="independently-platform/functiongemma-tasky", device="cuda")
24
- output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
25
- print(output["generated_text"])
26
- ```
27
 
28
- ## Training procedure
 
29
 
30
-
31
 
32
 
33
- This model was trained with SFT.
 
 
 
 
 
34
 
35
- ### Framework versions
36
 
37
- - TRL: 0.26.2
38
- - Transformers: 4.57.3
39
- - Pytorch: 2.9.1+cu128
40
- - Datasets: 4.4.2
41
- - Tokenizers: 0.22.2
42
 
43
- ## Citations
 
 
 
44
 
 
45
 
 
 
46
 
47
- Cite TRL as:
48
-
49
- ```bibtex
50
- @misc{vonwerra2022trl,
51
- title = {{TRL: Transformer Reinforcement Learning}},
52
- author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
53
- year = 2020,
54
- journal = {GitHub repository},
55
- publisher = {GitHub},
56
- howpublished = {\url{https://github.com/huggingface/trl}}
57
- }
58
- ```
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ - it
5
+ base_model: google/functiongemma-270m-it
6
+ tags:
7
+ - function-calling
8
+ - tool-calling
9
+ - task-management
10
+ - todo
11
+ - synthetic
12
+ - transformers
13
+ library_name: transformers
14
+ pipeline_tag: text-generation
15
+ ---
16
 
17
+ # Tasky
18
 
19
+ ## Model Summary
20
+ This model is a fine-tuned **function-calling assistant** for a todo/task application. It maps user requests to one of four tools and produces valid tool
21
+ arguments according to the schema in `AI-TRAINING-TOOLS.md`.
22
 
23
+ - **Base model:** `google/functiongemma-270m-it`
24
+ - **Primary languages:** English and Italian (with light spelling errors/typos to mimic real users)
25
+ - **Task:** Structured tool selection + argument generation
26
 
27
+ ## Intended Use
28
+ Use this model to translate natural language task requests into tool calls for:
29
+ - `create_tasks`
30
+ - `search_tasks`
31
+ - `update_tasks`
32
+ - `delete_tasks`
33
 
34
+ It is designed for **task/todo management** workflows and should be paired with strict validation of tool arguments before execution.
 
 
 
 
35
 
36
+ ### Example
37
+ **Input (user):**
38
 
39
+ Aggiungi un task per pagare la bolletta della luce domani mattina
40
 
41
 
42
+ **Expected output (model):**
43
+ ```json
44
+ {
45
+ "tool_name": "create_tasks",
46
+ "tool_arguments": "{\"tasks\":[{\"content\":\"pagare la bolletta della luce\",\"dueDate\":\"2026-01-13T09:00:00.000Z\"}]}"
47
+ }
48
 
49
+ ## Training Data
50
 
51
+ Synthetic, bilingual tool-calling data built from the tool schema, including:
 
 
 
 
52
 
53
+ - Multiple phrasings and paraphrases
54
+ - Mixed English/Italian prompts
55
+ - Light typos and user mistakes in user_content
56
+ - Broad coverage of optional parameters
57
 
58
+ Splits:
59
 
60
+ - Train: 1,500 examples
61
+ - Eval: 500 examples
62
 
63
+ ## Training Procedure
64
+
65
+ - Fine-tuning on synthetic tool-calling samples
66
+ - Deduplicated examples
67
+ - Balanced coverage of all tools and key parameters
68
+
69
+ ## Evaluation
70
+
71
+ Reported success rate: 99.5% on the 500‑example eval split vs 0% base model.
72
+ Success was measured as exact match on the predicted tool name and the JSON arguments after normalization.
73
+
74
+ ## Limitations
75
+
76
+ - Trained for a specific tool schema; not a general-purpose assistant.
77
+ - Outputs may include incorrect or incomplete tool arguments; validate before execution.
78
+ - Language coverage is strongest in English and Italian.
79
+ - Synthetic data may not capture all real-world user phrasing or ambiguity.