How to use from the
Use from the
Transformers library
# Use a pipeline as a high-level helper
from transformers import pipeline

pipe = pipeline("text-generation", model="CuriousDragon/functiongemma-270m-tiny-agent")
messages = [
    {"role": "user", "content": "Who are you?"},
]
pipe(messages)
# Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("CuriousDragon/functiongemma-270m-tiny-agent")
model = AutoModelForCausalLM.from_pretrained("CuriousDragon/functiongemma-270m-tiny-agent")
messages = [
    {"role": "user", "content": "Who are you?"},
]
inputs = tokenizer.apply_chat_template(
	messages,
	add_generation_prompt=True,
	tokenize=True,
	return_dict=True,
	return_tensors="pt",
).to(model.device)

outputs = model.generate(**inputs, max_new_tokens=40)
print(tokenizer.decode(outputs[0][inputs["input_ids"].shape[-1]:]))
Quick Links

Tiny Agent: FunctionGemma-270m-IT (Fine-Tuned)

This is a fine-tuned version of google/functiongemma-270m-it optimized for reliable function calling. It was trained as part of the "Tiny Agent Lab" project to distill the capabilities of larger models into a highly efficient 270M parameter model.

Model Description

  • Model Type: Causal LM (Gemma)
  • Language(s): English
  • License: Gemma Terms of Use
  • Finetuned from: google/functiongemma-270m-it

Capabilities

This model is designed to:

  1. Detect User Intent: Accurately identify when a tool call is needed.
  2. Generate Function Calls: Output valid <start_function_call> XML/JSON blocks.
  3. Refuse Out-of-Scope Requests: Politely decline requests for which no tool is available.
  4. Ask Clarification: Request missing parameter values interactively.

Performance (V4 Evaluation)

On a held-out test set of 100 diverse queries:

  • Overall Accuracy: 71%
  • Tool Selection Precision: 88%
  • Tool Selection Recall: 94%
  • F1 Score: 0.91

Usage

from transformers import AutoTokenizer, AutoModelForCausalLM
import torch

model_id = "CuriousDragon/functiongemma-270m-tiny-agent"

tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto", torch_dtype=torch.float16)

# ... (Add your inference code here)

Intended Use

This model is intended for research and educational purposes in building efficient agentic systems. It works best when provided with a clear system prompt defining the available tools.

Downloads last month
7
Safetensors
Model size
0.3B params
Tensor type
F16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for CuriousDragon/functiongemma-270m-tiny-agent

Finetuned
(408)
this model
Finetunes
1 model