marissaw's picture
Update README.md
aca3563 verified
metadata
license: gemma
base_model: google/functiongemma-270m-it
pipeline_tag: text-generation
library_name: litert-lm
tags:
  - gemma3
  - gemma
  - functiongemma
extra_gated_heading: Access Function Gemma 270M FT Tiny Garden on Hugging Face
extra_gated_prompt: >-
  To access Gemma 270M FT Tiny Garden on Hugging Face, you are required to
  review and agree to the gemma license. To do this, please ensure you are
  logged in to Hugging Face and click below. Requests are processed immediately.
extra_gated_button_content: Acknowledge licensed

litert-community/functiongemma-270m-ft-tiny-garden

Main Model Card: google/functiongemma-270m-it

This model card provides the Tiny Garden model that is ready for deployment on the Google AI Edge Gallery app.

Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models. This particular Gemma model is especially small so it is ideal for on-device use cases. By running this model on device, users can have private access to Generative AI technology without even requiring an internet connection.

Try it live

The Tiny Garden model is a finetune of google/functiongemma-270m-it. To try out Tiny Garden, download the Google AI Edge Gallery app on your phone, open the Tiny Garden panel and tap the download button.

Performance

Android

Benchmarked on S25 Ultra with 512 prefill tokens and 256 decode tokens.

Backend Quantization scheme Context length Prefill (tokens/sec) Decode (tokens/sec) Time-to-first-token Model size (MB) Peak RSS Memory (MB)

CPU

dynamic_int8

1024

2231 tk/s

153.6 tk/s

0.45 s

289 MB

513 MB

Notes:

  • Model Size: measured by the size of the file on disk.
  • The inference on CPU is accelerated via the LiteRT XNNPACK delegate with 4 threads
  • Benchmark is run with cache enabled and initialized. During the first run, the latency and memory usage may differ.