---
title: LiteRT-LM
emoji: 🚀
colorFrom: blue
colorTo: green
sdk: static
app_file: index.html
pinned: false
---
# LiteRT-LM
LiteRT-LM is Google's production-ready, high-performance, open-source inference
framework for deploying Large Language Models on edge devices.
🔗 [Product Website](https://ai.google.dev/edge/litert-lm)
## 🔥 What's New: Gemma 4 support with LiteRT-LM
Deploy [Gemma 4](https://blog.google/innovation-and-ai/technology/developers-tools/gemma-4/)
across a broad range of hardware with stellar performance
([blog](https://developers.googleblog.com/bring-state-of-the-art-agentic-skills-to-the-edge-with-gemma-4/)).
👉 Try on Linux, macOS, Windows (WSL) or Raspberry Pi with the
[LiteRT-LM CLI](https://ai.google.dev/edge/litert-lm/cli):
```bash
litert-lm run \
--from-huggingface-repo=litert-community/gemma-4-E2B-it-litert-lm \
gemma-4-E2B-it.litertlm \
--prompt="What is the capital of France?"
```
## 🌟 Key Features
- 📱 **Cross-Platform Support**: Android, iOS, Web, Desktop, and IoT (e.g.
Raspberry Pi).
- 🚀 **Hardware Acceleration**: Peak performance via GPU and NPU accelerators.
- 👁️ **Multi-Modality**: Support for vision and audio inputs.
- 🔧 **Tool Use**: Function calling support for agentic workflows.
- 📚 **Broad Model Support**: Gemma, Llama, Phi-4, Qwen, and more.

---
## 🚀 Production-Ready for Google's Products
LiteRT-LM powers on-device GenAI experiences in **Chrome**, **Chromebook Plus**,
**Pixel Watch**, and more.
You can also try the
[Google AI Edge Gallery](https://github.com/google-ai-edge/gallery) app to run
models immediately on your device.
| **Install the app today from Google Play** | **Install the app today from App Store** |
| :---: | :---: |
|
|
|
### 📰 Blogs & Announcements
| Link | Description |
| :--- | :--- |
| [Bring state-of-the-art agentic skills to the edge with Gemma 4](https://developers.googleblog.com/bring-state-of-the-art-agentic-skills-to-the-edge-with-gemma-4/) | Deploy Gemma 4 in-app and across a broader range of devices with stellar performance and broad reach using LiteRT-LM. |
| [On-device GenAI in Chrome, Chromebook Plus and Pixel Watch](https://developers.googleblog.com/on-device-genai-in-chrome-chromebook-plus-and-pixel-watch-with-litert-lm/) | Deploy language models on wearables and browser-based platforms using LiteRT-LM at scale. |
| [On-device Function Calling in Google AI Edge Gallery](https://developers.googleblog.com/on-device-function-calling-in-google-ai-edge-gallery/) | Explore how to fine-tune FunctionGemma and enable function calling capabilities powered by LiteRT-LM Tool Use APIs. |
| [Google AI Edge small language models, multimodality, and function calling](https://developers.googleblog.com/google-ai-edge-small-language-models-multimodality-rag-function-calling/) | Latest insights on RAG, multimodality, and function calling for edge language models. |
---
## 🏃 Quick Start
### 🔗 Key Links
- 👉 [Technical Overview](https://ai.google.dev/edge/litert-lm/overview) including performance benchmarks, model support, and more.
- 👉 [LiteRT-LM CLI Guide](https://ai.google.dev/edge/litert-lm/cli) including installation, getting started, and advanced usage.
### ⚡ Quick Try (No Code)
Try LiteRT-LM immediately from your terminal without writing a single line of code using [`uv`](https://docs.astral.sh/uv/getting-started/installation/):
```bash
uv tool install litert-lm
litert-lm run \
--from-huggingface-repo=google/gemma-3n-E2B-it-litert-lm \
gemma-3n-E2B-it-int4 \
--prompt="What is the capital of France?"
```
---
### 📚 Supported Language APIs
Ready to get started? Explore our language-specific guides and setup instructions.
| Language | Status | Best For... | Documentation |
| :--- | :--- | :--- | :--- |
| **Kotlin** | ✅ Stable | Android apps & JVM | [Android (Kotlin) Guide](https://ai.google.dev/edge/litert-lm/android) |
| **Python** | ✅ Stable | Prototyping & Scripting | [Python Guide](https://ai.google.dev/edge/litert-lm/python) |
| **C++** | ✅ Stable | High-performance native | [C++ Guide](https://ai.google.dev/edge/litert-lm/cpp) |
| **Swift** | 🚀 In Dev | Native iOS & macOS | (Coming Soon) |
#### 🏗️ Build From Source
This [guide](./docs/getting-started/build-and-run.md) shows how you can
compile LiteRT-LM from source. If you want to build the program from source,
you should checkout the stable [](https://github.com/google-ai-edge/LiteRT-LM/releases/latest) tag.
---
## 📦 Releases
- **v0.10.1**: Deploy [Gemma 4](https://blog.google/innovation-and-ai/technology/developers-tools/gemma-4/) with stellar performance ([blog](https://developers.googleblog.com/bring-state-of-the-art-agentic-skills-to-the-edge-with-gemma-4/)) and introduce [LiteRT-LM CLI](https://ai.google.dev/edge/litert-lm/cli).
- **v0.9.0**: Improvements to function calling capabilities, better app performance stability.
- **v0.8.0**: Desktop GPU support and Multi-Modality.
- **v0.7.0**: NPU acceleration for Gemma models.
For a full list of releases, see [GitHub Releases](https://github.com/google-ai-edge/LiteRT-LM/releases).
---