--- title: LiteRT-LM emoji: 🚀 colorFrom: blue colorTo: green sdk: static app_file: index.html pinned: false --- # LiteRT-LM LiteRT-LM is Google's production-ready, high-performance, open-source inference framework for deploying Large Language Models on edge devices. 🔗 [Product Website](https://ai.google.dev/edge/litert-lm) ## 🔥 What's New: Gemma 4 support with LiteRT-LM Deploy [Gemma 4](https://blog.google/innovation-and-ai/technology/developers-tools/gemma-4/) across a broad range of hardware with stellar performance ([blog](https://developers.googleblog.com/bring-state-of-the-art-agentic-skills-to-the-edge-with-gemma-4/)). 👉 Try on Linux, macOS, Windows (WSL) or Raspberry Pi with the [LiteRT-LM CLI](https://ai.google.dev/edge/litert-lm/cli): ```bash litert-lm run \ --from-huggingface-repo=litert-community/gemma-4-E2B-it-litert-lm \ gemma-4-E2B-it.litertlm \ --prompt="What is the capital of France?" ``` ## 🌟 Key Features - 📱 **Cross-Platform Support**: Android, iOS, Web, Desktop, and IoT (e.g. Raspberry Pi). - 🚀 **Hardware Acceleration**: Peak performance via GPU and NPU accelerators. - 👁️ **Multi-Modality**: Support for vision and audio inputs. - 🔧 **Tool Use**: Function calling support for agentic workflows. - 📚 **Broad Model Support**: Gemma, Llama, Phi-4, Qwen, and more. ![](./docs/api/kotlin/demo.gif) --- ## 🚀 Production-Ready for Google's Products LiteRT-LM powers on-device GenAI experiences in **Chrome**, **Chromebook Plus**, **Pixel Watch**, and more. You can also try the [Google AI Edge Gallery](https://github.com/google-ai-edge/gallery) app to run models immediately on your device. | **Install the app today from Google Play** | **Install the app today from App Store** | | :---: | :---: | | Get it on Google Play | Download on the App Store | ### 📰 Blogs & Announcements | Link | Description | | :--- | :--- | | [Bring state-of-the-art agentic skills to the edge with Gemma 4](https://developers.googleblog.com/bring-state-of-the-art-agentic-skills-to-the-edge-with-gemma-4/) | Deploy Gemma 4 in-app and across a broader range of devices with stellar performance and broad reach using LiteRT-LM. | | [On-device GenAI in Chrome, Chromebook Plus and Pixel Watch](https://developers.googleblog.com/on-device-genai-in-chrome-chromebook-plus-and-pixel-watch-with-litert-lm/) | Deploy language models on wearables and browser-based platforms using LiteRT-LM at scale. | | [On-device Function Calling in Google AI Edge Gallery](https://developers.googleblog.com/on-device-function-calling-in-google-ai-edge-gallery/) | Explore how to fine-tune FunctionGemma and enable function calling capabilities powered by LiteRT-LM Tool Use APIs. | | [Google AI Edge small language models, multimodality, and function calling](https://developers.googleblog.com/google-ai-edge-small-language-models-multimodality-rag-function-calling/) | Latest insights on RAG, multimodality, and function calling for edge language models. | --- ## 🏃 Quick Start ### 🔗 Key Links - 👉 [Technical Overview](https://ai.google.dev/edge/litert-lm/overview) including performance benchmarks, model support, and more. - 👉 [LiteRT-LM CLI Guide](https://ai.google.dev/edge/litert-lm/cli) including installation, getting started, and advanced usage. ### ⚡ Quick Try (No Code) Try LiteRT-LM immediately from your terminal without writing a single line of code using [`uv`](https://docs.astral.sh/uv/getting-started/installation/): ```bash uv tool install litert-lm litert-lm run \ --from-huggingface-repo=google/gemma-3n-E2B-it-litert-lm \ gemma-3n-E2B-it-int4 \ --prompt="What is the capital of France?" ``` --- ### 📚 Supported Language APIs Ready to get started? Explore our language-specific guides and setup instructions. | Language | Status | Best For... | Documentation | | :--- | :--- | :--- | :--- | | **Kotlin** | ✅ Stable | Android apps & JVM | [Android (Kotlin) Guide](https://ai.google.dev/edge/litert-lm/android) | | **Python** | ✅ Stable | Prototyping & Scripting | [Python Guide](https://ai.google.dev/edge/litert-lm/python) | | **C++** | ✅ Stable | High-performance native | [C++ Guide](https://ai.google.dev/edge/litert-lm/cpp) | | **Swift** | 🚀 In Dev | Native iOS & macOS | (Coming Soon) | #### 🏗️ Build From Source This [guide](./docs/getting-started/build-and-run.md) shows how you can compile LiteRT-LM from source. If you want to build the program from source, you should checkout the stable [![Latest Release](https://img.shields.io/github/v/release/google-ai-edge/LiteRT-LM)](https://github.com/google-ai-edge/LiteRT-LM/releases/latest) tag. --- ## 📦 Releases - **v0.10.1**: Deploy [Gemma 4](https://blog.google/innovation-and-ai/technology/developers-tools/gemma-4/) with stellar performance ([blog](https://developers.googleblog.com/bring-state-of-the-art-agentic-skills-to-the-edge-with-gemma-4/)) and introduce [LiteRT-LM CLI](https://ai.google.dev/edge/litert-lm/cli). - **v0.9.0**: Improvements to function calling capabilities, better app performance stability. - **v0.8.0**: Desktop GPU support and Multi-Modality. - **v0.7.0**: NPU acceleration for Gemma models. For a full list of releases, see [GitHub Releases](https://github.com/google-ai-edge/LiteRT-LM/releases). ---