shubeydoo's picture
Liquid AI LFM2.5-VL-1.6B-WebGPU Demo
e10c945
---
title: LFM2.5-VL-1.6B WebGPU
emoji: 🧠
colorFrom: purple
colorTo: blue
sdk: docker
pinned: false
models:
- LiquidAI/LFM2.5-VL-1.6B
- LiquidAI/LFM2.5-VL-1.6B-ONNX
short_description: In-browser vision-language inference with LFM2.5-VL-1.6B
---
# LFM2.5-VL-1.6B WebGPU Demo
In-browser vision-language inference with LFM2.5-VL-1.6B, powered by ONNX Runtime and WebGPU.
Everything runs entirely in your browser with WebGPU acceleration - no data is sent to a server.
## Features
- **Live Webcam Captioning**: Stream from your webcam with real-time AI-generated captions
- **Multiple Quantization Options**: Choose between Q4/Q4 (~1.5 GB), Q4/FP16 (~2.3 GB), or FP16/FP16 (~3.2 GB)
- **Browser Caching**: Models are cached locally after first download for faster subsequent loads
- **Adjustable Resolution**: Configure capture resolution (256-512px) for performance tuning
## Requirements
- WebGPU-enabled browser
- ~2-4 GB memory depending on quantization choice
- Webcam access for live captioning