Spaces:
Running
Running
metadata
title: LFM2.5-VL-1.6B WebGPU
emoji: 🧠
colorFrom: purple
colorTo: blue
sdk: docker
pinned: false
models:
- LiquidAI/LFM2.5-VL-1.6B
- LiquidAI/LFM2.5-VL-1.6B-ONNX
short_description: In-browser vision-language inference with LFM2.5-VL-1.6B
LFM2.5-VL-1.6B WebGPU Demo
In-browser vision-language inference with LFM2.5-VL-1.6B, powered by ONNX Runtime and WebGPU.
Everything runs entirely in your browser with WebGPU acceleration - no data is sent to a server.
Features
- Live Webcam Captioning: Stream from your webcam with real-time AI-generated captions
- Multiple Quantization Options: Choose between Q4/Q4 (
1.5 GB), Q4/FP16 (2.3 GB), or FP16/FP16 (~3.2 GB) - Browser Caching: Models are cached locally after first download for faster subsequent loads
- Adjustable Resolution: Configure capture resolution (256-512px) for performance tuning
Requirements
- WebGPU-enabled browser
- ~2-4 GB memory depending on quantization choice
- Webcam access for live captioning