shubeydoo's picture
Liquid AI LFM2.5-VL-1.6B-WebGPU Demo
e10c945
metadata
title: LFM2.5-VL-1.6B WebGPU
emoji: 🧠
colorFrom: purple
colorTo: blue
sdk: docker
pinned: false
models:
  - LiquidAI/LFM2.5-VL-1.6B
  - LiquidAI/LFM2.5-VL-1.6B-ONNX
short_description: In-browser vision-language inference with LFM2.5-VL-1.6B

LFM2.5-VL-1.6B WebGPU Demo

In-browser vision-language inference with LFM2.5-VL-1.6B, powered by ONNX Runtime and WebGPU.

Everything runs entirely in your browser with WebGPU acceleration - no data is sent to a server.

Features

  • Live Webcam Captioning: Stream from your webcam with real-time AI-generated captions
  • Multiple Quantization Options: Choose between Q4/Q4 (1.5 GB), Q4/FP16 (2.3 GB), or FP16/FP16 (~3.2 GB)
  • Browser Caching: Models are cached locally after first download for faster subsequent loads
  • Adjustable Resolution: Configure capture resolution (256-512px) for performance tuning

Requirements

  • WebGPU-enabled browser
  • ~2-4 GB memory depending on quantization choice
  • Webcam access for live captioning