mlabonne's picture
Update README.md
d963747 verified
metadata
title: LFM2.5-VL-1.6B WebGPU
emoji: 🧠
colorFrom: purple
colorTo: blue
sdk: docker
pinned: true
models:
  - LiquidAI/LFM2.5-VL-1.6B

LFM2.5-VL-1.6B WebGPU Demo

In-browser vision-language inference with LFM2.5-VL-1.6B, powered by ONNX Runtime and WebGPU.

Everything runs entirely in your browser with WebGPU acceleration - no data is sent to a server.

Features

  • Live Webcam Captioning: Stream from your webcam with real-time AI-generated captions
  • Multiple Quantization Options: Choose between Q4/Q4 (1.5 GB), Q4/FP16 (2.3 GB), or FP16/FP16 (~3.2 GB)
  • Browser Caching: Models are cached locally after first download for faster subsequent loads
  • Adjustable Resolution: Configure capture resolution (256-512px) for performance tuning

Requirements

  • WebGPU-enabled browser (Chrome 113+, Edge 113+)
  • ~2-4 GB memory depending on quantization choice
  • Webcam access for live captioning