Spaces:
Running
Running
| title: LFM2-VL-450M WebGPU | |
| emoji: ⚡️ | |
| colorFrom: purple | |
| colorTo: blue | |
| sdk: docker | |
| pinned: false | |
| models: | |
| - onnx-community/LFM2-VL-450M-ONNX | |
| - onnx-community/LFM2-VL-450M | |
| short_description: In-browser vision-language inference with LFM2-VL-450M | |
| # LFM2-VL-450M WebGPU Demo | |
| In-browser vision-language inference with LFM2-VL-450M, powered by ONNX Runtime and WebGPU. | |
| Everything runs entirely in your browser with WebGPU acceleration - no data is sent to a server. | |
| ## Features | |
| - **Live Webcam Captioning**: Stream from your webcam with real-time AI-generated captions | |
| - **Multiple Precision Options**: Choose between FP16 (~1.05 GB) or FP32 (~2.1 GB) | |
| - **Browser Caching**: Models are cached locally after first download for faster subsequent loads | |
| - **Adjustable Resolution**: Configure capture resolution (256-512px) for performance tuning | |
| ## Requirements | |
| - WebGPU-enabled browser | |
| - ~1-2 GB memory depending on precision choice | |
| - Webcam access for live captioning | |