First run will download ~2.1 GB (q4, smallest ONNX). Desktop Chrome/Edge recommended.
This will download model files (~2.1 GB, quantized q4) to your browser cache and run inference via WebGPU. Proceed?
WebGPU requires HTTPS or localhost. Chrome/Edge supported by default; Safari/Firefox may need experimental flags.