How to use webgpu/Phi-4-mini-instruct-ONNX-GQA with Transformers.js:
// npm i @huggingface/transformers import { pipeline } from '@huggingface/transformers'; // Allocate pipeline const pipe = await pipeline('text-generation', 'webgpu/Phi-4-mini-instruct-ONNX-GQA');
With fix of https://huggingface.co/onnx-community/Phi-4-mini-instruct-ONNX-GQA/discussions/4 and https://github.com/huggingface/transformers.js/issues/1460
https://huggingface.co/onnx-community/Phi-4-mini-instruct-ONNX-GQA/tree/refs%2Fpr%2F4
Base model