NPU - QNN
Collection
leading models optimized for NPU deployment on Qualcomm Snapdragon • 7 items • Updated
phi-3-mini-4k-onnx-qnn is an ONNX QNN int4 quantized version of Microsoft Phi-3-mini-instruct, providing a small fast NPU inference implementation, optimized for NPU deployment on Windows ARM64 AI PCs with Snapdragon Elite X NPU processors.