NPU - QNN
Collection
leading models optimized for NPU deployment on Qualcomm Snapdragon
•
7 items
•
Updated
phi-3-mini-4k-onnx-qnn is an ONNX QNN int4 quantized version of Microsoft Phi-3-mini-instruct, providing a small fast NPU inference implementation, optimized for NPU deployment on Windows ARM64 AI PCs with Snapdragon Elite X NPU processors.