NPU - OpenVINO
Collection
leading models optimized for OpenVINO NPU
•
24 items
•
Updated
llama-3.1-8b-instruct-npu-ov is an OpenVino int4 quantized version of Llama 3.1 Instruct, providing a very fast inference implementation, optimized for AI PCs using Intel NPU.
llama-3.1-instruct is a leading open source general foundation model from Meta.
Base model
meta-llama/Llama-3.1-8B