# CodeLlama-7b-Instruct-hf ## GenAI WebGPU Model Generated using `onnxruntime_genai.models.builder` - **onnxruntime-genai commit**: `41c4ce18fec1240c2f848725e31fbe8854010188` - **transformers version**: `4.57.0` - **Precision**: int4 - **Execution Provider**: webgpu ``` python -m onnxruntime_genai.models.builder -p int4 -e webgpu -m codellama/CodeLlama-7b-Instruct-hf -o E:\ai-models\CodeLlama-7b-Instruct-hf\onnx-webgpu\ --extra_options int4_algo_config=rtn_last int4_is_symmetric=true prune_lm_head=true enable_webgpu_graph=true ``` ## GGUF Model Downloaded from: [TheBloke/CodeLlama-7B-Instruct-GGUF](https://huggingface.co/TheBloke/CodeLlama-7B-Instruct-GGUF)