Model Summary

This repository hosts quantized versions of the Llama-3.3-70B-Instruct model.

Format: GGUF
Converter: llama.cpp 9ba399dfa7f115effc63d48e6860a94c9faa31b2
Quantizer: LM-Kit.NET 2024.12.11

For more detailed information on the base model, please visit the following link

Downloads last month
10
GGUF
Model size
71B params
Architecture
llama
Hardware compatibility
Log In to add your hardware

2-bit

3-bit

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support