hugging-quants/Meta-Llama-3.1-70B-Instruct-AWQ-INT4 Text Generation • Updated Aug 7, 2024 • 103k • 108
hugging-quants/Meta-Llama-3.1-8B-Instruct-GPTQ-INT4 Text Generation • 8B • Updated Aug 7, 2024 • 17.1k • 42
DavidAU/Maximizing-Model-Performance-All-Quants-Types-And-Full-Precision-by-Samplers_Parameters Updated Jul 27, 2025 • 171
DavidAU/L3.1-Dark-Reasoning-LewdPlay-evo-Hermes-R1-Uncensored-8B Text Generation • 8B • Updated Jul 28, 2025 • 770 • 44
DavidAU/L3.1-Dark-Reasoning-Unholy-Hermes-R1-Uncensored-8B Text Generation • 8B • Updated May 28, 2025 • 104 • 15
mradermacher/L3.1-Dark-Reasoning-LewdPlay-evo-Hermes-R1-Uncensored-8B-i1-GGUF 8B • Updated Jul 11, 2025 • 585 • 11
hugging-quants/Meta-Llama-3.1-405B-Instruct-AWQ-INT4 Text Generation • 410B • Updated Sep 13, 2024 • 1.33k • 36
hugging-quants/Meta-Llama-3.1-405B-Instruct-GPTQ-INT4 Text Generation • 410B • Updated Aug 7, 2024 • 83 • 16
hugging-quants/Meta-Llama-3.1-405B-Instruct-BNB-NF4 Text Generation • 423B • Updated Sep 16, 2024 • 12 • 5
hugging-quants/Meta-Llama-3.1-8B-Instruct-BNB-NF4 Text Generation • 8B • Updated Aug 8, 2024 • 2.85k • 8
ModelCloud/Meta-Llama-3.1-8B-Instruct-gptq-4bit Text Generation • 8B • Updated Jul 29, 2024 • 184 • 4
ModelCloud/Meta-Llama-3.1-70B-Instruct-gptq-4bit Text Generation • 71B • Updated Jul 27, 2024 • 78 • 4
hugging-quants/Meta-Llama-3.1-70B-Instruct-GPTQ-INT4 Text Generation • 71B • Updated Aug 7, 2024 • 4.3k • 23
sunnyyy/openbuddy-llama3.1-8b-v22.1-131k-Q4_K_M-GGUF Text Generation • 8B • Updated Jul 25, 2024 • 68
azhiboedova/Meta-Llama-3.1-8B-Instruct-AQLM-2Bit-1x16 Text Generation • 2B • Updated Aug 28, 2024 • 6 • 13