Hugging Face
Models
Datasets
Spaces
Buckets
new
Docs
Enterprise
Pricing
Log In
Sign Up
inference-optimization
/
Meta-Llama-3-8B-Instruct-NVFP4-GPTQ-Quant
like
0
Follow
Inference Optimization
20
Safetensors
llama
8-bit precision
compressed-tensors
Model card
Files
Files and versions
xet
Community
main
Meta-Llama-3-8B-Instruct-NVFP4-GPTQ-Quant
/
tokenizer.json
Commit History
Upload folder using huggingface_hub
d274493
verified
dsikka
commited on
Jan 29