EpsteinFilez-8B-GGUF

An 8B parameter GGUF model built on the Qwen3 architecture. Optimized for local inference with Q4_K_M quantization.

Available Quantizations

File Quant Size
EpsteinFilez-8B-Q4_K_M.gguf Q4_K_M ~4.7 GB

Usage

Works with any GGUF-compatible runtime:

# llama.cpp
./llama-cli -m EpsteinFilez-8B-Q4_K_M.gguf -p "Your prompt here"

# Ollama
ollama create epsteinfilez -f Modelfile
ollama run epsteinfilez

# LM Studio โ€” just drag and drop the .gguf file

Details

  • Architecture: Qwen3
  • Parameters: 8B
  • Format: GGUF (Q4_K_M)
  • License: Apache 2.0
Downloads last month
-
GGUF
Model size
8B params
Architecture
qwen3
Hardware compatibility
Log In to add your hardware

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support