Pauper Llama 3 8B

Fine-tuned version of meta-llama/Meta-Llama-3-8B-Instruct specialized for Magic: The Gathering's Pauper format using LoRA fine-tuning.

๐Ÿ“ฆ Available Formats

This repository contains both the full HuggingFace model and GGUF quantizations for various use cases.

HuggingFace Transformers (Full Precision)

Perfect for:

  • Further fine-tuning
  • Maximum quality inference
  • Integration with transformers library

GGUF Quantized Models (llama.cpp compatible)

Perfect for:

  • LM Studio, Ollama, llama.cpp
  • Local inference on consumer hardware
  • Faster inference with minimal quality loss
File Size Description Best For
gguf/pauper_llama3_q4km.gguf ~5GB 4-bit quantized Recommended - Best balance
gguf/pauper_llama3_q5km.gguf ~6GB 5-bit quantized Better quality
gguf/pauper_llama3_q8.gguf ~8GB 8-bit quantized Near-original quality
gguf/pauper_llama3_fp16.gguf ~15GB Full precision Maximum quality

๐Ÿš€ Usage

Option 1: HuggingFace Transformers

from transformers import AutoModelForCausalLM, AutoTokenizer
import torch

model = AutoModelForCausalLM.from_pretrained(
    "nmalinowski/pauper-llama3-8b",
    torch_dtype=torch.float16,
    device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained("nmalinowski/pauper-llama3-8b")

prompt = "What are the best cards in Pauper?"
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=256, temperature=0.7)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

Option 2: LM Studio (GGUF - Easiest!)

  1. Download gguf/pauper_llama3_q4km.gguf from Files tab
  2. Open LM Studio โ†’ Load Model
  3. Select the downloaded GGUF file
  4. Start chatting about Pauper!

Option 3: llama.cpp

# Download the quantized model
huggingface-cli download nmalinowski/pauper-llama3-8b gguf/pauper_llama3_q4km.gguf --local-dir ./

# Run inference
./llama-cli -m pauper_llama3_q4km.gguf \
    -p "What are the top Pauper decks in the current meta?" \
    -n 256 \
    --temp 0.7

Option 4: Ollama

# Create Modelfile
cat > Modelfile <<EOF
FROM ./gguf/pauper_llama3_q4km.gguf
PARAMETER temperature 0.7
PARAMETER top_p 0.9
SYSTEM "You are an expert on Magic: The Gathering's Pauper format."
EOF

# Create and run
ollama create pauper-llama3 -f Modelfile
ollama run pauper-llama3 "Explain the current Pauper meta"

๐ŸŽฏ Training Details

  • Base Model: Llama 3 8B Instruct
  • Training Method: LoRA (Low-Rank Adaptation)
  • Domain: Magic: The Gathering - Pauper format
  • LoRA Configuration:
    • Rank: 16
    • Alpha: 32
    • Target modules: q_proj, v_proj
    • Dropout: 0.05

๐Ÿ’ก Recommendations

  • For most users: Download gguf/pauper_llama3_q4km.gguf and use with LM Studio
  • For best quality: Use the full HuggingFace model with transformers
  • For low VRAM: Use Q4_K_M quantization (~5GB)
  • For high VRAM: Use Q8_0 or FP16 for better quality

๐Ÿ“Š Performance

The Q4_K_M quantization offers:

  • โœ… ~95% of full precision quality
  • โœ… 70% smaller file size
  • โœ… Faster inference on CPU and GPU
  • โœ… Runs on consumer hardware (16GB RAM recommended)

๐ŸŽฎ Example Prompts

"What are the best removal spells in Pauper?"
"Build me a Pauper deck around Monastery Swiftspear"
"Explain the differences between Affinity and Elves in Pauper"
"What are the current tier 1 Pauper decks?"

โš ๏ธ Limitations

  • Specialized for Pauper format - may not perform well on other MTG formats
  • May occasionally hallucinate card names or abilities
  • Knowledge cutoff: January 2025
  • Not suitable for medical, legal, or financial advice

๐Ÿ“„ License

This model inherits the Llama 3 Community License from Meta. See LICENSE for details.

๐Ÿ™ Acknowledgments

  • Base model: Meta's Llama 3 8B Instruct
  • Training framework: HuggingFace Transformers + PEFT
  • Quantization: llama.cpp

๐Ÿ“ž Issues & Feedback

If you encounter issues or have suggestions, please open an issue on the Community tab.

Downloads last month
186
Safetensors
Model size
8B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for nmalinowski/pauper-llama3-8b

Adapter
(2185)
this model
Adapters
2 models