DistillAgent-PaperQA-3B-GGUF

GGUF exports for QuantumCuddle/DistillAgent-PaperQA-3B.

Main HF model repo:

Files:

  • ...-F16.gguf (high quality, larger)
  • ...-Q4_K_M.gguf (smaller/faster, if uploaded)

Example (llama.cpp):

llama-cli -m qwen-qwen25-3b-instruct-qasper-qasper_qwen25_3b_sft_v2_fixed-F16.gguf -p "QUESTION: ..."

For metrics, training details, and limitations, see the main model card.

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for QuantumCuddle/DistillAgent-PaperQA-3B-GGUF

Base model

Qwen/Qwen2.5-3B
Finetuned
(1)
this model