vision_models all the image related models microsoft/Phi-3-vision-128k-instruct Text Generation • Updated Dec 10, 2025 • 94.5k • 971
Quantize-on-huggingface Running on A10G 1.88k GGUF My Repo 🦙 1.88k Quantize a Hugging Face model to GGUF and create a repo
mistral all the mistral models goes here TheBloke/OpenHermes-2.5-Mistral-7B-GGUF 7B • Updated Nov 2, 2023 • 8.16k • 275 TheBloke/CapybaraHermes-2.5-Mistral-7B-GGUF 7B • Updated Jan 31, 2024 • 5.45k • 125 TheBloke/Mistral-7B-Instruct-v0.2-AWQ Text Generation • 7B • Updated Dec 11, 2023 • 241k • 52 mistralai/Mistral-7B-Instruct-v0.3 7B • Updated Dec 3, 2025 • 1.67M • 2.46k
Find n GPUs to run an LLM Running Featured 1.04k Can You Run It? LLM version 🚀 1.04k Estimate GPU needs for running LLMs Runtime error 25 GGUF VRAM Calculator 📉 25
mistral all the mistral models goes here TheBloke/OpenHermes-2.5-Mistral-7B-GGUF 7B • Updated Nov 2, 2023 • 8.16k • 275 TheBloke/CapybaraHermes-2.5-Mistral-7B-GGUF 7B • Updated Jan 31, 2024 • 5.45k • 125 TheBloke/Mistral-7B-Instruct-v0.2-AWQ Text Generation • 7B • Updated Dec 11, 2023 • 241k • 52 mistralai/Mistral-7B-Instruct-v0.3 7B • Updated Dec 3, 2025 • 1.67M • 2.46k
vision_models all the image related models microsoft/Phi-3-vision-128k-instruct Text Generation • Updated Dec 10, 2025 • 94.5k • 971
Find n GPUs to run an LLM Running Featured 1.04k Can You Run It? LLM version 🚀 1.04k Estimate GPU needs for running LLMs Runtime error 25 GGUF VRAM Calculator 📉 25
Quantize-on-huggingface Running on A10G 1.88k GGUF My Repo 🦙 1.88k Quantize a Hugging Face model to GGUF and create a repo