GGUFLoader / README.md
Hussain2050's picture
Update README.md
97f4c39 verified
---
license: mit
---
**GGUF Loader is a GUI-first, plugin-based launcher for running local LLMs — featuring a built-in floating assistant that gives you instant AI access on top of any window.**
---
## 🧠 Problem
Open-source LLMs are becoming faster and smarter, but the tools around them are still broken. Most setups require:
- Command-line knowledge
- Model management across multiple platforms
- No way to extend with features like summarization or RAG
- No native way to interact with AI while working
This makes local AI powerful in theory, but unusable for most people in practice.
---
## 💡 Solution
**GGUF Loader** brings everything together:
- One-click model loading with GGUF support
- A modern desktop interface anyone can use
- Built-in **floating button** that allows AI to follow your workflow
- A growing **plugin system** — inspired by Blender — so users can extend, customize, and share features
From loading a Mistral model to summarizing your documents or running agents over folders, GGUF Loader is designed to be your personal AI operating system.
---
## 🎯 Vision
> We believe the future of AI is **local**, **modular**, and **user-owned**.
> GGUF Loader is building the user interface layer for that future — a plug-and-play AI desktop engine that works fully offline, built around your workflows and privacy.
---
## 🔧 What’s Next
- Public plugin SDK
- Pro features: drag-drop RAG builder, context memory, advanced floating tools
- Addon marketplace (free and paid)
- Community system for sharing tools
---
## 🛠️ Dev Philosophy
- 💻 Local-first, no cloud dependency
- 🧩 Modular and hackable
- 🖱️ Usable by non-devs
- ⚡ Fast to launch, fast to load
---
> GGUF Loader is open-source and in active development.
> If you're an investor, contributor, or power user — [get in touch](mailto:hussainnazary2@gmail.com) or star the repo to follow progress.
Update (v2.0.1): GGUF Loader now includes a powerful floating assistant button that lets you interact with any text instantly, system-wide. Plus, it now supports an addon system, allowing developers to extend its capabilities just like a plugin-based platform. These features make GGUF Loader even more flexible and interactive for local AI workflows. Try the new version now!
# 🧠 GGUF Loader Quickstart
### 📦 1. Install GGUF Loader via pip
```bash
pip install ggufloader
```
### 🚀 2. Launch the App
After installation, run the following command in your terminal:
```bash
ggufloader
```
This will start the GGUF Loader interface. You can now load and chat with any GGUF model locally.
## 🧩 🎬 Demo Video: Addon System + Floating Tool in Local LLM (v2.0.1 Update)
[![Watch the video](https://img.youtube.com/vi/5lQui7EeUe0/maxresdefault.jpg)](https://www.youtube.com/watch?v=5lQui7EeUe0)
> Discover how to supercharge your local AI workflows using the new floating addon system! No coding needed. Works offline.
> Let me know if you want to support GUI launching, system tray, or shortcuts too.
---
## 🔽 Download GGUF Models
> ⚡ Click a link below to download the model file directly (no Hugging Face page in between).
### 🧠 GPT-OSS Models (Open Source GPTs)
> High-quality, Apache 2.0 licensed, reasoning-focused models for local/enterprise use.
#### 🧠 GPT-OSS 120B (Dense)
- [⬇️ Download Q4_K (46.2 GB)](https://huggingface.co/lmstudio-community/gpt-oss-120b-GGUF/resolve/main/gpt-oss-120b-MXFP4-00001-of-00002.gguf)
#### 🧠 GPT-OSS 20B (Dense)
- [⬇️ Download Q4_K (7.34 GB)](https://huggingface.co/lmstudio-community/gpt-oss-20b-GGUF/resolve/main/gpt-oss-20b-MXFP4.gguf)
---
### 🧠 Mistral-7B Instruct
- [⬇️ Download Q4_0 (4.23 GB)](https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.1-GGUF/resolve/main/mistral-7b-instruct-v0.1.Q4_0.gguf)
- [⬇️ Download Q6_K (6.23 GB)](https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.1-GGUF/resolve/main/mistral-7b-instruct-v0.1.Q6_K.gguf)
### 🧠 Qwen 1.5-7B Chat
- [⬇️ Download Q4_K (4.88 GB)](https://huggingface.co/TheBloke/Qwen1.5-7B-Chat-GGUF/resolve/main/qwen1_5-7b-chat-q4_k.gguf)
- [⬇️ Download Q6_K (6.83 GB)](https://huggingface.co/TheBloke/Qwen1.5-7B-Chat-GGUF/resolve/main/qwen1_5-7b-chat-q6_k.gguf)
### 🧠 DeepSeek 7B Chat
- [⬇️ Download Q4_0 (4.87 GB)](https://huggingface.co/TheBloke/DeepSeek-7B-Chat-GGUF/resolve/main/deepseek-7b-chat.Q4_0.gguf)
- [⬇️ Download Q8_0 (9.33 GB)](https://huggingface.co/TheBloke/DeepSeek-7B-Chat-GGUF/resolve/main/deepseek-7b-chat.Q8_0.gguf)
### 🧠 LLaMA 3 8B Instruct
- [⬇️ Download Q4_0 (4.68 GB)](https://huggingface.co/TheBloke/Llama-3-8B-Instruct-GGUF/resolve/main/llama-3-8b-instruct.Q4_0.gguf)
- [⬇️ Download Q6_K (6.91 GB)](https://huggingface.co/TheBloke/Llama-3-8B-Instruct-GGUF/resolve/main/llama-3-8b-instruct.Q6_K.gguf)
---
### 🗂️ More Model Collections
- [🧠 TheBloke’s GGUF Model Collection](https://local-ai-zone.github.io)
- [🌍 GGUF Community Collection](https://local-ai-zone.github.io)