--- license: apache-2.0 language: - en base_model: - HyzeAI/HyzeMini pipeline_tag: text-generation library_name: transformers.js tags: - gguf - hyze - local - chat - Hitesh_V_Founder ---
Lightweight GGUF builds of HyzeMini for fast local inference
π hyzebot.vercel.app β’ π hyzedocs.vercel.app β’ π§ hyzecode.vercel.app
--- ## π Overview **HyzeMini (GGUF)** provides **quantized GGUF versions** of the HyzeMini model, optimized for **local execution** using tools like **llama.cpp**, **LM Studio**, **Ollama**, and other GGUF-compatible runtimes. This version keeps the same **Space + General Chat focus**, while enabling: - β‘ Faster inference - π§ Lower memory usage - π» CPU-friendly execution --- ## π§ Model Details - **Base model:** HyzeAI / HyzeMini - **Parameters:** ~0.1B - **Architecture:** Transformer (LLaMA-style) - **Format:** GGUF - **Language:** English - **License:** Apache-2.0 --- ## π§ͺ Available Quantizations *(Exact files may vary depending on upload)* Common GGUF variants include: - `Q2_K` β Ultra-low memory, fastest - `Q4_K_M` β Balanced quality & speed (recommended) - `Q5_K_M` β Higher quality, slightly slower - `Q8_0` β Best quality, highest memory usage > π‘ If youβre unsure, start with **Q4_K_M**. --- ## βοΈ Usage ### llama.cpp ```bash ./main -m HyzeMini-Q4_K_M.gguf -p "Tell me a cool space fact:" ---