metadata
license: apache-2.0
language:
- en
base_model:
- HyzeAI/HyzeMini
pipeline_tag: text-generation
library_name: transformers.js
tags:
- gguf
- hyze
- local
- chat
- Hitesh_V_Founder
HyzeMini (GGUF)
Lightweight GGUF builds of HyzeMini for fast local inference
π hyzebot.vercel.app β’ π hyzedocs.vercel.app β’ π§ hyzecode.vercel.app
π Overview
HyzeMini (GGUF) provides quantized GGUF versions of the HyzeMini model, optimized for local execution using tools like llama.cpp, LM Studio, Ollama, and other GGUF-compatible runtimes.
This version keeps the same Space + General Chat focus, while enabling:
- β‘ Faster inference
- π§ Lower memory usage
- π» CPU-friendly execution
π§ Model Details
- Base model: HyzeAI / HyzeMini
- Parameters: ~0.1B
- Architecture: Transformer (LLaMA-style)
- Format: GGUF
- Language: English
- License: Apache-2.0
π§ͺ Available Quantizations
(Exact files may vary depending on upload)
Common GGUF variants include:
Q2_Kβ Ultra-low memory, fastestQ4_K_Mβ Balanced quality & speed (recommended)Q5_K_Mβ Higher quality, slightly slowerQ8_0β Best quality, highest memory usage
π‘ If youβre unsure, start with Q4_K_M.
βοΈ Usage
llama.cpp
./main -m HyzeMini-Q4_K_M.gguf -p "Tell me a cool space fact:"
---