HyzeAI commited on
Commit
239437e
·
verified ·
1 Parent(s): a9f67aa

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +78 -3
README.md CHANGED
@@ -1,3 +1,78 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ license: apache-2.0
2
+ library_name: gguf
3
+ language:
4
+ - en
5
+ pipeline_tag: text-generation
6
+ tags:
7
+ - gguf
8
+ - llama.cpp
9
+ - local-llm
10
+ - space
11
+ - chat
12
+ - hyze
13
+ ---
14
+
15
+ <p align="center">
16
+ <img src="https://i.imgur.com/ePJMLNp.png" alt="Hyze Logo" width="320"/>
17
+ </p>
18
+
19
+ <h1 align="center">HyzeMini (GGUF)</h1>
20
+
21
+ <p align="center">
22
+ Lightweight GGUF builds of <b>HyzeMini</b> for fast local inference
23
+ </p>
24
+
25
+ <p align="center">
26
+ 🔗 <a href="https://hyzeai.vercel.app">hyzeai.vercel.app</a> •
27
+ 📘 <a href="https://hyzedocs.vercel.app">hyzedocs.vercel.app</a> •
28
+ 🧠 <a href="https://hyzecode.vercel.app">hyzecode.vercel.app</a>
29
+ </p>
30
+
31
+ ---
32
+
33
+ ## 🚀 Overview
34
+
35
+ **HyzeMini (GGUF)** provides **quantized GGUF versions** of the HyzeMini model, optimized for **local execution** using tools like **llama.cpp**, **LM Studio**, **Ollama**, and other GGUF-compatible runtimes.
36
+
37
+ This version keeps the same **Space + General Chat focus**, while enabling:
38
+
39
+ - ⚡ Faster inference
40
+ - 🧠 Lower memory usage
41
+ - 💻 CPU-friendly execution
42
+
43
+ ---
44
+
45
+ ## 🧠 Model Details
46
+
47
+ - **Base model:** HyzeAI / HyzeMini
48
+ - **Parameters:** ~0.1B
49
+ - **Architecture:** Transformer (LLaMA-style)
50
+ - **Format:** GGUF
51
+ - **Language:** English
52
+ - **License:** Apache-2.0
53
+
54
+ ---
55
+
56
+ ## 🧪 Available Quantizations
57
+
58
+ *(Exact files may vary depending on upload)*
59
+
60
+ Common GGUF variants include:
61
+
62
+ - `Q2_K` – Ultra-low memory, fastest
63
+ - `Q4_K_M` – Balanced quality & speed (recommended)
64
+ - `Q5_K_M` – Higher quality, slightly slower
65
+ - `Q8_0` – Best quality, highest memory usage
66
+
67
+ > 💡 If you’re unsure, start with **Q4_K_M**.
68
+
69
+ ---
70
+
71
+ ## ⚙️ Usage
72
+
73
+ ### llama.cpp
74
+
75
+ ```bash
76
+ ./main -m HyzeMini-Q4_K_M.gguf -p "Tell me a cool space fact:"
77
+
78
+ ---