sizzlebop commited on
Commit
7067e40
·
verified ·
1 Parent(s): b6f3826

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +335 -335
README.md CHANGED
@@ -1,336 +1,336 @@
1
- ---
2
- license: apache-2.0
3
- language:
4
- - en
5
- library_name: gguf
6
- pipeline_tag: text-generation
7
- tags:
8
- - mathematical-reasoning
9
- - qwen3
10
- - gguf
11
- - quantized
12
- - imatrix
13
- - importance-matrix
14
- - math
15
- - reasoning
16
- - fine-tuned
17
- base_model: PinkPixel/Crystal-Think-V2
18
- quantized_by: PinkPixel
19
- ---
20
-
21
- <div align="center">
22
- <img src="crystal-think-v2-logo.png" alt="Crystal Think V2 Logo" width="300"/>
23
- </div>
24
-
25
- # 🧠 Crystal Think V2 - GGUF Imatrix Quantized ✨
26
-
27
- **Premium Quality GGUF Quantizations with Importance Matrix Optimization**
28
-
29
- > **🔗 Original Model:** [PinkPixel/Crystal-Think-V2](https://huggingface.co/PinkPixel/Crystal-Think-V2)
30
- > **📦 Quantized by:** Pink Pixel
31
- > **🏷️ License:** Apache 2.0
32
- > **🎯 Special Feature:** Importance Matrix Enhanced
33
-
34
- ---
35
-
36
- ## 📋 About This Repository
37
-
38
- This repository contains **premium GGUF quantized versions** of Crystal Think V2, enhanced with **Importance Matrix (imatrix)** optimization. These quantizations use calibration data to intelligently preserve the most critical model activations, resulting in **superior quality** compared to standard quantizations.
39
-
40
- ### 🌟 **What is Importance Matrix?**
41
-
42
- **Importance Matrix** is an advanced quantization technique that:
43
- - 📊 **Analyzes activation patterns** using calibration data
44
- - 🎯 **Identifies critical neurons** that most impact model performance
45
- - 🔧 **Preserves precision** where it matters most
46
- - ⚡ **Maintains efficiency** while maximizing quality retention
47
-
48
- **Result:** Better mathematical reasoning performance at the same file sizes! 🚀
49
-
50
- ### 🎯 Original Model Features
51
- - 🧮 **Advanced Mathematical Reasoning** with enhanced chain-of-thought
52
- - 📐 **Multi-step Problem Solving** with clear explanations
53
- - 💻 **Mathematical Code Generation** and algorithm explanation
54
- - 🎯 **Enhanced `<think></think>` Reasoning Format**
55
- - 📊 **85.2% GSM8K accuracy** (+8.8% over base Qwen3-4B)
56
-
57
- ---
58
-
59
- ## 📦 Available Imatrix Quantizations
60
-
61
- | Quantization | File Size | Use Case | Memory Required | Quality vs Standard |
62
- |-------------|-----------|----------|-----------------|-------------------|
63
- | **IQ4_XS** | 2.1GB | Ultra-efficient | ~5.5GB RAM | +3-5% better |
64
- | **Q4_K_S** | 2.2GB | Small & fast | ~6GB RAM | +2-4% better |
65
- | **IQ4_NL** | 2.2GB | Natural language optimized | ~6GB RAM | +4-6% better |
66
- | **Q4_K_M** | 2.3GB | Balanced performance | ~6.5GB RAM | +3-5% better |
67
- | **Q5_K_S** | 2.6GB | High quality small | ~7GB RAM | +2-3% better |
68
- | **Q5_K_M** | 2.7GB | **RECOMMENDED** | ~7.5GB RAM | +2-4% better |
69
-
70
- ### 💡 **Quantization Guide:**
71
- - **IQ4_XS** - Smallest size with imatrix benefits
72
- - **IQ4_NL** - Optimized for natural language tasks (math word problems!)
73
- - **Q4_K_M** - **Best balance** of size and quality improvement
74
- - **Q5_K_M** - **Recommended choice** for most users - excellent quality retention
75
-
76
- ---
77
-
78
- ## 🚀 Quick Start
79
-
80
- ### Using llama.cpp
81
-
82
- ```bash
83
- # Download your preferred imatrix quantization
84
- wget https://huggingface.co/PinkPixel/Crystal-Think-V2-GGUF-Imatrix/resolve/main/crystal-think-v2-q4_k_m-imat.gguf
85
-
86
- # Run with llama.cpp
87
- ./llama.cpp/main -m crystal-think-v2-q4_k_m-imat.gguf -p "Solve this step by step: If x + 2y = 10 and 2x - y = 5, find x and y." -n 512
88
- ```
89
-
90
- ### Using llama-cpp-python
91
-
92
- ```python
93
- from llama_cpp import Llama
94
-
95
- # Load the imatrix model
96
- llm = Llama(
97
- model_path="crystal-think-v2-q5_k_m-imat.gguf",
98
- n_ctx=4096, # Context length
99
- n_threads=8, # CPU threads
100
- verbose=False
101
- )
102
-
103
- # Mathematical reasoning example
104
- prompt = """Solve this step by step:
105
- A circular garden has a radius of 8 meters. If you want to build a rectangular fence around it with 2 meters clearance on all sides, what's the area of the rectangular fence?
106
-
107
- Use <think></think> for your reasoning."""
108
-
109
- response = llm(
110
- prompt,
111
- max_tokens=512,
112
- temperature=0.7,
113
- stop=["</SOLUTION>", "<|endoftext|>"]
114
- )
115
-
116
- print(response["choices"][0]["text"])
117
- ```
118
-
119
- ### Using Ollama
120
-
121
- ```bash
122
- # Create Modelfile
123
- echo 'FROM ./crystal-think-v2-q5_k_m-imat.gguf' > Modelfile
124
-
125
- # Create Ollama model
126
- ollama create crystal-think-v2-imat -f Modelfile
127
-
128
- # Run the model
129
- ollama run crystal-think-v2-imat "What is the integral of sin(x)cos(x)?"
130
- ```
131
-
132
- ---
133
-
134
- ## 🎯 Enhanced Reasoning Format
135
-
136
- Crystal Think V2 uses a structured reasoning approach, perfectly preserved with imatrix:
137
-
138
- ```
139
- <think>
140
- [Step-by-step reasoning process]
141
- - Problem analysis and variable identification
142
- - Mathematical equation setup
143
- - Systematic solution steps
144
- - Verification and checking
145
- </think>
146
-
147
- <SOLUTION>
148
- [Final organized answer]
149
- 1) Clear results with explanations
150
- 2) Numerical values with proper units
151
- 3) Context and practical interpretation
152
- </SOLUTION>
153
- ```
154
-
155
- ---
156
-
157
- ## 📊 Performance Benchmarks
158
-
159
- ### Original Model Performance
160
- | Benchmark | Score | Improvement over Base |
161
- |-----------|-------|----------------------|
162
- | **GSM8K** | 85.2% | +8.8% |
163
- | **MATH** | 42.1% | +10.4% |
164
- | **Algebra** | 78.9% | +13.7% |
165
- | **Geometry** | 71.3% | +12.5% |
166
- | **Code Math** | 82.6% | +13.5% |
167
-
168
- ### Imatrix vs Standard GGUF Comparison
169
- | Quantization | Standard GGUF | Imatrix GGUF | Improvement |
170
- |-------------|---------------|--------------|-------------|
171
- | **Q4_K_M** | ~92% orig. | ~95-97% orig. | **+3-5%** |
172
- | **Q5_K_M** | ~95% orig. | ~97-99% orig. | **+2-4%** |
173
- | **IQ4_NL** | N/A | ~94-96% orig. | **New format** |
174
- | **IQ4_XS** | N/A | ~91-93% orig. | **Smallest size** |
175
-
176
- ### 🎯 **Why Imatrix is Better:**
177
- - **Smarter quantization** - Preserves critical mathematical reasoning paths
178
- - **Better accuracy** - Maintains performance on complex multi-step problems
179
- - **Consistent quality** - Less degradation on edge cases and difficult problems
180
-
181
- ---
182
-
183
- ## 💻 Hardware Requirements
184
-
185
- ### Minimum Requirements
186
- | Quantization | RAM | VRAM (GPU) | CPU |
187
- |-------------|-----|-----------|-----|
188
- | IQ4_XS | 5.5GB | 3.5GB | 4 cores |
189
- | Q4_K_S | 6GB | 4GB | 4 cores |
190
- | IQ4_NL | 6GB | 4GB | 4 cores |
191
- | Q4_K_M | 6.5GB | 4.5GB | 4 cores |
192
- | Q5_K_S | 7GB | 5GB | 6 cores |
193
- | Q5_K_M | 7.5GB | 5.5GB | 6 cores |
194
-
195
- ### Recommended for Best Performance
196
- - **CPU**: Modern 8+ core processor (AMD Ryzen 7/Intel i7 or better)
197
- - **RAM**: 16GB+ system memory
198
- - **GPU**: 8GB+ VRAM (RTX 4070/RX 7800 XT or better for GPU acceleration)
199
-
200
- ---
201
-
202
- ## 🔧 Installation & Dependencies
203
-
204
- ### llama.cpp (Latest Version Recommended)
205
- ```bash
206
- git clone https://github.com/ggerganov/llama.cpp
207
- cd llama.cpp
208
- make
209
- # For GPU support
210
- make LLAMA_CUBLAS=1
211
- ```
212
-
213
- ### llama-cpp-python
214
- ```bash
215
- pip install llama-cpp-python
216
- # For GPU support (CUDA)
217
- CMAKE_ARGS="-DLLAMA_CUBLAS=on" pip install llama-cpp-python
218
- # For GPU support (ROCm/AMD)
219
- CMAKE_ARGS="-DLLAMA_HIPBLAS=on" pip install llama-cpp-python
220
- ```
221
-
222
- ### Ollama
223
- ```bash
224
- # Install Ollama
225
- curl -fsSL https://ollama.com/install.sh | sh
226
- ```
227
-
228
- ---
229
-
230
- ## 📚 Advanced Usage Examples
231
-
232
- ### Complex Mathematical Reasoning
233
- ```
234
- Input: "A projectile is launched at 45° with initial velocity 50 m/s. Calculate the maximum height, range, and time of flight. Use g = 9.8 m/s²."
235
-
236
- Expected: Detailed physics solution with kinematic equations
237
- ```
238
-
239
- ### Multi-step Algebra
240
- ```
241
- Input: "Solve the system of equations: 2x + 3y - z = 7, x - 2y + 4z = -3, 3x + y + 2z = 10"
242
-
243
- Expected: Systematic solution using elimination or substitution
244
- ```
245
-
246
- ### Calculus Problem
247
- ```
248
- Input: "Find the area between the curves y = x² and y = 4x - x² from x = 0 to x = 4"
249
-
250
- Expected: Step-by-step integration with proper setup
251
- ```
252
-
253
- ---
254
-
255
- ## 🔍 Quality Comparison Test
256
-
257
- Test the imatrix advantage with this challenging problem:
258
-
259
- ```
260
- Prompt: "A cylindrical tank with radius 3m and height 8m is filled with water to 75% capacity. If water is drained at a rate of 2m³/min, how long will it take to empty the tank completely? Also calculate the water level after 30 minutes of draining."
261
-
262
- Expected Results:
263
- - Initial volume calculation: π × 3² × 8 × 0.75 = 54π m³
264
- - Time to empty: 27π minutes ≈ 84.8 minutes
265
- - Water level after 30 min: ~4.4 meters
266
-
267
- Imatrix models should show cleaner reasoning and more accurate intermediate steps!
268
- ```
269
-
270
- ---
271
-
272
- ## 🔗 Related Links
273
-
274
- - **🏠 Original Model:** [PinkPixel/Crystal-Think-V2](https://huggingface.co/PinkPixel/Crystal-Think-V2)
275
- - **📖 Model Documentation:** [Crystal Think V2 README](https://huggingface.co/PinkPixel/Crystal-Think-V2/blob/main/README.md)
276
- - **🔧 Standard GGUF:** [Crystal Think V2 GGUF](https://huggingface.co/PinkPixel/Crystal-Think-V2-GGUF)
277
- - **🛠️ llama.cpp:** [GitHub Repository](https://github.com/ggerganov/llama.cpp)
278
- - **🐍 llama-cpp-python:** [PyPI Package](https://pypi.org/project/llama-cpp-python/)
279
-
280
- ---
281
-
282
- ## ⚠️ Limitations
283
-
284
- - **Domain Focus**: Optimized for mathematical reasoning; may be less effective for general conversation
285
- - **Calibration Dependency**: Imatrix quality depends on calibration data relevance
286
- - **Language**: Primarily trained on English mathematical content
287
- - **Hardware Dependency**: Performance varies significantly with hardware specifications
288
-
289
- ---
290
-
291
- ## 🧪 Technical Details
292
-
293
- ### Imatrix Generation Process
294
- 1. **Calibration Data**: Used high-quality mathematical reasoning samples
295
- 2. **Activation Analysis**: Measured importance across all model layers
296
- 3. **Precision Mapping**: Applied higher precision to critical activations
297
- 4. **Quality Validation**: Tested on mathematical benchmarks
298
-
299
- ### Recommended Use Cases
300
- - **Mathematical tutoring systems**
301
- - **STEM education applications**
302
- - **Research and analysis tools**
303
- - **Competitive programming assistance**
304
- - **Physics and engineering calculations**
305
-
306
- ---
307
-
308
- ## 🤝 Contributing
309
-
310
- Found an issue with the imatrix quantizations or have suggestions for improvements? Please open an issue or reach out!
311
-
312
- ---
313
-
314
- ## 📧 Contact & Support
315
-
316
- - **Developer:** Pink Pixel
317
- - **GitHub:** [https://github.com/pinkpixel-dev](https://github.com/pinkpixel-dev)
318
- - **Website:** [https://pinkpixel.dev](https://pinkpixel.dev)
319
- - **Email:** [admin@pinkpixel.dev](mailto:admin@pinkpixel.dev)
320
-
321
- ---
322
-
323
- ## 🙏 Acknowledgments
324
-
325
- - **Original Model:** Crystal Think V2 by Pink Pixel
326
- - **Base Model:** [Qwen/Qwen3-4B](https://huggingface.co/Qwen/Qwen3-4B) by Qwen Team
327
- - **Quantization Tools:** [llama.cpp](https://github.com/ggerganov/llama.cpp) by Georgi Gerganov
328
- - **Imatrix Technique:** Advanced quantization methodology for preserving model quality
329
- - **Training Dataset:** [NVIDIA OpenMathReasoning](https://huggingface.co/datasets/nvidia/OpenMathReasoning)
330
-
331
- ---
332
-
333
- **Made with ❤️ by Pink Pixel** ✨
334
- *"Dream it, Pixel it"*
335
-
336
  > **💡 Pro Tip:** For the best mathematical reasoning experience, try the **Q5_K_M-imat** or **IQ4_NL-imat** variants - they offer excellent quality retention with the benefits of importance matrix optimization!
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ library_name: gguf
6
+ pipeline_tag: text-generation
7
+ tags:
8
+ - mathematical-reasoning
9
+ - qwen3
10
+ - gguf
11
+ - quantized
12
+ - imatrix
13
+ - importance-matrix
14
+ - math
15
+ - reasoning
16
+ - fine-tuned
17
+ base_model: PinkPixel/Crystal-Think-V2
18
+ quantized_by: PinkPixel
19
+ ---
20
+
21
+ <div align="center">
22
+ <img src="crystal-think-v2-logo.png" alt="Crystal Think V2 Logo" width="400"/>
23
+ </div>
24
+
25
+ # 🧠 Crystal Think V2 - GGUF Imatrix Quantized ✨
26
+
27
+ **Premium Quality GGUF Quantizations with Importance Matrix Optimization**
28
+
29
+ > **🔗 Original Model:** [PinkPixel/Crystal-Think-V2](https://huggingface.co/PinkPixel/Crystal-Think-V2)
30
+ > **📦 Quantized by:** Pink Pixel
31
+ > **🏷️ License:** Apache 2.0
32
+ > **🎯 Special Feature:** Importance Matrix Enhanced
33
+
34
+ ---
35
+
36
+ ## 📋 About This Repository
37
+
38
+ This repository contains **premium GGUF quantized versions** of Crystal Think V2, enhanced with **Importance Matrix (imatrix)** optimization. These quantizations use calibration data to intelligently preserve the most critical model activations, resulting in **superior quality** compared to standard quantizations.
39
+
40
+ ### 🌟 **What is Importance Matrix?**
41
+
42
+ **Importance Matrix** is an advanced quantization technique that:
43
+ - 📊 **Analyzes activation patterns** using calibration data
44
+ - 🎯 **Identifies critical neurons** that most impact model performance
45
+ - 🔧 **Preserves precision** where it matters most
46
+ - ⚡ **Maintains efficiency** while maximizing quality retention
47
+
48
+ **Result:** Better mathematical reasoning performance at the same file sizes! 🚀
49
+
50
+ ### 🎯 Original Model Features
51
+ - 🧮 **Advanced Mathematical Reasoning** with enhanced chain-of-thought
52
+ - 📐 **Multi-step Problem Solving** with clear explanations
53
+ - 💻 **Mathematical Code Generation** and algorithm explanation
54
+ - 🎯 **Enhanced `<think></think>` Reasoning Format**
55
+ - 📊 **85.2% GSM8K accuracy** (+8.8% over base Qwen3-4B)
56
+
57
+ ---
58
+
59
+ ## 📦 Available Imatrix Quantizations
60
+
61
+ | Quantization | File Size | Use Case | Memory Required | Quality vs Standard |
62
+ |-------------|-----------|----------|-----------------|-------------------|
63
+ | **IQ4_XS** | 2.1GB | Ultra-efficient | ~5.5GB RAM | +3-5% better |
64
+ | **Q4_K_S** | 2.2GB | Small & fast | ~6GB RAM | +2-4% better |
65
+ | **IQ4_NL** | 2.2GB | Natural language optimized | ~6GB RAM | +4-6% better |
66
+ | **Q4_K_M** | 2.3GB | Balanced performance | ~6.5GB RAM | +3-5% better |
67
+ | **Q5_K_S** | 2.6GB | High quality small | ~7GB RAM | +2-3% better |
68
+ | **Q5_K_M** | 2.7GB | **RECOMMENDED** | ~7.5GB RAM | +2-4% better |
69
+
70
+ ### 💡 **Quantization Guide:**
71
+ - **IQ4_XS** - Smallest size with imatrix benefits
72
+ - **IQ4_NL** - Optimized for natural language tasks (math word problems!)
73
+ - **Q4_K_M** - **Best balance** of size and quality improvement
74
+ - **Q5_K_M** - **Recommended choice** for most users - excellent quality retention
75
+
76
+ ---
77
+
78
+ ## 🚀 Quick Start
79
+
80
+ ### Using llama.cpp
81
+
82
+ ```bash
83
+ # Download your preferred imatrix quantization
84
+ wget https://huggingface.co/PinkPixel/Crystal-Think-V2-GGUF-Imatrix/resolve/main/crystal-think-v2-q4_k_m-imat.gguf
85
+
86
+ # Run with llama.cpp
87
+ ./llama.cpp/main -m crystal-think-v2-q4_k_m-imat.gguf -p "Solve this step by step: If x + 2y = 10 and 2x - y = 5, find x and y." -n 512
88
+ ```
89
+
90
+ ### Using llama-cpp-python
91
+
92
+ ```python
93
+ from llama_cpp import Llama
94
+
95
+ # Load the imatrix model
96
+ llm = Llama(
97
+ model_path="crystal-think-v2-q5_k_m-imat.gguf",
98
+ n_ctx=4096, # Context length
99
+ n_threads=8, # CPU threads
100
+ verbose=False
101
+ )
102
+
103
+ # Mathematical reasoning example
104
+ prompt = """Solve this step by step:
105
+ A circular garden has a radius of 8 meters. If you want to build a rectangular fence around it with 2 meters clearance on all sides, what's the area of the rectangular fence?
106
+
107
+ Use <think></think> for your reasoning."""
108
+
109
+ response = llm(
110
+ prompt,
111
+ max_tokens=512,
112
+ temperature=0.7,
113
+ stop=["</SOLUTION>", "<|endoftext|>"]
114
+ )
115
+
116
+ print(response["choices"][0]["text"])
117
+ ```
118
+
119
+ ### Using Ollama
120
+
121
+ ```bash
122
+ # Create Modelfile
123
+ echo 'FROM ./crystal-think-v2-q5_k_m-imat.gguf' > Modelfile
124
+
125
+ # Create Ollama model
126
+ ollama create crystal-think-v2-imat -f Modelfile
127
+
128
+ # Run the model
129
+ ollama run crystal-think-v2-imat "What is the integral of sin(x)cos(x)?"
130
+ ```
131
+
132
+ ---
133
+
134
+ ## 🎯 Enhanced Reasoning Format
135
+
136
+ Crystal Think V2 uses a structured reasoning approach, perfectly preserved with imatrix:
137
+
138
+ ```
139
+ <think>
140
+ [Step-by-step reasoning process]
141
+ - Problem analysis and variable identification
142
+ - Mathematical equation setup
143
+ - Systematic solution steps
144
+ - Verification and checking
145
+ </think>
146
+
147
+ <SOLUTION>
148
+ [Final organized answer]
149
+ 1) Clear results with explanations
150
+ 2) Numerical values with proper units
151
+ 3) Context and practical interpretation
152
+ </SOLUTION>
153
+ ```
154
+
155
+ ---
156
+
157
+ ## 📊 Performance Benchmarks
158
+
159
+ ### Original Model Performance
160
+ | Benchmark | Score | Improvement over Base |
161
+ |-----------|-------|----------------------|
162
+ | **GSM8K** | 85.2% | +8.8% |
163
+ | **MATH** | 42.1% | +10.4% |
164
+ | **Algebra** | 78.9% | +13.7% |
165
+ | **Geometry** | 71.3% | +12.5% |
166
+ | **Code Math** | 82.6% | +13.5% |
167
+
168
+ ### Imatrix vs Standard GGUF Comparison
169
+ | Quantization | Standard GGUF | Imatrix GGUF | Improvement |
170
+ |-------------|---------------|--------------|-------------|
171
+ | **Q4_K_M** | ~92% orig. | ~95-97% orig. | **+3-5%** |
172
+ | **Q5_K_M** | ~95% orig. | ~97-99% orig. | **+2-4%** |
173
+ | **IQ4_NL** | N/A | ~94-96% orig. | **New format** |
174
+ | **IQ4_XS** | N/A | ~91-93% orig. | **Smallest size** |
175
+
176
+ ### 🎯 **Why Imatrix is Better:**
177
+ - **Smarter quantization** - Preserves critical mathematical reasoning paths
178
+ - **Better accuracy** - Maintains performance on complex multi-step problems
179
+ - **Consistent quality** - Less degradation on edge cases and difficult problems
180
+
181
+ ---
182
+
183
+ ## 💻 Hardware Requirements
184
+
185
+ ### Minimum Requirements
186
+ | Quantization | RAM | VRAM (GPU) | CPU |
187
+ |-------------|-----|-----------|-----|
188
+ | IQ4_XS | 5.5GB | 3.5GB | 4 cores |
189
+ | Q4_K_S | 6GB | 4GB | 4 cores |
190
+ | IQ4_NL | 6GB | 4GB | 4 cores |
191
+ | Q4_K_M | 6.5GB | 4.5GB | 4 cores |
192
+ | Q5_K_S | 7GB | 5GB | 6 cores |
193
+ | Q5_K_M | 7.5GB | 5.5GB | 6 cores |
194
+
195
+ ### Recommended for Best Performance
196
+ - **CPU**: Modern 8+ core processor (AMD Ryzen 7/Intel i7 or better)
197
+ - **RAM**: 16GB+ system memory
198
+ - **GPU**: 8GB+ VRAM (RTX 4070/RX 7800 XT or better for GPU acceleration)
199
+
200
+ ---
201
+
202
+ ## 🔧 Installation & Dependencies
203
+
204
+ ### llama.cpp (Latest Version Recommended)
205
+ ```bash
206
+ git clone https://github.com/ggerganov/llama.cpp
207
+ cd llama.cpp
208
+ make
209
+ # For GPU support
210
+ make LLAMA_CUBLAS=1
211
+ ```
212
+
213
+ ### llama-cpp-python
214
+ ```bash
215
+ pip install llama-cpp-python
216
+ # For GPU support (CUDA)
217
+ CMAKE_ARGS="-DLLAMA_CUBLAS=on" pip install llama-cpp-python
218
+ # For GPU support (ROCm/AMD)
219
+ CMAKE_ARGS="-DLLAMA_HIPBLAS=on" pip install llama-cpp-python
220
+ ```
221
+
222
+ ### Ollama
223
+ ```bash
224
+ # Install Ollama
225
+ curl -fsSL https://ollama.com/install.sh | sh
226
+ ```
227
+
228
+ ---
229
+
230
+ ## 📚 Advanced Usage Examples
231
+
232
+ ### Complex Mathematical Reasoning
233
+ ```
234
+ Input: "A projectile is launched at 45° with initial velocity 50 m/s. Calculate the maximum height, range, and time of flight. Use g = 9.8 m/s²."
235
+
236
+ Expected: Detailed physics solution with kinematic equations
237
+ ```
238
+
239
+ ### Multi-step Algebra
240
+ ```
241
+ Input: "Solve the system of equations: 2x + 3y - z = 7, x - 2y + 4z = -3, 3x + y + 2z = 10"
242
+
243
+ Expected: Systematic solution using elimination or substitution
244
+ ```
245
+
246
+ ### Calculus Problem
247
+ ```
248
+ Input: "Find the area between the curves y = x² and y = 4x - x² from x = 0 to x = 4"
249
+
250
+ Expected: Step-by-step integration with proper setup
251
+ ```
252
+
253
+ ---
254
+
255
+ ## 🔍 Quality Comparison Test
256
+
257
+ Test the imatrix advantage with this challenging problem:
258
+
259
+ ```
260
+ Prompt: "A cylindrical tank with radius 3m and height 8m is filled with water to 75% capacity. If water is drained at a rate of 2m³/min, how long will it take to empty the tank completely? Also calculate the water level after 30 minutes of draining."
261
+
262
+ Expected Results:
263
+ - Initial volume calculation: π × 3² × 8 × 0.75 = 54π m³
264
+ - Time to empty: 27π minutes ≈ 84.8 minutes
265
+ - Water level after 30 min: ~4.4 meters
266
+
267
+ Imatrix models should show cleaner reasoning and more accurate intermediate steps!
268
+ ```
269
+
270
+ ---
271
+
272
+ ## 🔗 Related Links
273
+
274
+ - **🏠 Original Model:** [PinkPixel/Crystal-Think-V2](https://huggingface.co/PinkPixel/Crystal-Think-V2)
275
+ - **📖 Model Documentation:** [Crystal Think V2 README](https://huggingface.co/PinkPixel/Crystal-Think-V2/blob/main/README.md)
276
+ - **🔧 Standard GGUF:** [Crystal Think V2 GGUF](https://huggingface.co/PinkPixel/Crystal-Think-V2-GGUF)
277
+ - **🛠️ llama.cpp:** [GitHub Repository](https://github.com/ggerganov/llama.cpp)
278
+ - **🐍 llama-cpp-python:** [PyPI Package](https://pypi.org/project/llama-cpp-python/)
279
+
280
+ ---
281
+
282
+ ## ⚠️ Limitations
283
+
284
+ - **Domain Focus**: Optimized for mathematical reasoning; may be less effective for general conversation
285
+ - **Calibration Dependency**: Imatrix quality depends on calibration data relevance
286
+ - **Language**: Primarily trained on English mathematical content
287
+ - **Hardware Dependency**: Performance varies significantly with hardware specifications
288
+
289
+ ---
290
+
291
+ ## 🧪 Technical Details
292
+
293
+ ### Imatrix Generation Process
294
+ 1. **Calibration Data**: Used high-quality mathematical reasoning samples
295
+ 2. **Activation Analysis**: Measured importance across all model layers
296
+ 3. **Precision Mapping**: Applied higher precision to critical activations
297
+ 4. **Quality Validation**: Tested on mathematical benchmarks
298
+
299
+ ### Recommended Use Cases
300
+ - **Mathematical tutoring systems**
301
+ - **STEM education applications**
302
+ - **Research and analysis tools**
303
+ - **Competitive programming assistance**
304
+ - **Physics and engineering calculations**
305
+
306
+ ---
307
+
308
+ ## 🤝 Contributing
309
+
310
+ Found an issue with the imatrix quantizations or have suggestions for improvements? Please open an issue or reach out!
311
+
312
+ ---
313
+
314
+ ## 📧 Contact & Support
315
+
316
+ - **Developer:** Pink Pixel
317
+ - **GitHub:** [https://github.com/pinkpixel-dev](https://github.com/pinkpixel-dev)
318
+ - **Website:** [https://pinkpixel.dev](https://pinkpixel.dev)
319
+ - **Email:** [admin@pinkpixel.dev](mailto:admin@pinkpixel.dev)
320
+
321
+ ---
322
+
323
+ ## 🙏 Acknowledgments
324
+
325
+ - **Original Model:** Crystal Think V2 by Pink Pixel
326
+ - **Base Model:** [Qwen/Qwen3-4B](https://huggingface.co/Qwen/Qwen3-4B) by Qwen Team
327
+ - **Quantization Tools:** [llama.cpp](https://github.com/ggerganov/llama.cpp) by Georgi Gerganov
328
+ - **Imatrix Technique:** Advanced quantization methodology for preserving model quality
329
+ - **Training Dataset:** [NVIDIA OpenMathReasoning](https://huggingface.co/datasets/nvidia/OpenMathReasoning)
330
+
331
+ ---
332
+
333
+ **Made with ❤️ by Pink Pixel** ✨
334
+ *"Dream it, Pixel it"*
335
+
336
  > **💡 Pro Tip:** For the best mathematical reasoning experience, try the **Q5_K_M-imat** or **IQ4_NL-imat** variants - they offer excellent quality retention with the benefits of importance matrix optimization!