sizzlebop commited on
Commit
78db3c1
·
verified ·
1 Parent(s): d006c12

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +278 -278
README.md CHANGED
@@ -1,279 +1,279 @@
1
- ---
2
- license: apache-2.0
3
- language:
4
- - en
5
- library_name: gguf
6
- pipeline_tag: text-generation
7
- tags:
8
- - mathematical-reasoning
9
- - qwen3
10
- - gguf
11
- - quantized
12
- - math
13
- - reasoning
14
- - fine-tuned
15
- base_model: PinkPixel/Crystal-Think-V2
16
- quantized_by: PinkPixel
17
- ---
18
-
19
- <div align="center">
20
- <img src="crystal-think-v2-logo.png" alt="Crystal Think V2 Logo" width="300"/>
21
- </div>
22
-
23
- # 🧠 Crystal Think V2 - GGUF Quantized ✨
24
-
25
- **Optimized GGUF Quantizations for Efficient Mathematical Reasoning**
26
-
27
- > **🔗 Original Model:** [PinkPixel/Crystal-Think-V2](https://huggingface.co/PinkPixel/Crystal-Think-V2)
28
- > **📦 Quantized by:** Pink Pixel
29
- > **🏷️ License:** Apache 2.0
30
-
31
- ---
32
-
33
- ## 📋 About This Repository
34
-
35
- This repository contains **GGUF quantized versions** of Crystal Think V2, an advanced mathematical reasoning model based on Qwen3-4B. These quantized versions are optimized for **efficient inference** while maintaining excellent mathematical reasoning capabilities.
36
-
37
- ### 🎯 Original Model Features
38
- - 🧮 **Advanced Mathematical Reasoning** with enhanced chain-of-thought
39
- - 📐 **Multi-step Problem Solving** with clear explanations
40
- - 💻 **Mathematical Code Generation** and algorithm explanation
41
- - 🎯 **Enhanced `<think></think>` Reasoning Format**
42
- - 📊 **85.2% GSM8K accuracy** (+8.8% over base Qwen3-4B)
43
-
44
- ---
45
-
46
- ## 📦 Available Quantizations
47
-
48
- | Quantization | File Size | Use Case | Memory Required | Quality |
49
- |-------------|-----------|----------|-----------------|---------|
50
- | **Q4_K_M** | 2.3GB | Balanced efficiency | ~6GB RAM | Good |
51
- | **Q5_K_M** | 2.7GB | Better quality | ~7GB RAM | Very Good |
52
- | **Q6_K** | 3.1GB | High quality | ~8GB RAM | Excellent |
53
- | **Q8_0** | 4.0GB | Maximum quality | ~10GB RAM | Near-Original |
54
-
55
- ### 💡 **Quantization Guide:**
56
- - **Q4_K_M** - Best for limited hardware, good performance
57
- - **Q5_K_M** - Recommended balance of speed and quality
58
- - **Q6_K** - High quality with reasonable speed
59
- - **Q8_0** - Near-original quality, slower inference
60
-
61
- ---
62
-
63
- ## 🚀 Quick Start
64
-
65
- ### Using llama.cpp
66
-
67
- ```bash
68
- # Download your preferred quantization
69
- wget https://huggingface.co/PinkPixel/Crystal-Think-V2-GGUF/resolve/main/crystal-think-v2-q5_k_m.gguf
70
-
71
- # Run with llama.cpp
72
- ./llama.cpp/main -m crystal-think-v2-q5_k_m.gguf -p "Solve this step by step: If x + 2y = 10 and 2x - y = 5, find x and y." -n 512
73
- ```
74
-
75
- ### Using llama-cpp-python
76
-
77
- ```python
78
- from llama_cpp import Llama
79
-
80
- # Load the model
81
- llm = Llama(
82
- model_path="crystal-think-v2-q5_k_m.gguf",
83
- n_ctx=4096, # Context length
84
- n_threads=8, # CPU threads
85
- verbose=False
86
- )
87
-
88
- # Mathematical reasoning example
89
- prompt = """Solve this step by step:
90
- A rectangle has a length that is 3 more than twice its width. If the perimeter is 42 cm, what are the dimensions?
91
-
92
- Use <think></think> for your reasoning."""
93
-
94
- response = llm(
95
- prompt,
96
- max_tokens=512,
97
- temperature=0.7,
98
- stop=["</SOLUTION>", "<|endoftext|>"]
99
- )
100
-
101
- print(response["choices"][0]["text"])
102
- ```
103
-
104
- ### Using Ollama
105
-
106
- ```bash
107
- # Create Modelfile
108
- echo 'FROM ./crystal-think-v2-q5_k_m.gguf' > Modelfile
109
-
110
- # Create Ollama model
111
- ollama create crystal-think-v2 -f Modelfile
112
-
113
- # Run the model
114
- ollama run crystal-think-v2 "What is the derivative of x^3 + 2x^2 - 5?"
115
- ```
116
-
117
- ---
118
-
119
- ## 🎯 Enhanced Reasoning Format
120
-
121
- Crystal Think V2 uses a structured reasoning approach:
122
-
123
- ```
124
- <think>
125
- [Step-by-step reasoning process]
126
- - Variable definitions
127
- - Equation setup
128
- - Mathematical operations
129
- - Verification steps
130
- </think>
131
-
132
- <SOLUTION>
133
- [Final organized answer]
134
- 1) Specific results
135
- 2) Numerical values
136
- 3) Units and context
137
- </SOLUTION>
138
- ```
139
-
140
- ---
141
-
142
- ## 📊 Performance Benchmarks
143
-
144
- ### Original Model Performance
145
- | Benchmark | Score | Improvement over Base |
146
- |-----------|-------|----------------------|
147
- | **GSM8K** | 85.2% | +8.8% |
148
- | **MATH** | 42.1% | +10.4% |
149
- | **Algebra** | 78.9% | +13.7% |
150
- | **Geometry** | 71.3% | +12.5% |
151
- | **Code Math** | 82.6% | +13.5% |
152
-
153
- ### GGUF Quantization Impact
154
- - **Q8_0**: ~99% original performance
155
- - **Q6_K**: ~97% original performance
156
- - **Q5_K_M**: ~95% original performance
157
- - **Q4_K_M**: ~92% original performance
158
-
159
- ---
160
-
161
- ## 💻 Hardware Requirements
162
-
163
- ### Minimum Requirements
164
- | Quantization | RAM | VRAM (GPU) | CPU |
165
- |-------------|-----|-----------|-----|
166
- | Q4_K_M | 6GB | 4GB | 4 cores |
167
- | Q5_K_M | 7GB | 5GB | 4 cores |
168
- | Q6_K | 8GB | 6GB | 6 cores |
169
- | Q8_0 | 10GB | 8GB | 8 cores |
170
-
171
- ### Recommended for Best Performance
172
- - **CPU**: Modern 8+ core processor
173
- - **RAM**: 16GB+ system memory
174
- - **GPU**: 8GB+ VRAM (optional, for GPU acceleration)
175
-
176
- ---
177
-
178
- ## 🔧 Installation & Dependencies
179
-
180
- ### llama.cpp
181
- ```bash
182
- git clone https://github.com/ggerganov/llama.cpp
183
- cd llama.cpp
184
- make
185
- ```
186
-
187
- ### llama-cpp-python
188
- ```bash
189
- pip install llama-cpp-python
190
- # For GPU support (optional)
191
- CMAKE_ARGS="-DLLAMA_CUBLAS=on" pip install llama-cpp-python
192
- ```
193
-
194
- ### Ollama
195
- ```bash
196
- # Install Ollama
197
- curl -fsSL https://ollama.com/install.sh | sh
198
- ```
199
-
200
- ---
201
-
202
- ## 📚 Usage Examples
203
-
204
- ### Basic Mathematical Problem
205
- ```
206
- Input: "What is the integral of 2x + 3?"
207
- Expected: Step-by-step integration with explanation
208
- ```
209
-
210
- ### Complex Word Problem
211
- ```
212
- Input: "A train travels 120 miles in 2 hours, then 180 miles in 3 hours. What's the average speed?"
213
- Expected: Detailed solution with calculations
214
- ```
215
-
216
- ### Algebraic Reasoning
217
- ```
218
- Input: "Solve the system: 3x + 2y = 12, x - y = 1"
219
- Expected: Systematic solution using substitution or elimination
220
- ```
221
-
222
- ---
223
-
224
- ## 🔗 Related Links
225
-
226
- - **🏠 Original Model:** [PinkPixel/Crystal-Think-V2](https://huggingface.co/PinkPixel/Crystal-Think-V2)
227
- - **📖 Model Documentation:** [Crystal Think V2 README](https://huggingface.co/PinkPixel/Crystal-Think-V2/blob/main/README.md)
228
- - **🛠️ llama.cpp:** [GitHub Repository](https://github.com/ggerganov/llama.cpp)
229
- - **🐍 llama-cpp-python:** [PyPI Package](https://pypi.org/project/llama-cpp-python/)
230
-
231
- ---
232
-
233
- ## ⚠️ Limitations
234
-
235
- - **Domain Focus**: Optimized for mathematical reasoning; may be less effective for general conversation
236
- - **Quantization Trade-offs**: Lower quantizations may show reduced accuracy on complex problems
237
- - **Language**: Primarily trained on English mathematical content
238
- - **Hardware Dependency**: Performance varies significantly with hardware specifications
239
-
240
- ---
241
-
242
- ## 📈 Benchmarking Your Setup
243
-
244
- Test your quantization choice with this sample problem:
245
-
246
- ```
247
- Prompt: "A rectangular garden has a length that is 4 meters more than twice its width. The garden is surrounded by a walkway that is 2 meters wide on all sides. If the total area (garden + walkway) is 294 square meters, find the dimensions of the garden."
248
-
249
- Expected: The model should show step-by-step reasoning and arrive at width ≈ 8.13m, length ≈ 20.26m
250
- ```
251
-
252
- ---
253
-
254
- ## 🤝 Contributing
255
-
256
- Found an issue with the quantizations or have suggestions for improvements? Please open an issue or reach out!
257
-
258
- ---
259
-
260
- ## 📧 Contact & Support
261
-
262
- - **Developer:** Pink Pixel
263
- - **GitHub:** [https://github.com/pinkpixel-dev](https://github.com/pinkpixel-dev)
264
- - **Website:** [https://pinkpixel.dev](https://pinkpixel.dev)
265
- - **Email:** [admin@pinkpixel.dev](mailto:admin@pinkpixel.dev)
266
-
267
- ---
268
-
269
- ## 🙏 Acknowledgments
270
-
271
- - **Original Model:** Crystal Think V2 by Pink Pixel
272
- - **Base Model:** [Qwen/Qwen3-4B](https://huggingface.co/Qwen/Qwen3-4B) by Qwen Team
273
- - **Quantization Tools:** [llama.cpp](https://github.com/ggerganov/llama.cpp) by Georgi Gerganov
274
- - **Training Dataset:** [NVIDIA OpenMathReasoning](https://huggingface.co/datasets/nvidia/OpenMathReasoning)
275
-
276
- ---
277
-
278
- **Made with ❤️ by Pink Pixel** ✨
279
  *"Dream it, Pixel it"*
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ library_name: gguf
6
+ pipeline_tag: text-generation
7
+ tags:
8
+ - mathematical-reasoning
9
+ - qwen3
10
+ - gguf
11
+ - quantized
12
+ - math
13
+ - reasoning
14
+ - fine-tuned
15
+ base_model: PinkPixel/Crystal-Think-V2
16
+ quantized_by: PinkPixel
17
+ ---
18
+
19
+ <div align="center">
20
+ <img src="crystal-think-v2-logo.png" alt="Crystal Think V2 Logo" width="400"/>
21
+ </div>
22
+
23
+ # 🧠 Crystal Think V2 - GGUF Quantized ✨
24
+
25
+ **Optimized GGUF Quantizations for Efficient Mathematical Reasoning**
26
+
27
+ > **🔗 Original Model:** [PinkPixel/Crystal-Think-V2](https://huggingface.co/PinkPixel/Crystal-Think-V2)
28
+ > **📦 Quantized by:** Pink Pixel
29
+ > **🏷️ License:** Apache 2.0
30
+
31
+ ---
32
+
33
+ ## 📋 About This Repository
34
+
35
+ This repository contains **GGUF quantized versions** of Crystal Think V2, an advanced mathematical reasoning model based on Qwen3-4B. These quantized versions are optimized for **efficient inference** while maintaining excellent mathematical reasoning capabilities.
36
+
37
+ ### 🎯 Original Model Features
38
+ - 🧮 **Advanced Mathematical Reasoning** with enhanced chain-of-thought
39
+ - 📐 **Multi-step Problem Solving** with clear explanations
40
+ - 💻 **Mathematical Code Generation** and algorithm explanation
41
+ - 🎯 **Enhanced `<think></think>` Reasoning Format**
42
+ - 📊 **85.2% GSM8K accuracy** (+8.8% over base Qwen3-4B)
43
+
44
+ ---
45
+
46
+ ## 📦 Available Quantizations
47
+
48
+ | Quantization | File Size | Use Case | Memory Required | Quality |
49
+ |-------------|-----------|----------|-----------------|---------|
50
+ | **Q4_K_M** | 2.3GB | Balanced efficiency | ~6GB RAM | Good |
51
+ | **Q5_K_M** | 2.7GB | Better quality | ~7GB RAM | Very Good |
52
+ | **Q6_K** | 3.1GB | High quality | ~8GB RAM | Excellent |
53
+ | **Q8_0** | 4.0GB | Maximum quality | ~10GB RAM | Near-Original |
54
+
55
+ ### 💡 **Quantization Guide:**
56
+ - **Q4_K_M** - Best for limited hardware, good performance
57
+ - **Q5_K_M** - Recommended balance of speed and quality
58
+ - **Q6_K** - High quality with reasonable speed
59
+ - **Q8_0** - Near-original quality, slower inference
60
+
61
+ ---
62
+
63
+ ## 🚀 Quick Start
64
+
65
+ ### Using llama.cpp
66
+
67
+ ```bash
68
+ # Download your preferred quantization
69
+ wget https://huggingface.co/PinkPixel/Crystal-Think-V2-GGUF/resolve/main/crystal-think-v2-q5_k_m.gguf
70
+
71
+ # Run with llama.cpp
72
+ ./llama.cpp/main -m crystal-think-v2-q5_k_m.gguf -p "Solve this step by step: If x + 2y = 10 and 2x - y = 5, find x and y." -n 512
73
+ ```
74
+
75
+ ### Using llama-cpp-python
76
+
77
+ ```python
78
+ from llama_cpp import Llama
79
+
80
+ # Load the model
81
+ llm = Llama(
82
+ model_path="crystal-think-v2-q5_k_m.gguf",
83
+ n_ctx=4096, # Context length
84
+ n_threads=8, # CPU threads
85
+ verbose=False
86
+ )
87
+
88
+ # Mathematical reasoning example
89
+ prompt = """Solve this step by step:
90
+ A rectangle has a length that is 3 more than twice its width. If the perimeter is 42 cm, what are the dimensions?
91
+
92
+ Use <think></think> for your reasoning."""
93
+
94
+ response = llm(
95
+ prompt,
96
+ max_tokens=512,
97
+ temperature=0.7,
98
+ stop=["</SOLUTION>", "<|endoftext|>"]
99
+ )
100
+
101
+ print(response["choices"][0]["text"])
102
+ ```
103
+
104
+ ### Using Ollama
105
+
106
+ ```bash
107
+ # Create Modelfile
108
+ echo 'FROM ./crystal-think-v2-q5_k_m.gguf' > Modelfile
109
+
110
+ # Create Ollama model
111
+ ollama create crystal-think-v2 -f Modelfile
112
+
113
+ # Run the model
114
+ ollama run crystal-think-v2 "What is the derivative of x^3 + 2x^2 - 5?"
115
+ ```
116
+
117
+ ---
118
+
119
+ ## 🎯 Enhanced Reasoning Format
120
+
121
+ Crystal Think V2 uses a structured reasoning approach:
122
+
123
+ ```
124
+ <think>
125
+ [Step-by-step reasoning process]
126
+ - Variable definitions
127
+ - Equation setup
128
+ - Mathematical operations
129
+ - Verification steps
130
+ </think>
131
+
132
+ <SOLUTION>
133
+ [Final organized answer]
134
+ 1) Specific results
135
+ 2) Numerical values
136
+ 3) Units and context
137
+ </SOLUTION>
138
+ ```
139
+
140
+ ---
141
+
142
+ ## 📊 Performance Benchmarks
143
+
144
+ ### Original Model Performance
145
+ | Benchmark | Score | Improvement over Base |
146
+ |-----------|-------|----------------------|
147
+ | **GSM8K** | 85.2% | +8.8% |
148
+ | **MATH** | 42.1% | +10.4% |
149
+ | **Algebra** | 78.9% | +13.7% |
150
+ | **Geometry** | 71.3% | +12.5% |
151
+ | **Code Math** | 82.6% | +13.5% |
152
+
153
+ ### GGUF Quantization Impact
154
+ - **Q8_0**: ~99% original performance
155
+ - **Q6_K**: ~97% original performance
156
+ - **Q5_K_M**: ~95% original performance
157
+ - **Q4_K_M**: ~92% original performance
158
+
159
+ ---
160
+
161
+ ## 💻 Hardware Requirements
162
+
163
+ ### Minimum Requirements
164
+ | Quantization | RAM | VRAM (GPU) | CPU |
165
+ |-------------|-----|-----------|-----|
166
+ | Q4_K_M | 6GB | 4GB | 4 cores |
167
+ | Q5_K_M | 7GB | 5GB | 4 cores |
168
+ | Q6_K | 8GB | 6GB | 6 cores |
169
+ | Q8_0 | 10GB | 8GB | 8 cores |
170
+
171
+ ### Recommended for Best Performance
172
+ - **CPU**: Modern 8+ core processor
173
+ - **RAM**: 16GB+ system memory
174
+ - **GPU**: 8GB+ VRAM (optional, for GPU acceleration)
175
+
176
+ ---
177
+
178
+ ## 🔧 Installation & Dependencies
179
+
180
+ ### llama.cpp
181
+ ```bash
182
+ git clone https://github.com/ggerganov/llama.cpp
183
+ cd llama.cpp
184
+ make
185
+ ```
186
+
187
+ ### llama-cpp-python
188
+ ```bash
189
+ pip install llama-cpp-python
190
+ # For GPU support (optional)
191
+ CMAKE_ARGS="-DLLAMA_CUBLAS=on" pip install llama-cpp-python
192
+ ```
193
+
194
+ ### Ollama
195
+ ```bash
196
+ # Install Ollama
197
+ curl -fsSL https://ollama.com/install.sh | sh
198
+ ```
199
+
200
+ ---
201
+
202
+ ## 📚 Usage Examples
203
+
204
+ ### Basic Mathematical Problem
205
+ ```
206
+ Input: "What is the integral of 2x + 3?"
207
+ Expected: Step-by-step integration with explanation
208
+ ```
209
+
210
+ ### Complex Word Problem
211
+ ```
212
+ Input: "A train travels 120 miles in 2 hours, then 180 miles in 3 hours. What's the average speed?"
213
+ Expected: Detailed solution with calculations
214
+ ```
215
+
216
+ ### Algebraic Reasoning
217
+ ```
218
+ Input: "Solve the system: 3x + 2y = 12, x - y = 1"
219
+ Expected: Systematic solution using substitution or elimination
220
+ ```
221
+
222
+ ---
223
+
224
+ ## 🔗 Related Links
225
+
226
+ - **🏠 Original Model:** [PinkPixel/Crystal-Think-V2](https://huggingface.co/PinkPixel/Crystal-Think-V2)
227
+ - **📖 Model Documentation:** [Crystal Think V2 README](https://huggingface.co/PinkPixel/Crystal-Think-V2/blob/main/README.md)
228
+ - **🛠️ llama.cpp:** [GitHub Repository](https://github.com/ggerganov/llama.cpp)
229
+ - **🐍 llama-cpp-python:** [PyPI Package](https://pypi.org/project/llama-cpp-python/)
230
+
231
+ ---
232
+
233
+ ## ⚠️ Limitations
234
+
235
+ - **Domain Focus**: Optimized for mathematical reasoning; may be less effective for general conversation
236
+ - **Quantization Trade-offs**: Lower quantizations may show reduced accuracy on complex problems
237
+ - **Language**: Primarily trained on English mathematical content
238
+ - **Hardware Dependency**: Performance varies significantly with hardware specifications
239
+
240
+ ---
241
+
242
+ ## 📈 Benchmarking Your Setup
243
+
244
+ Test your quantization choice with this sample problem:
245
+
246
+ ```
247
+ Prompt: "A rectangular garden has a length that is 4 meters more than twice its width. The garden is surrounded by a walkway that is 2 meters wide on all sides. If the total area (garden + walkway) is 294 square meters, find the dimensions of the garden."
248
+
249
+ Expected: The model should show step-by-step reasoning and arrive at width ≈ 8.13m, length ≈ 20.26m
250
+ ```
251
+
252
+ ---
253
+
254
+ ## 🤝 Contributing
255
+
256
+ Found an issue with the quantizations or have suggestions for improvements? Please open an issue or reach out!
257
+
258
+ ---
259
+
260
+ ## 📧 Contact & Support
261
+
262
+ - **Developer:** Pink Pixel
263
+ - **GitHub:** [https://github.com/pinkpixel-dev](https://github.com/pinkpixel-dev)
264
+ - **Website:** [https://pinkpixel.dev](https://pinkpixel.dev)
265
+ - **Email:** [admin@pinkpixel.dev](mailto:admin@pinkpixel.dev)
266
+
267
+ ---
268
+
269
+ ## 🙏 Acknowledgments
270
+
271
+ - **Original Model:** Crystal Think V2 by Pink Pixel
272
+ - **Base Model:** [Qwen/Qwen3-4B](https://huggingface.co/Qwen/Qwen3-4B) by Qwen Team
273
+ - **Quantization Tools:** [llama.cpp](https://github.com/ggerganov/llama.cpp) by Georgi Gerganov
274
+ - **Training Dataset:** [NVIDIA OpenMathReasoning](https://huggingface.co/datasets/nvidia/OpenMathReasoning)
275
+
276
+ ---
277
+
278
+ **Made with ❤️ by Pink Pixel** ✨
279
  *"Dream it, Pixel it"*