SandLogicTechnologies commited on
Commit
d8e8372
·
verified ·
1 Parent(s): a44ffa9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +95 -3
README.md CHANGED
@@ -1,3 +1,95 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ base_model:
6
+ - Qwen/Qwen2.5-Math-1.5B
7
+ pipeline_tag: text-generation
8
+ tags:
9
+ - mathematics
10
+ - qwen2
11
+ - problem-solving
12
+ - education
13
+ ---
14
+
15
+
16
+ # Quantized Qwen2.5-Math-1.5B Model
17
+
18
+ This repository hosts the **Qwen2.5-Math-1.5B** language model an optimized transformer designed to handle advanced mathematical reasoning, symbolic problem solving, and step-by-step solution generation. Built for educational assistance, competitive mathematics settings, and research in formal reasoning, the model offers strong performance while maintaining efficient deployment requirements.
19
+
20
+ ## Model Overview
21
+
22
+ - **Base-Model**: Qwen2.5-Math-1.5B
23
+ - **Original-Model**: Qwen2.5-1.5B
24
+ - **Architecture**: Decoder-only transformer
25
+ - **Quantized Versions**:
26
+ - Q4_K_M (4-bit quantization)
27
+ - Q5_K_M (5-bit quantization)
28
+ - **Modalities**: Text
29
+ - **Developer**: Qwen
30
+ - **Language**: English
31
+ - **License**: [Apache 2.0](https://choosealicense.com/licenses/apache-2.0/)
32
+ - **Input/Output Format**: Instruction-tuned conversational format
33
+
34
+ ## Quantization Details
35
+
36
+ ### Q4_K_M Version
37
+ - Approx. ~70% size reduction
38
+ - Lower memory footprint (~940 MB)
39
+ - Best suited for deployment on edge devices or low-resource GPUs
40
+ - Slight performance degradation in complex reasoning scenarios
41
+
42
+ ### Q5_K_M Version
43
+ - Approx. ~66% size reduction
44
+ - Higher fidelity (~1.04 GB)
45
+ - Better performance retention, recommended when quality is a priority
46
+
47
+
48
+ ### Dataset & Training
49
+
50
+ - The model is trained on curated mathematics-focused datasets consisting of:
51
+ - Textbooks & structured solutions
52
+ - Problem-answer pairs and mathematical explanations
53
+ - High-difficulty reasoning tasks used in competitive examination preparation
54
+
55
+
56
+ ## Key Strengths
57
+
58
+ - Strong capability for multi-step reasoning and deriving structured solutions
59
+ - Generates stepwise explanations rather than single-answer outputs
60
+ - Suitable for high-performance inference on GPUs and high-end CPUs
61
+ - Rich instruction-following behavior for math problem sets and tutoring systems
62
+ - Works effectively with chain-of-thought prompting strategies
63
+
64
+
65
+ ## Intended Use
66
+
67
+ This model is designed for scenarios where mathematical reasoning is critical, such as:
68
+
69
+ - **Learning platforms & tutoring assistants** : Automated step-by-step math explainer systems
70
+ - **Academic research** : Algorithms and experiments involving symbolic reasoning
71
+ - **STEM educational tools** : Training models targeted at competitive exam preparation
72
+ - **Conversational reasoning agents** : Math-focused dialog systems for structured question answering
73
+
74
+
75
+ ### Usage
76
+
77
+ This model is meant for mathematical guidance and should not replace expert professional judgement in scientific or financial applications.
78
+
79
+ **llama.cpp (text-only)**
80
+ ```sh
81
+ ./llama-cli -hf SandLogicTechnologies/Qwen2.5-Math-1.5B-GGUF -p "Explain Taylor series"
82
+ ```
83
+
84
+ ## Acknowledgments
85
+
86
+ These quantized models are based on the original work by **Qwen** development team.
87
+
88
+ Special thanks to:
89
+ - The [Qwen](https://huggingface.co/Qwen) team for developing and releasing the [Qwen2.5-Math-1.5B](https://huggingface.co/Qwen/Qwen2.5-Math-1.5B) model.
90
+ - **Georgi Gerganov** and the entire [`llama.cpp`](https://github.com/ggerganov/llama.cpp) open-source community for enabling efficient model quantization and inference via the GGUF format.
91
+
92
+
93
+ ## Contact
94
+ For any inquiries or support, please contact us at support@sandlogic.com or visit our [Website](https://www.sandlogic.com/).
95
+