SandLogicTechnologies commited on
Commit
96966cc
·
verified ·
1 Parent(s): 53d0b69

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ Qwen2.5-Math-1.5B_Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
37
+ Qwen2.5-Math-1.5B_Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
Qwen2.5-Math-1.5B_Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b63f7d7c0b70e80b5fd2330c1d114118d2cdafa78acea157b8d628e2ad24ef4c
3
+ size 986048448
Qwen2.5-Math-1.5B_Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9213862d5bf73c713db734c754450fcbeea7975ec6e139beb4e0b4cf04333665
3
+ size 1125050304
Readme(Qweb-math).md ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: Apache-2.0
3
+ language:
4
+ - en
5
+ base_model:
6
+ - Qwen2.5-1.5B
7
+ tags:
8
+ - mathematics
9
+ - reasoning
10
+ - problem-solving
11
+ - education
12
+ - transformer
13
+ ---
14
+
15
+ # Qwen2.5-Math-1.5B Model
16
+
17
+ This repository hosts the **Qwen2.5-Math-1.5B** language model an optimized transformer designed to handle advanced mathematical reasoning, symbolic problem solving, and step-by-step solution generation. Built for educational assistance, competitive mathematics settings, and research in formal reasoning, the model offers strong performance while maintaining efficient deployment requirements.
18
+
19
+
20
+ ## Model Overview
21
+
22
+ - **Base-Model**: Qwen2.5-1.5B
23
+ - **Original-Model**: Qwen2.5-Math-1.5B
24
+ - **Architecture**: Decoder-only transformer
25
+ - **Quantized Versions**:
26
+ - Q4_K_M (4-bit quantization)
27
+ - Q5_K_M (5-bit quantization)
28
+ - **Modalities**: Text
29
+ - **Developer**: Qwen
30
+ - **Language**: English
31
+ - **License**: [Apache 2.0](https://choosealicense.com/licenses/apache-2.0/)
32
+ - **Input/Output Format**: Instruction-tuned conversational format
33
+
34
+
35
+ ## Quantization Details
36
+
37
+ ### Q4_K_M Version
38
+ - Approx. ~70% size reduction
39
+ - Lower memory footprint (~940 MB)
40
+ - Best suited for deployment on edge devices or low-resource GPUs
41
+ - Slight performance degradation in complex reasoning scenarios
42
+
43
+ ### Q5_K_M Version
44
+ - Approx. ~66% size reduction
45
+ - Higher fidelity (~1.04 GB)
46
+ - Better performance retention, recommended when quality is a priority
47
+
48
+
49
+ ### Dataset & Training
50
+
51
+ - The model is trained on curated mathematics-focused datasets consisting of:
52
+ - Textbooks & structured solutions
53
+ - Problem-answer pairs and mathematical explanations
54
+ - High-difficulty reasoning tasks used in competitive examination preparation
55
+
56
+
57
+ ## Key Strengths
58
+
59
+ - Strong capability for multi-step reasoning and deriving structured solutions
60
+ - Generates stepwise explanations rather than single-answer outputs
61
+ - Suitable for high-performance inference on GPUs and high-end CPUs
62
+ - Rich instruction-following behavior for math problem sets and tutoring systems
63
+ - Works effectively with chain-of-thought prompting strategies
64
+
65
+
66
+ ## Intended Use
67
+
68
+ This model is designed for scenarios where mathematical reasoning is critical, such as:
69
+
70
+ - **Learning platforms & tutoring assistants** : Automated step-by-step math explainer systems
71
+ - **Academic research** : Algorithms and experiments involving symbolic reasoning
72
+ - **STEM educational tools** : Training models targeted at competitive exam preparation
73
+ - **Conversational reasoning agents** : Math-focused dialog systems for structured question answering
74
+
75
+
76
+ ### Usage
77
+
78
+ This model is meant for mathematical guidance and should not replace expert professional judgement in scientific or financial applications.
79
+
80
+ **llama.cpp (text-only)**
81
+ ```sh
82
+ ./llama-cli -hf SandLogicTechnologies/Qwen2.5-Math-1.5B-GGUF -p "Explain Taylor series"
83
+ ```
84
+
85
+ ## Acknowledgments
86
+
87
+ These quantized models are based on the original work by **Qwen** development team.
88
+
89
+ Special thanks to:
90
+ - The [Qwen](https://huggingface.co/Qwen) team for developing and releasing the [Qwen2.5-Math-1.5B](https://huggingface.co/Qwen/Qwen2.5-Math-1.5B) model.
91
+ - **Georgi Gerganov** and the entire [`llama.cpp`](https://github.com/ggerganov/llama.cpp) open-source community for enabling efficient model quantization and inference via the GGUF format.
92
+
93
+
94
+ ## Contact
95
+ For any inquiries or support, please contact us at support@sandlogic.com or visit our [Website](https://www.sandlogic.com/).