SandLogicTechnologies commited on
Commit
a44ffa9
·
verified ·
1 Parent(s): 96966cc

Delete Readme(Qweb-math).md

Browse files
Files changed (1) hide show
  1. Readme(Qweb-math).md +0 -95
Readme(Qweb-math).md DELETED
@@ -1,95 +0,0 @@
1
- ---
2
- license: Apache-2.0
3
- language:
4
- - en
5
- base_model:
6
- - Qwen2.5-1.5B
7
- tags:
8
- - mathematics
9
- - reasoning
10
- - problem-solving
11
- - education
12
- - transformer
13
- ---
14
-
15
- # Qwen2.5-Math-1.5B Model
16
-
17
- This repository hosts the **Qwen2.5-Math-1.5B** language model an optimized transformer designed to handle advanced mathematical reasoning, symbolic problem solving, and step-by-step solution generation. Built for educational assistance, competitive mathematics settings, and research in formal reasoning, the model offers strong performance while maintaining efficient deployment requirements.
18
-
19
-
20
- ## Model Overview
21
-
22
- - **Base-Model**: Qwen2.5-1.5B
23
- - **Original-Model**: Qwen2.5-Math-1.5B
24
- - **Architecture**: Decoder-only transformer
25
- - **Quantized Versions**:
26
- - Q4_K_M (4-bit quantization)
27
- - Q5_K_M (5-bit quantization)
28
- - **Modalities**: Text
29
- - **Developer**: Qwen
30
- - **Language**: English
31
- - **License**: [Apache 2.0](https://choosealicense.com/licenses/apache-2.0/)
32
- - **Input/Output Format**: Instruction-tuned conversational format
33
-
34
-
35
- ## Quantization Details
36
-
37
- ### Q4_K_M Version
38
- - Approx. ~70% size reduction
39
- - Lower memory footprint (~940 MB)
40
- - Best suited for deployment on edge devices or low-resource GPUs
41
- - Slight performance degradation in complex reasoning scenarios
42
-
43
- ### Q5_K_M Version
44
- - Approx. ~66% size reduction
45
- - Higher fidelity (~1.04 GB)
46
- - Better performance retention, recommended when quality is a priority
47
-
48
-
49
- ### Dataset & Training
50
-
51
- - The model is trained on curated mathematics-focused datasets consisting of:
52
- - Textbooks & structured solutions
53
- - Problem-answer pairs and mathematical explanations
54
- - High-difficulty reasoning tasks used in competitive examination preparation
55
-
56
-
57
- ## Key Strengths
58
-
59
- - Strong capability for multi-step reasoning and deriving structured solutions
60
- - Generates stepwise explanations rather than single-answer outputs
61
- - Suitable for high-performance inference on GPUs and high-end CPUs
62
- - Rich instruction-following behavior for math problem sets and tutoring systems
63
- - Works effectively with chain-of-thought prompting strategies
64
-
65
-
66
- ## Intended Use
67
-
68
- This model is designed for scenarios where mathematical reasoning is critical, such as:
69
-
70
- - **Learning platforms & tutoring assistants** : Automated step-by-step math explainer systems
71
- - **Academic research** : Algorithms and experiments involving symbolic reasoning
72
- - **STEM educational tools** : Training models targeted at competitive exam preparation
73
- - **Conversational reasoning agents** : Math-focused dialog systems for structured question answering
74
-
75
-
76
- ### Usage
77
-
78
- This model is meant for mathematical guidance and should not replace expert professional judgement in scientific or financial applications.
79
-
80
- **llama.cpp (text-only)**
81
- ```sh
82
- ./llama-cli -hf SandLogicTechnologies/Qwen2.5-Math-1.5B-GGUF -p "Explain Taylor series"
83
- ```
84
-
85
- ## Acknowledgments
86
-
87
- These quantized models are based on the original work by **Qwen** development team.
88
-
89
- Special thanks to:
90
- - The [Qwen](https://huggingface.co/Qwen) team for developing and releasing the [Qwen2.5-Math-1.5B](https://huggingface.co/Qwen/Qwen2.5-Math-1.5B) model.
91
- - **Georgi Gerganov** and the entire [`llama.cpp`](https://github.com/ggerganov/llama.cpp) open-source community for enabling efficient model quantization and inference via the GGUF format.
92
-
93
-
94
- ## Contact
95
- For any inquiries or support, please contact us at support@sandlogic.com or visit our [Website](https://www.sandlogic.com/).