prithivMLmods commited on
Commit
1aaed8d
·
verified ·
1 Parent(s): 1328cb1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +94 -1
README.md CHANGED
@@ -9,4 +9,97 @@ library_name: transformers
9
  tags:
10
  - text-generation-inference
11
  - code
12
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9
  tags:
10
  - text-generation-inference
11
  - code
12
+ ---
13
+
14
+ ![IOP.png](https://cdn-uploads.huggingface.co/production/uploads/65bb837dbfb878f46c77de4c/X4wG8maYiZT68QLGW4NPn.png)
15
+
16
+ # Vulpecula-4B
17
+
18
+ > **Vulpecula-4B** is fine-tuned based on the traces of **SK1.1**, consisting of the same 1,000 entries of the **DeepSeek thinking trajectory**, along with fine-tuning on **Fine-Tome 100k** and **Open Math Reasoning** datasets. This specialized 4B parameter model is designed for enhanced mathematical reasoning, logical problem-solving, and structured content generation, optimized for precision and step-by-step explanation.
19
+
20
+ > \[!note]
21
+ > GGUF \[Q4\_K\_M] : [https://huggingface.co/prithivMLmods/Draconis-Qwen3\_Math-4B-Preview-Q4\_K\_M-GGUF](https://huggingface.co/prithivMLmods/Draconis-Qwen3_Math-4B-Preview-Q4_K_M-GGUF)
22
+
23
+ > \[!note]
24
+ > GGUF \[Q5\_K\_M] : [https://huggingface.co/prithivMLmods/Draconis-Qwen3\_Math-4B-Preview-Q5\_K\_M-GGUF](https://huggingface.co/prithivMLmods/Draconis-Qwen3_Math-4B-Preview-Q5_K_M-GGUF)
25
+
26
+ ## Key Features
27
+
28
+ 1. **Advanced Mathematical and Logical Reasoning**
29
+ Fine-tuned on DeepSeek trajectories and Open Math Reasoning to excel at symbolic logic, arithmetic, and complex multi-step math problems, ideal for STEM education and competitions.
30
+
31
+ 2. **Trace-Based Fine-Tuning**
32
+ Leverages SK1.1 trace dataset entries to model deep, interpretable reasoning paths, improving transparency and consistency in problem-solving.
33
+
34
+ 3. **Compact Code Understanding**
35
+ Capable of understanding and generating efficient code snippets in Python, JavaScript, and more, supporting algorithmic explanations and lightweight coding tasks.
36
+
37
+ 4. **Factual and Instructional Precision**
38
+ Trained on curated high-quality data with reasoning benchmarks to minimize hallucinations and strictly follow instructions for structured outputs (Markdown, JSON, tables).
39
+
40
+ 5. **Multilingual Capabilities**
41
+ Supports over 20 languages for technical reasoning and translation, enhancing multilingual educational applications.
42
+
43
+ 6. **Optimized Performance for Resource-Constrained Environments**
44
+ Balances reasoning capability with efficient resource use, suitable for deployment in environments with limited compute.
45
+
46
+ ## Quickstart with Transformers
47
+
48
+ ```python
49
+ from transformers import AutoModelForCausalLM, AutoTokenizer
50
+
51
+ model_name = "prithivMLmods/Vulpecula-4B"
52
+
53
+ model = AutoModelForCausalLM.from_pretrained(
54
+ model_name,
55
+ torch_dtype="auto",
56
+ device_map="auto"
57
+ )
58
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
59
+
60
+ prompt = "Solve the equation: 3x + 7 = 22. Show all steps."
61
+
62
+ messages = [
63
+ {"role": "system", "content": "You are a step-by-step math tutor."},
64
+ {"role": "user", "content": prompt}
65
+ ]
66
+
67
+ text = tokenizer.apply_chat_template(
68
+ messages,
69
+ tokenize=False,
70
+ add_generation_prompt=True
71
+ )
72
+
73
+ model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
74
+
75
+ generated_ids = model.generate(
76
+ **model_inputs,
77
+ max_new_tokens=512
78
+ )
79
+ generated_ids = [
80
+ output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
81
+ ]
82
+
83
+ response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
84
+ print(response)
85
+ ```
86
+
87
+ ## Intended Use
88
+
89
+ * Advanced mathematical and logical problem solving
90
+ * Education-centric STEM tutoring and explanations
91
+ * Code assistance and debugging for lightweight coding tasks
92
+ * Structured content generation including JSON, Markdown, and tables
93
+ * Multilingual reasoning and technical translation
94
+ * Efficient deployment in low-resource settings with a focus on accuracy and stepwise reasoning
95
+
96
+ ## Limitations
97
+
98
+ * Limited creativity in purely open-ended or fictional prompts
99
+ * May face challenges with ambiguous or multi-intent queries
100
+ * Smaller context window compared to larger 14B+ models
101
+ * Possible factual errors in complex edge cases or adversarial inputs
102
+
103
+ ## References
104
+
105
+ 1. [YaRN: Efficient Context Window Extension of Large Language Models](https://arxiv.org/pdf/2309.00071)