prithivMLmods commited on
Commit
87e367f
·
verified ·
1 Parent(s): e759204

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +87 -1
README.md CHANGED
@@ -14,4 +14,90 @@ tags:
14
  - code
15
  - synlogic
16
  - moe
17
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14
  - code
15
  - synlogic
16
  - moe
17
+ ---
18
+
19
+ # **Megatron-Bots-1.7B-Reasoning**
20
+
21
+ > **Megatron-Bots-1.7B-Reasoning** is a **logical reasoning and general-purpose thinking model** fine-tuned from **Qwen3-1.7B**, specifically designed for **advanced reasoning tasks and analytical problem-solving**. Built with data entries from the **SynLogic Dataset**, it excels at structured thinking, logical deduction, and comprehensive problem analysis in a compact yet powerful architecture.
22
+
23
+ > \[!note]
24
+ > GGUF: [https://huggingface.co/prithivMLmods/Megatron-Bots-1.7B-Reasoning-GGUF](https://huggingface.co/prithivMLmods/Megatron-Bots-1.7B-Reasoning-GGUF)
25
+
26
+ ## **Key Features**
27
+ 1. **Advanced Logical Reasoning**
28
+ Trained on the SynLogic Dataset to perform complex logical deductions, structured problem-solving, and analytical thinking across diverse domains with exceptional accuracy and clarity.
29
+
30
+ 2. **General-Purpose Thinking Engine**
31
+ Capable of handling multi-step reasoning, causal analysis, pattern recognition, and systematic problem decomposition for a wide range of cognitive tasks.
32
+
33
+ 3. **Compact High-Performance Architecture**
34
+ While only 1.7B parameters, this model delivers sophisticated reasoning capabilities with minimal resource requirements, making it ideal for deployment in resource-constrained environments.
35
+
36
+ 4. **SynLogic Dataset Foundation**
37
+ Built upon carefully curated synthetic logic problems and reasoning patterns, ensuring robust performance across mathematical reasoning, logical puzzles, and analytical challenges.
38
+
39
+ ## **Quickstart with Transformers**
40
+ ```python
41
+ from transformers import AutoModelForCausalLM, AutoTokenizer
42
+
43
+ model_name = "prithivMLmods/Megatron-Bots-1.7B-Reasoning"
44
+
45
+ model = AutoModelForCausalLM.from_pretrained(
46
+ model_name,
47
+ torch_dtype="auto",
48
+ device_map="auto"
49
+ )
50
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
51
+
52
+ prompt = "Solve this logic puzzle: If all A are B, and some B are C, what can we conclude about A and C?"
53
+
54
+ messages = [
55
+ {"role": "system", "content": "You are an advanced reasoning assistant specialized in logical analysis and problem-solving."},
56
+ {"role": "user", "content": prompt}
57
+ ]
58
+
59
+ text = tokenizer.apply_chat_template(
60
+ messages,
61
+ tokenize=False,
62
+ add_generation_prompt=True
63
+ )
64
+
65
+ model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
66
+
67
+ generated_ids = model.generate(
68
+ **model_inputs,
69
+ max_new_tokens=512,
70
+ temperature=0.1, # Lower temperature for more consistent reasoning
71
+ do_sample=True
72
+ )
73
+
74
+ generated_ids = [
75
+ output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
76
+ ]
77
+
78
+ response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
79
+ print(response)
80
+ ```
81
+
82
+ ## **Intended Use**
83
+ - **Educational Platforms**: Logical reasoning tutoring and step-by-step problem explanation for students.
84
+ - **Research Applications**: Automated logical analysis and hypothesis generation for academic research.
85
+ - **Decision Support Systems**: Structured analytical thinking for business and strategic decision-making.
86
+ - **Puzzle and Game AI**: Advanced reasoning for complex puzzles, strategy games, and logical challenges.
87
+ - **Code Analysis Tools**: Logical flow analysis and debugging assistance for software development.
88
+
89
+ ## **Limitations**
90
+ 1. **Reasoning Domain Specificity**:
91
+ While strong in logical reasoning, performance may vary on tasks requiring extensive domain-specific knowledge outside the training scope.
92
+
93
+ 2. **SynLogic Dataset Constraints**:
94
+ Training primarily on synthetic logic data may limit performance on real-world reasoning scenarios that require contextual understanding.
95
+
96
+ 3. **Parameter Scale Trade-offs**:
97
+ The 1.7B parameter size, while efficient, may struggle with extremely complex multi-step reasoning chains compared to larger models.
98
+
99
+ 4. **Base Model Inheritance**:
100
+ Inherits any limitations from Qwen3-1.7B's base architecture and potential biases from pretraining data.
101
+
102
+ 5. **Context Window Limitations**:
103
+ May face challenges with very long reasoning chains that exceed the model's context window capacity.