prithivMLmods commited on
Commit
6976c73
·
verified ·
1 Parent(s): 8d0e766

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +91 -1
README.md CHANGED
@@ -4,4 +4,94 @@ datasets:
4
  - sequelbox/Celestia3-DeepSeek-R1-0528
5
  base_model:
6
  - HuggingFaceTB/SmolLM2-135M-Instruct
7
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4
  - sequelbox/Celestia3-DeepSeek-R1-0528
5
  base_model:
6
  - HuggingFaceTB/SmolLM2-135M-Instruct
7
+ ---
8
+
9
+ # **SmolLM2-Rethink-135M**
10
+
11
+ > **SmolLM2-Rethink-135M** is an experimental lightweight model trained on the **Celestia3-DeepSeek-R1-0528** reasoning dataset. Based on the **SmolLM2-135M-Instruct** architecture, this model is specifically optimized for reasoning, structured outputs, and efficient small-scale deployment. Despite its compact size (135M parameters), it demonstrates strong capabilities in logical deduction, conversational coherence, and lightweight inference tasks.
12
+
13
+ ---
14
+
15
+ ## **Key Highlights**
16
+
17
+ 1. **Compact & Efficient**
18
+ Lightweight architecture (135M) suitable for fast inference, mobile applications, and edge deployment.
19
+
20
+ 2. **Reasoning-Centric Training**
21
+ Fine-tuned on high-quality reasoning and instruction datasets like **Celestia3-DeepSeek-R1-0528**, focusing on multi-step logical thinking.
22
+
23
+ 3. **Low-Resource Optimization**
24
+ Designed to run effectively on CPUs or single-GPU setups with minimal memory footprint.
25
+
26
+ 4. **Structured Outputs**
27
+ Supports generation of clean, structured content including lists, steps, tables, and JSON-like responses.
28
+
29
+ ---
30
+
31
+ ## **Quickstart with 🤗 Transformers**
32
+
33
+ ```python
34
+ %%capture
35
+ !pip install transformers
36
+ ```
37
+
38
+ ```python
39
+ from transformers import AutoModelForCausalLM, AutoTokenizer
40
+
41
+ checkpoint = "prithivMLmods/SmolLM2-Rethink-135M"
42
+ device = "cuda" # or "cpu"
43
+
44
+ tokenizer = AutoTokenizer.from_pretrained(checkpoint)
45
+ model = AutoModelForCausalLM.from_pretrained(checkpoint).to(device)
46
+
47
+ messages = [{"role": "user", "content": "What is gravity?"}]
48
+ input_text = tokenizer.apply_chat_template(messages, tokenize=False)
49
+ print(input_text)
50
+
51
+ inputs = tokenizer.encode(input_text, return_tensors="pt").to(device)
52
+ outputs = model.generate(
53
+ inputs,
54
+ max_new_tokens=1024,
55
+ temperature=0.2,
56
+ top_p=0.9,
57
+ do_sample=True
58
+ )
59
+
60
+ print(tokenizer.decode(outputs[0]))
61
+ ```
62
+
63
+ ---
64
+
65
+ ## **Intended Use**
66
+
67
+ * **Instruction Following & QA**
68
+ Good for answering simple questions, following short instructions, and general user interactions.
69
+
70
+ * **Educational Tools**
71
+ Suitable for lightweight tutoring bots or classroom assistants on low-compute setups.
72
+
73
+ * **Reasoning Tasks**
74
+ Performs well on logic puzzles, multi-step reasoning, and chain-of-thought queries.
75
+
76
+ * **Prototype Agents & Microservices**
77
+ Can be deployed in memory-efficient environments or as modular AI components.
78
+
79
+ ---
80
+
81
+ ## **Limitations**
82
+
83
+ 1. **Limited Knowledge Capacity**
84
+ Due to small parameter size, lacks the depth and breadth of large-scale models.
85
+
86
+ 2. **Short-Term Context Handling**
87
+ Performs best with short to moderate-length prompts; lacks extended context support.
88
+
89
+ 3. **Creative Generation Limitations**
90
+ Output may lack diversity or depth in open-ended storytelling or imaginative tasks.
91
+
92
+ 4. **Token Budget**
93
+ Smaller output range; optimized for shorter and structured completions.
94
+
95
+ 5. **Basic Multilingual Support**
96
+ Some support for multilingual input, but less accurate than larger multilingual models.
97
+