burak commited on
Update README.md
Browse files
README.md
CHANGED
|
@@ -17,6 +17,9 @@ parameters: 95.7M
|
|
| 17 |
|
| 18 |
This is the latest and most experimental version of the **SykoLLM** series. Developed and trained entirely by **Burak (15 years old)**, this model is designed to explore "Chain of Thought" (CoT) capabilities in small-scale Turkish Language Models.
|
| 19 |
|
|
|
|
|
|
|
|
|
|
| 20 |
## ⚠️ Important: Beta Status
|
| 21 |
This model is currently in a **strict Beta phase**.
|
| 22 |
- The training for the "thinking" mechanism is still ongoing/experimental.
|
|
|
|
| 17 |
|
| 18 |
This is the latest and most experimental version of the **SykoLLM** series. Developed and trained entirely by **Burak (15 years old)**, this model is designed to explore "Chain of Thought" (CoT) capabilities in small-scale Turkish Language Models.
|
| 19 |
|
| 20 |
+
## 🚀 Important Technical Distinction
|
| 21 |
+
**This model is **NOT a LoRA adapter** or a **simple copy of GPT-2.** It is a standalone, full-parameter fine-tuned model where the actual weights have been modified through training. The positional embeddings were manually expanded from 512 to 1024 tokens via a custom "weight surgery" process to support longer context natively.
|
| 22 |
+
|
| 23 |
## ⚠️ Important: Beta Status
|
| 24 |
This model is currently in a **strict Beta phase**.
|
| 25 |
- The training for the "thinking" mechanism is still ongoing/experimental.
|