burak commited on
Update README.md
Browse files
README.md
CHANGED
|
@@ -17,7 +17,7 @@ parameters: 95.7M
|
|
| 17 |
|
| 18 |
This is the latest and most experimental version of the **SykoLLM** series. Developed and trained entirely by **Burak (15 years old)**, this model is designed to explore "Chain of Thought" (CoT) capabilities in small-scale Turkish Language Models.
|
| 19 |
|
| 20 |
-
##
|
| 21 |
**This model is **NOT a LoRA adapter** or a **simple copy of GPT-2.** It is a standalone, full-parameter fine-tuned model where the actual weights have been modified through training. The positional embeddings were manually expanded from 512 to 1024 tokens via a custom "weight surgery" process to support longer context natively.
|
| 22 |
|
| 23 |
## ⚠️ Important: Beta Status
|
|
|
|
| 17 |
|
| 18 |
This is the latest and most experimental version of the **SykoLLM** series. Developed and trained entirely by **Burak (15 years old)**, this model is designed to explore "Chain of Thought" (CoT) capabilities in small-scale Turkish Language Models.
|
| 19 |
|
| 20 |
+
## Important Technical Distinction
|
| 21 |
**This model is **NOT a LoRA adapter** or a **simple copy of GPT-2.** It is a standalone, full-parameter fine-tuned model where the actual weights have been modified through training. The positional embeddings were manually expanded from 512 to 1024 tokens via a custom "weight surgery" process to support longer context natively.
|
| 22 |
|
| 23 |
## ⚠️ Important: Beta Status
|