burak commited on
Commit
4ee4e2d
·
verified ·
1 Parent(s): 2013941

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -0
README.md CHANGED
@@ -17,6 +17,9 @@ parameters: 95.7M
17
 
18
  This is the latest and most experimental version of the **SykoLLM** series. Developed and trained entirely by **Burak (15 years old)**, this model is designed to explore "Chain of Thought" (CoT) capabilities in small-scale Turkish Language Models.
19
 
 
 
 
20
  ## ⚠️ Important: Beta Status
21
  This model is currently in a **strict Beta phase**.
22
  - The training for the "thinking" mechanism is still ongoing/experimental.
 
17
 
18
  This is the latest and most experimental version of the **SykoLLM** series. Developed and trained entirely by **Burak (15 years old)**, this model is designed to explore "Chain of Thought" (CoT) capabilities in small-scale Turkish Language Models.
19
 
20
+ ## 🚀 Important Technical Distinction
21
+ **This model is **NOT a LoRA adapter** or a **simple copy of GPT-2.** It is a standalone, full-parameter fine-tuned model where the actual weights have been modified through training. The positional embeddings were manually expanded from 512 to 1024 tokens via a custom "weight surgery" process to support longer context natively.
22
+
23
  ## ⚠️ Important: Beta Status
24
  This model is currently in a **strict Beta phase**.
25
  - The training for the "thinking" mechanism is still ongoing/experimental.