File size: 2,474 Bytes
3695d52
 
668a2db
3695d52
 
 
668a2db
3695d52
668a2db
3695d52
668a2db
3695d52
668a2db
 
 
 
 
 
3695d52
668a2db
3695d52
668a2db
3695d52
668a2db
 
 
 
 
3695d52
668a2db
3695d52
668a2db
3695d52
668a2db
 
3695d52
 
668a2db
 
3695d52
 
668a2db
 
3695d52
 
668a2db
 
3695d52
 
 
668a2db
3695d52
668a2db
3695d52
668a2db
 
3695d52
668a2db
 
 
3695d52
668a2db
 
 
3695d52
668a2db
3695d52
668a2db
3695d52
668a2db
3695d52
668a2db
 
 
3695d52
 
668a2db
 
 
3695d52
668a2db
3695d52
668a2db
 
3695d52
ff9c2a1
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
---
library_name: transformers
license: apache-2.0
---


# Model Card: SykoLLM-V1-Turkish

SykoLLM-V1-Turkish is a custom-architected, lightweight Large Language Model (LLM) designed specifically for Turkish conversational tasks. Unlike standard pre-built models, this version features a custom configuration optimized for speed and efficiency in low-resource environments.

## Model Description

* **Developed by:** syko818121
* **Model Name:** SykoLLM-V1-Turkish
* **Model Type:** Causal Decoder-Only Custom Architecture
* **Language:** Turkish
* **Parameters:** ~50.8 Million
* **Training Data:** Turkish Wikipedia + Custom High-Quality Chat Dataset

## Architectural Specs

This model uses a custom configuration designed for Turkish linguistics:

* **Vocabulary Size:** 50,257
* **Hidden Dimension (n_embd):** 512
* **Number of Layers:** 8
* **Attention Heads:** 8
* **Context Window:** 512 tokens

## Fine-Tuning & Conversation Style

The model was fine-tuned on a high-quality, curated Turkish dataset  to ensure natural, human-like responses. The training data distribution was carefully balanced:

* 
**Greetings & Daily Talk (40%):** Natural openings and casual conversation.


* 
**Direct Question-Answering (30%):** Short and concise answers to general knowledge queries.


* 
**Brief Explanations (20%):** Simplified definitions for complex concepts.


* 
**Slang & Short Inputs (10%):** Robustness against one-word or incomplete messages.



## Usage

You can load and test SykoLLM-V1-Turkish using the following snippet:

```python
from transformers import AutoModelForCausalLM, AutoTokenizer

model_id = "syko818121/SykoLLM-V1-Turkish"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id, trust_remote_code=True)

prompt = "<user> Selam, naber?<assistant>"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=50, pad_token_id=tokenizer.eos_token_id)

print(tokenizer.decode(outputs[0], skip_special_tokens=True))

```

## Training Configuration

* **Learning Rate:** 5e-5
* 
**Scheduler:** Cosine 


* **Epochs:** 15
* **Batch Size:** 4
* **Precision:** FP16 (Mixed Precision)

## Limitations

* **Size:** As a 50.8M parameter model, it is a "micro-LLM." It excels at short chats but may hallucinate on highly complex logical tasks.
* **Response Length:** The model is intentionally biased toward concise and direct answers rather than long-form essays.

---