File size: 2,237 Bytes
be677ea
 
2cf0ca1
 
 
 
 
be677ea
 
563c732
be677ea
563c732
be677ea
563c732
be677ea
563c732
 
 
 
95def88
563c732
be677ea
 
563c732
be677ea
563c732
be677ea
563c732
 
be677ea
 
563c732
 
be677ea
 
563c732
 
be677ea
 
563c732
 
be677ea
 
 
563c732
be677ea
563c732
be677ea
563c732
 
be677ea
563c732
 
 
be677ea
563c732
 
 
be677ea
563c732
be677ea
563c732
be677ea
563c732
be677ea
563c732
 
 
be677ea
563c732
be677ea
95def88
563c732
be677ea
563c732
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
---
library_name: transformers
license: apache-2.0
datasets:
- TFLai/Turkish-Alpaca
language:
- tr
---

# Model Card: SykoLLM-V2.1-Turkish-Instruct

SykoLLM-V2.1-Turkish-Instruct is a custom-architected, lightweight Large Language Model (LLM) designed specifically for Turkish conversational tasks. Unlike standard pre-built models, this version features a custom configuration optimized for speed and efficiency in low-resource environments.

## Model Description

* **Developed by:** syko818121
* **Model Name:** SykoLLM-V2.1-Turkish-Instructt
* **Model Type:** Causal Decoder-Only Custom Architecture
* **Language:** Turkish
* **Parameters:** ~95.7 Million
* **Training Data:** Turkish Wikipedia + Custom High-Quality Chat Dataset


## Fine-Tuning & Conversation Style

The model was fine-tuned on a high-quality, curated Turkish dataset  to ensure natural, human-like responses. The training data distribution was carefully balanced:

* 
**Greetings & Daily Talk (40%):** Natural openings and casual conversation.


* 
**Direct Question-Answering (30%):** Short and concise answers to general knowledge queries.


* 
**Brief Explanations (20%):** Simplified definitions for complex concepts.


* 
**Slang & Short Inputs (10%):** Robustness against one-word or incomplete messages.



## Usage

You can load and test SykoLLM-V2.1-Turkish-Instruct using the following snippet:

```python
from transformers import AutoModelForCausalLM, AutoTokenizer

model_id = "SykoLLM-V2.1-Turkish-Instruct"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id, trust_remote_code=True)

prompt = "<user> Selam, naber?<assistant>"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=50, pad_token_id=tokenizer.eos_token_id)

print(tokenizer.decode(outputs[0], skip_special_tokens=True))

```

## Training Configuration

* **Learning Rate:** 5e-5
* 
**Scheduler:** Cosine 

## Limitations

* **Size:** As a 95.7M parameter model, it is a "mini-LLM." It excels at short chats but may hallucinate on highly complex logical tasks.
* **Response Length:** The model is intentionally biased toward concise and direct answers rather than long-form essays.

---