Upload README.md with huggingface_hub
Browse files
README.md
CHANGED
|
@@ -12,8 +12,6 @@ tags:
|
|
| 12 |
- instruction-tuned
|
| 13 |
- low-resource
|
| 14 |
- nlp
|
| 15 |
-
datasets:
|
| 16 |
-
- ogulcanaydogan/Turkish-LLM-v10-Training
|
| 17 |
pipeline_tag: text-generation
|
| 18 |
model-index:
|
| 19 |
- name: Turkish-LLM-14B-Instruct
|
|
@@ -26,8 +24,8 @@ An open-source 14.7 billion parameter language model fine-tuned for native Turki
|
|
| 26 |
|
| 27 |
<p align="center">
|
| 28 |
<a href="https://huggingface.co/spaces/ogulcanaydogan/Turkish-LLM-14B-Chat"><img src="https://img.shields.io/badge/Demo-Live_Chat-blue?style=for-the-badge&logo=huggingface" alt="Demo"></a>
|
|
|
|
| 29 |
<a href="https://github.com/ogulcanaydogan/Turkish-LLM"><img src="https://img.shields.io/badge/GitHub-Repository-black?style=for-the-badge&logo=github" alt="GitHub"></a>
|
| 30 |
-
<a href="https://huggingface.co/datasets/ogulcanaydogan/Turkish-LLM-v10-Training"><img src="https://img.shields.io/badge/Dataset-144K_samples-green?style=for-the-badge&logo=huggingface" alt="Dataset"></a>
|
| 31 |
<a href="https://huggingface.co/ogulcanaydogan/Turkish-LLM-7B-Instruct"><img src="https://img.shields.io/badge/Also_Available-7B_Model-yellow?style=for-the-badge&logo=huggingface" alt="7B"></a>
|
| 32 |
</p>
|
| 33 |
|
|
@@ -61,18 +59,17 @@ This model was developed to provide a **high-quality, open-source Turkish langua
|
|
| 61 |
|
| 62 |
### Model Family
|
| 63 |
|
| 64 |
-
This model is part of the **Turkish-LLM** family:
|
| 65 |
-
|
| 66 |
| Model | Parameters | Base | Method | Use Case |
|
| 67 |
|-------|-----------|------|--------|----------|
|
| 68 |
| **Turkish-LLM-14B-Instruct** (this) | 14.7B | Qwen2.5-14B-Instruct | SFT | Higher quality, complex reasoning |
|
|
|
|
| 69 |
| [Turkish-LLM-7B-Instruct](https://huggingface.co/ogulcanaydogan/Turkish-LLM-7B-Instruct) | 7B | Turkcell-LLM-7b-v1 | LoRA | Lightweight, faster inference |
|
| 70 |
|
| 71 |
## Training
|
| 72 |
|
| 73 |
### Dataset
|
| 74 |
|
| 75 |
-
Training data
|
| 76 |
|
| 77 |
| Domain | Examples | Purpose |
|
| 78 |
|--------|----------|---------|
|
|
@@ -95,13 +92,13 @@ Training data was sourced from the [Turkish-LLM-v10-Training](https://huggingfac
|
|
| 95 |
|
| 96 |
### Training Pipeline
|
| 97 |
|
| 98 |
-
Training was orchestrated using [LowResource-LLM-Forge](https://github.com/ogulcanaydogan/LowResource-LLM-Forge), a custom pipeline built for efficient fine-tuning of LLMs for low-resource languages.
|
| 99 |
|
| 100 |
```
|
| 101 |
-
Raw Turkish Data
|
| 102 |
-
(144K pairs)
|
| 103 |
-
|
| 104 |
-
|
| 105 |
```
|
| 106 |
|
| 107 |
### Design Decisions
|
|
@@ -158,9 +155,13 @@ vllm serve ogulcanaydogan/Turkish-LLM-14B-Instruct \
|
|
| 158 |
### Ollama (Local)
|
| 159 |
|
| 160 |
```bash
|
| 161 |
-
ollama run hf.co/ogulcanaydogan/Turkish-LLM-14B-Instruct
|
| 162 |
```
|
| 163 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 164 |
### Chat Template
|
| 165 |
|
| 166 |
This model uses the ChatML format:
|
|
@@ -182,7 +183,7 @@ Sen yardimci bir Turkce yapay zeka asistanisin.<|im_end|>
|
|
| 182 |
| INT8 | ~15 GB | RTX 4090, A10G |
|
| 183 |
| INT4 (GPTQ/AWQ) | ~8 GB | RTX 3090, RTX 4080, Apple M-series (24GB) |
|
| 184 |
|
| 185 |
-
For consumer hardware,
|
| 186 |
|
| 187 |
## Intended Use
|
| 188 |
|
|
@@ -218,8 +219,8 @@ This model is released under Apache 2.0 to support open research and development
|
|
| 218 |
|
| 219 |
| Resource | Link |
|
| 220 |
|----------|------|
|
|
|
|
| 221 |
| 7B Model | [Turkish-LLM-7B-Instruct](https://huggingface.co/ogulcanaydogan/Turkish-LLM-7B-Instruct) |
|
| 222 |
-
| Training Dataset (144K) | [Turkish-LLM-v10-Training](https://huggingface.co/datasets/ogulcanaydogan/Turkish-LLM-v10-Training) |
|
| 223 |
| Live Demo (14B) | [Turkish-LLM-14B-Chat](https://huggingface.co/spaces/ogulcanaydogan/Turkish-LLM-14B-Chat) |
|
| 224 |
| Live Demo (7B) | [Turkish-LLM-7B-Chat](https://huggingface.co/spaces/ogulcanaydogan/Turkish-LLM-7B-Chat) |
|
| 225 |
| Training Pipeline | [LowResource-LLM-Forge](https://github.com/ogulcanaydogan/LowResource-LLM-Forge) |
|
|
|
|
| 12 |
- instruction-tuned
|
| 13 |
- low-resource
|
| 14 |
- nlp
|
|
|
|
|
|
|
| 15 |
pipeline_tag: text-generation
|
| 16 |
model-index:
|
| 17 |
- name: Turkish-LLM-14B-Instruct
|
|
|
|
| 24 |
|
| 25 |
<p align="center">
|
| 26 |
<a href="https://huggingface.co/spaces/ogulcanaydogan/Turkish-LLM-14B-Chat"><img src="https://img.shields.io/badge/Demo-Live_Chat-blue?style=for-the-badge&logo=huggingface" alt="Demo"></a>
|
| 27 |
+
<a href="https://huggingface.co/ogulcanaydogan/Turkish-LLM-14B-Instruct-GGUF"><img src="https://img.shields.io/badge/GGUF-Quantized_Versions-orange?style=for-the-badge&logo=huggingface" alt="GGUF"></a>
|
| 28 |
<a href="https://github.com/ogulcanaydogan/Turkish-LLM"><img src="https://img.shields.io/badge/GitHub-Repository-black?style=for-the-badge&logo=github" alt="GitHub"></a>
|
|
|
|
| 29 |
<a href="https://huggingface.co/ogulcanaydogan/Turkish-LLM-7B-Instruct"><img src="https://img.shields.io/badge/Also_Available-7B_Model-yellow?style=for-the-badge&logo=huggingface" alt="7B"></a>
|
| 30 |
</p>
|
| 31 |
|
|
|
|
| 59 |
|
| 60 |
### Model Family
|
| 61 |
|
|
|
|
|
|
|
| 62 |
| Model | Parameters | Base | Method | Use Case |
|
| 63 |
|-------|-----------|------|--------|----------|
|
| 64 |
| **Turkish-LLM-14B-Instruct** (this) | 14.7B | Qwen2.5-14B-Instruct | SFT | Higher quality, complex reasoning |
|
| 65 |
+
| [Turkish-LLM-14B-Instruct-GGUF](https://huggingface.co/ogulcanaydogan/Turkish-LLM-14B-Instruct-GGUF) | 14.7B | This model | GGUF quantized | Local/edge deployment |
|
| 66 |
| [Turkish-LLM-7B-Instruct](https://huggingface.co/ogulcanaydogan/Turkish-LLM-7B-Instruct) | 7B | Turkcell-LLM-7b-v1 | LoRA | Lightweight, faster inference |
|
| 67 |
|
| 68 |
## Training
|
| 69 |
|
| 70 |
### Dataset
|
| 71 |
|
| 72 |
+
Training data consists of a curated collection of **144,000 Turkish instruction-response pairs**, with a focused SFT subset of approximately 2,600 high-quality examples selected for alignment.
|
| 73 |
|
| 74 |
| Domain | Examples | Purpose |
|
| 75 |
|--------|----------|---------|
|
|
|
|
| 92 |
|
| 93 |
### Training Pipeline
|
| 94 |
|
| 95 |
+
Training was orchestrated using [LowResource-LLM-Forge](https://github.com/ogulcanaydogan/LowResource-LLM-Forge), a custom pipeline built for efficient fine-tuning of LLMs for low-resource languages.
|
| 96 |
|
| 97 |
```
|
| 98 |
+
Raw Turkish Data --> Preprocessing --> SFT Training --> Evaluation --> Deployment
|
| 99 |
+
(144K pairs) (filtering, (A100 80GB, (manual + (HF Hub,
|
| 100 |
+
dedup, bf16 mixed qualitative) Spaces,
|
| 101 |
+
formatting) precision) vLLM)
|
| 102 |
```
|
| 103 |
|
| 104 |
### Design Decisions
|
|
|
|
| 155 |
### Ollama (Local)
|
| 156 |
|
| 157 |
```bash
|
| 158 |
+
ollama run hf.co/ogulcanaydogan/Turkish-LLM-14B-Instruct-GGUF:Q4_K_M
|
| 159 |
```
|
| 160 |
|
| 161 |
+
### GGUF (llama.cpp / LM Studio)
|
| 162 |
+
|
| 163 |
+
Quantized GGUF versions (Q4_K_M, Q5_K_M, Q8_0, F16) are available at [Turkish-LLM-14B-Instruct-GGUF](https://huggingface.co/ogulcanaydogan/Turkish-LLM-14B-Instruct-GGUF).
|
| 164 |
+
|
| 165 |
### Chat Template
|
| 166 |
|
| 167 |
This model uses the ChatML format:
|
|
|
|
| 183 |
| INT8 | ~15 GB | RTX 4090, A10G |
|
| 184 |
| INT4 (GPTQ/AWQ) | ~8 GB | RTX 3090, RTX 4080, Apple M-series (24GB) |
|
| 185 |
|
| 186 |
+
For consumer hardware, use the [GGUF versions](https://huggingface.co/ogulcanaydogan/Turkish-LLM-14B-Instruct-GGUF) for the best balance of quality and accessibility.
|
| 187 |
|
| 188 |
## Intended Use
|
| 189 |
|
|
|
|
| 219 |
|
| 220 |
| Resource | Link |
|
| 221 |
|----------|------|
|
| 222 |
+
| GGUF Versions | [Turkish-LLM-14B-Instruct-GGUF](https://huggingface.co/ogulcanaydogan/Turkish-LLM-14B-Instruct-GGUF) |
|
| 223 |
| 7B Model | [Turkish-LLM-7B-Instruct](https://huggingface.co/ogulcanaydogan/Turkish-LLM-7B-Instruct) |
|
|
|
|
| 224 |
| Live Demo (14B) | [Turkish-LLM-14B-Chat](https://huggingface.co/spaces/ogulcanaydogan/Turkish-LLM-14B-Chat) |
|
| 225 |
| Live Demo (7B) | [Turkish-LLM-7B-Chat](https://huggingface.co/spaces/ogulcanaydogan/Turkish-LLM-7B-Chat) |
|
| 226 |
| Training Pipeline | [LowResource-LLM-Forge](https://github.com/ogulcanaydogan/LowResource-LLM-Forge) |
|