ogulcanaydogan commited on
Commit
397efe7
·
verified ·
1 Parent(s): 4bd2a8b

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +15 -14
README.md CHANGED
@@ -12,8 +12,6 @@ tags:
12
  - instruction-tuned
13
  - low-resource
14
  - nlp
15
- datasets:
16
- - ogulcanaydogan/Turkish-LLM-v10-Training
17
  pipeline_tag: text-generation
18
  model-index:
19
  - name: Turkish-LLM-14B-Instruct
@@ -26,8 +24,8 @@ An open-source 14.7 billion parameter language model fine-tuned for native Turki
26
 
27
  <p align="center">
28
  <a href="https://huggingface.co/spaces/ogulcanaydogan/Turkish-LLM-14B-Chat"><img src="https://img.shields.io/badge/Demo-Live_Chat-blue?style=for-the-badge&logo=huggingface" alt="Demo"></a>
 
29
  <a href="https://github.com/ogulcanaydogan/Turkish-LLM"><img src="https://img.shields.io/badge/GitHub-Repository-black?style=for-the-badge&logo=github" alt="GitHub"></a>
30
- <a href="https://huggingface.co/datasets/ogulcanaydogan/Turkish-LLM-v10-Training"><img src="https://img.shields.io/badge/Dataset-144K_samples-green?style=for-the-badge&logo=huggingface" alt="Dataset"></a>
31
  <a href="https://huggingface.co/ogulcanaydogan/Turkish-LLM-7B-Instruct"><img src="https://img.shields.io/badge/Also_Available-7B_Model-yellow?style=for-the-badge&logo=huggingface" alt="7B"></a>
32
  </p>
33
 
@@ -61,18 +59,17 @@ This model was developed to provide a **high-quality, open-source Turkish langua
61
 
62
  ### Model Family
63
 
64
- This model is part of the **Turkish-LLM** family:
65
-
66
  | Model | Parameters | Base | Method | Use Case |
67
  |-------|-----------|------|--------|----------|
68
  | **Turkish-LLM-14B-Instruct** (this) | 14.7B | Qwen2.5-14B-Instruct | SFT | Higher quality, complex reasoning |
 
69
  | [Turkish-LLM-7B-Instruct](https://huggingface.co/ogulcanaydogan/Turkish-LLM-7B-Instruct) | 7B | Turkcell-LLM-7b-v1 | LoRA | Lightweight, faster inference |
70
 
71
  ## Training
72
 
73
  ### Dataset
74
 
75
- Training data was sourced from the [Turkish-LLM-v10-Training](https://huggingface.co/datasets/ogulcanaydogan/Turkish-LLM-v10-Training) dataset — a curated collection of **144,000 Turkish instruction-response pairs** — with a focused SFT subset of approximately 2,600 high-quality examples selected for alignment.
76
 
77
  | Domain | Examples | Purpose |
78
  |--------|----------|---------|
@@ -95,13 +92,13 @@ Training data was sourced from the [Turkish-LLM-v10-Training](https://huggingfac
95
 
96
  ### Training Pipeline
97
 
98
- Training was orchestrated using [LowResource-LLM-Forge](https://github.com/ogulcanaydogan/LowResource-LLM-Forge), a custom pipeline built for efficient fine-tuning of LLMs for low-resource languages. The pipeline handles data preprocessing, tokenization, training, and evaluation in a single workflow.
99
 
100
  ```
101
- Raw Turkish Data ──▶ Preprocessing ──▶ SFT Training ──▶ Evaluation ──▶ Deployment
102
- (144K pairs) (filtering, (A100 80GB, (manual + (HF Hub,
103
- dedup, bf16 mixed qualitative) Spaces,
104
- formatting) precision) vLLM)
105
  ```
106
 
107
  ### Design Decisions
@@ -158,9 +155,13 @@ vllm serve ogulcanaydogan/Turkish-LLM-14B-Instruct \
158
  ### Ollama (Local)
159
 
160
  ```bash
161
- ollama run hf.co/ogulcanaydogan/Turkish-LLM-14B-Instruct
162
  ```
163
 
 
 
 
 
164
  ### Chat Template
165
 
166
  This model uses the ChatML format:
@@ -182,7 +183,7 @@ Sen yardimci bir Turkce yapay zeka asistanisin.<|im_end|>
182
  | INT8 | ~15 GB | RTX 4090, A10G |
183
  | INT4 (GPTQ/AWQ) | ~8 GB | RTX 3090, RTX 4080, Apple M-series (24GB) |
184
 
185
- For consumer hardware, INT4 quantized versions provide the best balance of quality and accessibility.
186
 
187
  ## Intended Use
188
 
@@ -218,8 +219,8 @@ This model is released under Apache 2.0 to support open research and development
218
 
219
  | Resource | Link |
220
  |----------|------|
 
221
  | 7B Model | [Turkish-LLM-7B-Instruct](https://huggingface.co/ogulcanaydogan/Turkish-LLM-7B-Instruct) |
222
- | Training Dataset (144K) | [Turkish-LLM-v10-Training](https://huggingface.co/datasets/ogulcanaydogan/Turkish-LLM-v10-Training) |
223
  | Live Demo (14B) | [Turkish-LLM-14B-Chat](https://huggingface.co/spaces/ogulcanaydogan/Turkish-LLM-14B-Chat) |
224
  | Live Demo (7B) | [Turkish-LLM-7B-Chat](https://huggingface.co/spaces/ogulcanaydogan/Turkish-LLM-7B-Chat) |
225
  | Training Pipeline | [LowResource-LLM-Forge](https://github.com/ogulcanaydogan/LowResource-LLM-Forge) |
 
12
  - instruction-tuned
13
  - low-resource
14
  - nlp
 
 
15
  pipeline_tag: text-generation
16
  model-index:
17
  - name: Turkish-LLM-14B-Instruct
 
24
 
25
  <p align="center">
26
  <a href="https://huggingface.co/spaces/ogulcanaydogan/Turkish-LLM-14B-Chat"><img src="https://img.shields.io/badge/Demo-Live_Chat-blue?style=for-the-badge&logo=huggingface" alt="Demo"></a>
27
+ <a href="https://huggingface.co/ogulcanaydogan/Turkish-LLM-14B-Instruct-GGUF"><img src="https://img.shields.io/badge/GGUF-Quantized_Versions-orange?style=for-the-badge&logo=huggingface" alt="GGUF"></a>
28
  <a href="https://github.com/ogulcanaydogan/Turkish-LLM"><img src="https://img.shields.io/badge/GitHub-Repository-black?style=for-the-badge&logo=github" alt="GitHub"></a>
 
29
  <a href="https://huggingface.co/ogulcanaydogan/Turkish-LLM-7B-Instruct"><img src="https://img.shields.io/badge/Also_Available-7B_Model-yellow?style=for-the-badge&logo=huggingface" alt="7B"></a>
30
  </p>
31
 
 
59
 
60
  ### Model Family
61
 
 
 
62
  | Model | Parameters | Base | Method | Use Case |
63
  |-------|-----------|------|--------|----------|
64
  | **Turkish-LLM-14B-Instruct** (this) | 14.7B | Qwen2.5-14B-Instruct | SFT | Higher quality, complex reasoning |
65
+ | [Turkish-LLM-14B-Instruct-GGUF](https://huggingface.co/ogulcanaydogan/Turkish-LLM-14B-Instruct-GGUF) | 14.7B | This model | GGUF quantized | Local/edge deployment |
66
  | [Turkish-LLM-7B-Instruct](https://huggingface.co/ogulcanaydogan/Turkish-LLM-7B-Instruct) | 7B | Turkcell-LLM-7b-v1 | LoRA | Lightweight, faster inference |
67
 
68
  ## Training
69
 
70
  ### Dataset
71
 
72
+ Training data consists of a curated collection of **144,000 Turkish instruction-response pairs**, with a focused SFT subset of approximately 2,600 high-quality examples selected for alignment.
73
 
74
  | Domain | Examples | Purpose |
75
  |--------|----------|---------|
 
92
 
93
  ### Training Pipeline
94
 
95
+ Training was orchestrated using [LowResource-LLM-Forge](https://github.com/ogulcanaydogan/LowResource-LLM-Forge), a custom pipeline built for efficient fine-tuning of LLMs for low-resource languages.
96
 
97
  ```
98
+ Raw Turkish Data --> Preprocessing --> SFT Training --> Evaluation --> Deployment
99
+ (144K pairs) (filtering, (A100 80GB, (manual + (HF Hub,
100
+ dedup, bf16 mixed qualitative) Spaces,
101
+ formatting) precision) vLLM)
102
  ```
103
 
104
  ### Design Decisions
 
155
  ### Ollama (Local)
156
 
157
  ```bash
158
+ ollama run hf.co/ogulcanaydogan/Turkish-LLM-14B-Instruct-GGUF:Q4_K_M
159
  ```
160
 
161
+ ### GGUF (llama.cpp / LM Studio)
162
+
163
+ Quantized GGUF versions (Q4_K_M, Q5_K_M, Q8_0, F16) are available at [Turkish-LLM-14B-Instruct-GGUF](https://huggingface.co/ogulcanaydogan/Turkish-LLM-14B-Instruct-GGUF).
164
+
165
  ### Chat Template
166
 
167
  This model uses the ChatML format:
 
183
  | INT8 | ~15 GB | RTX 4090, A10G |
184
  | INT4 (GPTQ/AWQ) | ~8 GB | RTX 3090, RTX 4080, Apple M-series (24GB) |
185
 
186
+ For consumer hardware, use the [GGUF versions](https://huggingface.co/ogulcanaydogan/Turkish-LLM-14B-Instruct-GGUF) for the best balance of quality and accessibility.
187
 
188
  ## Intended Use
189
 
 
219
 
220
  | Resource | Link |
221
  |----------|------|
222
+ | GGUF Versions | [Turkish-LLM-14B-Instruct-GGUF](https://huggingface.co/ogulcanaydogan/Turkish-LLM-14B-Instruct-GGUF) |
223
  | 7B Model | [Turkish-LLM-7B-Instruct](https://huggingface.co/ogulcanaydogan/Turkish-LLM-7B-Instruct) |
 
224
  | Live Demo (14B) | [Turkish-LLM-14B-Chat](https://huggingface.co/spaces/ogulcanaydogan/Turkish-LLM-14B-Chat) |
225
  | Live Demo (7B) | [Turkish-LLM-7B-Chat](https://huggingface.co/spaces/ogulcanaydogan/Turkish-LLM-7B-Chat) |
226
  | Training Pipeline | [LowResource-LLM-Forge](https://github.com/ogulcanaydogan/LowResource-LLM-Forge) |