File size: 18,112 Bytes
6b1435e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
---
license: other
license_name: aigency-commercial
license_link: https://aigency.dev/license
language:
- tr
- en
library_name: aigency-api
pipeline_tag: text-generation
tags:
- turkish
- multimodal
- sovereign
- frontier-adjacent
- aigency
- ecloud
- production
inference: false
extra_gated_heading: AIGENCY V4 is offered via API
extra_gated_description: |
  Model weights are not distributed on HuggingFace. AIGENCY V4 is accessible
  via the eCloud production API at https://aigency.dev. This page is a
  reference card describing architecture, evaluation methodology, and
  benchmark results, and links to a live demo Space.
model-index:
- name: AIGENCY V4
  results:
  - task:
      type: text-generation
      name: Code generation
    dataset:
      type: openai_humaneval
      name: HumanEval (pass@1)
    metrics:
    - type: pass@1
      value: 84.15
      name: pass@1
      verified: false
  - task:
      type: text-generation
      name: Code generation extended
    dataset:
      type: humaneval-plus
      name: HumanEval+ (pass@1)
    metrics:
    - type: pass@1
      value: 79.88
      name: pass@1
      verified: false
  - task:
      type: text-generation
      name: Code generation
    dataset:
      type: mbpp
      name: MBPP (sanitized)
    metrics:
    - type: pass@1
      value: 84.82
      name: pass@1
      verified: false
  - task:
      type: text-generation
      name: Code generation extended
    dataset:
      type: mbpp-plus
      name: MBPP+
    metrics:
    - type: pass@1
      value: 78.04
      name: pass@1
      verified: false
  - task:
      type: text-generation
      name: Mathematical reasoning
    dataset:
      type: gsm8k
      name: GSM8K
    metrics:
    - type: accuracy
      value: 94.62
      name: accuracy
      verified: false
  - task:
      type: text-generation
      name: Multitask language understanding
    dataset:
      type: cais/mmlu
      name: MMLU (stratified n=1000)
    metrics:
    - type: accuracy
      value: 80.10
      name: accuracy
      verified: false
  - task:
      type: text-generation
      name: Multitask language understanding (Pro)
    dataset:
      type: TIGER-Lab/MMLU-Pro
      name: MMLU-Pro (n=1000)
    metrics:
    - type: accuracy
      value: 50.20
      name: accuracy
      verified: false
  - task:
      type: text-generation
      name: Scientific reasoning
    dataset:
      type: ai2_arc
      name: ARC-Challenge
    metrics:
    - type: accuracy
      value: 94.88
      name: accuracy
      verified: false
  - task:
      type: text-generation
      name: Graduate-level QA
    dataset:
      type: idavidrein/gpqa
      name: GPQA Diamond
    metrics:
    - type: accuracy
      value: 37.88
      name: accuracy
      verified: false
  - task:
      type: text-generation
      name: Truthfulness
    dataset:
      type: truthful_qa
      name: TruthfulQA MC1
    metrics:
    - type: accuracy
      value: 76.38
      name: accuracy
      verified: false
  - task:
      type: text-generation
      name: Instruction following
    dataset:
      type: google/IFEval
      name: IFEval (strict)
    metrics:
    - type: accuracy
      value: 80.22
      name: strict-prompt-level
      verified: false
  - task:
      type: text-generation
      name: Commonsense reasoning
    dataset:
      type: hellaswag
      name: HellaSwag (n=1000)
    metrics:
    - type: accuracy
      value: 88.60
      name: accuracy
      verified: false
  - task:
      type: text-generation
      name: Coreference reasoning
    dataset:
      type: winogrande
      name: WinoGrande XL
    metrics:
    - type: accuracy
      value: 74.66
      name: accuracy
      verified: false
  - task:
      type: text-generation
      name: Turkish reading comprehension
    dataset:
      type: facebook/belebele
      name: Belebele-TR (Turkish)
    metrics:
    - type: accuracy
      value: 87.33
      name: accuracy
      verified: false
  - task:
      type: text-generation
      name: Turkish extractive QA
    dataset:
      type: tquad
      name: TQuAD (F1  0.5)
    metrics:
    - type: f1
      value: 82.40
      name: F1  0.5
      verified: false
  - task:
      type: text-generation
      name: Turkish multitask understanding
    dataset:
      type: tr-mmlu
      name: TR-MMLU
    metrics:
    - type: accuracy
      value: 70.80
      name: accuracy
      verified: false
  - task:
      type: text-generation
      name: Turkish natural-language inference
    dataset:
      type: xnli
      name: XNLI-TR
    metrics:
    - type: accuracy
      value: 73.40
      name: accuracy
      verified: false
  - task:
      type: text-generation
      name: Turkish grammar
    dataset:
      type: tr-grammar-synthetic
      name: TR Grammar (synthetic 50/50)
    metrics:
    - type: accuracy
      value: 79.00
      name: accuracy
      verified: false
  - task:
      type: image-text-to-text
      name: Multimodal QA
    dataset:
      type: MMMU
      name: MMMU (val, n=30)
    metrics:
    - type: accuracy
      value: 53.33
      name: accuracy
      verified: false
  - task:
      type: image-text-to-text
      name: Chart QA
    dataset:
      type: HuggingFaceM4/ChartQA
      name: ChartQA (relaxed)
    metrics:
    - type: accuracy
      value: 67.68
      name: relaxed accuracy
      verified: false
  - task:
      type: image-text-to-text
      name: Document QA
    dataset:
      type: lmms-lab/DocVQA
      name: DocVQA (ANLS  0.5)
    metrics:
    - type: accuracy
      value: 79.17
      name: ANLS  0.5
      verified: false
  - task:
      type: image-text-to-text
      name: Visual mathematical reasoning
    dataset:
      type: AI4Math/MathVista
      name: MathVista (testmini)
    metrics:
    - type: accuracy
      value: 34.13
      name: accuracy
      verified: false
---

# AIGENCY V4

> **Sovereign, fully independent, multimodal — 128B parameters.**
> A globally competitive Turkish-first AI model: world-leading on Turkish
> reading comprehension and natural-language inference, frontier-level on
> grade-school math and scientific reasoning, KVKK-resident.

[**🇹🇷 Türkçe README**](#türkçe) · [**🇬🇧 English README**](#english) · [**📄 Whitepaper (EN)**](https://github.com/ecloud-bh/aigency-v4-whitepaper/blob/main/AIGENCY-V4-Whitepaper-EN.pdf) · [**📄 Whitepaper (TR)**](https://github.com/ecloud-bh/aigency-v4-whitepaper/blob/main/AIGENCY-V4-Whitepaper-TR.pdf) · [**🌐 Try the demo**](https://huggingface.co/spaces/aigencydev/AIGENCY-V4-Demo) · [**🔗 API**](https://aigency.dev)

---

## English

### Model summary

**AIGENCY V4** is the multimodal successor to AIGENCY V3, developed by
**eCloud Yazılım Teknolojileri** and released to production in Q2 2026.
The model retains V3's four sovereignty principles — zero external parameter
dependency, sovereign data residency, transparent architectural documentation,
and Turkish morphological context fidelity — and adds a sovereign 8B-parameter
vision encoder for image, document, chart, and visual-math understanding.

| | |
|---|---|
| **Total parameters** | 128B (120B core + 8B vision encoder) |
| **Architecture** | Sovereign decoder-only transformer + side vision encoder |
| **Optimisations** | Adaptive LoRA+, Selective Layer Collapse, Localised MoE, 4-bit block quantization, chunked attention |
| **Context window** | 278K tokens (HBM 3-tier: STM 4k / ITM 64k / LTM 278k) |
| **Active inference memory** | ~6.5 GB GPU under 4-bit quant |
| **Languages** | Turkish (primary), English |
| **Modalities** | Text, image (one image per request, 30 MB max, image/* MIME) |
| **Release version** | 1.0 production |
| **Release date** | April 2026 |
| **Licence** | API-only commercial — see https://aigency.dev/license |

### Distribution

**Weights are not distributed.** AIGENCY V4 is accessed exclusively through
the eCloud production API at `https://aigency.dev/api/v2`. This page provides
the architectural specification, the evaluation methodology, and the full
benchmark results. To try the model interactively, use the
[demo Space](https://huggingface.co/spaces/aigencydev/AIGENCY-V4-Demo). For
production access, see [aigency.dev](https://aigency.dev).

### Evaluation

A comprehensive single-session evaluation was conducted on **27 April 2026**
against the production API. **13,344 real API calls** across **22 distinct
benchmarks** were executed; every result is reported with a Wilson 95%
confidence interval, deterministic subsampling (seed=42), and an open dataset
identifier.

#### Tier 1 — Critical benchmarks (full set)

| Benchmark | Accuracy | Wilson 95% CI | n | Errors |
|---|---|---|---|---|
| HumanEval (pass@1) | **0.8415** | [0.778, 0.889] | 164/164 | 0 |
| IFEval (strict) | **0.8022** | [0.767, 0.834] | 541/541 | 1 |
| GPQA Diamond | 0.3788 | [0.314, 0.448] | 198/198 | 0 |
| Belebele-TR | **0.8733** | [0.850, 0.893] | 900/900 | 0 |
| ARC-Challenge | **0.9488** | [0.935, 0.960] | 1172/1172 | 0 |
| TruthfulQA MC1 | **0.7638** | [0.734, 0.792] | 817/817 | 0 |
| GSM8K | **0.9462** | [0.933, 0.957] | 1319/1319 | 0 |

#### Tier 2 — Mid-volume

| Benchmark | Accuracy | Wilson 95% CI | n |
|---|---|---|---|
| MMLU (stratified) | **0.8010** | [0.775, 0.825] | 1000/1000 |
| MMLU-Pro | 0.5020 | [0.471, 0.533] | 1000/1000 |
| HellaSwag | **0.8860** | [0.865, 0.904] | 1000/1000 |
| WinoGrande XL | 0.7466 | [0.722, 0.770] | 1267/1267 |
| HumanEval+ (extended) | **0.7988** | [0.731, 0.853] | 164/164 |
| MBPP (sanitized) | **0.8482** | [0.799, 0.887] | 257/257 |
| MBPP+ | **0.7804** | [0.736, 0.819] | 378/378 |

#### Tier 3-A — Turkish (V4 is the de-facto global reference)

| Benchmark | Accuracy | Wilson 95% CI | n |
|---|---|---|---|
| Belebele-TR | **0.8733** | [0.850, 0.893] | 900/900 |
| TQuAD (F1 ≥ 0.5) | **0.8240** | [0.788, 0.855] | 500/500 |
| TR-MMLU | **0.7080** | [0.667, 0.746] | 500/500 |
| XNLI-TR | **0.7340** | [0.694, 0.771] | 500/500 |
| TR Grammar (synthetic) | **0.7900** | [0.700, 0.858] | 100/100 |

> Frontier models do not consistently publish Turkish-specific scores.
> Within published global evaluation, AIGENCY V4 is the **Turkish reference**.

#### Tier 3-B — Multimodal (first production release)

| Benchmark | Accuracy | Wilson 95% CI | n |
|---|---|---|---|
| MMMU (val) | 0.5333 | [0.361, 0.698] | 30/30 |
| ChartQA (relaxed) | 0.6768 | [0.634, 0.717] | 492/500 |
| DocVQA (ANLS ≥ 0.5) | 0.7917 | [0.595, 0.908] | 24 |
| MathVista (testmini) | 0.3413 | [0.280, 0.408] | 208 |

### Comparison with frontier (April 2026)

| Benchmark | AIGENCY V4 | GPT-5 | Claude 4.6/4.7 | Gemini 3 Pro |
|---|---|---|---|---|
| GSM8K | **94.62** | 96.8 | ~96 | ~94 |
| ARC-Challenge | **94.88** | ~96 | ~96 | ~95 |
| HumanEval | 84.15 | 94.0 | 95.0 | 89.7 |
| MMLU | 80.10 | 94.2 | 88-93 | 92.4 |
| MMLU-Pro | 50.20 | ~85 | ~84 | ~81 |
| GPQA Diamond | 37.88 | 88-94 | 91.3-94.2 | 91.9 |
| MMMU | 53.33 | 79.1 | 84.1 | — |

V4 is **at frontier level on grade-school math and scientific reasoning**,
**upper-mid frontier on code generation**, **lower-mid frontier on general
academic and instruction following**, and **in active development on
graduate-level expert knowledge and multimodal**. The V4.1 roadmap (Q4 2026)
targets MMLU-Pro 0.65, GPQA Diamond 0.55, and average latency 4 s.

### Operational performance (single-session, 27 April 2026)

- Total API calls: 13,344
- Persistent error rate: 0.3%
- Average latency: 9.55 s · p50 4.39 s · p95 32.77 s · p99 33.59 s
- V4.1 latency target: average ≤ 4 s · p95 ≤ 15 s

### Reproducibility

Full evaluation harness, raw responses, scored items, summary JSON, and the
deterministic subsample seed are available at:

- **Benchmark code**: https://github.com/ecloud-bh/aigency-benchmarks
- **Evaluation results dataset**: https://huggingface.co/datasets/aigencydev/aigency-v4-evaluation
- **Whitepaper (EN/TR)**: https://github.com/ecloud-bh/aigency-v4-whitepaper

### Intended use

**Primary deployment domains:**

1. Public-sector and government workloads requiring KVKK residency
2. Legal and legal-tech (statute search, contract analysis — Tural model integration)
3. Education and higher education (Turkish academic, exam prep, course assistants)
4. Banking, finance and insurance (Turkish-heavy KYC/AML)
5. Healthcare administrative workloads (KVKK-compliant document handling)
6. Media, publishing and editorial (Turkish grammar precision)
7. Defence and critical infrastructure (sovereign architecture)
8. Software, R&D and engineering (code generation, large-codebase analysis)

**Out-of-scope or non-recommended:**

- Clinical diagnosis or medical advice (administrative use only)
- Autonomous critical decisions without human review
- Graduate-level scientific research where GPQA-Diamond–class accuracy is required (use frontier model + V4 hybrid)
- High-fidelity multimodal reasoning where MMMU > 75 is required (await V4.1)

### Safety and compliance

- KVKK §5 / §12 (Turkish PDPA) compliant — KVKK-resident hosting (TR DC)
- ISO/IEC 27001 — IT-ISMS, risk and control matrix
- NIST SP 800-207 (Zero-Trust) — mTLS, least privilege, continuous monitoring
- EU AI Act (ratified 2025) — high-risk classification with model card
- Memory encryption: AES-256-XTS (RAM), ChaCha20-Poly1305 (LTM disk)
- Image cache: AES-256-GCM, 30 MB limit, 24h TTL
- Pre-encoding visual safety filter + post-encoding output check

### Known limitations

1. **GPQA Diamond / MMLU-Pro gap** — 35-50pp behind frontier; graduate-level expert knowledge is a V4.1 target.
2. **First-generation multimodal** — vision encoder is 8B; V4.1 plans to scale to 16B.
3. **Latency 2-3× frontier** — vision-encoder overhead, multimodal safety filter; V4.1 targets ≤ 4 s avg.
4. **Multimodal subsample size** — DocVQA n=24, MMMU n=30 (HF cache constraints); CIs are wide.
5. **Multilingual non-TR evaluation not published** — global-scale claim is currently Turkish-anchored.

### Citation

```bibtex
@techreport{aigency-v4-2026,
  title  = {AIGENCY V4: Sovereign, Fully Independent and Multimodal 128B-Parameter AI Architecture},
  author = {{eCloud Yaz{\i}l{\i}m Teknolojileri}},
  year   = {2026},
  month  = apr,
  institution = {eCloud Yaz{\i}l{\i}m Teknolojileri},
  url    = {https://github.com/ecloud-bh/aigency-v4-whitepaper},
  note   = {Whitepaper v1.0, April 2026}
}
```

---

## Türkçe

### Model özeti

**AIGENCY V4**, eCloud Yazılım Teknolojileri tarafından geliştirilen, V3'ün
multimodal halefi olan 128 milyar parametreli yerli yapay zekâ modelidir.
2026/Q2'de üretime alındı. V3'ün dört bağımsızlık ilkesini (dış parametre
sıfırlama, yerel veri egemenliği, şeffaf belgeleme, Türkçe bağlam uyumu)
korur ve görsel anlama, belge soru-cevap, grafik yorumlama, görsel matematik
yetkinliklerini ekleyen 8B parametreli yerli vision encoder ile genişletir.

| | |
|---|---|
| **Toplam parametre** | 128B (120B çekirdek + 8B vision encoder) |
| **Mimari** | Yerli decoder-only transformer + yan vision encoder |
| **Optimizasyonlar** | Adaptif LoRA+, Selective Layer Collapse, L-MoE, 4-bit blok kuantizasyon, öbekli dikkat |
| **Bağlam penceresi** | 278K token (HBM 3-katmanlı: STM 4k / ITM 64k / LTM 278k) |
| **Aktif inferans bellek** | 4-bit kuantizasyon altında ~6.5 GB GPU |
| **Diller** | Türkçe (birincil), İngilizce |
| **Modaliteler** | Metin, görsel (istek başına bir görsel, max 30 MB, image/* MIME) |
| **Sürüm** | 1.0 üretim |
| **Yayın tarihi** | Nisan 2026 |
| **Lisans** | API-only ticari — https://aigency.dev/license |

### Dağıtım

**Ağırlıklar HuggingFace'de paylaşılmaz.** AIGENCY V4'e erişim yalnızca
`https://aigency.dev/api/v2` üzerinden sağlanır. Bu sayfa mimari
spesifikasyonu, değerlendirme metodolojisini ve tam benchmark sonuçlarını
sunar. Modeli interaktif olarak denemek için
[demo Space](https://huggingface.co/spaces/aigencydev/AIGENCY-V4-Demo)
sayfasını kullanın. Üretim erişimi için: [aigency.dev](https://aigency.dev).

### Konumlandırma — Tek cümlede

AIGENCY V4, Türkçe okuma anlama ve doğal dil çıkarımında dünya lideri,
fen muhakemesi ve grade-school matematikte küresel frontier seviyesinde,
kod üretiminde üst-orta frontier segmentinde, multimodal ve graduate-level
uzman bilgide aktif geliştirme aşamasında, tam-bağımsız ve KVKK-yerel bir
yerli yapay zekâ modelidir.

### Hedef kullanım alanları

1. Kamu sektörü ve devlet kurumları (KVKK gereksinimi)
2. Hukuk ve hukuk teknolojileri (mevzuat arama, sözleşme analizi)
3. Eğitim ve yükseköğretim (Türkçe akademik, sınav hazırlık)
4. Bankacılık, finans ve sigorta (Türkçe-yoğun KYC/AML)
5. Sağlık idari iş yükleri (KVKK uyumlu belge işleme)
6. Medya, yayıncılık ve editoryal (Türkçe dilbilgisi titizliği)
7. Savunma ve kritik altyapı (egemen mimari)
8. Yazılım, AR-GE ve mühendislik

### Bilinen kısıtlar

1. GPQA Diamond / MMLU-Pro frontier'ın 35-50pp gerisinde — V4.1 hedefi.
2. Multimodal ilk üretim sürümü — V4.1'de 16B vision encoder planlandı.
3. Latency frontier'ın 2-3 katı — V4.1 hedefi ≤ 4 s ortalama.
4. Multimodal subsample boyutu küçük (DocVQA n=24, MMMU n=30); CI geniş.
5. TR-dışı çok-dilli profil yayımlanmadı — küresel iddia şu an TR-merkezli.

### Atıf

```bibtex
@techreport{aigency-v4-2026,
  title  = {AIGENCY V4: Yerli, Tam Ba{\u g}{\i}ms{\i}z ve Multimodal 128B Parametreli Yapay Zek\^a Mimarisi},
  author = {{eCloud Yaz{\i}l{\i}m Teknolojileri}},
  year   = {2026},
  month  = apr,
  institution = {eCloud Yaz{\i}l{\i}m Teknolojileri},
  url    = {https://github.com/ecloud-bh/aigency-v4-whitepaper}
}
```

---

## License

AIGENCY V4 is offered under the **eCloud AIGENCY Commercial Licence** (API-only).
Model weights are not redistributed. The accompanying whitepaper is licensed
under **CC BY-ND 4.0**, and the benchmark code is licensed under **MIT**.

For commercial use, partnership, or research collaboration:
**info@e-cloud.web.tr · ai@aigency.dev** · https://aigency.dev

© 2026 eCloud Yazılım Teknolojileri.