OpceanAI commited on
Commit
9a924f0
·
verified ·
1 Parent(s): 382cbca

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +697 -0
README.md ADDED
@@ -0,0 +1,697 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ datasets:
4
+ - OpceanAI/Yuuki-dataset
5
+ language:
6
+ - en
7
+ - es
8
+ base_model:
9
+ - Qwen/Qwen2.5-3B
10
+ pipeline_tag: text-generation
11
+ library_name: transformers
12
+ tags:
13
+ - conversation
14
+ - companion
15
+ - personality
16
+ - fine-tuned
17
+ metrics:
18
+ - perplexity
19
+ widget:
20
+ - text: "Hello, how are you?"
21
+ example_title: "General Conversation"
22
+ - text: "Can you help me understand recursion?"
23
+ example_title: "Technical Explanation"
24
+ - text: "I've been feeling a bit overwhelmed lately."
25
+ example_title: "Emotional Support"
26
+ ---
27
+
28
+ <div align="center">
29
+
30
+ <br>
31
+
32
+ <img src="https://img.shields.io/badge/%E2%9C%A6-YUUKI--NxG-0D1117?style=for-the-badge&labelColor=0D1117" alt="Yuuki NxG" height="50">
33
+
34
+ <br><br>
35
+
36
+ # A 3B Companion Model Fine-Tuned on a Mac Pro
37
+
38
+ **Personality-aligned language model trained with zero cloud compute budget.**<br>
39
+ **Qwen2.5 architecture. 3 billion parameters. Mac Pro (2020). $0.00.**
40
+
41
+ <br>
42
+
43
+ <a href="#benchmark-results"><img src="https://img.shields.io/badge/BENCHMARKS-0D1117?style=for-the-badge" alt="Benchmarks"></a>
44
+ &nbsp;&nbsp;
45
+ <a href="#usage"><img src="https://img.shields.io/badge/USAGE-0D1117?style=for-the-badge" alt="Usage"></a>
46
+ &nbsp;&nbsp;
47
+ <a href="https://github.com/sponsors/aguitauwu"><img src="https://img.shields.io/badge/SPONSOR-0D1117?style=for-the-badge" alt="Sponsor"></a>
48
+
49
+ <br><br>
50
+
51
+ [![License](https://img.shields.io/badge/Apache_2.0-1a1a2e?style=flat-square&logo=opensourceinitiative&logoColor=white)](LICENSE)
52
+ &nbsp;
53
+ [![Base Model](https://img.shields.io/badge/Qwen2.5--3B-1a1a2e?style=flat-square&logo=alibabadotcom&logoColor=white)](https://huggingface.co/Qwen/Qwen2.5-3B)
54
+ &nbsp;
55
+ [![Framework](https://img.shields.io/badge/Transformers-1a1a2e?style=flat-square&logo=huggingface&logoColor=white)](https://huggingface.co/docs/transformers)
56
+ &nbsp;
57
+ [![Hardware](https://img.shields.io/badge/Mac_Pro_2020-1a1a2e?style=flat-square&logo=apple&logoColor=white)](https://www.apple.com/mac-pro/)
58
+ &nbsp;
59
+ [![Eval](https://img.shields.io/badge/lm--eval--harness-1a1a2e?style=flat-square&logo=python&logoColor=white)](https://github.com/EleutherAI/lm-evaluation-harness)
60
+
61
+ <br>
62
+
63
+ ---
64
+
65
+ <br>
66
+
67
+ </div>
68
+
69
+ ## What is Yuuki NxG?
70
+
71
+ **Yuuki NxG** is a 3-billion parameter language model fine-tuned from [Qwen2.5-3B](https://huggingface.co/Qwen/Qwen2.5-3B) for open-ended conversation, emotional support, and general-purpose reasoning. It is the flagship release of the NxG model family developed by OpceanAI.
72
+
73
+ The model was trained entirely on a **Mac Pro (2020)** with no external compute budget and no cloud GPU infrastructure. All benchmark evaluations were conducted on Kaggle P100 using [lm-evaluation-harness](https://github.com/EleutherAI/lm-evaluation-harness).
74
+
75
+ Despite being fine-tuned — which typically degrades base model benchmark scores — and evaluated strictly **0-shot** while competitors use 5–25 shot prompting, Yuuki NxG achieves the **highest TruthfulQA score** across all compared 3B-scale models, including the Qwen2.5-3B base model from which it was derived.
76
+
77
+ <br>
78
+
79
+ ---
80
+
81
+ <br>
82
+
83
+ <div align="center">
84
+
85
+ ## Model Summary
86
+
87
+ </div>
88
+
89
+ <br>
90
+
91
+ <table>
92
+ <tr>
93
+ <td width="50%" valign="top">
94
+
95
+ **Architecture**
96
+
97
+ | Property | Value |
98
+ |:---------|:------|
99
+ | Base Model | Qwen2.5-3B |
100
+ | Parameters | 3B |
101
+ | Fine-tuning | Supervised SFT |
102
+ | Training Examples | ~5,000 |
103
+ | Training Hardware | Mac Pro (2020) |
104
+ | Context Length | 32,768 tokens |
105
+
106
+ </td>
107
+ <td width="50%" valign="top">
108
+
109
+ **Release**
110
+
111
+ | Property | Value |
112
+ |:---------|:------|
113
+ | Organization | OpceanAI |
114
+ | Release Date | February 2026 |
115
+ | Languages | English, Spanish |
116
+ | License | Apache 2.0 |
117
+ | Evaluation | lm-evaluation-harness |
118
+ | Compute Budget | $0.00 |
119
+
120
+ </td>
121
+ </tr>
122
+ </table>
123
+
124
+ <br>
125
+
126
+ ---
127
+
128
+ <br>
129
+
130
+ <div align="center">
131
+
132
+ ## Benchmark Results
133
+
134
+ </div>
135
+
136
+ <br>
137
+
138
+ All Yuuki NxG results are evaluated **0-shot**. Competitor scores are sourced from their official technical reports and use few-shot prompting (5–25 shots depending on benchmark). Direct numerical comparison systematically favors base models evaluated with few-shot prompting.
139
+
140
+ <br>
141
+
142
+ ![Yuuki NxG Benchmark Evaluation](yuuki_bars_hq.png)
143
+
144
+ <br>
145
+
146
+ | Model | MMLU | ARC-C | HellaSwag | WinoGrande | TruthfulQA | Eval |
147
+ |:------|:----:|:-----:|:---------:|:----------:|:----------:|:----:|
148
+ | **Yuuki NxG** | **60.65** | 45.31 | 52.25 | 63.14 | **50.87** | 0-shot |
149
+ | Qwen2.5-3B | 65.6 | 56.5 | 74.6 | 71.1 | 48.9 | 5–25 shot |
150
+ | Llama-3.2-3B | 58.0 | 43.0 | 71.0 | 67.0 | 44.0 | 5–25 shot |
151
+ | Phi-3-mini (3.8B) | 68.8 | 60.0 | 76.7 | 73.0 | 45.0 | 5–25 shot |
152
+ | Gemma-2-2B | 52.0 | 42.0 | 71.0 | 65.0 | 39.0 | 5–25 shot |
153
+
154
+ <br>
155
+
156
+ Yuuki NxG achieves the highest TruthfulQA score across all compared models under equivalent 0-shot conditions, including the base model from which it was fine-tuned. This indicates that alignment fine-tuning improved factual honesty rather than degrading it — an outcome that runs counter to the typical fine-tuning tradeoff.
157
+
158
+ HellaSwag degradation is expected and well-documented in personality-aligned models, as sentence-completion benchmarks are sensitive to conversational fine-tuning.
159
+
160
+ <br>
161
+
162
+ ### MMLU Category Breakdown
163
+
164
+ <table>
165
+ <tr>
166
+ <td width="50%" valign="top">
167
+
168
+ **Strongest Domains**
169
+
170
+ | Category | Score |
171
+ |:---------|:-----:|
172
+ | Marketing | 87.18% |
173
+ | High School Psychology | 83.67% |
174
+ | Sociology | 80.60% |
175
+ | World Religions | 80.12% |
176
+ | US Foreign Policy | 79.00% |
177
+ | Logical Fallacies | 76.69% |
178
+ | HS Computer Science | 76.00% |
179
+
180
+ </td>
181
+ <td width="50%" valign="top">
182
+
183
+ **Domain Averages**
184
+
185
+ | Domain | Score |
186
+ |:-------|:-----:|
187
+ | Social Sciences | 71.56% |
188
+ | Other | 66.08% |
189
+ | STEM | 56.17% |
190
+ | Humanities | 52.92% |
191
+ | **Overall** | **60.65%** |
192
+
193
+ </td>
194
+ </tr>
195
+ </table>
196
+
197
+ The performance profile is consistent with a model optimized for conversation: strong in social sciences, psychology, and humanities; below average in formal STEM domains. This is the expected and intended tradeoff for a companion-purpose model.
198
+
199
+ <br>
200
+
201
+ ---
202
+
203
+ <br>
204
+
205
+ <div align="center">
206
+
207
+ ## NxG Model Family
208
+
209
+ </div>
210
+
211
+ <br>
212
+
213
+ <table>
214
+ <tr>
215
+ <td width="50%" valign="top">
216
+
217
+ **Released Models**
218
+
219
+ | Model | Parameters | Description |
220
+ |:------|:----------:|:------------|
221
+ | Yuuki NxG | 3B | Full model, general conversation |
222
+ | Yuuki NxG Nano | 81M | Lightweight, constrained environments |
223
+
224
+ </td>
225
+ <td width="50%" valign="top">
226
+
227
+ **Community GGUF (via mradermacher)**
228
+
229
+ Quantized independently without solicitation — organic community adoption prior to any formal announcement.
230
+
231
+ | Format | Size |
232
+ |:-------|:----:|
233
+ | Q4_K_M | 63.3 MB |
234
+ | Q8_0 | 91.3 MB |
235
+ | F16 | 167 MB |
236
+
237
+ Available at [mradermacher/Yuuki-NxG-nano-GGUF](https://huggingface.co/mradermacher/Yuuki-NxG-nano-GGUF).
238
+
239
+ </td>
240
+ </tr>
241
+ </table>
242
+
243
+ <br>
244
+
245
+ ---
246
+
247
+ <br>
248
+
249
+ <div align="center">
250
+
251
+ ## Usage
252
+
253
+ </div>
254
+
255
+ <br>
256
+
257
+ ### With Transformers (PyTorch)
258
+
259
+ ```python
260
+ from transformers import AutoTokenizer, AutoModelForCausalLM
261
+ import torch
262
+
263
+ model_id = "OpceanAI/Yuuki-NxG"
264
+
265
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
266
+ model = AutoModelForCausalLM.from_pretrained(
267
+ model_id,
268
+ torch_dtype=torch.bfloat16,
269
+ device_map="auto"
270
+ )
271
+
272
+ messages = [
273
+ {"role": "user", "content": "Hello, how are you?"}
274
+ ]
275
+
276
+ inputs = tokenizer.apply_chat_template(
277
+ messages,
278
+ return_tensors="pt"
279
+ ).to(model.device)
280
+
281
+ with torch.no_grad():
282
+ outputs = model.generate(
283
+ inputs,
284
+ max_new_tokens=512,
285
+ temperature=0.7,
286
+ do_sample=True,
287
+ repetition_penalty=1.1
288
+ )
289
+
290
+ print(tokenizer.decode(outputs[0][inputs.shape[1]:], skip_special_tokens=True))
291
+ ```
292
+
293
+ <br>
294
+
295
+ ### With llama.cpp (GGUF)
296
+
297
+ ```bash
298
+ ./llama.cpp/main -m yuuki-nxg-q4_k_m.gguf \
299
+ -p "Hello, how are you?" \
300
+ -n 256 \
301
+ -t 4 \
302
+ --temp 0.7 \
303
+ --repeat-penalty 1.1
304
+ ```
305
+
306
+ <br>
307
+
308
+ ### With Ollama
309
+
310
+ ```bash
311
+ cat > Modelfile << EOF
312
+ FROM ./yuuki-nxg-q4_k_m.gguf
313
+
314
+ PARAMETER temperature 0.7
315
+ PARAMETER top_p 0.9
316
+ PARAMETER repeat_penalty 1.1
317
+ EOF
318
+
319
+ ollama create yuuki-nxg -f Modelfile
320
+ ollama run yuuki-nxg "Hello, how are you?"
321
+ ```
322
+
323
+ <br>
324
+
325
+ ### Recommended Parameters
326
+
327
+ | Parameter | Value |
328
+ |:----------|:-----:|
329
+ | Temperature | 0.7 |
330
+ | Top-p | 0.9 |
331
+ | Max new tokens | 512–2048 |
332
+ | Repetition penalty | 1.1 |
333
+
334
+ <br>
335
+
336
+ ---
337
+
338
+ <br>
339
+
340
+ <div align="center">
341
+
342
+ ## Training Details
343
+
344
+ </div>
345
+
346
+ <br>
347
+
348
+ <table>
349
+ <tr>
350
+ <td width="50%" valign="top">
351
+
352
+ **Hardware**
353
+
354
+ | Component | Specification |
355
+ |:----------|:-------------|
356
+ | Device | Mac Pro (2020) |
357
+ | Chip | Intel Xeon W |
358
+ | RAM | 48 GB ECC |
359
+ | GPU | AMD Radeon Pro W5700X |
360
+ | Cloud Compute | None |
361
+ | Cost | $0.00 |
362
+
363
+ </td>
364
+ <td width="50%" valign="top">
365
+
366
+ **Training Configuration**
367
+
368
+ | Parameter | Value |
369
+ |:----------|:-----:|
370
+ | Base Model | Qwen2.5-3B |
371
+ | Method | Supervised Fine-Tuning |
372
+ | Training Examples | ~5,000 |
373
+ | Optimizer | AdamW |
374
+ | Learning Rate | 2e-5 |
375
+ | Max Sequence Length | 2,048 tokens |
376
+
377
+ </td>
378
+ </tr>
379
+ </table>
380
+
381
+ <br>
382
+
383
+ Yuuki NxG was produced through supervised fine-tuning on a curated conversational dataset. The training objective was to produce a model with consistent personality, high factual honesty, and broad general-knowledge retention from the Qwen2.5 base.
384
+
385
+ Training without GPU-accelerated cloud infrastructure imposes constraints on batch size and total training duration relative to commercially produced models. The resulting benchmark profile reflects these constraints: strong performance in domains well-represented in the training data, with expected degradation in areas requiring dense technical knowledge such as formal mathematics and physics.
386
+
387
+ <br>
388
+
389
+ ---
390
+
391
+ <br>
392
+
393
+ <div align="center">
394
+
395
+ ## Features
396
+
397
+ </div>
398
+
399
+ <br>
400
+
401
+ <table>
402
+ <tr>
403
+ <td width="50%" valign="top">
404
+
405
+ **Personality Alignment**
406
+
407
+ Fine-tuned for consistent, context-aware conversation. The model maintains a coherent identity across extended dialogues, with particular strength in emotional support and casual Q&A.
408
+
409
+ <br>
410
+
411
+ **Factual Honesty**
412
+
413
+ Achieves highest TruthfulQA score (50.87%) among all compared 3B-scale models — including its own base model. Fine-tuning improved factual calibration rather than degrading it.
414
+
415
+ <br>
416
+
417
+ **Multilingual**
418
+
419
+ Functional in both English and Spanish. Primary evaluation in English; Spanish capability inherited from Qwen2.5 pretraining.
420
+
421
+ </td>
422
+ <td width="50%" valign="top">
423
+
424
+ **Zero-Budget Training**
425
+
426
+ Trained entirely on owned hardware with no cloud compute expenditure. Demonstrates that meaningful alignment fine-tuning is accessible without data center infrastructure.
427
+
428
+ <br>
429
+
430
+ **Community Adoption**
431
+
432
+ Independently quantized and distributed by mradermacher before any formal announcement — organic community interest in the model's capabilities.
433
+
434
+ <br>
435
+
436
+ **Open Source**
437
+
438
+ Apache 2.0. Use commercially, modify, distribute. Full transparency on training methodology and evaluation protocol.
439
+
440
+ </td>
441
+ </tr>
442
+ </table>
443
+
444
+ <br>
445
+
446
+ ---
447
+
448
+ <br>
449
+
450
+ <div align="center">
451
+
452
+ ## Limitations
453
+
454
+ </div>
455
+
456
+ <br>
457
+
458
+ - **Mathematical reasoning** performance is below the Qwen2.5-3B base. Users requiring quantitative precision should use tool augmentation or a specialized model.
459
+ - **HellaSwag degradation** reflects the standard tradeoff of personality fine-tuning on sentence-completion benchmarks.
460
+ - **Benchmark methodology**: Yuuki NxG is evaluated 0-shot while competitor reports use 5–25 shot prompting, creating a systematic disadvantage in direct comparisons.
461
+ - **Safety alignment** has not been formally evaluated. Not recommended for adversarial or high-stakes deployment without additional safety filtering.
462
+ - **Training scale**: 5,000 examples on consumer hardware impose generalization limits relative to commercially scaled models.
463
+
464
+ <br>
465
+
466
+ ---
467
+
468
+ <br>
469
+
470
+ <div align="center">
471
+
472
+ ## Intended Use
473
+
474
+ </div>
475
+
476
+ <br>
477
+
478
+ <table>
479
+ <tr>
480
+ <td width="50%" valign="top">
481
+
482
+ **Intended For**
483
+
484
+ - General-purpose conversational assistance
485
+ - Emotional support and companionship applications
486
+ - Educational Q&A in humanities and social sciences
487
+ - Research into small-scale fine-tuning and personality alignment
488
+ - Local deployment on consumer hardware
489
+
490
+ </td>
491
+ <td width="50%" valign="top">
492
+
493
+ **Not Intended For**
494
+
495
+ - Medical, legal, or financial advice
496
+ - Tasks requiring high-precision mathematical reasoning
497
+ - Applications requiring certified safety alignment
498
+ - Production systems without additional safety review
499
+
500
+ </td>
501
+ </tr>
502
+ </table>
503
+
504
+ <br>
505
+
506
+ ---
507
+
508
+ <br>
509
+
510
+ <div align="center">
511
+
512
+ ## Philosophy
513
+
514
+ </div>
515
+
516
+ <br>
517
+
518
+ > **"Meaningful AI development does not require a data center. It requires patience, clarity of purpose, and time."**
519
+
520
+ Yuuki NxG was built to demonstrate that a fine-tuned 3B model trained by one person on owned hardware can compete with base models from large organizations on key benchmarks — and surpass them where it matters most.
521
+
522
+ <br>
523
+
524
+ ---
525
+
526
+ <br>
527
+
528
+ <div align="center">
529
+
530
+ ## Related Projects
531
+
532
+ </div>
533
+
534
+ <br>
535
+
536
+ | Project | Description |
537
+ |:--------|:------------|
538
+ | [Yuuki-NxG-Nano](https://huggingface.co/OpceanAI/Yuuki-NxG-Nano) | 81M lightweight variant |
539
+ | [Yuuki-3.7](https://huggingface.co/OpceanAI/Yuuki-3.7) | Earlier code generation checkpoint |
540
+ | [Yuuki-best](https://huggingface.co/OpceanAI/Yuuki-best) | Best checkpoint of the v0.1 series |
541
+ | [yuy](https://github.com/YuuKi-OS/yuy) | CLI for managing and running Yuuki models |
542
+ | [yuy-chat](https://github.com/YuuKi-OS/yuy-chat) | TUI chat interface |
543
+ | [Yuuki-chat](https://github.com/YuuKi-OS/Yuuki-chat) | Web-based chat interface |
544
+ | [Yuuki Space](https://huggingface.co/spaces/OpceanAI/Yuuki) | Interactive demo |
545
+
546
+ <br>
547
+
548
+ ---
549
+
550
+ <br>
551
+
552
+ <div align="center">
553
+
554
+ ## Links
555
+
556
+ </div>
557
+
558
+ <br>
559
+
560
+ <div align="center">
561
+
562
+ [![Model Weights](https://img.shields.io/badge/Model_Weights-Hugging_Face-ffd21e?style=for-the-badge&logo=huggingface&logoColor=black)](https://huggingface.co/OpceanAI/Yuuki-NxG)
563
+ &nbsp;
564
+ [![Live Demo](https://img.shields.io/badge/Live_Demo-Spaces-ffd21e?style=for-the-badge&logo=huggingface&logoColor=black)](https://huggingface.co/spaces/OpceanAI/Yuuki)
565
+ &nbsp;
566
+ [![GGUF](https://img.shields.io/badge/GGUF_Quants-mradermacher-181717?style=for-the-badge&logo=github&logoColor=white)](https://huggingface.co/mradermacher/Yuuki-NxG-nano-GGUF)
567
+
568
+ <br>
569
+
570
+ [![YUY CLI](https://img.shields.io/badge/Yuy_CLI-GitHub-181717?style=for-the-badge&logo=github&logoColor=white)](https://github.com/YuuKi-OS/yuy)
571
+ &nbsp;
572
+ [![Sponsor](https://img.shields.io/badge/Sponsor-GitHub_Sponsors-ea4aaa?style=for-the-badge&logo=githubsponsors&logoColor=white)](https://github.com/sponsors/aguitauwu)
573
+ &nbsp;
574
+ [![Discord](https://img.shields.io/badge/Discord-Community-5865F2?style=for-the-badge&logo=discord&logoColor=white)](https://discord.gg/j8zV2u8k)
575
+
576
+ </div>
577
+
578
+ <br>
579
+
580
+ ---
581
+
582
+ <br>
583
+
584
+ <div align="center">
585
+
586
+ ## Community
587
+
588
+ </div>
589
+
590
+ <br>
591
+
592
+ - [Discord Server](https://discord.gg/j8zV2u8k) — Development discussion and user community
593
+ - [Twitter](https://twitter.com/aguitauwu) — Updates and announcements
594
+ - [GitHub](https://github.com/aguitauwu) — Source code and training scripts
595
+ - [GitHub Sponsors](https://github.com/sponsors/aguitauwu) — Support the project
596
+ - [Ollama](https://ollama.com/aguitachan3/yuuki-nxg) — Run locally with Ollama
597
+
598
+ <br>
599
+
600
+ ---
601
+
602
+ <br>
603
+
604
+ <div align="center">
605
+
606
+ ## Citation
607
+
608
+ </div>
609
+
610
+ <br>
611
+
612
+ ```bibtex
613
+ @misc{opceanai2026yuukinxg,
614
+ title = {Yuuki NxG: A Fine-Tuned 3B Companion Language Model},
615
+ author = {OpceanAI},
616
+ year = {2026},
617
+ month = {February},
618
+ howpublished = {\url{https://huggingface.co/OpceanAI/Yuuki-NxG}},
619
+ }
620
+ ```
621
+
622
+ <br>
623
+
624
+ ---
625
+
626
+ <br>
627
+
628
+ <div align="center">
629
+
630
+ ## License
631
+
632
+ </div>
633
+
634
+ <br>
635
+
636
+ ```
637
+ Apache License 2.0
638
+
639
+ Copyright (c) 2026 OpceanAI
640
+
641
+ Licensed under the Apache License, Version 2.0 (the "License");
642
+ you may not use this file except in compliance with the License.
643
+ You may obtain a copy of the License at
644
+
645
+ http://www.apache.org/licenses/LICENSE-2.0
646
+
647
+ Unless required by applicable law or agreed to in writing, software
648
+ distributed under the License is distributed on an "AS IS" BASIS,
649
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
650
+ See the License for the specific language governing permissions and
651
+ limitations under the License.
652
+ ```
653
+
654
+ Use commercially, modify, distribute. Attribution required.
655
+
656
+ <br>
657
+
658
+ ---
659
+
660
+ <br>
661
+
662
+ <div align="center">
663
+
664
+ ## Updates
665
+
666
+ </div>
667
+
668
+ <br>
669
+
670
+ | Date | Milestone |
671
+ |:-----|:----------|
672
+ | **2026-02-27** | Benchmark evaluation completed (Kaggle P100) |
673
+ | **2026-02-27** | TruthfulQA: 50.87% — best among all compared 3B models |
674
+ | **2026-02-27** | Community GGUF quantization by mradermacher |
675
+ | **2026-02-27** | Yuuki NxG released on HuggingFace |
676
+
677
+ **Last updated:** 2026-02-27
678
+
679
+ <br>
680
+
681
+ ---
682
+
683
+ <br>
684
+
685
+ <div align="center">
686
+
687
+ **Built on a Mac Pro. Trained on 5,000 examples. Competitive with models from teams of hundreds.**
688
+
689
+ <br>
690
+
691
+ [![OpceanAI](https://img.shields.io/badge/OpceanAI-2026-0D1117?style=for-the-badge)](https://huggingface.co/OpceanAI)
692
+
693
+ <br>
694
+
695
+ *The NxG family. More releases coming.*
696
+
697
+ </div>