File size: 15,673 Bytes
ac9b283
391573b
 
 
 
 
 
 
 
 
 
 
 
df03cbf
 
ac9b283
df03cbf
 
ac9b283
df03cbf
 
ac9b283
df03cbf
 
 
 
 
ac9b283
df03cbf
 
 
 
ac9b283
df03cbf
 
ac9b283
df03cbf
 
ac9b283
391573b
ac9b283
df03cbf
 
 
 
ac9b283
391573b
ac9b283
df03cbf
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ac9b283
391573b
ac9b283
df03cbf
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
391573b
 
 
 
 
 
 
 
 
 
 
 
 
 
df03cbf
 
 
 
 
 
391573b
df03cbf
 
 
 
 
391573b
ac9b283
391573b
 
 
 
 
ac9b283
391573b
ac9b283
391573b
ac9b283
391573b
ac9b283
391573b
 
 
ac9b283
 
391573b
 
 
 
 
 
ac9b283
391573b
ac9b283
 
391573b
 
 
 
 
 
 
 
 
 
 
 
 
 
ac9b283
391573b
ac9b283
391573b
 
 
 
 
 
 
 
 
ac9b283
391573b
 
ac9b283
391573b
ac9b283
 
391573b
ac9b283
391573b
ac9b283
391573b
 
 
 
ac9b283
391573b
ac9b283
391573b
 
 
 
 
 
ac9b283
391573b
 
 
 
 
 
 
 
 
 
ac9b283
391573b
 
 
 
 
 
 
ac9b283
391573b
 
 
 
 
 
 
ac9b283
391573b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ac9b283
 
 
 
391573b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
---
license: other
license_name: trillion
license_link: LICENSE
tags:
- finetuned
- chat
language:
- en
- ko
- ja
pipeline_tag: text-generation
library_name: transformers
extra_gated_prompt: '**TRILLION LABS AI MODEL LICENSE AGREEMENT** Tri- Model Series
  Version Effective Date: February 1, 2025

  "**Agreement**" means the terms and conditions for use, reproduction, distribution
  and modification of the Trillion Labs AI Model series set forth herein.

  "**Documentation**" means the specifications, manuals and documentation accompanying
  the Tri- Model series distributed by Trillion Labs.

  "**Licensee**" or "**you**" means you, or your employer or any other person or entity
  (if you are entering into this Agreement on such person or entity''s behalf), of
  the age required under applicable laws, rules or regulations to provide legal consent
  and that has legal authority to bind your employer or such other person or entity
  if you are entering in this Agreement on their behalf.

  "**Model**" means the artificial intelligence model series provided by Licensor
  ("Tri-" series), including software, algorithms, machine learning models, and related
  components provided by Licensor, including all updates, enhancements, improvements,
  bug fixes, patches, or other modifications.

  "**Trillion Labs**" or "**we**" means Trillion Labs, the owner, developer, and provider
  of the Model, holding all rights, title, and interest in the Model.

  By clicking "I Accept" below or by using or distributing any portion or element
  of the Model, you agree to be bound by this Agreement.

  1\. **License Grant and Redistribution**.

  a. Grant of Rights. You are granted a limited, non-exclusive, non-transferable,
  worldwide, revocable license under Trillion Labs'' intellectual property or other
  rights to use, reproduce, distribute, and make modifications to the Model for research
  purposes.

  b. Redistribution and Use.

  i. If you distribute or make available the Model (or any derivative works thereof),
  or a product or service that contains any of them, you shall (A) provide a copy
  of this Agreement with any such Model; and (B) prominently display "Built with Tri-"
  on a related website, user interface, blogpost, about page, or product documentation.
  If you use the Model to create, train, fine tune, or otherwise improve an AI model,
  which is distributed or made available, you shall also include "Tri-" followed by
  the original Model version at the beginning of any such AI model name.

  ii. You must retain in all copies of the Model that you distribute the following
  attribution notice within a "Notice" text file distributed as a part of such copies:
  "Tri- Model Series is licensed under the Trillion Labs AI Model License Agreement,
  Copyright © Trillion Labs. All Rights Reserved."

  iii. Your use of the Model must comply with applicable laws and regulations (including
  trade compliance laws and regulations).

  2\. **Additional Commercial Terms**. If the monthly active users of the products
  or services made available by or for Licensee, or Licensee''s affiliates, is greater
  than 1 million monthly active users OR Annual Recurring Revenue is greater than
  $10 million USD, you must request a commercial license from Trillion Labs, and you
  are not authorized to exercise any commercial rights under this Agreement unless
  or until Trillion Labs otherwise expressly grants you such rights.

  3\. **Disclaimer of Warranty**. THE MODEL, DERIVATIVES, AND OUTPUT ARE PROVIDED
  ON AN "AS IS" BASIS, WITHOUT WARRANTIES OF ANY KIND, AND TRILLION LABS DISCLAIMS
  ALL WARRANTIES OF ANY KIND, BOTH EXPRESS AND IMPLIED, INCLUDING, WITHOUT LIMITATION,
  ANY WARRANTIES OF TITLE, NON-INFRINGEMENT, MERCHANTABILITY, OR FITNESS FOR A PARTICULAR
  PURPOSE.

  4\. **Limitation of Liability**. IN NO EVENT WILL TRILLION LABS BE LIABLE UNDER
  ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, TORT, NEGLIGENCE, PRODUCTS LIABILITY,
  OR OTHERWISE, ARISING OUT OF THIS AGREEMENT, FOR ANY LOST PROFITS OR ANY INDIRECT,
  SPECIAL, CONSEQUENTIAL, INCIDENTAL, EXEMPLARY OR PUNITIVE DAMAGES.

  5\. **Intellectual Property**.

  a. No trademark licenses are granted under this Agreement, and in connection with
  the Model, neither Trillion Labs nor Licensee may use any name or mark owned by
  or associated with the other or any of its affiliates, except as required for reasonable
  and customary use in describing and redistributing the Model or as set forth in
  this Section 5(a).

  b. All rights, title, and interest in the Model, including modifications, Derivatives,
  and documentation, remain exclusively with Trillion Labs.

  6\. **Term and Termination**. The term of this Agreement will commence upon your
  acceptance of this Agreement or access to the Model and will continue in full force
  and effect until terminated in accordance with the terms and conditions herein.
  Trillion Labs may terminate this Agreement if you are in breach of any term or condition
  of this Agreement. Upon termination of this Agreement, you shall delete and cease
  use of the Model. Sections 3, 4 and 5 shall survive the termination of this Agreement.

  7\. **Governing Law and Jurisdiction**. This Agreement will be governed and construed
  under the laws of the State of California without regard to choice of law principles.
  The courts of California shall have exclusive jurisdiction of any dispute arising
  out of this Agreement.'
extra_gated_fields:
  First Name: text
  Last Name: text
  Affiliation: text
  Job title:
    type: select
    options:
    - Student
    - Research Graduate
    - AI researcher
    - AI developer/engineer
    - Reporter
    - Other
  geo: ip_location
  ? By clicking Submit below I accept the terms of the license and acknowledge that
    the information I provide will be collected stored processed and shared in accordance
    with the Trillion Labs Privacy Policy
  : checkbox
extra_gated_description: The information you provide will be collected, stored, processed
  and shared in accordance with the Trillion Labs Privacy Policy.
extra_gated_button_content: Submit
extra_gated_heading: Please be sure to provide your full legal name, date of birth,
  and full organization name with all corporate identifiers. Avoid the use of acronyms
  and special characters. Failure to follow these instructions may prevent you from
  accessing this model and others on Hugging Face. You will not have the ability to
  edit this form after submission, so please ensure all information is accurate.
---

<p align="center">
<picture>
  <img src="https://raw.githubusercontent.com/trillion-labs/.github/main/Tri-21B.png" alt="Tri-21B", style="width: 80%;">
</picture>
</p>

## Introduction

**Tri-21B**를 4bit으로 양자화한 모델

We introduce **Tri-21B**, our flagship large language model that redefines the efficiency frontier in LLM training. By achieving state-of-the-art performance with only 2.3T training tokens, we demonstrate that exceptional capabilities don't require excessive computational resources.

<p align="center">
<img src="https://raw.githubusercontent.com/trillion-labs/.github/main/pareto-2507.png" alt="Average Performance vs. Approximate Training FLOPs" style="width: 100%; max-width: 1400px;">
</p>


### Key Highlights
* **Unprecedented Training Efficiency**: Trained on just 2.3T tokens—significantly less than comparable models—while achieving 70.3% average accuracy across MMLU/KMMLU/Global MMLU benchmarks
* **Pushing the Pareto Frontier**: With only 2.95E+23 FLOPs, Tri-21B outperforms models requiring 2-10x more compute, setting a new standard for efficient scaling
* **Enhanced Reasoning**: Modified training dataset mixture specifically optimized for reasoning capabilities
* **Advanced Post-Training**: Significantly improved RL training pipeline focusing on mathematical reasoning and everyday usage
* **Multi-lingual**: Specially optimized for Korean, English, and Japanese.

Our **Tri-21B** represents a paradigm shift in efficient model development. When comparing performance to training FLOPs, our model dramatically pushes the Pareto frontier—achieving performance comparable to or exceeding models like Qwen2.5-32B (74.6% at 3.46E+24 FLOPs) and Gemma 3 IT 27B (67.6% at 2.27E+24 FLOPs) while using approximately 8-12x fewer computational resources.


### Model Specifications

#### Tri-21B
- Type: Causal Language Model
- Training Stage: Pre-training & Post-training
- Architecture: Transformer Decoder with RoPE, SwiGLU, RMSNorm, and GQA
- Number of Parameters: 20.73B
- Number of Layers: 32
- Number of Attention Heads: 32 (Query) / 8 (Key, Value)
- Context Length: 8,192
- Number of Tokens Seen: 2.3T
- Vocab Size: 124,416 

## Training Efficiency Analysis

Our approach to training efficiency sets new benchmarks in the field. The following comparison demonstrates how Tri-21B achieves superior performance per FLOP compared to other state-of-the-art models of similar scale:

| Model | FLOPs | Avg. Accuracy¹ | Efficiency Ratio² |
|:------|:------|:--------------|:-----------------|
| **Tri-21B** | **2.95E+23** | **70.3%** | **1.00x (baseline)** |
| Gemma2-9b | 4.42E+23 | 61.5% | 0.48x |
| Qwen2.5-7B | 8.22E+23 | 63.4% | 0.29x |
| Exaone-3.5-32B | 1.25E+24 | 58.5% | 0.19x |
| Gemma 3 IT 27B | 2.27E+24 | 67.6% | 0.11x |
| Qwen2.5-32B | 3.46E+24 | 74.6% | 0.10x |
| Qwen3-32B | 5.77E+24 | 73.5% | 0.06x |

¹ Average of MMLU / KMMLU / Global MMLU (ja)  
² Performance per FLOP relative to Tri-21B

This efficiency breakthrough enables organizations to deploy state-of-the-art language models without the traditional computational barriers, democratizing access to advanced AI capabilities.


## Quickstart

Here is a code snippet with `apply_chat_template` that demonstrates how to load the tokenizer and model and generate text.

### Tri-21B Usage
```python
import torch
from transformers import AutoModelForCausalLM, AutoTokenizer

model_name = "trillionlabs/Tri-21B"

model = AutoModelForCausalLM.from_pretrained(
    model_name,
    torch_dtype=torch.bfloat16,
    device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained(model_name)

prompt = "Explain the concept of quantum computing in simple terms."
messages = [
    {"role": "user", "content": prompt}
]
text = tokenizer.apply_chat_template(
    messages,
    tokenize=False,
    add_generation_prompt=True
)
model_inputs = tokenizer([text], return_tensors="pt").to(model.device)

generated_ids = model.generate(
    **model_inputs,
    max_new_tokens=512
)
generated_ids = [
    output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
]

response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
print(response)
```

### vLLM, SGLang Deployment

Tri-21B is also available with [vLLM](https://docs.vllm.ai/en/latest/) and [SGLang](https://docs.sglang.ai/)!

```bash
# vLLM
vllm serve trillionlabs/Tri-21B --dtype bfloat16 --max-model-len 8192

# vLLM with custom options
vllm serve trillionlabs/Tri-21B \
    --dtype bfloat16 \
    --max-model-len 8192 \
    --gpu-memory-utilization 0.95 \
    --port 8000
````

```bash
# SGLang
python3 -m sglang.launch_server --model-path trillionlabs/Tri-21B --dtype bfloat16

# SGLang with custom options
python3 -m sglang.launch_server \
    --model-path trillionlabs/Tri-21B \
    --dtype bfloat16 \
    --context-length 8192 \
    --port 30000 \
    --host 0.0.0.0
```

## Evaluation


We evaluated Tri-21B across a comprehensive suite of benchmarks assessing general reasoning, knowledge recall, coding abilities, mathematical reasoning, and instruction-following capabilities. We compare our model against state-of-the-art models of similar scale: Gemmma-3-IT-27B and Qwen3-32B to demonstrate its competitive performance.

<details>
<summary> Full evaluation settings </summary>
# Benchmark Evaluation Settings

| Benchmark | Language | Evaluation Setting | Metric |
|:----------|:---------|:------------------|:-------|
| **General Reasoning and Factuality** | | | |
| • HellaSwag | English | 0-shot | accuracy |
| • ARC:C | English | 0-shot | accuracy |
| • HAERAE | Korean | 3-shot | accuracy |
| • CLIcK | Korean | 0-shot | accuracy |
| • KoBEST | Korean | 5-shot | accuracy |
| **Knowledge and Reasoning** | | | |
| • KMMLU | Korean | 5-shot (0-shot, CoT) | accuracy (exact-match) |
| • MMLU | English | 5-shot (0-shot, CoT) | accuracy (exact-match) |
| • MMLU-Pro | English | 0-shot, CoT | exact-match |
| • Global-MMLU-Lite-ja | Japaneses | 5-shot | accuracy |
| **Coding** | | | |
| • HumanEval | English | 0-shot | pass@1 |
| • MBPPPlus | English | 0-shot | pass@1 |
| **Mathematical Reasoning** | | | |
| • GSM8k | English | 0-shot, CoT | exact-match |
| • MATH | English | 0-shot, CoT | exact-match |
| • GPQA | English | 4-shot | accuracy |
| • GPQA Diamond | English | 0-shot, CoT | accuracy |
| • HRM8k | Korean | 0-shot, CoT | exact-match |
| **Instruction Following and Chat** | | | |
| • IFEval | English | 0-shot | strict-average |
| • koIFEval | Korean | 0-shot | strict-average |
| • MT-Bench | English | LLM-as-a-judge (gpt-4o) | LLM score |
| • KO-MT-Bench | Korean | LLM-as-a-judge (gpt-4o) | LLM score |
| • systemIFEval | English | 0-shot | strict-average |

- *Note that koIFEval, systemIFEval, and KoRuler are our in-house evaluation benchmarks adapted for Korean to better assess model capabilities in Korean language tasks.
- **Note that MT-Bench, KO-MT-Bench, and LogicKor use a 10-point scale.

</details>

### Benchmark Results

Models compared:

- **Tri-21B**: Our flagship 21B parameter model
- **Qwen3-32B**: Qwen's 32B parameter model
- **Gemma3-IT-27B**: Google's Gemma 3 instruction-tuned 27B model


### General Reasoning and Factuality

| Benchmark | Tri-21B | Qwen3-32B | Gemma3-IT-27B |
| --- | --- | --- | --- |
| HAERAE | 86.16 | 71.67 | 78.09 | 
| KoBEST | 85.92 | 83.39 | 87.66 | 
| CLIcK | 72.32 | 66.89 | 67.54 | 
| KMMLU | 61.89 (69.90) | 61.73 (67.55)| 55.03 (60.61)| 
| MMLU | 77.62 (85.02) | 81.86 (84.46) | 77.42 (84.09) | 
| MMLU-Pro | 64.74 | 70.53 | 64.26 | 
| Global-MMLU-Lite-ja | 70.25 | 77.00 | 72.00 | 

### Coding

| Benchmark | Tri-21B | Qwen3-32B | Gemma3-IT-27B |
| --- | --- | --- | --- |
| HumanEval | 75.61 | 74.39 | 87.80 | 
| MBPPPlus | 73.02 | 74.40 | 84.92 | 

### Mathematical Reasoning

| Benchmark | Tri-21B | Qwen3-32B | Gemma3-IT-27B |
| --- | --- | --- | --- |
| GSM8k | 87.95 | 86.66 | 90.52 | 
| MATH | 77.60 | 81.40 | 85.00 | 
| GPQA | 39.73 | 41.07 | 37.95 | 
| GPQA-Diamond | 44.95 | 54.04 | 44.44 | 
| HRM8k | 56.70 | 66.24 | 63.90 | 

### Instruction Following and Chat

| Benchmark | Tri-21B | Qwen3-32B | Gemma3-IT-27B |
| --- | --- | --- | --- |
| IFEval | 80.75 | 86.08 | 80.78 | 
| koIFEval | 66.51 | 62.93 | 69.24 | 
| MT-Bench | 8.21 | 8.52 | 8.53 | 
| KO-MT-Bench | 7.79 | 8.47 | 8.46 | 
| systemIFEval | 77.40 | 77.92 | 77.94 | 

### Base Model Evaluation

The following table shows the performance of Tri-21B base model (before instruction tuning) on key benchmarks:

| Benchmark | Tri-21B Base |
| --- | --- |
| MMLU | 76.99 |
| KMMLU | 62.37 |
| KoBEST | 85.07 |
| BBH | 77.19 |
| GSM8K | 70.36 |
| MBPPPlus | 75.40 | 

## Limitations

- Language Support: The models are optimized for English, Korean, and Japanese. Usage with other languages may result in degraded performance.
- Knowledge Cutoff: The model's information is limited to data available up to Febuary, 2025.

## License
This model repository is licensed under the Trillion License.

## Contact
For inquiries, please contact: info@trillionlabs.co