|
|
--- |
|
|
license: apache-2.0 |
|
|
tags: |
|
|
- finetuned |
|
|
- chat |
|
|
language: |
|
|
- en |
|
|
- ko |
|
|
- ja |
|
|
pipeline_tag: text-generation |
|
|
library_name: transformers |
|
|
extra_gated_fields: |
|
|
Full Name: text |
|
|
Email: text |
|
|
Organization: text |
|
|
--- |
|
|
|
|
|
<p align="center"> |
|
|
<picture> |
|
|
<img src="https://raw.githubusercontent.com/trillion-labs/.github/main/Tri-7B.png" alt="Tri-7B", style="width: 80%;"> |
|
|
</picture> |
|
|
</p> |
|
|
|
|
|
# Tri-7B |
|
|
|
|
|
## Introduction |
|
|
|
|
|
We introduce **Tri-7B**, the next generation model following Trillion-7B-preview, that continues to push the boundaries of efficient training while achieving exceptional performance at the 7B parameter scale. |
|
|
|
|
|
|
|
|
<p align="center"> |
|
|
<img src="https://raw.githubusercontent.com/trillion-labs/.github/main/pareto-2507.png" alt="Average Performance vs. Approximate Training FLOPs" style="width: 100%; max-width: 1400px;"> |
|
|
</p> |
|
|
|
|
|
|
|
|
### Key Highlights |
|
|
* **Enhanced Reasoning**: Modified training dataset mixture specifically optimized for reasoning capabilities |
|
|
* **Advanced Post-Training**: Significantly improved RL training pipeline focusing on mathematical reasoning and everyday usage |
|
|
* **Extended Context**: Supports up to 32K context length for long-form understanding |
|
|
* **Multi-lingual**: Specially optimized for Korean, English, and Japanese. |
|
|
|
|
|
Our **Tri-7B** model represents a significant advancement over Trillion-7B-preview, achieving substantial performance improvements across all evaluated domains while maintaining the same efficient parameter count. |
|
|
|
|
|
### Model Specifications |
|
|
|
|
|
#### Tri-7B |
|
|
- Type: Causal Language Model |
|
|
- Training Stage: Pre-training & Post-training |
|
|
- Architecture: Transformer Decoder with RoPE, SwiGLU, RMSNorm |
|
|
- Number of Parameters: 7.76B |
|
|
- Number of Layers: 32 |
|
|
- Number of Attention Heads: 32 |
|
|
- Context Length: 32,768 |
|
|
- Vocab Size: 128,256 |
|
|
|
|
|
|
|
|
## Quickstart |
|
|
|
|
|
Here is a code snippet with `apply_chat_template` that demonstrates how to load the tokenizer and model and generate text. |
|
|
|
|
|
### Tri-7B Usage |
|
|
```python |
|
|
import torch |
|
|
from transformers import AutoModelForCausalLM, AutoTokenizer |
|
|
|
|
|
model_name = "trillionlabs/Tri-7B" |
|
|
|
|
|
model = AutoModelForCausalLM.from_pretrained( |
|
|
model_name, |
|
|
torch_dtype=torch.bfloat16, |
|
|
device_map="auto" |
|
|
) |
|
|
tokenizer = AutoTokenizer.from_pretrained(model_name) |
|
|
|
|
|
prompt = "Explain the concept of quantum computing in simple terms." |
|
|
messages = [ |
|
|
{"role": "user", "content": prompt} |
|
|
] |
|
|
text = tokenizer.apply_chat_template( |
|
|
messages, |
|
|
tokenize=False, |
|
|
add_generation_prompt=True |
|
|
) |
|
|
model_inputs = tokenizer([text], return_tensors="pt").to(model.device) |
|
|
|
|
|
generated_ids = model.generate( |
|
|
**model_inputs, |
|
|
max_new_tokens=512 |
|
|
) |
|
|
generated_ids = [ |
|
|
output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids) |
|
|
] |
|
|
|
|
|
response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0] |
|
|
print(response) |
|
|
``` |
|
|
|
|
|
Tri-7B is also available with vLLM and SGLang! |
|
|
|
|
|
```bash |
|
|
# vLLM |
|
|
vllm serve trillionlabs/Tri-7B --dtype bfloat16 --max-model-len 32768 |
|
|
|
|
|
# vLLM with custom options |
|
|
vllm serve trillionlabs/Tri-7B \ |
|
|
--dtype bfloat16 \ |
|
|
--max-model-len 32768 \ |
|
|
--gpu-memory-utilization 0.95 \ |
|
|
--port 8000 |
|
|
``` |
|
|
|
|
|
```bash |
|
|
# SGLang |
|
|
python3 -m sglang.launch_server --model-path trillionlabs/Tri-7B --dtype bfloat16 |
|
|
|
|
|
# SGLang with custom options |
|
|
python3 -m sglang.launch_server \ |
|
|
--model-path trillionlabs/Tri-7B \ |
|
|
--dtype bfloat16 \ |
|
|
--context-length 32768 \ |
|
|
--port 30000 \ |
|
|
--host 0.0.0.0 |
|
|
``` |
|
|
|
|
|
## Evaluation |
|
|
|
|
|
We evaluated Tri-7B across a comprehensive suite of benchmarks assessing general reasoning, knowledge recall, coding abilities, mathematical reasoning, and instruction-following capabilities. Compared to our previous generation model Trillion-7B-preview, Tri-7B achieves significant gains across all domains. |
|
|
|
|
|
<details> |
|
|
<summary> Full evaluation settings </summary> |
|
|
|
|
|
| Benchmark | Language | Evaluation Setting | Metric | |
|
|
|:----------|:---------|:------------------|:-------| |
|
|
| **General Reasoning and Factuality** | | | | |
|
|
| • HellaSwag | English | 0-shot | accuracy | |
|
|
| • ARC:C | English | 0-shot | accuracy | |
|
|
| • HAERAE | Korean | 3-shot | accuracy | |
|
|
| • CLIcK | Korean | 0-shot | accuracy | |
|
|
| • KoBEST | Korean | 5-shot | accuracy | |
|
|
| **Knowledge and Reasoning** | | | | |
|
|
| • KMMLU | Korean | 5-shot (0-shot, CoT) | accuracy | |
|
|
| • MMLU | English | 5-shot (0-shot, CoT) | accuracy | |
|
|
| • Global-MMLU-Lite-ja | English | 5-shot | accuracy | |
|
|
| **Coding** | | | | |
|
|
| • HumanEval | English | 0-shot | pass@1 | |
|
|
| • MBPPPlus | English | 0-shot | pass@1 | |
|
|
| **Mathematical Reasoning** | | | | |
|
|
| • GSM8k | English | 0-shot, CoT | exact-match | |
|
|
| • MATH | English | 0-shot, CoT | exact-match | |
|
|
| • GPQA | English | 4-shot | accuracy | |
|
|
| • HRM8k | Korean | 0-shot, CoT | exact-match | |
|
|
| **Instruction Following and Chat** | | | | |
|
|
| • IFEval | English | 0-shot | strict-average | |
|
|
| • koIFEval | Korean | 0-shot | strict-average | |
|
|
| • MT-Bench | English | LLM-as-a-judge (gpt-4o) | LLM score | |
|
|
| • KO-MT-Bench | Korean | LLM-as-a-judge (gpt-4o) | LLM score | |
|
|
| • systemIFEval | English | 0-shot | strict-average | |
|
|
|
|
|
- *Note that koIFEval, systemIFEval, and KoRuler are our in-house evaluation benchmarks adapted for Korean to better assess model capabilities in Korean language tasks. |
|
|
- **Note that MT-Bench, KO-MT-Bench, and LogicKor use a 10-point scale. |
|
|
|
|
|
</details> |
|
|
|
|
|
### Benchmark Results |
|
|
|
|
|
Models compared: |
|
|
- **Tri-7B** (Next Generation) |
|
|
- **Trillion-7B-preview** (Previous Generation) |
|
|
|
|
|
### General Reasoning and Factuality |
|
|
|
|
|
| Benchmark | Tri-7B | Trillion-7B-preview | Improvement | |
|
|
| --- | --- | --- | --- | |
|
|
| HellaSwag | 59.52 | 58.94 | +0.58 | |
|
|
| ARC:C | 58.28 | 54.44 | +3.84 | |
|
|
| HAERAE | 82.49 | 80.02 | +2.47 | |
|
|
| KoBEST | 82.72 | 79.61 | +3.11 | |
|
|
| CLIcK | 64.43 | 60.41 | +4.02 | |
|
|
| KMMLU | 51.74 (53.51) | 48.09 | +3.65 | |
|
|
| MMLU | 68.16 (74.67) | 63.52 | +4.64 | |
|
|
| Global-MMLU-Lite-ja | 59.25 | 60.75 | -1.50 | |
|
|
|
|
|
|
|
|
### Coding |
|
|
|
|
|
| Benchmark | Tri-7B | Trillion-7B-preview | Improvement | |
|
|
| --- | --- | --- | --- | |
|
|
| HumanEval | 53.66 | 55.48 | -1.82 | |
|
|
| MBPPPlus | 64.29 | 58.99 | +5.30 | |
|
|
|
|
|
### Mathematical Reasoning |
|
|
|
|
|
| Benchmark | Tri-7B | Trillion-7B-preview | Improvement | |
|
|
| --- | --- | --- | --- | |
|
|
| GSM8k | 77.94 | 72.25 | +5.69 | |
|
|
| MATH | 49.40 | 32.70 | +16.70 | |
|
|
| GPQA | 34.15 | 32.81 | +1.34 | |
|
|
| HRM8k | 39.08 | 30.10 | +8.98 | |
|
|
|
|
|
### Instruction Following and Chat |
|
|
|
|
|
| Benchmark | Tri-7B | Trillion-7B-preview | Improvement | |
|
|
| --- | --- | --- | --- | |
|
|
| IFEval | 79.26 | 79.13 | +0.13 | |
|
|
| koIFEval | 76.63 | 66.58 | +10.05 | |
|
|
| MT-Bench | 7.82 | 6.53 | +1.29 | |
|
|
| KO-MT-Bench | 7.64 | 6.27 | +1.37 | |
|
|
| systemIFEval | 66.43 | 27.28 | +39.15 | |
|
|
|
|
|
## Limitations |
|
|
|
|
|
- Language Support: The model is optimized for English, Korean, and Japanese. Usage with other languages may result in degraded performance. |
|
|
- Knowledge Cutoff: The model's information is limited to data available up to Febuary, 2025. |
|
|
|
|
|
## License |
|
|
This model is licensed under the Apache License 2.0. |
|
|
|
|
|
|
|
|
## Contact |
|
|
For inquiries, please contact: info@trillionlabs.co |