juyoung-trl commited on
Commit
d5d5f3a
·
verified ·
1 Parent(s): 7b0f643

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,6 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ assets/Signiture_Black_White_BG_resized.jpg filter=lfs diff=lfs merge=lfs -text
37
+ assets/Signiture_Trillion_BlackBG.png filter=lfs diff=lfs merge=lfs -text
38
+ assets/Signiture_Trillion_WhiteBG.png filter=lfs diff=lfs merge=lfs -text
LICENSE ADDED
@@ -0,0 +1,201 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Apache License
2
+ Version 2.0, January 2004
3
+ http://www.apache.org/licenses/
4
+
5
+ TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION
6
+
7
+ 1. Definitions.
8
+
9
+ "License" shall mean the terms and conditions for use, reproduction,
10
+ and distribution as defined by Sections 1 through 9 of this document.
11
+
12
+ "Licensor" shall mean the copyright owner or entity authorized by
13
+ the copyright owner that is granting the License.
14
+
15
+ "Legal Entity" shall mean the union of the acting entity and all
16
+ other entities that control, are controlled by, or are under common
17
+ control with that entity. For the purposes of this definition,
18
+ "control" means (i) the power, direct or indirect, to cause the
19
+ direction or management of such entity, whether by contract or
20
+ otherwise, or (ii) ownership of fifty percent (50%) or more of the
21
+ outstanding shares, or (iii) beneficial ownership of such entity.
22
+
23
+ "You" (or "Your") shall mean an individual or Legal Entity
24
+ exercising permissions granted by this License.
25
+
26
+ "Source" form shall mean the preferred form for making modifications,
27
+ including but not limited to software source code, documentation
28
+ source, and configuration files.
29
+
30
+ "Object" form shall mean any form resulting from mechanical
31
+ transformation or translation of a Source form, including but
32
+ not limited to compiled object code, generated documentation,
33
+ and conversions to other media types.
34
+
35
+ "Work" shall mean the work of authorship, whether in Source or
36
+ Object form, made available under the License, as indicated by a
37
+ copyright notice that is included in or attached to the work
38
+ (an example is provided in the Appendix below).
39
+
40
+ "Derivative Works" shall mean any work, whether in Source or Object
41
+ form, that is based on (or derived from) the Work and for which the
42
+ editorial revisions, annotations, elaborations, or other modifications
43
+ represent, as a whole, an original work of authorship. For the purposes
44
+ of this License, Derivative Works shall not include works that remain
45
+ separable from, or merely link (or bind by name) to the interfaces of,
46
+ the Work and Derivative Works thereof.
47
+
48
+ "Contribution" shall mean any work of authorship, including
49
+ the original version of the Work and any modifications or additions
50
+ to that Work or Derivative Works thereof, that is intentionally
51
+ submitted to Licensor for inclusion in the Work by the copyright owner
52
+ or by an individual or Legal Entity authorized to submit on behalf of
53
+ the copyright owner. For the purposes of this definition, "submitted"
54
+ means any form of electronic, verbal, or written communication sent
55
+ to the Licensor or its representatives, including but not limited to
56
+ communication on electronic mailing lists, source code control systems,
57
+ and issue tracking systems that are managed by, or on behalf of, the
58
+ Licensor for the purpose of discussing and improving the Work, but
59
+ excluding communication that is conspicuously marked or otherwise
60
+ designated in writing by the copyright owner as "Not a Contribution."
61
+
62
+ "Contributor" shall mean Licensor and any individual or Legal Entity
63
+ on behalf of whom a Contribution has been received by Licensor and
64
+ subsequently incorporated within the Work.
65
+
66
+ 2. Grant of Copyright License. Subject to the terms and conditions of
67
+ this License, each Contributor hereby grants to You a perpetual,
68
+ worldwide, non-exclusive, no-charge, royalty-free, irrevocable
69
+ copyright license to reproduce, prepare Derivative Works of,
70
+ publicly display, publicly perform, sublicense, and distribute the
71
+ Work and such Derivative Works in Source or Object form.
72
+
73
+ 3. Grant of Patent License. Subject to the terms and conditions of
74
+ this License, each Contributor hereby grants to You a perpetual,
75
+ worldwide, non-exclusive, no-charge, royalty-free, irrevocable
76
+ (except as stated in this section) patent license to make, have made,
77
+ use, offer to sell, sell, import, and otherwise transfer the Work,
78
+ where such license applies only to those patent claims licensable
79
+ by such Contributor that are necessarily infringed by their
80
+ Contribution(s) alone or by combination of their Contribution(s)
81
+ with the Work to which such Contribution(s) was submitted. If You
82
+ institute patent litigation against any entity (including a
83
+ cross-claim or counterclaim in a lawsuit) alleging that the Work
84
+ or a Contribution incorporated within the Work constitutes direct
85
+ or contributory patent infringement, then any patent licenses
86
+ granted to You under this License for that Work shall terminate
87
+ as of the date such litigation is filed.
88
+
89
+ 4. Redistribution. You may reproduce and distribute copies of the
90
+ Work or Derivative Works thereof in any medium, with or without
91
+ modifications, and in Source or Object form, provided that You
92
+ meet the following conditions:
93
+
94
+ (a) You must give any other recipients of the Work or
95
+ Derivative Works a copy of this License; and
96
+
97
+ (b) You must cause any modified files to carry prominent notices
98
+ stating that You changed the files; and
99
+
100
+ (c) You must retain, in the Source form of any Derivative Works
101
+ that You distribute, all copyright, patent, trademark, and
102
+ attribution notices from the Source form of the Work,
103
+ excluding those notices that do not pertain to any part of
104
+ the Derivative Works; and
105
+
106
+ (d) If the Work includes a "NOTICE" text file as part of its
107
+ distribution, then any Derivative Works that You distribute must
108
+ include a readable copy of the attribution notices contained
109
+ within such NOTICE file, excluding those notices that do not
110
+ pertain to any part of the Derivative Works, in at least one
111
+ of the following places: within a NOTICE text file distributed
112
+ as part of the Derivative Works; within the Source form or
113
+ documentation, if provided along with the Derivative Works; or,
114
+ within a display generated by the Derivative Works, if and
115
+ wherever such third-party notices normally appear. The contents
116
+ of the NOTICE file are for informational purposes only and
117
+ do not modify the License. You may add Your own attribution
118
+ notices within Derivative Works that You distribute, alongside
119
+ or as an addendum to the NOTICE text from the Work, provided
120
+ that such additional attribution notices cannot be construed
121
+ as modifying the License.
122
+
123
+ You may add Your own copyright statement to Your modifications and
124
+ may provide additional or different license terms and conditions
125
+ for use, reproduction, or distribution of Your modifications, or
126
+ for any such Derivative Works as a whole, provided Your use,
127
+ reproduction, and distribution of the Work otherwise complies with
128
+ the conditions stated in this License.
129
+
130
+ 5. Submission of Contributions. Unless You explicitly state otherwise,
131
+ any Contribution intentionally submitted for inclusion in the Work
132
+ by You to the Licensor shall be under the terms and conditions of
133
+ this License, without any additional terms or conditions.
134
+ Notwithstanding the above, nothing herein shall supersede or modify
135
+ the terms of any separate license agreement you may have executed
136
+ with Licensor regarding such Contributions.
137
+
138
+ 6. Trademarks. This License does not grant permission to use the trade
139
+ names, trademarks, service marks, or product names of the Licensor,
140
+ except as required for reasonable and customary use in describing the
141
+ origin of the Work and reproducing the content of the NOTICE file.
142
+
143
+ 7. Disclaimer of Warranty. Unless required by applicable law or
144
+ agreed to in writing, Licensor provides the Work (and each
145
+ Contributor provides its Contributions) on an "AS IS" BASIS,
146
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
147
+ implied, including, without limitation, any warranties or conditions
148
+ of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A
149
+ PARTICULAR PURPOSE. You are solely responsible for determining the
150
+ appropriateness of using or redistributing the Work and assume any
151
+ risks associated with Your exercise of permissions under this License.
152
+
153
+ 8. Limitation of Liability. In no event and under no legal theory,
154
+ whether in tort (including negligence), contract, or otherwise,
155
+ unless required by applicable law (such as deliberate and grossly
156
+ negligent acts) or agreed to in writing, shall any Contributor be
157
+ liable to You for damages, including any direct, indirect, special,
158
+ incidental, or consequential damages of any character arising as a
159
+ result of this License or out of the use or inability to use the
160
+ Work (including but not limited to damages for loss of goodwill,
161
+ work stoppage, computer failure or malfunction, or any and all
162
+ other commercial damages or losses), even if such Contributor
163
+ has been advised of the possibility of such damages.
164
+
165
+ 9. Accepting Warranty or Additional Liability. While redistributing
166
+ the Work or Derivative Works thereof, You may choose to offer,
167
+ and charge a fee for, acceptance of support, warranty, indemnity,
168
+ or other liability obligations and/or rights consistent with this
169
+ License. However, in accepting such obligations, You may act only
170
+ on Your own behalf and on Your sole responsibility, not on behalf
171
+ of any other Contributor, and only if You agree to indemnify,
172
+ defend, and hold each Contributor harmless for any liability
173
+ incurred by, or claims asserted against, such Contributor by reason
174
+ of your accepting any such warranty or additional liability.
175
+
176
+ END OF TERMS AND CONDITIONS
177
+
178
+ APPENDIX: How to apply the Apache License to your work.
179
+
180
+ To apply the Apache License to your work, attach the following
181
+ boilerplate notice, with the fields enclosed by brackets "[]"
182
+ replaced with your own identifying information. (Don't include
183
+ the brackets!) The text should be enclosed in the appropriate
184
+ comment syntax for the file format. We also recommend that a
185
+ file or class name and description of purpose be included on the
186
+ same "printed page" as the copyright notice for easier
187
+ identification within third-party archives.
188
+
189
+ Copyright [yyyy] [name of copyright owner]
190
+
191
+ Licensed under the Apache License, Version 2.0 (the "License");
192
+ you may not use this file except in compliance with the License.
193
+ You may obtain a copy of the License at
194
+
195
+ http://www.apache.org/licenses/LICENSE-2.0
196
+
197
+ Unless required by applicable law or agreed to in writing, software
198
+ distributed under the License is distributed on an "AS IS" BASIS,
199
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
200
+ See the License for the specific language governing permissions and
201
+ limitations under the License.
README.md CHANGED
@@ -1,199 +1,209 @@
1
  ---
 
 
 
 
 
 
 
 
 
2
  library_name: transformers
3
- tags: []
4
  ---
5
 
6
- # Model Card for Model ID
7
-
8
- <!-- Provide a quick summary of what the model is/does. -->
9
-
10
-
11
-
12
- ## Model Details
13
-
14
- ### Model Description
15
-
16
- <!-- Provide a longer summary of what this model is. -->
17
-
18
- This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
19
-
20
- - **Developed by:** [More Information Needed]
21
- - **Funded by [optional]:** [More Information Needed]
22
- - **Shared by [optional]:** [More Information Needed]
23
- - **Model type:** [More Information Needed]
24
- - **Language(s) (NLP):** [More Information Needed]
25
- - **License:** [More Information Needed]
26
- - **Finetuned from model [optional]:** [More Information Needed]
27
-
28
- ### Model Sources [optional]
29
-
30
- <!-- Provide the basic links for the model. -->
31
-
32
- - **Repository:** [More Information Needed]
33
- - **Paper [optional]:** [More Information Needed]
34
- - **Demo [optional]:** [More Information Needed]
35
-
36
- ## Uses
37
-
38
- <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
-
40
- ### Direct Use
41
-
42
- <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
-
44
- [More Information Needed]
45
-
46
- ### Downstream Use [optional]
47
-
48
- <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
-
50
- [More Information Needed]
51
-
52
- ### Out-of-Scope Use
53
-
54
- <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
-
56
- [More Information Needed]
57
-
58
- ## Bias, Risks, and Limitations
59
-
60
- <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
-
62
- [More Information Needed]
63
-
64
- ### Recommendations
65
-
66
- <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
-
68
- Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
-
70
- ## How to Get Started with the Model
71
-
72
- Use the code below to get started with the model.
73
-
74
- [More Information Needed]
75
-
76
- ## Training Details
77
-
78
- ### Training Data
79
-
80
- <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
-
82
- [More Information Needed]
83
-
84
- ### Training Procedure
85
-
86
- <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
-
88
- #### Preprocessing [optional]
89
-
90
- [More Information Needed]
91
-
92
-
93
- #### Training Hyperparameters
94
-
95
- - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
-
97
- #### Speeds, Sizes, Times [optional]
98
-
99
- <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
-
101
- [More Information Needed]
 
 
102
 
103
  ## Evaluation
104
 
105
- <!-- This section describes the evaluation protocols and provides the results. -->
106
-
107
- ### Testing Data, Factors & Metrics
108
-
109
- #### Testing Data
110
-
111
- <!-- This should link to a Dataset Card if possible. -->
112
-
113
- [More Information Needed]
114
-
115
- #### Factors
116
-
117
- <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
-
119
- [More Information Needed]
120
-
121
- #### Metrics
122
-
123
- <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
-
125
- [More Information Needed]
126
-
127
- ### Results
128
-
129
- [More Information Needed]
130
-
131
- #### Summary
132
-
133
-
134
-
135
- ## Model Examination [optional]
136
-
137
- <!-- Relevant interpretability work for the model goes here -->
138
-
139
- [More Information Needed]
140
-
141
- ## Environmental Impact
142
-
143
- <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
-
145
- Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
-
147
- - **Hardware Type:** [More Information Needed]
148
- - **Hours used:** [More Information Needed]
149
- - **Cloud Provider:** [More Information Needed]
150
- - **Compute Region:** [More Information Needed]
151
- - **Carbon Emitted:** [More Information Needed]
152
-
153
- ## Technical Specifications [optional]
154
-
155
- ### Model Architecture and Objective
156
-
157
- [More Information Needed]
158
-
159
- ### Compute Infrastructure
160
-
161
- [More Information Needed]
162
-
163
- #### Hardware
164
-
165
- [More Information Needed]
166
-
167
- #### Software
168
-
169
- [More Information Needed]
170
-
171
- ## Citation [optional]
172
-
173
- <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
-
175
- **BibTeX:**
176
-
177
- [More Information Needed]
178
-
179
- **APA:**
180
-
181
- [More Information Needed]
182
-
183
- ## Glossary [optional]
184
-
185
- <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
-
187
- [More Information Needed]
188
-
189
- ## More Information [optional]
190
-
191
- [More Information Needed]
192
-
193
- ## Model Card Authors [optional]
194
-
195
- [More Information Needed]
196
-
197
- ## Model Card Contact
198
-
199
- [More Information Needed]
 
1
  ---
2
+ license: apache-2.0
3
+ tags:
4
+ - finetuned
5
+ - chat
6
+ language:
7
+ - en
8
+ - ko
9
+ - ja
10
+ pipeline_tag: text-generation
11
  library_name: transformers
 
12
  ---
13
 
14
+ # Tri-7B
15
+
16
+ <p align="center">
17
+ <picture>
18
+ <img src="assets/Signiture_Black_White_BG_resized.jpg" alt="logo", width="300", style="margin: 40 auto;">
19
+ </picture>
20
+
21
+ ## Introduction
22
+
23
+ We introduce **Tri-7B**, the next generation model following Trillion-7B-preview, designed to achieve exceptional performance with efficient training at the 7B parameter scale.
24
+
25
+ ### Key Highlights
26
+ * **Enhanced Reasoning**: Modified training dataset mixture specifically optimized for reasoning capabilities
27
+ * **Advanced Post-Training**: Significantly improved RL training pipeline focusing on mathematical reasoning and everyday usage
28
+ * **Extended Context**: Supports up to 32K context length for long-form understanding
29
+
30
+ Our **Tri-7B** model represents a significant advancement over Trillion-7B-preview, achieving substantial performance improvements across all evaluated domains while maintaining the same efficient parameter count.
31
+
32
+ ### Model Specifications
33
+
34
+ #### Tri-7B
35
+ - Type: Causal Language Model
36
+ - Training Stage: Pre-training & Post-training
37
+ - Architecture: Transformer Decoder with RoPE, SwiGLU, RMSNorm
38
+ - Number of Parameters: 7.76B
39
+ - Number of Layers: 32
40
+ - Number of Attention Heads: 32
41
+ - Context Length: 32,768
42
+ - Vocab Size: 128,256
43
+
44
+
45
+ ## Quickstart
46
+
47
+ Here is a code snippet with `apply_chat_template` that demonstrates how to load the tokenizer and model and generate text.
48
+
49
+ ### Tri-7B Usage
50
+ ```python
51
+ import torch
52
+ from transformers import AutoModelForCausalLM, AutoTokenizer
53
+
54
+ model_name = "trillionlabs/Tri-7B"
55
+
56
+ model = AutoModelForCausalLM.from_pretrained(
57
+ model_name,
58
+ torch_dtype=torch.bfloat16,
59
+ device_map="auto"
60
+ )
61
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
62
+
63
+ prompt = "Explain the concept of quantum computing in simple terms."
64
+ messages = [
65
+ {"role": "user", "content": prompt}
66
+ ]
67
+ text = tokenizer.apply_chat_template(
68
+ messages,
69
+ tokenize=False,
70
+ add_generation_prompt=True
71
+ )
72
+ model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
73
+
74
+ generated_ids = model.generate(
75
+ **model_inputs,
76
+ max_new_tokens=512
77
+ )
78
+ generated_ids = [
79
+ output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
80
+ ]
81
+
82
+ response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
83
+ print(response)
84
+ ```
85
+
86
+ Tri-7B is also available with vLLM and SGLang!
87
+
88
+ ```bash
89
+ # vLLM
90
+ vllm serve trillionlabs/Tri-7B --dtype bfloat16 --max-model-len 32768
91
+
92
+ # vLLM with custom options
93
+ vllm serve trillionlabs/Tri-7B \
94
+ --dtype bfloat16 \
95
+ --max-model-len 32768 \
96
+ --gpu-memory-utilization 0.95 \
97
+ --port 8000
98
+ ```
99
+
100
+ ```bash
101
+ # SGLang
102
+ python3 -m sglang.launch_server --model-path trillionlabs/Tri-7B --dtype bfloat16
103
+
104
+ # SGLang with custom options
105
+ python3 -m sglang.launch_server \
106
+ --model-path trillionlabs/Tri-7B \
107
+ --dtype bfloat16 \
108
+ --context-length 32768 \
109
+ --port 30000 \
110
+ --host 0.0.0.0
111
+ ```
112
 
113
  ## Evaluation
114
 
115
+ We evaluated Tri-7B across a comprehensive suite of benchmarks assessing general reasoning, knowledge recall, coding abilities, mathematical reasoning, and instruction-following capabilities. Compared to our previous generation model Trillion-7B-preview, Tri-7B achieves significant gains across all domains.
116
+
117
+ <details>
118
+ <summary> Full evaluation settings </summary>
119
+
120
+ | Benchmark | Language | Evaluation Setting | Metric |
121
+ |:----------|:---------|:------------------|:-------|
122
+ | **General Reasoning and Factuality** | | | |
123
+ | HellaSwag | English | 0-shot | accuracy |
124
+ | • ARC:C | English | 0-shot | accuracy |
125
+ | • HAERAE | Korean | 3-shot | accuracy |
126
+ | • CLIcK | Korean | 0-shot | accuracy |
127
+ | KoBEST | Korean | 5-shot | accuracy |
128
+ | • BBH | English | 0-shot, CoT | accuracy |
129
+ | **Knowledge and Reasoning** | | | |
130
+ | • KMMLU | Korean | 5-shot | accuracy |
131
+ | • MMLU | English | 5-shot | accuracy |
132
+ | **Coding** | | | |
133
+ | HumanEval | English | 0-shot | pass@1 |
134
+ | • MBPPPlus | English | 0-shot | pass@1 |
135
+ | **Mathematical Reasoning** | | | |
136
+ | • GSM8k | English | 0-shot, CoT | exact-match |
137
+ | • MATH | English | 0-shot, CoT | exact-match |
138
+ | • GPQA | English | 4-shot | accuracy |
139
+ | HRM8k | Korean | 0-shot, CoT | exact-match |
140
+ | **Instruction Following and Chat** | | | |
141
+ | • IFEval | English | 0-shot | strict-average |
142
+ | • koIFEval | Korean | 0-shot | strict-average |
143
+ | • MT-Bench | English | LLM-as-a-judge (gpt-4o-2024-08-06) | LLM score |
144
+ | • KO-MT-Bench | Korean | LLM-as-a-judge (gpt-4o-2024-08-06) | LLM score |
145
+ | LogicKor | Korean | LLM-as-a-judge (gpt-4o-2024-08-06) | LLM score |
146
+ | • systemIFEval | English | 0-shot | strict-average |
147
+ | **Long Context** | | | |
148
+ | • KoRuler4K | Korean | 0-shot | exact-match |
149
+ | KoRuler8K | Korean | 0-shot | exact-match |
150
+ | • Ruler4K | English | 0-shot | exact-match |
151
+ | Ruler8K | English | 0-shot | exact-match |
152
+
153
+ - *Note that koIFEval, systemIFEval, and KoRuler are our in-house evaluation benchmarks adapted for Korean to better assess model capabilities in Korean language tasks.
154
+ - **Note that MT-Bench, KO-MT-Bench, and LogicKor use a 10-point scale.
155
+
156
+ </details>
157
+
158
+ ### Benchmark Results
159
+
160
+ Models compared:
161
+ - **Tri-7B** (Next Generation)
162
+ - **Trillion-7B-preview** (Previous Generation)
163
+
164
+ ### General Reasoning and Factuality
165
+
166
+ | Benchmark | Tri-7B | Trillion-7B-preview | Improvement |
167
+ | --- | --- | --- | --- |
168
+ | HAERAE | 82.49 | 80.02 | +2.47 |
169
+ | KoBEST | 82.72 | 79.61 | +3.11 |
170
+ | CLIcK | 64.43 | - | - |
171
+ | KMMLU | 51.74 | 48.09 | +3.65 |
172
+ | MMLU | 68.16 | 63.52 | +4.64 |
173
+
174
+ ### Coding
175
+
176
+ | Benchmark | Tri-7B | Trillion-7B-preview | Improvement |
177
+ | --- | --- | --- | --- |
178
+ | HumanEval | 53.66 | 55.48 | -1.82 |
179
+ | MBPPPlus | 64.29 | - | - |
180
+
181
+ ### Mathematical Reasoning
182
+
183
+ | Benchmark | Tri-7B | Trillion-7B-preview | Improvement |
184
+ | --- | --- | --- | --- |
185
+ | GSM8k | 77.94 | 72.25 | +5.69 |
186
+ | MATH | 49.40 | 32.70 | +16.70 |
187
+
188
+ ### Instruction Following and Chat
189
+
190
+ | Benchmark | Tri-7B | Trillion-7B-preview | Improvement |
191
+ | --- | --- | --- | --- |
192
+ | IFEval | 79.26 | 79.13 | +0.13 |
193
+ | koIFEval | 76.63 | 66.58 | +10.05 |
194
+ | MT-Bench | 7.82 | 6.53 | +1.29 |
195
+ | KO-MT-Bench | 7.64 | 6.27 | +1.37 |
196
+ | LogicKor | 8.20 | 8.14 | +0.06 |
197
+ | systemIFEval | 66.43 | - | - |
198
+
199
+ ## Limitations
200
+
201
+ - Language Support: The model is optimized for English, Korean, and Japanese. Usage with other languages may result in degraded performance.
202
+ - Knowledge Cutoff: The model's information is limited to data available up to Febuary, 2025.
203
+
204
+ ## License
205
+ This model is licensed under the Apache License 2.0.
206
+
207
+
208
+ ## Contact
209
+ For inquiries, please contact: info@trillionlabs.co
assets/Signiture_Black_White_BG_resized.jpg ADDED

Git LFS Details

  • SHA256: 7b26b8b82c0957d3b888c5c4d92b3c2cad5ae8084cf7ddd29deb3c74a75d84d6
  • Pointer size: 131 Bytes
  • Size of remote file: 303 kB
assets/Signiture_Trillion_BlackBG.png ADDED

Git LFS Details

  • SHA256: 1937cad1995db07ec65f853ef89cabd47beb8c641425a96bc9b050da6ddc9d19
  • Pointer size: 131 Bytes
  • Size of remote file: 717 kB
assets/Signiture_Trillion_WhiteBG.png ADDED

Git LFS Details

  • SHA256: e6fd124d4171fe2c050d129d5b525610b101257e2364de684076c3b024377caa
  • Pointer size: 131 Bytes
  • Size of remote file: 713 kB