Improve language tag

#2
by lbourdois - opened
Files changed (1) hide show
  1. README.md +202 -191
README.md CHANGED
@@ -1,191 +1,202 @@
1
- ---
2
- language:
3
- - th
4
- - en
5
- license: apache-2.0
6
- library_name: transformers
7
- base_model:
8
- - Qwen/Qwen2.5-7B-Instruct
9
- - Qwen/Qwen2.5-7B
10
- pipeline_tag: text-generation
11
- model-index:
12
- - name: Tsunami-0.5x-7B-Instruct
13
- results:
14
- - task:
15
- type: text-generation
16
- name: Text Generation
17
- dataset:
18
- name: IFEval (0-Shot)
19
- type: HuggingFaceH4/ifeval
20
- args:
21
- num_few_shot: 0
22
- metrics:
23
- - type: inst_level_strict_acc and prompt_level_strict_acc
24
- value: 70.99
25
- name: strict accuracy
26
- source:
27
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Tsunami-th/Tsunami-0.5x-7B-Instruct
28
- name: Open LLM Leaderboard
29
- - task:
30
- type: text-generation
31
- name: Text Generation
32
- dataset:
33
- name: BBH (3-Shot)
34
- type: BBH
35
- args:
36
- num_few_shot: 3
37
- metrics:
38
- - type: acc_norm
39
- value: 37.36
40
- name: normalized accuracy
41
- source:
42
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Tsunami-th/Tsunami-0.5x-7B-Instruct
43
- name: Open LLM Leaderboard
44
- - task:
45
- type: text-generation
46
- name: Text Generation
47
- dataset:
48
- name: MATH Lvl 5 (4-Shot)
49
- type: hendrycks/competition_math
50
- args:
51
- num_few_shot: 4
52
- metrics:
53
- - type: exact_match
54
- value: 4.83
55
- name: exact match
56
- source:
57
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Tsunami-th/Tsunami-0.5x-7B-Instruct
58
- name: Open LLM Leaderboard
59
- - task:
60
- type: text-generation
61
- name: Text Generation
62
- dataset:
63
- name: GPQA (0-shot)
64
- type: Idavidrein/gpqa
65
- args:
66
- num_few_shot: 0
67
- metrics:
68
- - type: acc_norm
69
- value: 8.61
70
- name: acc_norm
71
- source:
72
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Tsunami-th/Tsunami-0.5x-7B-Instruct
73
- name: Open LLM Leaderboard
74
- - task:
75
- type: text-generation
76
- name: Text Generation
77
- dataset:
78
- name: MuSR (0-shot)
79
- type: TAUR-Lab/MuSR
80
- args:
81
- num_few_shot: 0
82
- metrics:
83
- - type: acc_norm
84
- value: 18.57
85
- name: acc_norm
86
- source:
87
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Tsunami-th/Tsunami-0.5x-7B-Instruct
88
- name: Open LLM Leaderboard
89
- - task:
90
- type: text-generation
91
- name: Text Generation
92
- dataset:
93
- name: MMLU-PRO (5-shot)
94
- type: TIGER-Lab/MMLU-Pro
95
- config: main
96
- split: test
97
- args:
98
- num_few_shot: 5
99
- metrics:
100
- - type: acc
101
- value: 38.42
102
- name: accuracy
103
- source:
104
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Tsunami-th/Tsunami-0.5x-7B-Instruct
105
- name: Open LLM Leaderboard
106
- ---
107
-
108
- <img src="./Tsunami.webp" alt="Tsunami Model" width="800" style="margin-left:'auto' margin-right:'auto' display:'block'"/>
109
-
110
- # Tsunami-0.5x-7B-Instruct
111
- **TSUNAMI**: Transformative Semantic Understanding and Natural Augmentation Model for Intelligence.
112
-
113
- **TSUNAMI** full name was created by ChatGPT.
114
-
115
- ---
116
-
117
- ### infomation
118
- **Tsunami-0.5x-7B-Instruct** is Thai Large Language Model that fine-tuned from **Qwen2.5-7B** around **100,000** rows in Thai dataset.
119
-
120
- ---
121
-
122
- ### Prompt Template
123
-
124
- This model uses `ChatML` prompt template:
125
-
126
- ```
127
- <|im_start|>system
128
- {System}<|im_end|>
129
- <|im_start|>user
130
- {User}<|im_end|>
131
- <|im_start|>assistant
132
- {Assistant}
133
- ````
134
-
135
- ### How to use
136
-
137
-
138
- ```python
139
-
140
- from transformers import AutoModelForCausalLM, AutoTokenizer
141
- import torch
142
-
143
- model_name = "Tsunami-th/Tsunami-0.5x-7B-Instruct"
144
-
145
- model = AutoModelForCausalLM.from_pretrained(
146
- model_name,
147
- torch_dtype="auto",
148
- device_map="auto"
149
- )
150
- tokenizer = AutoTokenizer.from_pretrained(model_name)
151
-
152
- messages = [
153
- {"role": "system", "content": "You are a helpful assistant."},
154
- {"role": "user", "content": "สวัสดีครับ"}
155
- ]
156
- text = tokenizer.apply_chat_template(
157
- messages,
158
- tokenize=False,
159
- add_generation_prompt=True
160
- )
161
-
162
- inputs = tokenizer(text, return_tensors="pt")
163
- inputs = inputs.to(model.device)
164
- with torch.no_grad():
165
- output = model.generate(**inputs, max_new_tokens=512)
166
-
167
- response = tokenizer.decode(output[0, len(inputs['input_ids'][0]):], skip_special_tokens=True)
168
- ```
169
-
170
- ---
171
-
172
- ### Author
173
- - Pollakrit Lorprasertkul | game.pollakrit@gmail.com
174
-
175
- ---
176
-
177
- - **Tsunami-0.5x-7B-Instruct** is the version 0.5x that did not train on the whole dataset.
178
- - **Tsunami-1.0-7B-Instruct** is coming soon.
179
- # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
180
- Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Tsunami-th__Tsunami-0.5x-7B-Instruct)
181
-
182
- | Metric |Value|
183
- |-------------------|----:|
184
- |Avg. |29.80|
185
- |IFEval (0-Shot) |70.99|
186
- |BBH (3-Shot) |37.36|
187
- |MATH Lvl 5 (4-Shot)| 4.83|
188
- |GPQA (0-shot) | 8.61|
189
- |MuSR (0-shot) |18.57|
190
- |MMLU-PRO (5-shot) |38.42|
191
-
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - zho
4
+ - eng
5
+ - fra
6
+ - spa
7
+ - por
8
+ - deu
9
+ - ita
10
+ - rus
11
+ - jpn
12
+ - kor
13
+ - vie
14
+ - tha
15
+ - ara
16
+ license: apache-2.0
17
+ library_name: transformers
18
+ base_model:
19
+ - Qwen/Qwen2.5-7B-Instruct
20
+ - Qwen/Qwen2.5-7B
21
+ pipeline_tag: text-generation
22
+ model-index:
23
+ - name: Tsunami-0.5x-7B-Instruct
24
+ results:
25
+ - task:
26
+ type: text-generation
27
+ name: Text Generation
28
+ dataset:
29
+ name: IFEval (0-Shot)
30
+ type: HuggingFaceH4/ifeval
31
+ args:
32
+ num_few_shot: 0
33
+ metrics:
34
+ - type: inst_level_strict_acc and prompt_level_strict_acc
35
+ value: 70.99
36
+ name: strict accuracy
37
+ source:
38
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Tsunami-th/Tsunami-0.5x-7B-Instruct
39
+ name: Open LLM Leaderboard
40
+ - task:
41
+ type: text-generation
42
+ name: Text Generation
43
+ dataset:
44
+ name: BBH (3-Shot)
45
+ type: BBH
46
+ args:
47
+ num_few_shot: 3
48
+ metrics:
49
+ - type: acc_norm
50
+ value: 37.36
51
+ name: normalized accuracy
52
+ source:
53
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Tsunami-th/Tsunami-0.5x-7B-Instruct
54
+ name: Open LLM Leaderboard
55
+ - task:
56
+ type: text-generation
57
+ name: Text Generation
58
+ dataset:
59
+ name: MATH Lvl 5 (4-Shot)
60
+ type: hendrycks/competition_math
61
+ args:
62
+ num_few_shot: 4
63
+ metrics:
64
+ - type: exact_match
65
+ value: 4.83
66
+ name: exact match
67
+ source:
68
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Tsunami-th/Tsunami-0.5x-7B-Instruct
69
+ name: Open LLM Leaderboard
70
+ - task:
71
+ type: text-generation
72
+ name: Text Generation
73
+ dataset:
74
+ name: GPQA (0-shot)
75
+ type: Idavidrein/gpqa
76
+ args:
77
+ num_few_shot: 0
78
+ metrics:
79
+ - type: acc_norm
80
+ value: 8.61
81
+ name: acc_norm
82
+ source:
83
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Tsunami-th/Tsunami-0.5x-7B-Instruct
84
+ name: Open LLM Leaderboard
85
+ - task:
86
+ type: text-generation
87
+ name: Text Generation
88
+ dataset:
89
+ name: MuSR (0-shot)
90
+ type: TAUR-Lab/MuSR
91
+ args:
92
+ num_few_shot: 0
93
+ metrics:
94
+ - type: acc_norm
95
+ value: 18.57
96
+ name: acc_norm
97
+ source:
98
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Tsunami-th/Tsunami-0.5x-7B-Instruct
99
+ name: Open LLM Leaderboard
100
+ - task:
101
+ type: text-generation
102
+ name: Text Generation
103
+ dataset:
104
+ name: MMLU-PRO (5-shot)
105
+ type: TIGER-Lab/MMLU-Pro
106
+ config: main
107
+ split: test
108
+ args:
109
+ num_few_shot: 5
110
+ metrics:
111
+ - type: acc
112
+ value: 38.42
113
+ name: accuracy
114
+ source:
115
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Tsunami-th/Tsunami-0.5x-7B-Instruct
116
+ name: Open LLM Leaderboard
117
+ ---
118
+
119
+ <img src="./Tsunami.webp" alt="Tsunami Model" width="800" style="margin-left:'auto' margin-right:'auto' display:'block'"/>
120
+
121
+ # Tsunami-0.5x-7B-Instruct
122
+ **TSUNAMI**: Transformative Semantic Understanding and Natural Augmentation Model for Intelligence.
123
+
124
+ **TSUNAMI** full name was created by ChatGPT.
125
+
126
+ ---
127
+
128
+ ### infomation
129
+ **Tsunami-0.5x-7B-Instruct** is Thai Large Language Model that fine-tuned from **Qwen2.5-7B** around **100,000** rows in Thai dataset.
130
+
131
+ ---
132
+
133
+ ### Prompt Template
134
+
135
+ This model uses `ChatML` prompt template:
136
+
137
+ ```
138
+ <|im_start|>system
139
+ {System}<|im_end|>
140
+ <|im_start|>user
141
+ {User}<|im_end|>
142
+ <|im_start|>assistant
143
+ {Assistant}
144
+ ````
145
+
146
+ ### How to use
147
+
148
+
149
+ ```python
150
+
151
+ from transformers import AutoModelForCausalLM, AutoTokenizer
152
+ import torch
153
+
154
+ model_name = "Tsunami-th/Tsunami-0.5x-7B-Instruct"
155
+
156
+ model = AutoModelForCausalLM.from_pretrained(
157
+ model_name,
158
+ torch_dtype="auto",
159
+ device_map="auto"
160
+ )
161
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
162
+
163
+ messages = [
164
+ {"role": "system", "content": "You are a helpful assistant."},
165
+ {"role": "user", "content": "สวัสดีครับ"}
166
+ ]
167
+ text = tokenizer.apply_chat_template(
168
+ messages,
169
+ tokenize=False,
170
+ add_generation_prompt=True
171
+ )
172
+
173
+ inputs = tokenizer(text, return_tensors="pt")
174
+ inputs = inputs.to(model.device)
175
+ with torch.no_grad():
176
+ output = model.generate(**inputs, max_new_tokens=512)
177
+
178
+ response = tokenizer.decode(output[0, len(inputs['input_ids'][0]):], skip_special_tokens=True)
179
+ ```
180
+
181
+ ---
182
+
183
+ ### Author
184
+ - Pollakrit Lorprasertkul | game.pollakrit@gmail.com
185
+
186
+ ---
187
+
188
+ - **Tsunami-0.5x-7B-Instruct** is the version 0.5x that did not train on the whole dataset.
189
+ - **Tsunami-1.0-7B-Instruct** is coming soon.
190
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
191
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Tsunami-th__Tsunami-0.5x-7B-Instruct)
192
+
193
+ | Metric |Value|
194
+ |-------------------|----:|
195
+ |Avg. |29.80|
196
+ |IFEval (0-Shot) |70.99|
197
+ |BBH (3-Shot) |37.36|
198
+ |MATH Lvl 5 (4-Shot)| 4.83|
199
+ |GPQA (0-shot) | 8.61|
200
+ |MuSR (0-shot) |18.57|
201
+ |MMLU-PRO (5-shot) |38.42|
202
+