lbourdois commited on
Commit
ad9f16e
·
verified ·
1 Parent(s): ed320c1

Improve language tag

Browse files

Hi! As the model is multilingual, this is a PR to add other languages than English to the language tag to improve the referencing. Note that 29 languages are announced in the README, but only 13 are explicitly listed. I was therefore only able to add these 13 languages.

Files changed (1) hide show
  1. README.md +99 -88
README.md CHANGED
@@ -1,89 +1,100 @@
1
- ---
2
- language:
3
- - th
4
- - en
5
- license: apache-2.0
6
- library_name: transformers
7
- base_model:
8
- - Qwen/Qwen2.5-14B-Instruct
9
- - Qwen/Qwen2.5-14B
10
- pipeline_tag: text-generation
11
- ---
12
- <img src="./Tsunami.webp" alt="Tsunami Model" width="800" style="margin-left:'auto' margin-right:'auto' display:'block'"/>
13
-
14
- # Tsunami-1.0-14B-Instruct
15
- **TSUNAMI**: Transformative Semantic Understanding and Natural Augmentation Model for Intelligence.
16
-
17
- **TSUNAMI** full name was created by ChatGPT.
18
-
19
- ---
20
-
21
- ### infomation
22
- **Tsunami-1.0-14B-Instruct** is Thai Large Language Model that fine-tuned from **Qwen2.5-14B** in Thai dataset.
23
-
24
- ---
25
-
26
- ### Author
27
- - Pollakrit Lorprasertkul | game.pollakrit@gmail.com
28
-
29
- ---
30
-
31
- ### Performance Evaluation
32
-
33
- Below are the benchmark results of **Tsunami-1.0-14B-Instruct** compared to similar models in its class:
34
-
35
- | Model | Average | Thai Exam | M3Exam |
36
- | --- | --- | --- | --- |
37
- | Qwen2.5-14B-Instruct | 58.45 | 57.35 | 59.55 |
38
- | Meta-Llama-3.1-70B-Instruct | 59.38 | 58.23 | 60.52 |
39
- | llama-3-typhoon-v1.5x-70b-instruct | 59.34 | 58.76 | 59.92 |
40
- | openthaigpt1.5-14b-instruct | 60.41 | 58.41 | 62.41 |
41
- | **Tsunami-1.0-14B-Instruct** | **62.05** | **61.06** | **63.05** |
42
-
43
- ---
44
-
45
- ### Prompt Template
46
-
47
- This model uses `ChatML` prompt template:
48
-
49
- ```
50
- <|im_start|>system
51
- {System}<|im_end|>
52
- <|im_start|>user
53
- {User}<|im_end|>
54
- <|im_start|>assistant
55
- {Assistant}
56
- ````
57
-
58
- ---
59
-
60
- ### How to use
61
-
62
-
63
- ```python
64
- from transformers import AutoModelForCausalLM, AutoTokenizer
65
- import torch
66
- model_name = "Tsunami-th/Tsunami-1.0-14B-Instruct"
67
- model = AutoModelForCausalLM.from_pretrained(
68
- model_name,
69
- torch_dtype="auto",
70
- device_map="auto"
71
- )
72
- tokenizer = AutoTokenizer.from_pretrained(model_name)
73
- messages = [
74
- {"role": "system", "content": "You are a helpful assistant."},
75
- {"role": "user", "content": "สวัสดีครับ"}
76
- ]
77
- text = tokenizer.apply_chat_template(
78
- messages,
79
- tokenize=False,
80
- add_generation_prompt=True
81
- )
82
- inputs = tokenizer(text, return_tensors="pt")
83
- inputs = inputs.to(model.device)
84
- with torch.no_grad():
85
- output = model.generate(**inputs, max_new_tokens=512)
86
- response = tokenizer.decode(output[0, len(inputs['input_ids'][0]):], skip_special_tokens=True)
87
- ```
88
-
 
 
 
 
 
 
 
 
 
 
 
89
  ---
 
1
+ ---
2
+ language:
3
+ - zho
4
+ - eng
5
+ - fra
6
+ - spa
7
+ - por
8
+ - deu
9
+ - ita
10
+ - rus
11
+ - jpn
12
+ - kor
13
+ - vie
14
+ - tha
15
+ - ara
16
+ license: apache-2.0
17
+ library_name: transformers
18
+ base_model:
19
+ - Qwen/Qwen2.5-14B-Instruct
20
+ - Qwen/Qwen2.5-14B
21
+ pipeline_tag: text-generation
22
+ ---
23
+ <img src="./Tsunami.webp" alt="Tsunami Model" width="800" style="margin-left:'auto' margin-right:'auto' display:'block'"/>
24
+
25
+ # Tsunami-1.0-14B-Instruct
26
+ **TSUNAMI**: Transformative Semantic Understanding and Natural Augmentation Model for Intelligence.
27
+
28
+ **TSUNAMI** full name was created by ChatGPT.
29
+
30
+ ---
31
+
32
+ ### infomation
33
+ **Tsunami-1.0-14B-Instruct** is Thai Large Language Model that fine-tuned from **Qwen2.5-14B** in Thai dataset.
34
+
35
+ ---
36
+
37
+ ### Author
38
+ - Pollakrit Lorprasertkul | game.pollakrit@gmail.com
39
+
40
+ ---
41
+
42
+ ### Performance Evaluation
43
+
44
+ Below are the benchmark results of **Tsunami-1.0-14B-Instruct** compared to similar models in its class:
45
+
46
+ | Model | Average | Thai Exam | M3Exam |
47
+ | --- | --- | --- | --- |
48
+ | Qwen2.5-14B-Instruct | 58.45 | 57.35 | 59.55 |
49
+ | Meta-Llama-3.1-70B-Instruct | 59.38 | 58.23 | 60.52 |
50
+ | llama-3-typhoon-v1.5x-70b-instruct | 59.34 | 58.76 | 59.92 |
51
+ | openthaigpt1.5-14b-instruct | 60.41 | 58.41 | 62.41 |
52
+ | **Tsunami-1.0-14B-Instruct** | **62.05** | **61.06** | **63.05** |
53
+
54
+ ---
55
+
56
+ ### Prompt Template
57
+
58
+ This model uses `ChatML` prompt template:
59
+
60
+ ```
61
+ <|im_start|>system
62
+ {System}<|im_end|>
63
+ <|im_start|>user
64
+ {User}<|im_end|>
65
+ <|im_start|>assistant
66
+ {Assistant}
67
+ ````
68
+
69
+ ---
70
+
71
+ ### How to use
72
+
73
+
74
+ ```python
75
+ from transformers import AutoModelForCausalLM, AutoTokenizer
76
+ import torch
77
+ model_name = "Tsunami-th/Tsunami-1.0-14B-Instruct"
78
+ model = AutoModelForCausalLM.from_pretrained(
79
+ model_name,
80
+ torch_dtype="auto",
81
+ device_map="auto"
82
+ )
83
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
84
+ messages = [
85
+ {"role": "system", "content": "You are a helpful assistant."},
86
+ {"role": "user", "content": "สวัสดีครับ"}
87
+ ]
88
+ text = tokenizer.apply_chat_template(
89
+ messages,
90
+ tokenize=False,
91
+ add_generation_prompt=True
92
+ )
93
+ inputs = tokenizer(text, return_tensors="pt")
94
+ inputs = inputs.to(model.device)
95
+ with torch.no_grad():
96
+ output = model.generate(**inputs, max_new_tokens=512)
97
+ response = tokenizer.decode(output[0, len(inputs['input_ids'][0]):], skip_special_tokens=True)
98
+ ```
99
+
100
  ---