lbourdois commited on
Commit
e58d725
Β·
verified Β·
1 Parent(s): 2c6934a

Improve language tag

Browse files

Hi! As the model is multilingual, this is a PR to add other languages than English to the language tag to improve the referencing. Note that 29 languages are announced in the README, but only 13 are explicitly listed. I was therefore only able to add these 13 languages.

Files changed (1) hide show
  1. README.md +111 -99
README.md CHANGED
@@ -1,99 +1,111 @@
1
- ---
2
- license: apache-2.0
3
- language:
4
- - ko
5
- base_model:
6
- - Qwen/Qwen2.5-14B-Instruct
7
- ---
8
-
9
- # Announcing OLAFv2: The Next Step in Korean Language Understanding πŸš€
10
-
11
- We are thrilled to announce the release of **OLAFv2**, our state-of-the-art Korean language model, now available on Hugging Face! πŸŽ‰ Designed to excel in complex reasoning, mathematical problem-solving, and general language understanding, OLAFv2 represents a significant leap forward in NLP capabilities for the Korean language.
12
-
13
-
14
- ![image/png](https://cdn-uploads.huggingface.co/production/uploads/650c0029987b1ae4e51fa2d4/KxAsxe10pZkaqC6x82qH3.png)
15
-
16
- ## Key Features of OLAFv2 🌟
17
-
18
- ### **Two Model Sizes for Flexibility**
19
- OLAFv2 is available in two parameter sizes:
20
- - **14B (Billion) Parameters**: For maximum performance. πŸ‹οΈβ€β™‚οΈ
21
- - **1.5B (Billion) Parameters**: For lightweight applications and hardware-constrained environments. πŸͺΆ
22
-
23
- ### **Reasoning Mode for Complex Tasks** πŸ€”
24
- One of OLAFv2's standout features is its **Reasoning Mode**, specifically designed for:
25
- - Complex mathematical problem-solving. βœ–οΈβž—
26
- - STEM (Science, Technology, Engineering, Mathematics) applications. πŸ”¬πŸ“
27
- - Tasks requiring detailed step-by-step reasoning. 🧠
28
-
29
- This mode can be effectively utilized for **Test-Time Scaling**, enabling the model to harness additional computational resources during inference. This approach enhances output detail and accuracy, achieving performance levels that surpass GPT-4o. πŸ“ˆ
30
-
31
-
32
- ![image/png](https://cdn-uploads.huggingface.co/production/uploads/650c0029987b1ae4e51fa2d4/aZlD94ZAqxePTaGdb4TQ8.png)
33
-
34
-
35
- ### **Long Context Support** πŸ“œ
36
- With support for up to **32K tokens**, OLAFv2 is perfect for:
37
- - Retrieval-Augmented Generation (RAG). πŸ› οΈ
38
- - Tasks requiring long-context understanding and reasoning. 🧡
39
-
40
- ## Benchmarks and Performance πŸ“Š
41
-
42
- We share evaluation results across three benchmarks, KMMLU, HRM8K and LogicKor.
43
-
44
- <div style="text-align: center;">
45
- <img src="https://cdn-uploads.huggingface.co/production/uploads/650c0029987b1ae4e51fa2d4/rCloMEgq16D8-UuCkM8Pa.png" width="700px" height="450px" title="polyglot_budget" alt="polyglot_budget"/>
46
- </div>
47
-
48
-
49
- We also share results with inference-time scaling. For more details have a look into our [blog](https://www.onelineai.com/blog/test-time-scaling).
50
-
51
- <!-- ![alt-text-1](https://cdn-uploads.huggingface.co/production/uploads/650c0029987b1ae4e51fa2d4/Qa2-91s0nvwIsx0cjRM-W.png "title-1") ![alt-text-2](https://cdn-uploads.huggingface.co/production/uploads/650c0029987b1ae4e51fa2d4/8eATw4sMb-OrgqhRujR0z.png "title-2") -->
52
-
53
- <div style="display: flex; justify-content: space-between; align-items: center;">
54
- <img src="https://cdn-uploads.huggingface.co/production/uploads/650c0029987b1ae4e51fa2d4/Qa2-91s0nvwIsx0cjRM-W.png" alt="alt-text-1" title="title-1" style="width: 48%;"/>
55
- <img src="https://cdn-uploads.huggingface.co/production/uploads/650c0029987b1ae4e51fa2d4/8eATw4sMb-OrgqhRujR0z.png" alt="alt-text-2" title="title-2" style="width: 48%;"/>
56
- </div>
57
-
58
-
59
- ## Getting Started πŸš€
60
- OLAFv2 is now available on Hugging Face! You can start using it by accessing our repository:
61
-
62
- ```python
63
- # pip install transformers
64
- from transformers import AutoModelForCausalLM, AutoTokenizer
65
-
66
- model_name = "OLAResearch/OLAF2-14B"
67
-
68
- model = AutoModelForCausalLM.from_pretrained(
69
- model_name,
70
- torch_dtype="auto",
71
- device_map="auto"
72
- )
73
- tokenizer = AutoTokenizer.from_pretrained(model_name)
74
-
75
- prompt = "introduce yourself!"
76
- messages = [
77
- {"role": "system", "content": "You're name is OLAF. A large language model made by OneLineAI, specializing in Korean culture and finance."},
78
- # for reasoning mode
79
- #{"role": "system", "content": "You're name is OLAF. A large language model made by OneLineAI, specializing in Korean culture and finance.Perform two-step reasoning. Return your answers in \\boxed{N} format."},
80
- {"role": "user", "content": prompt}
81
- ]
82
- text = tokenizer.apply_chat_template(
83
- messages,
84
- tokenize=False,
85
- add_generation_prompt=True
86
- )
87
- model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
88
-
89
- generated_ids = model.generate(
90
- **model_inputs,
91
- max_new_tokens=512
92
- )
93
- generated_ids = [
94
- output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
95
- ]
96
-
97
- response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
98
- ```
99
-
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - zho
5
+ - eng
6
+ - fra
7
+ - spa
8
+ - por
9
+ - deu
10
+ - ita
11
+ - rus
12
+ - jpn
13
+ - kor
14
+ - vie
15
+ - tha
16
+ - ara
17
+ base_model:
18
+ - Qwen/Qwen2.5-14B-Instruct
19
+ ---
20
+
21
+ # Announcing OLAFv2: The Next Step in Korean Language Understanding πŸš€
22
+
23
+ We are thrilled to announce the release of **OLAFv2**, our state-of-the-art Korean language model, now available on Hugging Face! πŸŽ‰ Designed to excel in complex reasoning, mathematical problem-solving, and general language understanding, OLAFv2 represents a significant leap forward in NLP capabilities for the Korean language.
24
+
25
+
26
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/650c0029987b1ae4e51fa2d4/KxAsxe10pZkaqC6x82qH3.png)
27
+
28
+ ## Key Features of OLAFv2 🌟
29
+
30
+ ### **Two Model Sizes for Flexibility**
31
+ OLAFv2 is available in two parameter sizes:
32
+ - **14B (Billion) Parameters**: For maximum performance. πŸ‹οΈβ€β™‚οΈ
33
+ - **1.5B (Billion) Parameters**: For lightweight applications and hardware-constrained environments. πŸͺΆ
34
+
35
+ ### **Reasoning Mode for Complex Tasks** πŸ€”
36
+ One of OLAFv2's standout features is its **Reasoning Mode**, specifically designed for:
37
+ - Complex mathematical problem-solving. βœ–οΈβž—
38
+ - STEM (Science, Technology, Engineering, Mathematics) applications. πŸ”¬πŸ“
39
+ - Tasks requiring detailed step-by-step reasoning. 🧠
40
+
41
+ This mode can be effectively utilized for **Test-Time Scaling**, enabling the model to harness additional computational resources during inference. This approach enhances output detail and accuracy, achieving performance levels that surpass GPT-4o. πŸ“ˆ
42
+
43
+
44
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/650c0029987b1ae4e51fa2d4/aZlD94ZAqxePTaGdb4TQ8.png)
45
+
46
+
47
+ ### **Long Context Support** πŸ“œ
48
+ With support for up to **32K tokens**, OLAFv2 is perfect for:
49
+ - Retrieval-Augmented Generation (RAG). πŸ› οΈ
50
+ - Tasks requiring long-context understanding and reasoning. 🧡
51
+
52
+ ## Benchmarks and Performance πŸ“Š
53
+
54
+ We share evaluation results across three benchmarks, KMMLU, HRM8K and LogicKor.
55
+
56
+ <div style="text-align: center;">
57
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/650c0029987b1ae4e51fa2d4/rCloMEgq16D8-UuCkM8Pa.png" width="700px" height="450px" title="polyglot_budget" alt="polyglot_budget"/>
58
+ </div>
59
+
60
+
61
+ We also share results with inference-time scaling. For more details have a look into our [blog](https://www.onelineai.com/blog/test-time-scaling).
62
+
63
+ <!-- ![alt-text-1](https://cdn-uploads.huggingface.co/production/uploads/650c0029987b1ae4e51fa2d4/Qa2-91s0nvwIsx0cjRM-W.png "title-1") ![alt-text-2](https://cdn-uploads.huggingface.co/production/uploads/650c0029987b1ae4e51fa2d4/8eATw4sMb-OrgqhRujR0z.png "title-2") -->
64
+
65
+ <div style="display: flex; justify-content: space-between; align-items: center;">
66
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/650c0029987b1ae4e51fa2d4/Qa2-91s0nvwIsx0cjRM-W.png" alt="alt-text-1" title="title-1" style="width: 48%;"/>
67
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/650c0029987b1ae4e51fa2d4/8eATw4sMb-OrgqhRujR0z.png" alt="alt-text-2" title="title-2" style="width: 48%;"/>
68
+ </div>
69
+
70
+
71
+ ## Getting Started πŸš€
72
+ OLAFv2 is now available on Hugging Face! You can start using it by accessing our repository:
73
+
74
+ ```python
75
+ # pip install transformers
76
+ from transformers import AutoModelForCausalLM, AutoTokenizer
77
+
78
+ model_name = "OLAResearch/OLAF2-14B"
79
+
80
+ model = AutoModelForCausalLM.from_pretrained(
81
+ model_name,
82
+ torch_dtype="auto",
83
+ device_map="auto"
84
+ )
85
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
86
+
87
+ prompt = "introduce yourself!"
88
+ messages = [
89
+ {"role": "system", "content": "You're name is OLAF. A large language model made by OneLineAI, specializing in Korean culture and finance."},
90
+ # for reasoning mode
91
+ #{"role": "system", "content": "You're name is OLAF. A large language model made by OneLineAI, specializing in Korean culture and finance.Perform two-step reasoning. Return your answers in \\boxed{N} format."},
92
+ {"role": "user", "content": prompt}
93
+ ]
94
+ text = tokenizer.apply_chat_template(
95
+ messages,
96
+ tokenize=False,
97
+ add_generation_prompt=True
98
+ )
99
+ model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
100
+
101
+ generated_ids = model.generate(
102
+ **model_inputs,
103
+ max_new_tokens=512
104
+ )
105
+ generated_ids = [
106
+ output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
107
+ ]
108
+
109
+ response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
110
+ ```
111
+