lbourdois commited on
Commit
26ed770
·
verified ·
1 Parent(s): 190f9dd

Improve language tag

Browse files

Hi! As the model is multilingual, this is a PR to add other languages than English to the language tag to improve the referencing. Note that 29 languages are announced in the README, but only 13 are explicitly listed. I was therefore only able to add these 13 languages.

Files changed (1) hide show
  1. README.md +78 -66
README.md CHANGED
@@ -1,67 +1,79 @@
1
- ---
2
- license: mit
3
- language:
4
- - en
5
- base_model:
6
- - Qwen/Qwen2.5-32B-Instruct
7
- pipeline_tag: text-generation
8
- ---
9
-
10
- # Apollo Model
11
-
12
- This is an experimental hybrid reasoning model built on Qwen2.5-32B-Instruct
13
-
14
- # GGUF
15
-
16
- mradermacher/Apollo-v3-32B-GGUF
17
-
18
- thanks mradermacher for this gguf
19
-
20
- ### Merge Method
21
-
22
- This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [Qwen/Qwen2.5-32B-Instruct](https://huggingface.co/Qwen/Qwen2.5-32B-Instruct) as a base.
23
-
24
-
25
- ### Enable reasoning
26
-
27
- prompt the LLM with think deeper and step by step
28
-
29
- ### Example code
30
-
31
- ```
32
-
33
- from transformers import AutoModelForCausalLM, AutoTokenizer
34
-
35
- model_name = "rootxhacker/Apollo-v3-32B"
36
-
37
- model = AutoModelForCausalLM.from_pretrained(
38
- model_name,
39
- torch_dtype="auto",
40
- device_map="auto"
41
- )
42
- tokenizer = AutoTokenizer.from_pretrained(model_name)
43
-
44
- prompt = "How many r's are in the word strawberry"
45
- messages = [
46
- {"role": "user", "content": prompt}
47
- ]
48
- text = tokenizer.apply_chat_template(
49
- messages,
50
- tokenize=False,
51
- add_generation_prompt=True
52
- )
53
-
54
- model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
55
-
56
- generated_ids = model.generate(
57
- **model_inputs,
58
- max_new_tokens=32768
59
- )
60
- generated_ids = [
61
- output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
62
- ]
63
-
64
- response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
65
- print(response)
66
-
 
 
 
 
 
 
 
 
 
 
 
 
67
  ```
 
1
+ ---
2
+ license: mit
3
+ language:
4
+ - zho
5
+ - eng
6
+ - fra
7
+ - spa
8
+ - por
9
+ - deu
10
+ - ita
11
+ - rus
12
+ - jpn
13
+ - kor
14
+ - vie
15
+ - tha
16
+ - ara
17
+ base_model:
18
+ - Qwen/Qwen2.5-32B-Instruct
19
+ pipeline_tag: text-generation
20
+ ---
21
+
22
+ # Apollo Model
23
+
24
+ This is an experimental hybrid reasoning model built on Qwen2.5-32B-Instruct
25
+
26
+ # GGUF
27
+
28
+ mradermacher/Apollo-v3-32B-GGUF
29
+
30
+ thanks mradermacher for this gguf
31
+
32
+ ### Merge Method
33
+
34
+ This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [Qwen/Qwen2.5-32B-Instruct](https://huggingface.co/Qwen/Qwen2.5-32B-Instruct) as a base.
35
+
36
+
37
+ ### Enable reasoning
38
+
39
+ prompt the LLM with think deeper and step by step
40
+
41
+ ### Example code
42
+
43
+ ```
44
+
45
+ from transformers import AutoModelForCausalLM, AutoTokenizer
46
+
47
+ model_name = "rootxhacker/Apollo-v3-32B"
48
+
49
+ model = AutoModelForCausalLM.from_pretrained(
50
+ model_name,
51
+ torch_dtype="auto",
52
+ device_map="auto"
53
+ )
54
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
55
+
56
+ prompt = "How many r's are in the word strawberry"
57
+ messages = [
58
+ {"role": "user", "content": prompt}
59
+ ]
60
+ text = tokenizer.apply_chat_template(
61
+ messages,
62
+ tokenize=False,
63
+ add_generation_prompt=True
64
+ )
65
+
66
+ model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
67
+
68
+ generated_ids = model.generate(
69
+ **model_inputs,
70
+ max_new_tokens=32768
71
+ )
72
+ generated_ids = [
73
+ output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
74
+ ]
75
+
76
+ response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
77
+ print(response)
78
+
79
  ```