Text Generation
Transformers
Safetensors
qwen2
conversational
text-generation-inference

Improve language tag

#7
by lbourdois - opened
Files changed (1) hide show
  1. README.md +56 -81
README.md CHANGED
@@ -1,82 +1,57 @@
1
- ---
2
- license: mit
3
- language:
4
- - am
5
- - ar
6
- - bn
7
- - zh
8
- - cs
9
- - nl
10
- - en
11
- - fr
12
- - de
13
- - el
14
- - ha
15
- - he
16
- - hi
17
- - id
18
- - it
19
- - ja
20
- - jv
21
- - km
22
- - ko
23
- - lo
24
- - ms
25
- - mr
26
- - fa
27
- - pl
28
- - pt
29
- - ro
30
- - ru
31
- - es
32
- - sw
33
- - sv
34
- - tl
35
- - ta
36
- - te
37
- - th
38
- - tr
39
- - uk
40
- - ur
41
- - vi
42
- datasets:
43
- - simplescaling/s1K
44
- - lightblue/reasoning-multilingual-R1-Llama-70B-train
45
- base_model:
46
- - Qwen/Qwen2.5-1.5B-Instruct
47
- library_name: transformers
48
- ---
49
- It's a 1.5B model.
50
-
51
- It's a distill model like s1 and deepseek-r1-distill.
52
-
53
- It's test model. I hope I can reproduce a rl model like RL-Zero.
54
-
55
- This model is a mini-step.
56
-
57
- Thanks for evveryone in the open community.
58
-
59
- how to use:
60
- ```
61
- from vllm import LLM, SamplingParams
62
- from transformers import AutoTokenizer
63
-
64
- model = LLM(
65
- "Amu/t1-1.5B"
66
- )
67
- tok = AutoTokenizer.from_pretrained("simplescaling/s1-32B")
68
-
69
- stop_token_ids = tok("<|im_end|>")["input_ids"]
70
-
71
- sampling_params = SamplingParams(
72
- max_tokens=32768,
73
- min_tokens=0,
74
- stop_token_ids=stop_token_ids,
75
- )
76
-
77
- prompt = "How many r in raspberry"
78
- prompt = "<|im_start|>system\nYou are t1, created by Amu. You are a helpful assistant.<|im_end|>\n<|im_start|>user\n" + prompt + "<|im_end|>\n<|im_start|>assistant\n"
79
-
80
- o = model.generate(prompt, sampling_params=sampling_params)
81
- print(o[0].outputs[0].text)
82
  ```
 
1
+ ---
2
+ license: mit
3
+ language:
4
+ - zho
5
+ - eng
6
+ - fra
7
+ - spa
8
+ - por
9
+ - deu
10
+ - ita
11
+ - rus
12
+ - jpn
13
+ - kor
14
+ - vie
15
+ - tha
16
+ - ara
17
+ datasets:
18
+ - simplescaling/s1K
19
+ - lightblue/reasoning-multilingual-R1-Llama-70B-train
20
+ base_model:
21
+ - Qwen/Qwen2.5-1.5B-Instruct
22
+ library_name: transformers
23
+ ---
24
+ It's a 1.5B model.
25
+
26
+ It's a distill model like s1 and deepseek-r1-distill.
27
+
28
+ It's test model. I hope I can reproduce a rl model like RL-Zero.
29
+
30
+ This model is a mini-step.
31
+
32
+ Thanks for evveryone in the open community.
33
+
34
+ how to use:
35
+ ```
36
+ from vllm import LLM, SamplingParams
37
+ from transformers import AutoTokenizer
38
+
39
+ model = LLM(
40
+ "Amu/t1-1.5B"
41
+ )
42
+ tok = AutoTokenizer.from_pretrained("simplescaling/s1-32B")
43
+
44
+ stop_token_ids = tok("<|im_end|>")["input_ids"]
45
+
46
+ sampling_params = SamplingParams(
47
+ max_tokens=32768,
48
+ min_tokens=0,
49
+ stop_token_ids=stop_token_ids,
50
+ )
51
+
52
+ prompt = "How many r in raspberry"
53
+ prompt = "<|im_start|>system\nYou are t1, created by Amu. You are a helpful assistant.<|im_end|>\n<|im_start|>user\n" + prompt + "<|im_end|>\n<|im_start|>assistant\n"
54
+
55
+ o = model.generate(prompt, sampling_params=sampling_params)
56
+ print(o[0].outputs[0].text)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
57
  ```