lbourdois commited on
Commit
29ebcaf
·
verified ·
1 Parent(s): 96a3bdc

Improve language tag

Browse files

Hi! As the model is multilingual, this is a PR to add other languages than English to the language tag to improve the referencing. Note that 29 languages are announced in the README, but only 13 are explicitly listed. I was therefore only able to add these 13 languages.

Files changed (1) hide show
  1. README.md +88 -74
README.md CHANGED
@@ -1,75 +1,89 @@
1
- ---
2
- base_model:
3
- - Qwen/Qwen2.5-72B
4
- tags:
5
- - roleplay
6
- - storywriting
7
- - qwen2.5
8
- - finetune
9
- - transformers
10
- - pytorch
11
- ---
12
-
13
- # Zeus Labs ~ Chronos-Platinum-72B
14
-
15
- ![image/png](https://cdn-uploads.huggingface.co/production/uploads/630417380907b9a115c6aa9f/G05mAhqcp4S_WBfE2vBLl.png)
16
-
17
- Qwen 2.5 72B base model, trained for two epochs on the Chronos Divergence dataset using ChatML. It works well for roleplaying and storywriting as well as general assistant tasks.
18
-
19
- ## Instruct Template
20
-
21
- This model uses `ChatML` - below is an example. It is a preset in many frontends.
22
-
23
- ```
24
- <|im_start|>system
25
- You are a helpful assistant<|im_end|>
26
- <|im_start|>user
27
- Hello there!<|im_end|>
28
- <|im_start|>assistant
29
- Hi! I'm an AI assistant, designed to help people like you with all sorts of tasks. Is there anything you need help with?<|im_end|>
30
- <|im_start|>user
31
- I was wondering how transformers work?<|im_end|>
32
- <|im_start|>assistant
33
- ```
34
-
35
- ## Quantizations
36
- Please note that we tested this model with a 5.0bpw EXL2 quant. Results are not expected to be the same when going below this quanitzation. Thanks to our model quanters!
37
-
38
- #### LlamaCPP (GGUF)
39
- [bartowski](https://huggingface.co/bartowski/Chronos-Platinum-72B-GGUF)
40
-
41
- [mradermacher](https://huggingface.co/mradermacher/Chronos-Platinum-72B-i1-GGUF)
42
-
43
- #### Exllama2
44
- [bartowski](https://huggingface.co/bartowski/Chronos-Platinum-72B-exl2)
45
-
46
- ## Sampling Settings
47
- Here are some settings that work well with this model:
48
- ```
49
- Temp -> 0.7 - 1.2
50
- Min P -> 0.025 - 0.05 [temp in order, not last]
51
- Presence Penalty -> 1.0
52
- Repetition Penalty range -> 4000
53
- ```
54
- Higher temp gives more uniqueness and less repetition. Please do not take these settings as the "best" - your system prompt matters significantly, and if you're roleplaying
55
- use the Basic system prompt in SillyTavern. You can also try other samplers like Top P.
56
-
57
- **Note that Presence Penalty works with Repetition Penalty Range.**
58
-
59
- ## Credit
60
- Thank you to my team consisting of [@ToastyPigeon](https://huggingface.co/ToastyPigeon), [@Fizzarolli](https://huggingface.co/Fizzarolli), and myself [@elinas](https://huggingface.co/elinas).
61
-
62
- Additional thanks to [@AlpinDale](https://huggingface.co/AlpinDale) and the rest of the PygmalionAI team for graciously providing the compute to finetune this model!
63
- Thank you to [anthracite-org](https://huggingface.co/anthracite-org) as well for sponsoring this model.
64
-
65
- ## Additional Details 
66
-
67
- We used a combination of provided logs and WizardLM evol both cleaned up and de-slopped.
68
-
69
- Thanks to Anthropic and OpenAI for the models used to generate synthetic and partially synthetic data to train this model.
70
-
71
- Thanks Elon Musk for being based enough to train AI that compares to the top models.
72
-
73
- If you have any questions or concerns, please post in the community tab.
74
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
75
  DISCLAIMER: Outputs generated by the model are not reflective of our views.
 
1
+ ---
2
+ base_model:
3
+ - Qwen/Qwen2.5-72B
4
+ tags:
5
+ - roleplay
6
+ - storywriting
7
+ - qwen2.5
8
+ - finetune
9
+ - transformers
10
+ - pytorch
11
+ language:
12
+ - zho
13
+ - eng
14
+ - fra
15
+ - spa
16
+ - por
17
+ - deu
18
+ - ita
19
+ - rus
20
+ - jpn
21
+ - kor
22
+ - vie
23
+ - tha
24
+ - ara
25
+ ---
26
+
27
+ # Zeus Labs ~ Chronos-Platinum-72B
28
+
29
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/630417380907b9a115c6aa9f/G05mAhqcp4S_WBfE2vBLl.png)
30
+
31
+ Qwen 2.5 72B base model, trained for two epochs on the Chronos Divergence dataset using ChatML. It works well for roleplaying and storywriting as well as general assistant tasks.
32
+
33
+ ## Instruct Template
34
+
35
+ This model uses `ChatML` - below is an example. It is a preset in many frontends.
36
+
37
+ ```
38
+ <|im_start|>system
39
+ You are a helpful assistant<|im_end|>
40
+ <|im_start|>user
41
+ Hello there!<|im_end|>
42
+ <|im_start|>assistant
43
+ Hi! I'm an AI assistant, designed to help people like you with all sorts of tasks. Is there anything you need help with?<|im_end|>
44
+ <|im_start|>user
45
+ I was wondering how transformers work?<|im_end|>
46
+ <|im_start|>assistant
47
+ ```
48
+
49
+ ## Quantizations
50
+ Please note that we tested this model with a 5.0bpw EXL2 quant. Results are not expected to be the same when going below this quanitzation. Thanks to our model quanters!
51
+
52
+ #### LlamaCPP (GGUF)
53
+ [bartowski](https://huggingface.co/bartowski/Chronos-Platinum-72B-GGUF)
54
+
55
+ [mradermacher](https://huggingface.co/mradermacher/Chronos-Platinum-72B-i1-GGUF)
56
+
57
+ #### Exllama2
58
+ [bartowski](https://huggingface.co/bartowski/Chronos-Platinum-72B-exl2)
59
+
60
+ ## Sampling Settings
61
+ Here are some settings that work well with this model:
62
+ ```
63
+ Temp -> 0.7 - 1.2
64
+ Min P -> 0.025 - 0.05 [temp in order, not last]
65
+ Presence Penalty -> 1.0
66
+ Repetition Penalty range -> 4000
67
+ ```
68
+ Higher temp gives more uniqueness and less repetition. Please do not take these settings as the "best" - your system prompt matters significantly, and if you're roleplaying
69
+ use the Basic system prompt in SillyTavern. You can also try other samplers like Top P.
70
+
71
+ **Note that Presence Penalty works with Repetition Penalty Range.**
72
+
73
+ ## Credit
74
+ Thank you to my team consisting of [@ToastyPigeon](https://huggingface.co/ToastyPigeon), [@Fizzarolli](https://huggingface.co/Fizzarolli), and myself [@elinas](https://huggingface.co/elinas).
75
+
76
+ Additional thanks to [@AlpinDale](https://huggingface.co/AlpinDale) and the rest of the PygmalionAI team for graciously providing the compute to finetune this model!
77
+ Thank you to [anthracite-org](https://huggingface.co/anthracite-org) as well for sponsoring this model.
78
+
79
+ ## Additional Details 
80
+
81
+ We used a combination of provided logs and WizardLM evol both cleaned up and de-slopped.
82
+
83
+ Thanks to Anthropic and OpenAI for the models used to generate synthetic and partially synthetic data to train this model.
84
+
85
+ Thanks Elon Musk for being based enough to train AI that compares to the top models.
86
+
87
+ If you have any questions or concerns, please post in the community tab.
88
+
89
  DISCLAIMER: Outputs generated by the model are not reflective of our views.