Improve language tag
Browse filesHi! As the model is multilingual, this is a PR to add other languages than English to the language tag to improve the referencing. Note that 29 languages are announced in the README, but only 13 are explicitly listed. I was therefore only able to add these 13 languages.
README.md
CHANGED
|
@@ -1,75 +1,89 @@
|
|
| 1 |
-
---
|
| 2 |
-
base_model:
|
| 3 |
-
- Qwen/Qwen2.5-72B
|
| 4 |
-
tags:
|
| 5 |
-
- roleplay
|
| 6 |
-
- storywriting
|
| 7 |
-
- qwen2.5
|
| 8 |
-
- finetune
|
| 9 |
-
- transformers
|
| 10 |
-
- pytorch
|
| 11 |
-
|
| 12 |
-
|
| 13 |
-
|
| 14 |
-
|
| 15 |
-
|
| 16 |
-
|
| 17 |
-
|
| 18 |
-
|
| 19 |
-
|
| 20 |
-
|
| 21 |
-
|
| 22 |
-
|
| 23 |
-
|
| 24 |
-
|
| 25 |
-
|
| 26 |
-
|
| 27 |
-
|
| 28 |
-
|
| 29 |
-
|
| 30 |
-
|
| 31 |
-
|
| 32 |
-
|
| 33 |
-
|
| 34 |
-
|
| 35 |
-
|
| 36 |
-
|
| 37 |
-
|
| 38 |
-
|
| 39 |
-
|
| 40 |
-
|
| 41 |
-
|
| 42 |
-
|
| 43 |
-
|
| 44 |
-
|
| 45 |
-
|
| 46 |
-
|
| 47 |
-
|
| 48 |
-
|
| 49 |
-
|
| 50 |
-
|
| 51 |
-
|
| 52 |
-
|
| 53 |
-
|
| 54 |
-
|
| 55 |
-
|
| 56 |
-
|
| 57 |
-
|
| 58 |
-
|
| 59 |
-
|
| 60 |
-
|
| 61 |
-
|
| 62 |
-
|
| 63 |
-
|
| 64 |
-
|
| 65 |
-
|
| 66 |
-
|
| 67 |
-
|
| 68 |
-
|
| 69 |
-
|
| 70 |
-
|
| 71 |
-
|
| 72 |
-
|
| 73 |
-
|
| 74 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 75 |
DISCLAIMER: Outputs generated by the model are not reflective of our views.
|
|
|
|
| 1 |
+
---
|
| 2 |
+
base_model:
|
| 3 |
+
- Qwen/Qwen2.5-72B
|
| 4 |
+
tags:
|
| 5 |
+
- roleplay
|
| 6 |
+
- storywriting
|
| 7 |
+
- qwen2.5
|
| 8 |
+
- finetune
|
| 9 |
+
- transformers
|
| 10 |
+
- pytorch
|
| 11 |
+
language:
|
| 12 |
+
- zho
|
| 13 |
+
- eng
|
| 14 |
+
- fra
|
| 15 |
+
- spa
|
| 16 |
+
- por
|
| 17 |
+
- deu
|
| 18 |
+
- ita
|
| 19 |
+
- rus
|
| 20 |
+
- jpn
|
| 21 |
+
- kor
|
| 22 |
+
- vie
|
| 23 |
+
- tha
|
| 24 |
+
- ara
|
| 25 |
+
---
|
| 26 |
+
|
| 27 |
+
# Zeus Labs ~ Chronos-Platinum-72B
|
| 28 |
+
|
| 29 |
+

|
| 30 |
+
|
| 31 |
+
Qwen 2.5 72B base model, trained for two epochs on the Chronos Divergence dataset using ChatML. It works well for roleplaying and storywriting as well as general assistant tasks.
|
| 32 |
+
|
| 33 |
+
## Instruct Template
|
| 34 |
+
|
| 35 |
+
This model uses `ChatML` - below is an example. It is a preset in many frontends.
|
| 36 |
+
|
| 37 |
+
```
|
| 38 |
+
<|im_start|>system
|
| 39 |
+
You are a helpful assistant<|im_end|>
|
| 40 |
+
<|im_start|>user
|
| 41 |
+
Hello there!<|im_end|>
|
| 42 |
+
<|im_start|>assistant
|
| 43 |
+
Hi! I'm an AI assistant, designed to help people like you with all sorts of tasks. Is there anything you need help with?<|im_end|>
|
| 44 |
+
<|im_start|>user
|
| 45 |
+
I was wondering how transformers work?<|im_end|>
|
| 46 |
+
<|im_start|>assistant
|
| 47 |
+
```
|
| 48 |
+
|
| 49 |
+
## Quantizations
|
| 50 |
+
Please note that we tested this model with a 5.0bpw EXL2 quant. Results are not expected to be the same when going below this quanitzation. Thanks to our model quanters!
|
| 51 |
+
|
| 52 |
+
#### LlamaCPP (GGUF)
|
| 53 |
+
[bartowski](https://huggingface.co/bartowski/Chronos-Platinum-72B-GGUF)
|
| 54 |
+
|
| 55 |
+
[mradermacher](https://huggingface.co/mradermacher/Chronos-Platinum-72B-i1-GGUF)
|
| 56 |
+
|
| 57 |
+
#### Exllama2
|
| 58 |
+
[bartowski](https://huggingface.co/bartowski/Chronos-Platinum-72B-exl2)
|
| 59 |
+
|
| 60 |
+
## Sampling Settings
|
| 61 |
+
Here are some settings that work well with this model:
|
| 62 |
+
```
|
| 63 |
+
Temp -> 0.7 - 1.2
|
| 64 |
+
Min P -> 0.025 - 0.05 [temp in order, not last]
|
| 65 |
+
Presence Penalty -> 1.0
|
| 66 |
+
Repetition Penalty range -> 4000
|
| 67 |
+
```
|
| 68 |
+
Higher temp gives more uniqueness and less repetition. Please do not take these settings as the "best" - your system prompt matters significantly, and if you're roleplaying
|
| 69 |
+
use the Basic system prompt in SillyTavern. You can also try other samplers like Top P.
|
| 70 |
+
|
| 71 |
+
**Note that Presence Penalty works with Repetition Penalty Range.**
|
| 72 |
+
|
| 73 |
+
## Credit
|
| 74 |
+
Thank you to my team consisting of [@ToastyPigeon](https://huggingface.co/ToastyPigeon), [@Fizzarolli](https://huggingface.co/Fizzarolli), and myself [@elinas](https://huggingface.co/elinas).
|
| 75 |
+
|
| 76 |
+
Additional thanks to [@AlpinDale](https://huggingface.co/AlpinDale) and the rest of the PygmalionAI team for graciously providing the compute to finetune this model!
|
| 77 |
+
Thank you to [anthracite-org](https://huggingface.co/anthracite-org) as well for sponsoring this model.
|
| 78 |
+
|
| 79 |
+
## Additional Details
|
| 80 |
+
|
| 81 |
+
We used a combination of provided logs and WizardLM evol both cleaned up and de-slopped.
|
| 82 |
+
|
| 83 |
+
Thanks to Anthropic and OpenAI for the models used to generate synthetic and partially synthetic data to train this model.
|
| 84 |
+
|
| 85 |
+
Thanks Elon Musk for being based enough to train AI that compares to the top models.
|
| 86 |
+
|
| 87 |
+
If you have any questions or concerns, please post in the community tab.
|
| 88 |
+
|
| 89 |
DISCLAIMER: Outputs generated by the model are not reflective of our views.
|