Initial upload of fine‑tuned Gemma + custom tokenizer
Browse files
README.md
CHANGED
|
@@ -50,7 +50,8 @@ There are three variants of the model for now:
|
|
| 50 |
| **Example w/o inputs** | ```text\nDESCRIPTION\n<start_of_turn>OUTPUT1<end_of_turn>\n<start_of_turn>OUTPUT2<end_of_turn>``` | ```text\n<start_of_turn>description\nDESCRIPTION<end_of_turn>\n<start_of_turn>output\nOUTPUT1<end_of_turn>\n<start_of_turn>output\nOUTPUT2<end_of_turn>``` | ```text\n<start_of_turn>user\nGenerate …\nDescription: DESCRIPTION\n\nGenerate.<end_of_turn>\n<start_of_turn>model\nOUTPUT1<end_of_turn>\n<start_of_turn>user\nGenerate.<end_of_turn>\n<start_of_turn>model\nOUTPUT2<end_of_turn>``` |
|
| 51 |
|
| 52 |
At the moment, I recommend:
|
| 53 |
-
- [special](https://huggingface.co/tsor13/special12b)
|
|
|
|
| 54 |
- [chat](https://huggingface.co/tsor13/chat12b) is a good fit for chat-style data or conversations.
|
| 55 |
|
| 56 |
|
|
|
|
| 50 |
| **Example w/o inputs** | ```text\nDESCRIPTION\n<start_of_turn>OUTPUT1<end_of_turn>\n<start_of_turn>OUTPUT2<end_of_turn>``` | ```text\n<start_of_turn>description\nDESCRIPTION<end_of_turn>\n<start_of_turn>output\nOUTPUT1<end_of_turn>\n<start_of_turn>output\nOUTPUT2<end_of_turn>``` | ```text\n<start_of_turn>user\nGenerate …\nDescription: DESCRIPTION\n\nGenerate.<end_of_turn>\n<start_of_turn>model\nOUTPUT1<end_of_turn>\n<start_of_turn>user\nGenerate.<end_of_turn>\n<start_of_turn>model\nOUTPUT2<end_of_turn>``` |
|
| 51 |
|
| 52 |
At the moment, I recommend:
|
| 53 |
+
- [special](https://huggingface.co/tsor13/special12b) for most use cases (token-efficient and gets best loss on training data)
|
| 54 |
+
- [extra](https://huggingface.co/tsor13/extra12b) for when generation quality is more important than token efficiency
|
| 55 |
- [chat](https://huggingface.co/tsor13/chat12b) is a good fit for chat-style data or conversations.
|
| 56 |
|
| 57 |
|