File size: 1,407 Bytes
9200fe2 7c478b4 9200fe2 7c478b4 9200fe2 7c478b4 9200fe2 7c478b4 9200fe2 7c478b4 9200fe2 7c478b4 9200fe2 7c478b4 9200fe2 7c478b4 9200fe2 7c478b4 ebe8539 7c478b4 b11abda 7c478b4 9200fe2 7c478b4 9200fe2 7c478b4 9200fe2 7c478b4 9200fe2 7c478b4 b11abda 7c478b4 9200fe2 7c478b4 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 | ---
base_model: ServiceNow-AI/Apriel-Nemotron-15b-Thinker
datasets:
- Undi95/R1-RP-ShareGPT3
- PJMixers-Dev/Gryphe-Aesir-RPG-Charcards-Opus-Mixed-split-v3-0324
---
### Apriel R1P V.2
Day 2 RP finetune of Apriel 15B, with several iterative improvements from the first version.
I have fully converted the model to use the Phi 3 format; this comes at the slight tradeoff of the `<|end|>` tag not always tokenizing exactly the same way in a few niche scenarios.
Further attempts were also made to fix formatting issues with asterisks on the base model.
NOTE: THIS IS THE NON-THINKING VERSION
Upon further testing, I discovered that while merging back onto the instruct improved thinking mode, this came at the cost of degraded non-thinking outputs.
Notably, the model had worse characterization and nuance.
As such I have decided upload this version which was not merged onto the instruct.
Use the [thinking version](https://huggingface.co/ConicCat/Apriel-R1PV.2-Thinking) instead if you want a thinking model.
## Settings
The chat template has been converted to a Phi 3 template as the model seemed to respond best to this format.
This model does prefer having character cards placed in user messages, not the system prompt.
Add `/no_think` to the system prompt
## Special Thanks:
Undi95 for portions of their dataset and inspiration.
PJMixers-Dev for their dataset curation and creation efforts |