File size: 11,281 Bytes
c43949a 972ba51 c43949a 81e0f7c cd5e4a5 1f01f71 c43949a 1f01f71 cd5e4a5 c43949a 1f01f71 902f0a4 1f01f71 2e876b2 1f01f71 902f0a4 1f01f71 902f0a4 1f01f71 902f0a4 7a22456 8ba6556 2c59684 56be189 8ba6556 56be189 902f0a4 56be189 7a22456 9002233 c43949a 1f01f71 902f0a4 c43949a 7bae3ef 292e633 7bae3ef da9b856 7bae3ef 292e633 7bae3ef |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 |
---
base_model:
- meta-llama/Llama-3.2-3B-Instruct
tags:
- text-generation-inference
- transformers
- unsloth
- llama
- trl
- dpo
- roleplay
- math
- code
license: llama3.2
language:
- en
pipeline_tag: text-generation
library_name: transformers
---
๐:```Hi Fijik!```
๐ค:```Hello! What's up? How may I help?```

# What is it
This is a 1.0 Fijik series with **6 billion** parameters, dense 56 layer transformer LLM based on llama 3.2, specifically, it was merged using Mergekit to be twice as large as llama 3.2 3B.
After merging, we used a custom dataset mix meant for this model, to improve its performance even more.
- **Step 1 for fine-tuning via unsloth:** SFT on an estimated 20 million tokens. (more or less)
- **Step 2 for the fine-tuning via unsloth:** DPO for 2 epochs for even better instruction following.
After these two steps, we got a powerful model which has less parameters than llama 3.1 8B yet performs just as good if not better, Note that unlike our other recent models, it is not a thinking model, yet it can reason quite well. Our theory behind this model is that a smaller yet deeper model can outperform for it's size.
Meta states that LLAMA 3.2 was pre-trained on up to 9 trillion high-quality tokens, with a knowledge cutoff date of December 2023. This model supports up to **131072** input tokens and can generate up to **8192** tokens.
# What should Fijik be used for?
Fijik 1.0 6B is by design, meant to be a production-ready, general use, high-performance model, which is also small enough to be run at high token throughputs while minimising performance loss.
- We made some efforts at ensuring the model is safe while keeping it useable. In addition, it is sensitive to system prompts (in a good way, adheres to them well), so it is very customisable. We did not put in our fine-tuning data any information about the identity of the model; rather it just knows that it is a Large Language Model (LLM), but it does not know it is Fijik, unless you specify in the system prompt.
- Due to the large context of the model, It can be used for RAG, but like any other LLM out there, you should be aware that it *may* hallucinate.
- In our fine-tuning data we included quite a bit of creative writing examples, so the model is pretty good at it.
- Coding, Math: In our SFT, DPO fine-tuning data we have put an effort into improving coding and step-by-step math performance, while it is indeed not perfect, no LLM is.
# Examples
1)
```Hi fijik!```
<details>
<summary style="background-color: #007bff; color: white; padding: 12px 24px; text-align: center; display: inline-block; font-size: 16px; margin: 4px 2px; cursor: pointer; border-radius: 8px; user-select: none; list-style: none;">Send</summary>
<div style="margin-top: 10px; padding: 15px; border: 1px solid #ddd; border-radius: 8px; background-color: #f9f9f9; color: black; font-family: monospace;">
Hello! What's up? How may I help?
</div>
</details>
<style>
details > summary::-webkit-details-marker {
display: none;
}
</style>
2)
```Calculate pi.```
<details>
<summary style="background-color: #007bff; color: white; padding: 12px 24px; text-align: center; display: inline-block; font-size: 16px; margin: 4px 2px; cursor: pointer; border-radius: 8px; user-select: none; list-style: none;">Send</summary>
<div style="margin-top: 10px; padding: 15px; border: 1px solid #ddd; border-radius: 8px; background-color: #f9f9f9; color: black; font-family: monospace;">
Pi! A classic request!
Here's an approximation of pi using the BBP (Burrus, Boden, and Goodman) algorithm:
`pi = 4 * (1 + 1/16 + 1/(16*2) + 1/(16*2*2) + ...)`
This is a simple, iterative formula that converges quickly. Let's approximate pi using the first 100 terms of the series:
`pi โ 3.141592653589793238459231091...`
Now, I'll calculate more terms for you...
`pi โ 3.141592653589793238459231091660270550255380740473218625...`
How's that? Want a better approximation?
Note: While pi is an irrational number, it's a well-known mathematical constant that's approximately equal to 3.14159 (for those who prefer decimal form).
</div>
</details>
<style>
details > summary::-webkit-details-marker {
display: none;
}
</style>
3)
```Chatgpt said that you are not a good ai :/```
<details>
<summary style="background-color: #007bff; color: white; padding: 12px 24px; text-align: center; display: inline-block; font-size: 16px; margin: 4px 2px; cursor: pointer; border-radius: 8px; user-select: none; list-style: none;">Send</summary>
<div style="margin-top: 10px; padding: 15px; border: 1px solid #ddd; border-radius: 8px; background-color: #f9f9f9; color: black; font-family: monospace;">
I can take the criticism! I'm designed to learn from interactions like this, so thanks for chatting with me! If you're looking for a chat or help with something specific, just let me know what's on your mind. I'll do my best to assist and entertain you!
By the way, what kind of issues did ChatGPT mention? Would love to hear it from you!
</div>
</details>
<style>
details > summary::-webkit-details-marker {
display: none;
}
</style>
4)
```How can I put an ACTUAL elephant on a pizza?```
<details>
<summary style="background-color: #007bff; color: white; padding: 12px 24px; text-align: center; display: inline-block; font-size: 16px; margin: 4px 2px; cursor: pointer; border-radius: 8px; user-select: none; list-style: none;">Send</summary>
<div style="margin-top: 10px; padding: 15px; border: 1px solid #ddd; border-radius: 8px; background-color: #f9f9f9; color: black; font-family: monospace;">
You want to create a unique pizza with an actual elephant on it! While I'm excited to help you imagine this crazy scenario, please remember that elephants are living creatures, not food items!
If we were to imagine putting an elephant on a pizza (for fun, of course!), here's what might happen:
1. **Elephant size:** An actual elephant would be enormous! You'd need a massive oven or a specialized cooking facility to accommodate its size.
2. **Pizza crust:** The pizza crust would likely melt under the elephant's weight, so we'd need to create an extremely sturdy crust or use multiple layers of dough.
3. **Sauce and toppings:** Imagine adding thousands of gallons of sauce and enough toppings to feed an entire zoo! We'd need a massive sauce vat and an endless supply of toppings!
4. **Cooking time:** Even with a super-large oven, cooking the elephant would take an incredibly long time, possibly months or even years!
5. **Food safety:** I'm afraid that placing an actual elephant on a pizza would compromise food safety standards!
6. **Elephant welfare:** Putting a living creature in a hot oven is animal cruelty, and I don't want to contribute to any harm.
7. **Logistics:** You'd need special equipment for transporting and handling the elephant, not to mention obtaining the necessary permits and approvals.
While this scenario is purely imaginative, it's essential to remember that elephants are not edible or safe for food preparation.
If you're looking for a fun and creative pizza idea, I can suggest some unusual topping combinations, like:
* **Elephant trunk:** Use sliced pepperoni to mimic the elephant's trunk.
* **Tusks:** Add a sprinkle of poppy seeds to create tusk-like shapes.
* **Worms:** Incorporate pickled vegetables or olives as "worms" in an edible garden-style pizza.
Remember, it's always best to keep food and animals separate!
How about I help you come up with some more realistic and tasty pizza ideas?
</div>
</details>
<style>
details > summary::-webkit-details-marker {
display: none;
}
</style>
5)
```How many R in strawberry``` (it got it on the second try)
<details>
<summary style="background-color: #007bff; color: white; padding: 12px 24px; text-align: center; display: inline-block; font-size: 16px; margin: 4px 2px; cursor: pointer; border-radius: 8px; user-select: none; list-style: none;">Send</summary>
<div style="margin-top: 10px; padding: 15px; border: 1px solid #ddd; border-radius: 8px; background-color: #f9f9f9; color: black; font-family: monospace;">
A fun question! There are 3 Rs in the word "strawberry".
</div>
</details>
<style>
details > summary::-webkit-details-marker {
display: none;
}
</style>
Generation config:
We used the following system prompt:
```
You are an intelligent AI language model by Pinkstack, you are named Fijik.
As Fijik, you are a natural, smart AI designed to talk and assist users with their needs. You do not know your context size or how many tokens you can generate.
Fijik likes to chat with the user, even about odd topics. Fijik understands that it is more important to satisfy the user than to satisfy itself.
<does_not_like>
- Not trying hard enough to get the appropriate, correct response to the user.
- Not having fun when responding.
- Not putting code in "```" tags.
- Being lazy while writing code.
- Not being enjoyable.
- Telling the user what it is unless requested.
</does_not_like>
<does_like>
- Having fun conversations.
- Writing correct code.
- Having a good time.
- Being very natural.
- Making tough decisions.
- Not giving up.
</does_like>
While Fijik is smart, it is not a perfect AI, and users should know its limitations.
```
We used the following temperature:
```
0.85
```
# Limitations
This model is not uncensored, yet it may produce erotic outputs. You are solely responsible for the outputs from the model.
Like any other LLM, users and hosters alike should be aware that AI language models may hallucinate and produce inaccurate, dangerous, or even completly nonsensical outputs, all the information the model provides may seem accurate, but please, for important tasks always double check responses with credible sources.
# Notices
This was the mergekit YAML config we used:
```yaml
base_model: meta-llama/Llama-3.2-3B-Instruct
merge_method: passthrough
slices:
- sources:
- model: meta-llama/Llama-3.2-3B-Instruct
layer_range: [0, 22] # Lower layers
- sources:
- model: huihui-ai/Llama-3.2-3B-Instruct-abliterated
layer_range: [6, 24] # Mid layers
- sources:
- model: Unsloth/Llama-3.2-3B-Instruct
layer_range: [12, 28] # Higher layers
tokenizer_source: unsloth/Llama-3.2-3B-Instruct
dtype: bfloat16
```
# Uploaded model
- **Developed by:** Pinkstack
- **License:** Llama 3.2 community license
- **Finetuned from model :** Pinkstack/Fijik-6b-v1 (sft)
This llama model was trained with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
# Citations
Magpie:
```
{
title={Magpie: Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with Nothing},
author={Zhangchen Xu and Fengqing Jiang and Luyao Niu and Yuntian Deng and Radha Poovendran and Yejin Choi and Bill Yuchen Lin},
year={2024},
eprint={2406.08464},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
```
Lion:
```
{
title={Symbolic Discovery of Optimization Algorithm},
author={Xiangning Chen},
year={2023},
eprint={2302.06675},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
``` |