Update README.md
Browse files
README.md
CHANGED
|
@@ -7,17 +7,32 @@ tags:
|
|
| 7 |
- llama
|
| 8 |
- trl
|
| 9 |
- dpo
|
| 10 |
-
license:
|
| 11 |
language:
|
| 12 |
- en
|
|
|
|
| 13 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 14 |
|
| 15 |
# Uploaded model
|
| 16 |
|
| 17 |
- **Developed by:** Pinkstack
|
| 18 |
-
- **License:**
|
| 19 |
- **Finetuned from model :** Pinkstack/Fijik-6b-v1
|
| 20 |
|
| 21 |
-
This llama model was trained
|
| 22 |
-
|
| 23 |
-
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
|
|
|
|
| 7 |
- llama
|
| 8 |
- trl
|
| 9 |
- dpo
|
| 10 |
+
license: llama3.2
|
| 11 |
language:
|
| 12 |
- en
|
| 13 |
+
pipeline_tag: text-generation
|
| 14 |
---
|
| 15 |
+
😁:```Hi Fijik!```
|
| 16 |
+
|
| 17 |
+
🤖:```Hello! What's up? How may I help?```
|
| 18 |
+

|
| 19 |
+
# What is it
|
| 20 |
+
Fijik is a **6 billion** parameter, dense 56 layer transformer LLM based on llama 3.2, specifically, it was merged using Mergekit to be twice as large as llama 3.2 3B.
|
| 21 |
+
|
| 22 |
+
After merging, we used a custom dataset mix meant for this model, to improve its performance even more.
|
| 23 |
+
- **Step 1 for fine-tuning via unsloth:** SFT on an estimated 20 million tokens. (more or less)
|
| 24 |
+
- **Step 2 for the fine-tuning via unsloth:** DPO for 2 epochs for even better instruction following.
|
| 25 |
+
After these two steps, we got a powerful model which has less parameters than llama 3.1 8B yet performs just as good if not better, Note that unlike our other models, it is not a thinking model. our theory behind this model is that a smaller yet deeper model can outperform for it's size.
|
| 26 |
+
|
| 27 |
+
Meta states that LLAMA 3.2 was pre-trained on up to 9 trillion high quality tokens.
|
| 28 |
+
|
| 29 |
+
# What should Fijik be used for?
|
| 30 |
+
Fijik
|
| 31 |
|
| 32 |
# Uploaded model
|
| 33 |
|
| 34 |
- **Developed by:** Pinkstack
|
| 35 |
+
- **License:** Llama 3.2 community license
|
| 36 |
- **Finetuned from model :** Pinkstack/Fijik-6b-v1
|
| 37 |
|
| 38 |
+
This llama model was trained with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
|
|
|
|
|
|