Update README.md
Browse files
README.md
CHANGED
|
@@ -17,6 +17,15 @@ language:
|
|
| 17 |
- **License:** apache-2.0
|
| 18 |
- **Finetuned from model :** unsloth/llama-3.2-3b-unsloth-bnb-4bit
|
| 19 |
|
| 20 |
-
|
|
|
|
|
|
|
| 21 |
|
| 22 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 17 |
- **License:** apache-2.0
|
| 18 |
- **Finetuned from model :** unsloth/llama-3.2-3b-unsloth-bnb-4bit
|
| 19 |
|
| 20 |
+
# Datasets Used
|
| 21 |
+
- **1.** https://huggingface.co/datasets/mlabonne/FineTome-100k
|
| 22 |
+
- **2.** https://huggingface.co/datasets/anon8231489123/ShareGPT_Vicuna_unfiltered
|
| 23 |
|
| 24 |
+
# Updates
|
| 25 |
+
- This version now prints the tokenizer bs I was talking about in my previous version, it could be due to the dataset or how I've prepared the data.
|
| 26 |
+
Regardless, multi-turn datasets were too much for me to handle anyways.
|
| 27 |
+
|
| 28 |
+
- model does not refuse to anything yet.
|
| 29 |
+
|
| 30 |
+
# To Do
|
| 31 |
+
- Scrap or finetune this, not sure anyways
|