Update README.md
#2
by
cherry0328 - opened
README.md
CHANGED
|
@@ -7,6 +7,8 @@ tags:
|
|
| 7 |
- unsloth
|
| 8 |
- transformers
|
| 9 |
- tinyllama
|
|
|
|
|
|
|
| 10 |
---
|
| 11 |
|
| 12 |
# Finetune Mistral, Gemma, Llama 2-5x faster with 70% less memory via Unsloth!
|
|
@@ -35,4 +37,4 @@ All notebooks are **beginner friendly**! Add your dataset, click "Run All", and
|
|
| 35 |
|
| 36 |
- This [conversational notebook](https://colab.research.google.com/drive/1Aau3lgPzeZKQ-98h69CCu1UJcvIBLmy2?usp=sharing) is useful for ShareGPT ChatML / Vicuna templates.
|
| 37 |
- This [text completion notebook](https://colab.research.google.com/drive/1ef-tab5bhkvWmBOObepl1WgJvfvSzn5Q?usp=sharing) is for raw text. This [DPO notebook](https://colab.research.google.com/drive/15vttTpzzVXv_tJwEk-hIcQ0S9FcEWvwP?usp=sharing) replicates Zephyr.
|
| 38 |
-
- \* Kaggle has 2x T4s, but we use 1. Due to overhead, 1x T4 is 5x faster.
|
|
|
|
| 7 |
- unsloth
|
| 8 |
- transformers
|
| 9 |
- tinyllama
|
| 10 |
+
base_model:
|
| 11 |
+
- TinyLlama/TinyLlama_v1.1
|
| 12 |
---
|
| 13 |
|
| 14 |
# Finetune Mistral, Gemma, Llama 2-5x faster with 70% less memory via Unsloth!
|
|
|
|
| 37 |
|
| 38 |
- This [conversational notebook](https://colab.research.google.com/drive/1Aau3lgPzeZKQ-98h69CCu1UJcvIBLmy2?usp=sharing) is useful for ShareGPT ChatML / Vicuna templates.
|
| 39 |
- This [text completion notebook](https://colab.research.google.com/drive/1ef-tab5bhkvWmBOObepl1WgJvfvSzn5Q?usp=sharing) is for raw text. This [DPO notebook](https://colab.research.google.com/drive/15vttTpzzVXv_tJwEk-hIcQ0S9FcEWvwP?usp=sharing) replicates Zephyr.
|
| 40 |
+
- \* Kaggle has 2x T4s, but we use 1. Due to overhead, 1x T4 is 5x faster.
|