Update README.md
Browse files
README.md
CHANGED
|
@@ -5,11 +5,18 @@ This model is a part of Project InterACT (Multi model AI system) involving an ob
|
|
| 5 |
|
| 6 |
This is a model built by finetuning the Llama-2-7b-chat model on custom dataset: Jithendra-k/InterACT_LLM.
|
| 7 |
|
| 8 |
-
Points to consider for Finetuning Llama-2_7B_chat model
|
| 9 |
-
=> Free Google Colab offers a 15GB Graphics Card (Limited Resources --> Barely enough to store Llama 2–7b’s weights)
|
| 10 |
-
=> We also considered the overhead due to optimizer states, gradients, and forward activations
|
| 11 |
-
=> Full fine-tuning is not possible in our case due to computation: we used parameter-efficient fine-tuning (PEFT) techniques like LoRA or QLoRA
|
| 12 |
-
=> To drastically reduce the VRAM usage, we fine-tuned the model in 4-bit precision, which is why we've used QLoRA technique
|
| 13 |
-
=> We only trained with 5 epochs considering our computation, time and early stopping
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 14 |
|
| 15 |
Code to finetune a Llama-2_7B_chat model: https://colab.research.google.com/drive/1ZTdSKu2mgvQ1uNs0Wl7T7gniuoZJWs24?usp=sharing
|
|
|
|
| 5 |
|
| 6 |
This is a model built by finetuning the Llama-2-7b-chat model on custom dataset: Jithendra-k/InterACT_LLM.
|
| 7 |
|
| 8 |
+
Points to consider for Finetuning Llama-2_7B_chat model:<br>
|
| 9 |
+
=> Free Google Colab offers a 15GB Graphics Card (Limited Resources --> Barely enough to store Llama 2–7b’s weights)<br>
|
| 10 |
+
=> We also considered the overhead due to optimizer states, gradients, and forward activations<br>
|
| 11 |
+
=> Full fine-tuning is not possible in our case due to computation: we used parameter-efficient fine-tuning (PEFT) techniques like LoRA or QLoRA.<br>
|
| 12 |
+
=> To drastically reduce the VRAM usage, we fine-tuned the model in 4-bit precision, which is why we've used QLoRA technique.<br>
|
| 13 |
+
=> We only trained with 5 epochs considering our computation, time and early stopping.<br>
|
| 14 |
+
|
| 15 |
+
Here are some plots of model performance during training:<br>
|
| 16 |
+
|
| 17 |
+
|
| 18 |
+
|
| 19 |
+
Here is an Example Input/Output:<br>
|
| 20 |
+
<img src="https://drive.google.com/file/d/1E0z3MAlJXu05bc8E9yDID0CVEbhowuca/view?usp=sharing"><br>
|
| 21 |
|
| 22 |
Code to finetune a Llama-2_7B_chat model: https://colab.research.google.com/drive/1ZTdSKu2mgvQ1uNs0Wl7T7gniuoZJWs24?usp=sharing
|