Update README.md
Browse files
README.md
CHANGED
|
@@ -1,3 +1,15 @@
|
|
| 1 |
---
|
| 2 |
license: mit
|
| 3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
---
|
| 2 |
license: mit
|
| 3 |
---
|
| 4 |
+
This model is a part of Project InterACT (Multi model AI system) involving an object detection model and an LLM
|
| 5 |
+
|
| 6 |
+
This is a model built by finetuning the Llama-2-7b-chat model on custom dataset: Jithendra-k/InterACT_LLM.
|
| 7 |
+
|
| 8 |
+
Points to consider for Finetuning Llama-2_7B_chat model:
|
| 9 |
+
=> Free Google Colab offers a 15GB Graphics Card (Limited Resources --> Barely enough to store Llama 2–7b’s weights)
|
| 10 |
+
=> We also considered the overhead due to optimizer states, gradients, and forward activations
|
| 11 |
+
=> Full fine-tuning is not possible in our case due to computation: we used parameter-efficient fine-tuning (PEFT) techniques like LoRA or QLoRA.
|
| 12 |
+
=> To drastically reduce the VRAM usage, we fine-tuned the model in 4-bit precision, which is why we've used QLoRA technique.
|
| 13 |
+
=> We only trained with 5 epochs considering our computation, time and early stopping.
|
| 14 |
+
|
| 15 |
+
Code to finetune a Llama-2_7B_chat model: https://colab.research.google.com/drive/1ZTdSKu2mgvQ1uNs0Wl7T7gniuoZJWs24?usp=sharing
|