Update README.md
Browse files
README.md
CHANGED
|
@@ -24,7 +24,9 @@ inference:
|
|
| 24 |
max_new_tokens: 1024
|
| 25 |
|
| 26 |
---
|
| 27 |
-
|
|
|
|
|
|
|
| 28 |
|
| 29 |
This model was fine-tuned using [Predibase](https://predibase.com/), the first low-code AI platform for engineers.
|
| 30 |
I fine-tuned base Llama-2-7b using LoRA with 4 bit quantization on a single T4 GPU, which cost approximately $3 to train
|
|
@@ -34,6 +36,8 @@ Dataset and training parameters are borrowed from: https://github.com/sahil28011
|
|
| 34 |
but all of these parameters including DeepSpeed can be directly used with [Ludwig](https://ludwig.ai/latest/), the open-source
|
| 35 |
toolkit for LLMs that Predibase is built on.
|
| 36 |
|
|
|
|
|
|
|
| 37 |
To use these weights:
|
| 38 |
```python
|
| 39 |
from peft import PeftModel, PeftConfig
|
|
|
|
| 24 |
max_new_tokens: 1024
|
| 25 |
|
| 26 |
---
|
| 27 |
+
# QLoRA weights using Llama-2-7b for the Code Alpaca Dataset
|
| 28 |
+
|
| 29 |
+
# Fine-Tuning on Predibase
|
| 30 |
|
| 31 |
This model was fine-tuned using [Predibase](https://predibase.com/), the first low-code AI platform for engineers.
|
| 32 |
I fine-tuned base Llama-2-7b using LoRA with 4 bit quantization on a single T4 GPU, which cost approximately $3 to train
|
|
|
|
| 36 |
but all of these parameters including DeepSpeed can be directly used with [Ludwig](https://ludwig.ai/latest/), the open-source
|
| 37 |
toolkit for LLMs that Predibase is built on.
|
| 38 |
|
| 39 |
+
# How To Use The Model
|
| 40 |
+
|
| 41 |
To use these weights:
|
| 42 |
```python
|
| 43 |
from peft import PeftModel, PeftConfig
|