|
|
--- |
|
|
license: apache-2.0 |
|
|
base_model: Llama-2-7B-bf16-sharded |
|
|
model-index: |
|
|
- name: llama_ft |
|
|
results: [] |
|
|
--- |
|
|
|
|
|
|
|
|
# llama_ft |
|
|
|
|
|
This model is a fine-tuned version of [Llama-2-7B-bf16-sharded](https://huggingface.co/TinyPixel/Llama-2-7B-bf16-sharded) on a grocery cart dataset. |
|
|
|
|
|
## Intended uses & limitations |
|
|
|
|
|
The model helps to tell to what type of grocery does the following items belong to. |
|
|
|
|
|
|
|
|
## Training procedure |
|
|
Fine tuning techniques like Qlora and PEFT have been used to train the model on the dataset on a single gpu , and the adapters are then finally merged with the model. |
|
|
|
|
|
load_in_4bit=True, |
|
|
bnb_4bit_quant_type="nf4", |
|
|
bnb_4bit_compute_dtype=torch.float16 |
|
|
|
|
|
The loading configurations of the model |
|
|
|
|
|
### Training hyperparameters |
|
|
|
|
|
The following are the LORA configs--> |
|
|
|
|
|
lora_alpha = 16 |
|
|
lora_dropout = 0.1 |
|
|
lora_r = 64 |
|
|
|
|
|
peft_config = LoraConfig( |
|
|
lora_alpha=lora_alpha, |
|
|
lora_dropout=lora_dropout, |
|
|
r=lora_r, |
|
|
bias="none", |
|
|
task_type="CAUSAL_LM", |
|
|
target_modules=["q_proj","v_proj"] |
|
|
) |
|
|
|
|
|
The following are the training configs --> |
|
|
|
|
|
per_device_train_batch_size = 4 |
|
|
gradient_accumulation_steps = 4 |
|
|
optim = "paged_adamw_32bit" |
|
|
save_steps = 10 |
|
|
logging_steps = 1 |
|
|
learning_rate = 2e-4 |
|
|
max_grad_norm = 0.3 |
|
|
max_steps = 120 |
|
|
warmup_ratio = 0.03 |
|
|
lr_scheduler_type = "constant" |
|
|
|
|
|
|