sweetpablo commited on
Commit
b5df2ae
·
1 Parent(s): 6189c17

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +57 -0
README.md ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: Llama-2-7B-bf16-sharded
4
+ model-index:
5
+ - name: llama_ft
6
+ results: []
7
+ ---
8
+
9
+
10
+ # llama_ft
11
+
12
+ This model is a fine-tuned version of [Llama-2-7B-bf16-sharded](https://huggingface.co/TinyPixel/Llama-2-7B-bf16-sharded) on a grocery cart dataset.
13
+
14
+ ## Intended uses & limitations
15
+
16
+ The model helps to tell to what type of grocery does the following items belong to.
17
+
18
+
19
+ ## Training procedure
20
+ Fine tuning techniques like Qlora and PEFT have been used to train the model on the dataset on a single gpu , and the adapters are then finally merged with the model.
21
+
22
+ load_in_4bit=True,
23
+ bnb_4bit_quant_type="nf4",
24
+ bnb_4bit_compute_dtype=torch.float16
25
+
26
+ The loading configurations of the model
27
+
28
+ ### Training hyperparameters
29
+
30
+ The following are the LORA configs-->
31
+
32
+ lora_alpha = 16
33
+ lora_dropout = 0.1
34
+ lora_r = 64
35
+
36
+ peft_config = LoraConfig(
37
+ lora_alpha=lora_alpha,
38
+ lora_dropout=lora_dropout,
39
+ r=lora_r,
40
+ bias="none",
41
+ task_type="CAUSAL_LM",
42
+ target_modules=["q_proj","v_proj"]
43
+ )
44
+
45
+ The following are the training configs -->
46
+
47
+ per_device_train_batch_size = 4
48
+ gradient_accumulation_steps = 4
49
+ optim = "paged_adamw_32bit"
50
+ save_steps = 10
51
+ logging_steps = 1
52
+ learning_rate = 2e-4
53
+ max_grad_norm = 0.3
54
+ max_steps = 120
55
+ warmup_ratio = 0.03
56
+ lr_scheduler_type = "constant"
57
+