noeloco commited on
Commit
d544a81
·
verified ·
1 Parent(s): 0cda28b

End of training

Browse files
Files changed (2) hide show
  1. README.md +29 -21
  2. adapter_model.bin +2 -2
README.md CHANGED
@@ -25,8 +25,8 @@ is_llama_derived_model: true
25
 
26
  hub_model_id: noeloco/camel-lora
27
 
28
- load_in_8bit: false
29
- load_in_4bit: true
30
  strict: false
31
 
32
  datasets:
@@ -44,13 +44,21 @@ sequence_len: 4096
44
  sample_packing: false
45
  pad_to_sequence_len: true
46
 
47
- adapter: qlora
48
  lora_model_dir:
49
- lora_r: 32
50
  lora_alpha: 16
51
  lora_dropout: 0.05
52
- lora_target_linear: true
53
  lora_fan_in_fan_out:
 
 
 
 
 
 
 
 
54
 
55
  wandb_project: runpod1
56
  wandb_entity:
@@ -98,9 +106,9 @@ special_tokens:
98
 
99
  # camel-lora
100
 
101
- This model is a fine-tuned version of [codellama/CodeLlama-7b-hf](https://huggingface.co/codellama/CodeLlama-7b-hf) on the None dataset.
102
  It achieves the following results on the evaluation set:
103
- - Loss: 0.0290
104
 
105
  ## Model description
106
 
@@ -134,20 +142,20 @@ The following hyperparameters were used during training:
134
 
135
  | Training Loss | Epoch | Step | Validation Loss |
136
  |:-------------:|:-----:|:----:|:---------------:|
137
- | 1.7685 | 0.06 | 1 | 2.5524 |
138
- | 1.8762 | 0.29 | 5 | 2.4927 |
139
- | 1.215 | 0.57 | 10 | 1.4546 |
140
- | 0.484 | 0.86 | 15 | 0.7250 |
141
- | 0.3667 | 1.14 | 20 | 0.4146 |
142
- | 0.1638 | 1.43 | 25 | 0.2123 |
143
- | 0.2948 | 1.71 | 30 | 0.0980 |
144
- | 0.2003 | 2.0 | 35 | 0.0629 |
145
- | 0.0888 | 2.29 | 40 | 0.0577 |
146
- | 0.0918 | 2.57 | 45 | 0.0414 |
147
- | 0.0931 | 2.86 | 50 | 0.0363 |
148
- | 0.0982 | 3.14 | 55 | 0.0304 |
149
- | 0.0849 | 3.43 | 60 | 0.0289 |
150
- | 0.0511 | 3.71 | 65 | 0.0290 |
151
 
152
 
153
  ### Framework versions
 
25
 
26
  hub_model_id: noeloco/camel-lora
27
 
28
+ load_in_8bit: true
29
+ load_in_4bit: false
30
  strict: false
31
 
32
  datasets:
 
44
  sample_packing: false
45
  pad_to_sequence_len: true
46
 
47
+ adapter: lora
48
  lora_model_dir:
49
+ lora_r: 8
50
  lora_alpha: 16
51
  lora_dropout: 0.05
52
+ lora_target_linear: false
53
  lora_fan_in_fan_out:
54
+ lora_target_modules:
55
+ - q_proj
56
+ - v_proj
57
+ - k_proj
58
+ - o_proj
59
+ - gate_proj
60
+ - down_proj
61
+ - up_proj
62
 
63
  wandb_project: runpod1
64
  wandb_entity:
 
106
 
107
  # camel-lora
108
 
109
+ This model is a fine-tuned version of [codellama/CodeLlama-7b-hf](https://huggingface.co/codellama/CodeLlama-7b-hf) on an unknown dataset.
110
  It achieves the following results on the evaluation set:
111
+ - Loss: 0.0294
112
 
113
  ## Model description
114
 
 
142
 
143
  | Training Loss | Epoch | Step | Validation Loss |
144
  |:-------------:|:-----:|:----:|:---------------:|
145
+ | 1.7211 | 0.06 | 1 | 2.5058 |
146
+ | 1.834 | 0.29 | 5 | 2.4238 |
147
+ | 1.1688 | 0.57 | 10 | 1.3647 |
148
+ | 0.483 | 0.86 | 15 | 0.7108 |
149
+ | 0.3742 | 1.14 | 20 | 0.3942 |
150
+ | 0.1581 | 1.43 | 25 | 0.2196 |
151
+ | 0.2905 | 1.71 | 30 | 0.0822 |
152
+ | 0.1803 | 2.0 | 35 | 0.0548 |
153
+ | 0.0799 | 2.29 | 40 | 0.0543 |
154
+ | 0.0932 | 2.57 | 45 | 0.0390 |
155
+ | 0.0851 | 2.86 | 50 | 0.0328 |
156
+ | 0.096 | 3.14 | 55 | 0.0287 |
157
+ | 0.086 | 3.43 | 60 | 0.0289 |
158
+ | 0.0459 | 3.71 | 65 | 0.0294 |
159
 
160
 
161
  ### Framework versions
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:430885f7d5e76c656bb16d0c30097c03805f898ac3586031af9f6c6c1d88520a
3
- size 160069834
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:076ff87239662d4acee0599b11ba12bf120bc15669aa1bc54da375fe3d51e040
3
+ size 80115210