End of training
Browse files- README.md +29 -21
- adapter_model.bin +2 -2
README.md
CHANGED
|
@@ -25,8 +25,8 @@ is_llama_derived_model: true
|
|
| 25 |
|
| 26 |
hub_model_id: noeloco/camel-lora
|
| 27 |
|
| 28 |
-
load_in_8bit:
|
| 29 |
-
load_in_4bit:
|
| 30 |
strict: false
|
| 31 |
|
| 32 |
datasets:
|
|
@@ -44,13 +44,21 @@ sequence_len: 4096
|
|
| 44 |
sample_packing: false
|
| 45 |
pad_to_sequence_len: true
|
| 46 |
|
| 47 |
-
adapter:
|
| 48 |
lora_model_dir:
|
| 49 |
-
lora_r:
|
| 50 |
lora_alpha: 16
|
| 51 |
lora_dropout: 0.05
|
| 52 |
-
lora_target_linear:
|
| 53 |
lora_fan_in_fan_out:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 54 |
|
| 55 |
wandb_project: runpod1
|
| 56 |
wandb_entity:
|
|
@@ -98,9 +106,9 @@ special_tokens:
|
|
| 98 |
|
| 99 |
# camel-lora
|
| 100 |
|
| 101 |
-
This model is a fine-tuned version of [codellama/CodeLlama-7b-hf](https://huggingface.co/codellama/CodeLlama-7b-hf) on
|
| 102 |
It achieves the following results on the evaluation set:
|
| 103 |
-
- Loss: 0.
|
| 104 |
|
| 105 |
## Model description
|
| 106 |
|
|
@@ -134,20 +142,20 @@ The following hyperparameters were used during training:
|
|
| 134 |
|
| 135 |
| Training Loss | Epoch | Step | Validation Loss |
|
| 136 |
|:-------------:|:-----:|:----:|:---------------:|
|
| 137 |
-
| 1.
|
| 138 |
-
| 1.
|
| 139 |
-
| 1.
|
| 140 |
-
| 0.
|
| 141 |
-
| 0.
|
| 142 |
-
| 0.
|
| 143 |
-
| 0.
|
| 144 |
-
| 0.
|
| 145 |
-
| 0.
|
| 146 |
-
| 0.
|
| 147 |
-
| 0.
|
| 148 |
-
| 0.
|
| 149 |
-
| 0.
|
| 150 |
-
| 0.
|
| 151 |
|
| 152 |
|
| 153 |
### Framework versions
|
|
|
|
| 25 |
|
| 26 |
hub_model_id: noeloco/camel-lora
|
| 27 |
|
| 28 |
+
load_in_8bit: true
|
| 29 |
+
load_in_4bit: false
|
| 30 |
strict: false
|
| 31 |
|
| 32 |
datasets:
|
|
|
|
| 44 |
sample_packing: false
|
| 45 |
pad_to_sequence_len: true
|
| 46 |
|
| 47 |
+
adapter: lora
|
| 48 |
lora_model_dir:
|
| 49 |
+
lora_r: 8
|
| 50 |
lora_alpha: 16
|
| 51 |
lora_dropout: 0.05
|
| 52 |
+
lora_target_linear: false
|
| 53 |
lora_fan_in_fan_out:
|
| 54 |
+
lora_target_modules:
|
| 55 |
+
- q_proj
|
| 56 |
+
- v_proj
|
| 57 |
+
- k_proj
|
| 58 |
+
- o_proj
|
| 59 |
+
- gate_proj
|
| 60 |
+
- down_proj
|
| 61 |
+
- up_proj
|
| 62 |
|
| 63 |
wandb_project: runpod1
|
| 64 |
wandb_entity:
|
|
|
|
| 106 |
|
| 107 |
# camel-lora
|
| 108 |
|
| 109 |
+
This model is a fine-tuned version of [codellama/CodeLlama-7b-hf](https://huggingface.co/codellama/CodeLlama-7b-hf) on an unknown dataset.
|
| 110 |
It achieves the following results on the evaluation set:
|
| 111 |
+
- Loss: 0.0294
|
| 112 |
|
| 113 |
## Model description
|
| 114 |
|
|
|
|
| 142 |
|
| 143 |
| Training Loss | Epoch | Step | Validation Loss |
|
| 144 |
|:-------------:|:-----:|:----:|:---------------:|
|
| 145 |
+
| 1.7211 | 0.06 | 1 | 2.5058 |
|
| 146 |
+
| 1.834 | 0.29 | 5 | 2.4238 |
|
| 147 |
+
| 1.1688 | 0.57 | 10 | 1.3647 |
|
| 148 |
+
| 0.483 | 0.86 | 15 | 0.7108 |
|
| 149 |
+
| 0.3742 | 1.14 | 20 | 0.3942 |
|
| 150 |
+
| 0.1581 | 1.43 | 25 | 0.2196 |
|
| 151 |
+
| 0.2905 | 1.71 | 30 | 0.0822 |
|
| 152 |
+
| 0.1803 | 2.0 | 35 | 0.0548 |
|
| 153 |
+
| 0.0799 | 2.29 | 40 | 0.0543 |
|
| 154 |
+
| 0.0932 | 2.57 | 45 | 0.0390 |
|
| 155 |
+
| 0.0851 | 2.86 | 50 | 0.0328 |
|
| 156 |
+
| 0.096 | 3.14 | 55 | 0.0287 |
|
| 157 |
+
| 0.086 | 3.43 | 60 | 0.0289 |
|
| 158 |
+
| 0.0459 | 3.71 | 65 | 0.0294 |
|
| 159 |
|
| 160 |
|
| 161 |
### Framework versions
|
adapter_model.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:076ff87239662d4acee0599b11ba12bf120bc15669aa1bc54da375fe3d51e040
|
| 3 |
+
size 80115210
|