error577 commited on
Commit
25e2068
·
verified ·
1 Parent(s): eeab954

End of training

Browse files
README.md CHANGED
@@ -42,7 +42,7 @@ early_stopping_patience: null
42
  eval_max_new_tokens: 128
43
  eval_table_size: null
44
  evals_per_epoch: 1
45
- flash_attention: false
46
  fp16: null
47
  fsdp: null
48
  fsdp_config: null
@@ -102,7 +102,7 @@ xformers_attention: null
102
 
103
  This model is a fine-tuned version of [unsloth/Meta-Llama-3.1-8B](https://huggingface.co/unsloth/Meta-Llama-3.1-8B) on the None dataset.
104
  It achieves the following results on the evaluation set:
105
- - Loss: nan
106
 
107
  ## Model description
108
 
@@ -136,11 +136,11 @@ The following hyperparameters were used during training:
136
 
137
  | Training Loss | Epoch | Step | Validation Loss |
138
  |:-------------:|:------:|:----:|:---------------:|
139
- | 0.0 | 0.0004 | 1 | nan |
140
- | 0.0 | 0.0103 | 25 | nan |
141
- | 0.0 | 0.0206 | 50 | nan |
142
- | 0.0 | 0.0309 | 75 | nan |
143
- | 0.0 | 0.0412 | 100 | nan |
144
 
145
 
146
  ### Framework versions
 
42
  eval_max_new_tokens: 128
43
  eval_table_size: null
44
  evals_per_epoch: 1
45
+ flash_attention: true
46
  fp16: null
47
  fsdp: null
48
  fsdp_config: null
 
102
 
103
  This model is a fine-tuned version of [unsloth/Meta-Llama-3.1-8B](https://huggingface.co/unsloth/Meta-Llama-3.1-8B) on the None dataset.
104
  It achieves the following results on the evaluation set:
105
+ - Loss: 0.8628
106
 
107
  ## Model description
108
 
 
136
 
137
  | Training Loss | Epoch | Step | Validation Loss |
138
  |:-------------:|:------:|:----:|:---------------:|
139
+ | 1.4349 | 0.0004 | 1 | 1.4960 |
140
+ | 1.0719 | 0.0103 | 25 | 0.9809 |
141
+ | 0.8261 | 0.0206 | 50 | 0.8903 |
142
+ | 0.8859 | 0.0309 | 75 | 0.8677 |
143
+ | 0.8202 | 0.0412 | 100 | 0.8628 |
144
 
145
 
146
  ### Framework versions
adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
 
 
23
  "down_proj",
24
  "q_proj",
25
  "gate_proj",
26
- "up_proj",
27
- "o_proj",
28
- "v_proj",
29
- "k_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "o_proj",
24
+ "k_proj",
25
+ "up_proj",
26
  "down_proj",
27
  "q_proj",
28
  "gate_proj",
29
+ "v_proj"
 
 
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:16184e613928f3396f83c0c7ff31330d342233fdbecb20944b82a69ed0829da3
3
  size 335706186
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92ed922a0f732a1e202018c98c4068cda3ed9dc18911402fa2073540e3ddb3b6
3
  size 335706186
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b6283686037b29c419488a0dfab5a8ac24663cbe9678b492717761a6b3996ff6
3
  size 335604696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:12565efd5f64bc493473760b2fa639964bcbbe89ace12120fc59dbaa190cc379
3
  size 335604696
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3888676450e93ddaf7e7bf415d416189cffc6565f5de5935ce9cb7a326a69c8d
3
  size 6776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd170f249f9caf7a2a3855411a8014d5fd024df787e2726693ab5077b1224fb5
3
  size 6776