shirwu commited on
Commit
ce50a54
·
verified ·
1 Parent(s): deed762

Training in progress, step 2

Browse files
adapter_config.json CHANGED
@@ -24,12 +24,12 @@
24
  "revision": null,
25
  "target_modules": [
26
  "o_proj",
27
- "up_proj",
28
- "q_proj",
29
- "gate_proj",
30
  "down_proj",
31
  "v_proj",
32
- "k_proj"
 
 
 
33
  ],
34
  "task_type": "CAUSAL_LM",
35
  "use_dora": false,
 
24
  "revision": null,
25
  "target_modules": [
26
  "o_proj",
 
 
 
27
  "down_proj",
28
  "v_proj",
29
+ "k_proj",
30
+ "gate_proj",
31
+ "q_proj",
32
+ "up_proj"
33
  ],
34
  "task_type": "CAUSAL_LM",
35
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6b51aa8e69ce1d1ab206d4c03f778aef4b0c7fb46718c22901c5c0e403e7af48
3
  size 335604696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84ffef693e76e44506b26d938714d20a6b455ee71c516a59ad9d1697b87868cf
3
  size 335604696
all_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 0.008888888888888889,
3
+ "eval_accuracy": 0.30833333333333335,
4
+ "eval_loss": 0.69873046875,
5
+ "eval_runtime": 12.9332,
6
+ "eval_samples_per_second": 15.464,
7
+ "eval_steps_per_second": 1.933
8
+ }
eval_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 0.008888888888888889,
3
+ "eval_accuracy": 0.30833333333333335,
4
+ "eval_loss": 0.69873046875,
5
+ "eval_runtime": 12.9332,
6
+ "eval_samples_per_second": 15.464,
7
+ "eval_steps_per_second": 1.933
8
+ }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e993bf94036cb23f4188321b7ae2d7af3049c9f5b86a163732d611892881a708
3
  size 5432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ed6741bda4959c7a4525e886843c26cd1f59fbb620407eeae07f38a2965fb27
3
  size 5432