Kibalama commited on
Commit
9107788
·
verified ·
1 Parent(s): 9129c51

Training in progress, epoch 1

Browse files
adapter_config.json CHANGED
@@ -26,13 +26,13 @@
26
  "rank_pattern": {},
27
  "revision": null,
28
  "target_modules": [
 
 
29
  "gate_proj",
30
  "k_proj",
31
- "o_proj",
32
- "down_proj",
33
- "up_proj",
34
  "q_proj",
35
- "v_proj"
 
36
  ],
37
  "task_type": "CAUSAL_LM",
38
  "use_dora": false,
 
26
  "rank_pattern": {},
27
  "revision": null,
28
  "target_modules": [
29
+ "o_proj",
30
+ "v_proj",
31
  "gate_proj",
32
  "k_proj",
 
 
 
33
  "q_proj",
34
+ "up_proj",
35
+ "down_proj"
36
  ],
37
  "task_type": "CAUSAL_LM",
38
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a618361b3a82ad91cec14975adb4ebe590ed0a4000e3494fe9bd481a64f82387
3
  size 2468152240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae81da57d5d2e449482f6fbdcefb4b4277bb4abadfa162b0fa2904e24f58f9b3
3
  size 2468152240
runs/Jul09_11-08-47_ed11a66b1901/events.out.tfevents.1752059395.ed11a66b1901.1867.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21772ca7f89ffaaefce68963fe6513f4f1c6f5a46bf3b8567321cee74a2c656d
3
+ size 35536
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:750ddf541fcdea5976b2685e1d11af1b685e5a39af666e6787164c4bb74dd0a5
3
  size 5624
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb711f8e77b79390cd1943ae231d1b666ff1075bf434ab403ac1b8824c8ab0c2
3
  size 5624