JakeOh commited on
Commit
71a40f1
·
verified ·
1 Parent(s): b4b898d

Upload folder using huggingface_hub

Browse files
demo/checkpoint-7/adapter_config.json CHANGED
@@ -29,8 +29,8 @@
29
  "rank_pattern": {},
30
  "revision": null,
31
  "target_modules": [
32
- "q_proj",
33
- "v_proj"
34
  ],
35
  "target_parameters": null,
36
  "task_type": "CAUSAL_LM",
 
29
  "rank_pattern": {},
30
  "revision": null,
31
  "target_modules": [
32
+ "v_proj",
33
+ "q_proj"
34
  ],
35
  "target_parameters": null,
36
  "task_type": "CAUSAL_LM",
demo/checkpoint-7/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:89ae05c972a5f9aba93958d02f2ac186c5c17c1387557fd445b24308d1fcb600
3
  size 6304960
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:558c232cb815df2032928af45dfec313a17a3c21a86257af49669151c4c63886
3
  size 6304960
demo/checkpoint-7/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ee52089e44f5467afbaa4ea6a51026af8395523f54eaaf05506dbd20196ed16b
3
  size 12665547
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:628aa4a92a28cfd8dcb73b1e0c17507c1b408c8200c124542d689e03a59c31bf
3
  size 12665547
demo/checkpoint-7/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8568b303ea99fb716e32c2d6b2f98429cf8f071b8b191bda2b3e11cb1e51f308
3
  size 14645
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5beecedfb6e959ba27f2e8ebbc08b2d75c9209d149ac1527923b923ec8b69b20
3
  size 14645
demo/checkpoint-7/trainer_state.json CHANGED
@@ -12,10 +12,10 @@
12
  {
13
  "entropy": 2.546337890625,
14
  "epoch": 0.8,
15
- "grad_norm": 1.4311243295669556,
16
  "learning_rate": 8.000000000000001e-07,
17
- "loss": 2.356211471557617,
18
- "mean_token_accuracy": 0.508102647960186,
19
  "num_tokens": 35895.0,
20
  "step": 5
21
  }
 
12
  {
13
  "entropy": 2.546337890625,
14
  "epoch": 0.8,
15
+ "grad_norm": 1.5417859554290771,
16
  "learning_rate": 8.000000000000001e-07,
17
+ "loss": 2.356197547912598,
18
+ "mean_token_accuracy": 0.5081676602363586,
19
  "num_tokens": 35895.0,
20
  "step": 5
21
  }