JakeOh commited on
Commit
b4a151c
·
verified ·
1 Parent(s): 47f39f0

Upload folder using huggingface_hub

Browse files
demo/checkpoint-7/adapter_config.json CHANGED
@@ -29,8 +29,8 @@
29
  "rank_pattern": {},
30
  "revision": null,
31
  "target_modules": [
32
- "q_proj",
33
- "v_proj"
34
  ],
35
  "target_parameters": null,
36
  "task_type": "CAUSAL_LM",
 
29
  "rank_pattern": {},
30
  "revision": null,
31
  "target_modules": [
32
+ "v_proj",
33
+ "q_proj"
34
  ],
35
  "target_parameters": null,
36
  "task_type": "CAUSAL_LM",
demo/checkpoint-7/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d0262e996639ef50ff94a47f0f73c0afb42b5820cfa7308eb54468fe54b6c37e
3
  size 6304960
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9ffd945ad81454e9c049ba59b2c793a8820b12774d1d806055fb90bc3fa3286
3
  size 6304960
demo/checkpoint-7/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:34ad5c7fc2a2ef2eb6cce133e499e79cfb5da1adc4cead051d9e26215d5d4c84
3
  size 12665547
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fce6cc11c17a2d7e866f560b48d49df4d580ed4c524d95db6ffb6eaa465c1fda
3
  size 12665547
demo/checkpoint-7/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:497dfae9a097b88aab676f67ae75b1179292ca6a804df78a0c103f0c9c7e64b5
3
  size 14645
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5414978c52d77815cb1ed101a59575eef3f341e1808e51b3afe0665083c162c2
3
  size 14645
demo/checkpoint-7/trainer_state.json CHANGED
@@ -12,10 +12,10 @@
12
  {
13
  "entropy": 2.546337890625,
14
  "epoch": 0.8,
15
- "grad_norm": 1.4163397550582886,
16
  "learning_rate": 8.000000000000001e-07,
17
- "loss": 2.3561832427978517,
18
- "mean_token_accuracy": 0.5081296697258949,
19
  "num_tokens": 35895.0,
20
  "step": 5
21
  }
 
12
  {
13
  "entropy": 2.546337890625,
14
  "epoch": 0.8,
15
+ "grad_norm": 1.3540213108062744,
16
  "learning_rate": 8.000000000000001e-07,
17
+ "loss": 2.3561912536621095,
18
+ "mean_token_accuracy": 0.5081938788294792,
19
  "num_tokens": 35895.0,
20
  "step": 5
21
  }