bimabk commited on
Commit
c22314a
·
verified ·
1 Parent(s): d42a5db

Upload task output 1

Browse files
adapter_config.json CHANGED
@@ -29,13 +29,13 @@
29
  "rank_pattern": {},
30
  "revision": null,
31
  "target_modules": [
32
- "up_proj",
33
  "q_proj",
34
- "gate_proj",
35
  "k_proj",
 
 
36
  "v_proj",
37
- "down_proj",
38
- "o_proj"
39
  ],
40
  "target_parameters": null,
41
  "task_type": "CAUSAL_LM",
 
29
  "rank_pattern": {},
30
  "revision": null,
31
  "target_modules": [
 
32
  "q_proj",
33
+ "o_proj",
34
  "k_proj",
35
+ "up_proj",
36
+ "gate_proj",
37
  "v_proj",
38
+ "down_proj"
 
39
  ],
40
  "target_parameters": null,
41
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f7749a3248d4c22e7e19fc9ca88c990ff93c7187a4a736c35eccac6adc250de6
3
  size 70430032
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2fad742fb80ea2fa475f786a4b34488030e4f7e71fd4ea99b8421783125c9cec
3
  size 70430032
loss.txt CHANGED
@@ -1 +1 @@
1
- 1,no_eval_ppo_save_only
 
1
+ 2,no_eval_ppo_save_only
ppo_save_info.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "step": 1,
3
  "save_type": "model_only",
4
  "best_reward": null,
5
  "reward_history": []
 
1
  {
2
+ "step": 2,
3
  "save_type": "model_only",
4
  "best_reward": null,
5
  "reward_history": []
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6a5a830cfee2e939f8d5446d6e135abbf5d812a5a8885dfc1390454b3854a5cc
3
  size 5525
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ae3aad25c0cd91f6773cfe1689ded4f83098772cc4bcfd6a168b7cadf38e960
3
  size 5525