auphong2707 commited on
Commit
b081ae3
·
verified ·
1 Parent(s): e50a679

Add files using upload-large-folder tool

Browse files
qwen25_0.5b_lora_official_5e-05/adapter_config.json CHANGED
@@ -26,10 +26,10 @@
26
  "revision": null,
27
  "target_modules": [
28
  "down_proj",
29
- "v_proj",
30
  "q_proj",
31
  "k_proj",
32
- "up_proj",
33
  "gate_proj"
34
  ],
35
  "task_type": "CAUSAL_LM",
 
26
  "revision": null,
27
  "target_modules": [
28
  "down_proj",
29
+ "up_proj",
30
  "q_proj",
31
  "k_proj",
32
+ "v_proj",
33
  "gate_proj"
34
  ],
35
  "task_type": "CAUSAL_LM",
qwen25_0.5b_lora_official_5e-05/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a590d93194183b7ecc702e7aee6dd5973cc692961742910c4032e256fe99f5ee
3
  size 576045776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d2811766a06e1b7100757892236fb1054545124bbf2dd8a85aee940cfa89d7cb
3
  size 576045776
qwen25_0.5b_lora_official_5e-05/complete_results.json CHANGED
The diff for this file is too large to render. See raw diff