bluespy commited on
Commit
10b2415
·
verified ·
1 Parent(s): 57a1f04

Upload folder using huggingface_hub

Browse files
Files changed (3) hide show
  1. lora.pt +3 -0
  2. meta.json +75 -0
  3. model.pt +3 -0
lora.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1ee28c0076564643f6e2d2c2ff5f235d6c3e549a326b01b6648e9b0a5126fdf
3
+ size 3438829
meta.json ADDED
@@ -0,0 +1,75 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "type": "sft_baseline",
3
+ "epoch": 3,
4
+ "global_step": 27000,
5
+ "base_model_path": "/home/jingyuan.feng/.llama/checkpoints/Llama3.2-1B-instruct",
6
+ "lora": {
7
+ "num_modules": 64,
8
+ "module_names": [
9
+ "layers.0.self_attn.q_proj",
10
+ "layers.0.self_attn.k_proj",
11
+ "layers.0.self_attn.v_proj",
12
+ "layers.0.self_attn.o_proj",
13
+ "layers.1.self_attn.q_proj",
14
+ "layers.1.self_attn.k_proj",
15
+ "layers.1.self_attn.v_proj",
16
+ "layers.1.self_attn.o_proj",
17
+ "layers.2.self_attn.q_proj",
18
+ "layers.2.self_attn.k_proj",
19
+ "layers.2.self_attn.v_proj",
20
+ "layers.2.self_attn.o_proj",
21
+ "layers.3.self_attn.q_proj",
22
+ "layers.3.self_attn.k_proj",
23
+ "layers.3.self_attn.v_proj",
24
+ "layers.3.self_attn.o_proj",
25
+ "layers.4.self_attn.q_proj",
26
+ "layers.4.self_attn.k_proj",
27
+ "layers.4.self_attn.v_proj",
28
+ "layers.4.self_attn.o_proj",
29
+ "layers.5.self_attn.q_proj",
30
+ "layers.5.self_attn.k_proj",
31
+ "layers.5.self_attn.v_proj",
32
+ "layers.5.self_attn.o_proj",
33
+ "layers.6.self_attn.q_proj",
34
+ "layers.6.self_attn.k_proj",
35
+ "layers.6.self_attn.v_proj",
36
+ "layers.6.self_attn.o_proj",
37
+ "layers.7.self_attn.q_proj",
38
+ "layers.7.self_attn.k_proj",
39
+ "layers.7.self_attn.v_proj",
40
+ "layers.7.self_attn.o_proj",
41
+ "layers.8.self_attn.q_proj",
42
+ "layers.8.self_attn.k_proj",
43
+ "layers.8.self_attn.v_proj",
44
+ "layers.8.self_attn.o_proj",
45
+ "layers.9.self_attn.q_proj",
46
+ "layers.9.self_attn.k_proj",
47
+ "layers.9.self_attn.v_proj",
48
+ "layers.9.self_attn.o_proj",
49
+ "layers.10.self_attn.q_proj",
50
+ "layers.10.self_attn.k_proj",
51
+ "layers.10.self_attn.v_proj",
52
+ "layers.10.self_attn.o_proj",
53
+ "layers.11.self_attn.q_proj",
54
+ "layers.11.self_attn.k_proj",
55
+ "layers.11.self_attn.v_proj",
56
+ "layers.11.self_attn.o_proj",
57
+ "layers.12.self_attn.q_proj",
58
+ "layers.12.self_attn.k_proj",
59
+ "layers.12.self_attn.v_proj",
60
+ "layers.12.self_attn.o_proj",
61
+ "layers.13.self_attn.q_proj",
62
+ "layers.13.self_attn.k_proj",
63
+ "layers.13.self_attn.v_proj",
64
+ "layers.13.self_attn.o_proj",
65
+ "layers.14.self_attn.q_proj",
66
+ "layers.14.self_attn.k_proj",
67
+ "layers.14.self_attn.v_proj",
68
+ "layers.14.self_attn.o_proj",
69
+ "layers.15.self_attn.q_proj",
70
+ "layers.15.self_attn.k_proj",
71
+ "layers.15.self_attn.v_proj",
72
+ "layers.15.self_attn.o_proj"
73
+ ]
74
+ }
75
+ }
model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7931dc45ee51cd94ded686dceeed5d9d34c1d2e94b98813ac920d9270f71d0ba
3
+ size 250871808