PEFT
Safetensors
xretr0 commited on
Commit
e13152c
·
verified ·
1 Parent(s): 2904497

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -2,10 +2,9 @@
2
  base_model: Qwen/Qwen3-0.6B
3
  library_name: peft
4
  ---
 
5
 
6
- # Model Card for Model ID
7
-
8
- <!-- Provide a quick summary of what the model is/does. -->
9
 
10
 
11
 
 
2
  base_model: Qwen/Qwen3-0.6B
3
  library_name: peft
4
  ---
5
+ ### Framework versions
6
 
7
+ - PEFT 0.15.1ide a quick summary of what the model is/does. -->
 
 
8
 
9
 
10
 
adapter_config.json CHANGED
@@ -24,13 +24,13 @@
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
 
27
  "gate_proj",
 
28
  "down_proj",
 
29
  "k_proj",
30
- "q_proj",
31
- "up_proj",
32
- "v_proj",
33
- "o_proj"
34
  ],
35
  "task_type": "CAUSAL_LM",
36
  "trainable_token_indices": null,
 
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
27
+ "v_proj",
28
  "gate_proj",
29
+ "q_proj",
30
  "down_proj",
31
+ "o_proj",
32
  "k_proj",
33
+ "up_proj"
 
 
 
34
  ],
35
  "task_type": "CAUSAL_LM",
36
  "trainable_token_indices": null,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c213b339d3beff756c65a7624aea9cfacc0f041499406e83f01dbc2c149356e3
3
  size 161533160
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37cca7d0e06ebdcc365613800bb8f220b076275e2298986dbf2e77b3d8c91700
3
  size 161533160
trainer_state.json CHANGED
@@ -11,107 +11,107 @@
11
  "log_history": [
12
  {
13
  "epoch": 0.2,
14
- "grad_norm": 4.880182266235352,
15
  "learning_rate": 1e-05,
16
  "loss": 0.9092,
17
  "step": 1
18
  },
19
  {
20
  "epoch": 0.4,
21
- "grad_norm": 5.186452388763428,
22
  "learning_rate": 9.890738003669029e-06,
23
- "loss": 0.9494,
24
  "step": 2
25
  },
26
  {
27
  "epoch": 0.6,
28
- "grad_norm": 5.036950588226318,
29
  "learning_rate": 9.567727288213005e-06,
30
- "loss": 0.7849,
31
  "step": 3
32
  },
33
  {
34
  "epoch": 0.8,
35
- "grad_norm": 5.341607570648193,
36
  "learning_rate": 9.045084971874738e-06,
37
- "loss": 0.7692,
38
  "step": 4
39
  },
40
  {
41
  "epoch": 1.0,
42
- "grad_norm": 4.708967685699463,
43
  "learning_rate": 8.345653031794292e-06,
44
- "loss": 0.6502,
45
  "step": 5
46
  },
47
  {
48
  "epoch": 1.2,
49
- "grad_norm": 3.691345453262329,
50
  "learning_rate": 7.500000000000001e-06,
51
- "loss": 0.5581,
52
  "step": 6
53
  },
54
  {
55
  "epoch": 1.4,
56
- "grad_norm": 2.7508931159973145,
57
  "learning_rate": 6.545084971874738e-06,
58
- "loss": 0.5363,
59
  "step": 7
60
  },
61
  {
62
  "epoch": 1.6,
63
- "grad_norm": 1.858952283859253,
64
  "learning_rate": 5.522642316338268e-06,
65
- "loss": 0.5077,
66
  "step": 8
67
  },
68
  {
69
  "epoch": 1.8,
70
- "grad_norm": 1.805176019668579,
71
  "learning_rate": 4.477357683661734e-06,
72
- "loss": 0.5521,
73
  "step": 9
74
  },
75
  {
76
  "epoch": 2.0,
77
- "grad_norm": 1.4771124124526978,
78
  "learning_rate": 3.4549150281252635e-06,
79
- "loss": 0.4747,
80
  "step": 10
81
  },
82
  {
83
  "epoch": 2.2,
84
- "grad_norm": 1.5449007749557495,
85
  "learning_rate": 2.5000000000000015e-06,
86
- "loss": 0.5101,
87
  "step": 11
88
  },
89
  {
90
  "epoch": 2.4,
91
- "grad_norm": 1.393797755241394,
92
  "learning_rate": 1.6543469682057105e-06,
93
- "loss": 0.4949,
94
  "step": 12
95
  },
96
  {
97
  "epoch": 2.6,
98
- "grad_norm": 1.2324782609939575,
99
  "learning_rate": 9.549150281252633e-07,
100
- "loss": 0.4521,
101
  "step": 13
102
  },
103
  {
104
  "epoch": 2.8,
105
- "grad_norm": 1.2213858366012573,
106
  "learning_rate": 4.322727117869951e-07,
107
- "loss": 0.432,
108
  "step": 14
109
  },
110
  {
111
  "epoch": 3.0,
112
- "grad_norm": 1.4015122652053833,
113
  "learning_rate": 1.0926199633097156e-07,
114
- "loss": 0.4952,
115
  "step": 15
116
  }
117
  ],
 
11
  "log_history": [
12
  {
13
  "epoch": 0.2,
14
+ "grad_norm": 4.834155559539795,
15
  "learning_rate": 1e-05,
16
  "loss": 0.9092,
17
  "step": 1
18
  },
19
  {
20
  "epoch": 0.4,
21
+ "grad_norm": 5.137935638427734,
22
  "learning_rate": 9.890738003669029e-06,
23
+ "loss": 0.9497,
24
  "step": 2
25
  },
26
  {
27
  "epoch": 0.6,
28
+ "grad_norm": 4.987179279327393,
29
  "learning_rate": 9.567727288213005e-06,
30
+ "loss": 0.7857,
31
  "step": 3
32
  },
33
  {
34
  "epoch": 0.8,
35
+ "grad_norm": 5.295909881591797,
36
  "learning_rate": 9.045084971874738e-06,
37
+ "loss": 0.7714,
38
  "step": 4
39
  },
40
  {
41
  "epoch": 1.0,
42
+ "grad_norm": 4.681474208831787,
43
  "learning_rate": 8.345653031794292e-06,
44
+ "loss": 0.6515,
45
  "step": 5
46
  },
47
  {
48
  "epoch": 1.2,
49
+ "grad_norm": 3.703080892562866,
50
  "learning_rate": 7.500000000000001e-06,
51
+ "loss": 0.56,
52
  "step": 6
53
  },
54
  {
55
  "epoch": 1.4,
56
+ "grad_norm": 2.7864272594451904,
57
  "learning_rate": 6.545084971874738e-06,
58
+ "loss": 0.5379,
59
  "step": 7
60
  },
61
  {
62
  "epoch": 1.6,
63
+ "grad_norm": 1.8693835735321045,
64
  "learning_rate": 5.522642316338268e-06,
65
+ "loss": 0.509,
66
  "step": 8
67
  },
68
  {
69
  "epoch": 1.8,
70
+ "grad_norm": 1.8035385608673096,
71
  "learning_rate": 4.477357683661734e-06,
72
+ "loss": 0.5536,
73
  "step": 9
74
  },
75
  {
76
  "epoch": 2.0,
77
+ "grad_norm": 1.4770323038101196,
78
  "learning_rate": 3.4549150281252635e-06,
79
+ "loss": 0.4764,
80
  "step": 10
81
  },
82
  {
83
  "epoch": 2.2,
84
+ "grad_norm": 1.537730097770691,
85
  "learning_rate": 2.5000000000000015e-06,
86
+ "loss": 0.5115,
87
  "step": 11
88
  },
89
  {
90
  "epoch": 2.4,
91
+ "grad_norm": 1.3944271802902222,
92
  "learning_rate": 1.6543469682057105e-06,
93
+ "loss": 0.4966,
94
  "step": 12
95
  },
96
  {
97
  "epoch": 2.6,
98
+ "grad_norm": 1.2273967266082764,
99
  "learning_rate": 9.549150281252633e-07,
100
+ "loss": 0.4532,
101
  "step": 13
102
  },
103
  {
104
  "epoch": 2.8,
105
+ "grad_norm": 1.2184977531433105,
106
  "learning_rate": 4.322727117869951e-07,
107
+ "loss": 0.4331,
108
  "step": 14
109
  },
110
  {
111
  "epoch": 3.0,
112
+ "grad_norm": 1.390999436378479,
113
  "learning_rate": 1.0926199633097156e-07,
114
+ "loss": 0.4964,
115
  "step": 15
116
  }
117
  ],