TobDeBer commited on
Commit
2aee4e4
·
verified ·
1 Parent(s): d7d6230

Upload folder using huggingface_hub

Browse files
Files changed (2) hide show
  1. config.json +1 -121
  2. model.safetensors +2 -2
config.json CHANGED
@@ -38,212 +38,92 @@
38
  "model.layers.0.mlp.down_proj": 313,
39
  "model.layers.0.mlp.gate_proj": 322,
40
  "model.layers.0.mlp.up_proj": 322,
41
- "model.layers.0.self_attn.k_proj": 77,
42
- "model.layers.0.self_attn.o_proj": 121,
43
- "model.layers.0.self_attn.q_proj": 103,
44
- "model.layers.0.self_attn.v_proj": 88,
45
  "model.layers.1.mlp.down_proj": 318,
46
  "model.layers.1.mlp.gate_proj": 323,
47
  "model.layers.1.mlp.up_proj": 323,
48
- "model.layers.1.self_attn.k_proj": 102,
49
- "model.layers.1.self_attn.o_proj": 195,
50
- "model.layers.1.self_attn.q_proj": 244,
51
- "model.layers.1.self_attn.v_proj": 102,
52
  "model.layers.10.mlp.down_proj": 316,
53
  "model.layers.10.mlp.gate_proj": 321,
54
  "model.layers.10.mlp.up_proj": 324,
55
- "model.layers.10.self_attn.k_proj": 101,
56
- "model.layers.10.self_attn.o_proj": 224,
57
- "model.layers.10.self_attn.q_proj": 241,
58
- "model.layers.10.self_attn.v_proj": 106,
59
  "model.layers.11.mlp.down_proj": 315,
60
  "model.layers.11.mlp.gate_proj": 320,
61
  "model.layers.11.mlp.up_proj": 323,
62
- "model.layers.11.self_attn.k_proj": 100,
63
- "model.layers.11.self_attn.o_proj": 216,
64
- "model.layers.11.self_attn.q_proj": 241,
65
- "model.layers.11.self_attn.v_proj": 105,
66
  "model.layers.12.mlp.down_proj": 315,
67
  "model.layers.12.mlp.gate_proj": 319,
68
  "model.layers.12.mlp.up_proj": 322,
69
- "model.layers.12.self_attn.k_proj": 97,
70
- "model.layers.12.self_attn.o_proj": 216,
71
- "model.layers.12.self_attn.q_proj": 221,
72
- "model.layers.12.self_attn.v_proj": 105,
73
  "model.layers.13.mlp.down_proj": 309,
74
  "model.layers.13.mlp.gate_proj": 316,
75
  "model.layers.13.mlp.up_proj": 321,
76
- "model.layers.13.self_attn.k_proj": 97,
77
- "model.layers.13.self_attn.o_proj": 224,
78
- "model.layers.13.self_attn.q_proj": 229,
79
- "model.layers.13.self_attn.v_proj": 105,
80
  "model.layers.14.mlp.down_proj": 309,
81
  "model.layers.14.mlp.gate_proj": 316,
82
  "model.layers.14.mlp.up_proj": 321,
83
- "model.layers.14.self_attn.k_proj": 91,
84
- "model.layers.14.self_attn.o_proj": 226,
85
- "model.layers.14.self_attn.q_proj": 205,
86
- "model.layers.14.self_attn.v_proj": 103,
87
  "model.layers.15.mlp.down_proj": 309,
88
  "model.layers.15.mlp.gate_proj": 316,
89
  "model.layers.15.mlp.up_proj": 320,
90
- "model.layers.15.self_attn.k_proj": 90,
91
- "model.layers.15.self_attn.o_proj": 207,
92
- "model.layers.15.self_attn.q_proj": 202,
93
- "model.layers.15.self_attn.v_proj": 105,
94
  "model.layers.16.mlp.down_proj": 305,
95
  "model.layers.16.mlp.gate_proj": 317,
96
  "model.layers.16.mlp.up_proj": 321,
97
- "model.layers.16.self_attn.k_proj": 97,
98
- "model.layers.16.self_attn.o_proj": 219,
99
- "model.layers.16.self_attn.q_proj": 227,
100
- "model.layers.16.self_attn.v_proj": 103,
101
  "model.layers.17.mlp.down_proj": 303,
102
  "model.layers.17.mlp.gate_proj": 317,
103
  "model.layers.17.mlp.up_proj": 321,
104
- "model.layers.17.self_attn.k_proj": 89,
105
- "model.layers.17.self_attn.o_proj": 210,
106
- "model.layers.17.self_attn.q_proj": 192,
107
- "model.layers.17.self_attn.v_proj": 100,
108
  "model.layers.18.mlp.down_proj": 306,
109
  "model.layers.18.mlp.gate_proj": 315,
110
  "model.layers.18.mlp.up_proj": 319,
111
- "model.layers.18.self_attn.k_proj": 86,
112
- "model.layers.18.self_attn.o_proj": 181,
113
- "model.layers.18.self_attn.q_proj": 179,
114
- "model.layers.18.self_attn.v_proj": 105,
115
  "model.layers.19.mlp.down_proj": 306,
116
  "model.layers.19.mlp.gate_proj": 317,
117
  "model.layers.19.mlp.up_proj": 321,
118
- "model.layers.19.self_attn.k_proj": 95,
119
- "model.layers.19.self_attn.o_proj": 205,
120
- "model.layers.19.self_attn.q_proj": 204,
121
- "model.layers.19.self_attn.v_proj": 105,
122
  "model.layers.2.mlp.down_proj": 317,
123
  "model.layers.2.mlp.gate_proj": 324,
124
  "model.layers.2.mlp.up_proj": 324,
125
- "model.layers.2.self_attn.k_proj": 102,
126
- "model.layers.2.self_attn.o_proj": 200,
127
- "model.layers.2.self_attn.q_proj": 238,
128
- "model.layers.2.self_attn.v_proj": 102,
129
  "model.layers.20.mlp.down_proj": 309,
130
  "model.layers.20.mlp.gate_proj": 315,
131
  "model.layers.20.mlp.up_proj": 319,
132
- "model.layers.20.self_attn.k_proj": 87,
133
- "model.layers.20.self_attn.o_proj": 216,
134
- "model.layers.20.self_attn.q_proj": 185,
135
- "model.layers.20.self_attn.v_proj": 105,
136
  "model.layers.21.mlp.down_proj": 308,
137
  "model.layers.21.mlp.gate_proj": 314,
138
  "model.layers.21.mlp.up_proj": 318,
139
- "model.layers.21.self_attn.k_proj": 91,
140
- "model.layers.21.self_attn.o_proj": 213,
141
- "model.layers.21.self_attn.q_proj": 201,
142
- "model.layers.21.self_attn.v_proj": 103,
143
  "model.layers.22.mlp.down_proj": 306,
144
  "model.layers.22.mlp.gate_proj": 315,
145
  "model.layers.22.mlp.up_proj": 318,
146
- "model.layers.22.self_attn.k_proj": 92,
147
- "model.layers.22.self_attn.o_proj": 215,
148
- "model.layers.22.self_attn.q_proj": 194,
149
- "model.layers.22.self_attn.v_proj": 103,
150
  "model.layers.23.mlp.down_proj": 309,
151
  "model.layers.23.mlp.gate_proj": 313,
152
  "model.layers.23.mlp.up_proj": 317,
153
- "model.layers.23.self_attn.k_proj": 82,
154
- "model.layers.23.self_attn.o_proj": 201,
155
- "model.layers.23.self_attn.q_proj": 166,
156
- "model.layers.23.self_attn.v_proj": 93,
157
  "model.layers.24.mlp.down_proj": 311,
158
  "model.layers.24.mlp.gate_proj": 315,
159
  "model.layers.24.mlp.up_proj": 318,
160
- "model.layers.24.self_attn.k_proj": 85,
161
- "model.layers.24.self_attn.o_proj": 237,
162
- "model.layers.24.self_attn.q_proj": 195,
163
- "model.layers.24.self_attn.v_proj": 95,
164
  "model.layers.25.mlp.down_proj": 311,
165
  "model.layers.25.mlp.gate_proj": 316,
166
  "model.layers.25.mlp.up_proj": 319,
167
- "model.layers.25.self_attn.k_proj": 97,
168
- "model.layers.25.self_attn.o_proj": 241,
169
- "model.layers.25.self_attn.q_proj": 227,
170
- "model.layers.25.self_attn.v_proj": 106,
171
  "model.layers.26.mlp.down_proj": 311,
172
  "model.layers.26.mlp.gate_proj": 317,
173
  "model.layers.26.mlp.up_proj": 319,
174
- "model.layers.26.self_attn.k_proj": 93,
175
- "model.layers.26.self_attn.o_proj": 238,
176
- "model.layers.26.self_attn.q_proj": 198,
177
- "model.layers.26.self_attn.v_proj": 107,
178
  "model.layers.27.mlp.down_proj": 310,
179
  "model.layers.27.mlp.gate_proj": 316,
180
  "model.layers.27.mlp.up_proj": 319,
181
- "model.layers.27.self_attn.k_proj": 95,
182
- "model.layers.27.self_attn.o_proj": 237,
183
- "model.layers.27.self_attn.q_proj": 222,
184
- "model.layers.27.self_attn.v_proj": 110,
185
  "model.layers.28.mlp.down_proj": 311,
186
  "model.layers.28.mlp.gate_proj": 318,
187
  "model.layers.28.mlp.up_proj": 320,
188
- "model.layers.28.self_attn.k_proj": 87,
189
- "model.layers.28.self_attn.o_proj": 237,
190
- "model.layers.28.self_attn.q_proj": 199,
191
- "model.layers.28.self_attn.v_proj": 107,
192
  "model.layers.29.mlp.down_proj": 306,
193
  "model.layers.29.mlp.gate_proj": 314,
194
  "model.layers.29.mlp.up_proj": 316,
195
- "model.layers.29.self_attn.k_proj": 95,
196
- "model.layers.29.self_attn.o_proj": 227,
197
- "model.layers.29.self_attn.q_proj": 209,
198
- "model.layers.29.self_attn.v_proj": 108,
199
  "model.layers.3.mlp.down_proj": 315,
200
  "model.layers.3.mlp.gate_proj": 324,
201
  "model.layers.3.mlp.up_proj": 325,
202
- "model.layers.3.self_attn.k_proj": 100,
203
- "model.layers.3.self_attn.o_proj": 204,
204
- "model.layers.3.self_attn.q_proj": 229,
205
- "model.layers.3.self_attn.v_proj": 104,
206
  "model.layers.4.mlp.down_proj": 312,
207
  "model.layers.4.mlp.gate_proj": 323,
208
  "model.layers.4.mlp.up_proj": 325,
209
- "model.layers.4.self_attn.k_proj": 99,
210
- "model.layers.4.self_attn.o_proj": 218,
211
- "model.layers.4.self_attn.q_proj": 243,
212
- "model.layers.4.self_attn.v_proj": 106,
213
  "model.layers.5.mlp.down_proj": 312,
214
  "model.layers.5.mlp.gate_proj": 322,
215
  "model.layers.5.mlp.up_proj": 324,
216
- "model.layers.5.self_attn.k_proj": 98,
217
- "model.layers.5.self_attn.o_proj": 217,
218
- "model.layers.5.self_attn.q_proj": 234,
219
- "model.layers.5.self_attn.v_proj": 105,
220
  "model.layers.6.mlp.down_proj": 313,
221
  "model.layers.6.mlp.gate_proj": 322,
222
  "model.layers.6.mlp.up_proj": 324,
223
- "model.layers.6.self_attn.k_proj": 100,
224
- "model.layers.6.self_attn.o_proj": 224,
225
- "model.layers.6.self_attn.q_proj": 235,
226
- "model.layers.6.self_attn.v_proj": 104,
227
  "model.layers.7.mlp.down_proj": 313,
228
  "model.layers.7.mlp.gate_proj": 321,
229
  "model.layers.7.mlp.up_proj": 323,
230
- "model.layers.7.self_attn.k_proj": 96,
231
- "model.layers.7.self_attn.o_proj": 210,
232
- "model.layers.7.self_attn.q_proj": 222,
233
- "model.layers.7.self_attn.v_proj": 102,
234
  "model.layers.8.mlp.down_proj": 314,
235
  "model.layers.8.mlp.gate_proj": 321,
236
  "model.layers.8.mlp.up_proj": 323,
237
- "model.layers.8.self_attn.k_proj": 94,
238
- "model.layers.8.self_attn.o_proj": 228,
239
- "model.layers.8.self_attn.q_proj": 220,
240
- "model.layers.8.self_attn.v_proj": 104,
241
  "model.layers.9.mlp.down_proj": 316,
242
  "model.layers.9.mlp.gate_proj": 321,
243
- "model.layers.9.mlp.up_proj": 324,
244
- "model.layers.9.self_attn.k_proj": 96,
245
- "model.layers.9.self_attn.o_proj": 221,
246
- "model.layers.9.self_attn.q_proj": 219,
247
- "model.layers.9.self_attn.v_proj": 100
248
  }
249
  }
 
38
  "model.layers.0.mlp.down_proj": 313,
39
  "model.layers.0.mlp.gate_proj": 322,
40
  "model.layers.0.mlp.up_proj": 322,
 
 
 
 
41
  "model.layers.1.mlp.down_proj": 318,
42
  "model.layers.1.mlp.gate_proj": 323,
43
  "model.layers.1.mlp.up_proj": 323,
 
 
 
 
44
  "model.layers.10.mlp.down_proj": 316,
45
  "model.layers.10.mlp.gate_proj": 321,
46
  "model.layers.10.mlp.up_proj": 324,
 
 
 
 
47
  "model.layers.11.mlp.down_proj": 315,
48
  "model.layers.11.mlp.gate_proj": 320,
49
  "model.layers.11.mlp.up_proj": 323,
 
 
 
 
50
  "model.layers.12.mlp.down_proj": 315,
51
  "model.layers.12.mlp.gate_proj": 319,
52
  "model.layers.12.mlp.up_proj": 322,
 
 
 
 
53
  "model.layers.13.mlp.down_proj": 309,
54
  "model.layers.13.mlp.gate_proj": 316,
55
  "model.layers.13.mlp.up_proj": 321,
 
 
 
 
56
  "model.layers.14.mlp.down_proj": 309,
57
  "model.layers.14.mlp.gate_proj": 316,
58
  "model.layers.14.mlp.up_proj": 321,
 
 
 
 
59
  "model.layers.15.mlp.down_proj": 309,
60
  "model.layers.15.mlp.gate_proj": 316,
61
  "model.layers.15.mlp.up_proj": 320,
 
 
 
 
62
  "model.layers.16.mlp.down_proj": 305,
63
  "model.layers.16.mlp.gate_proj": 317,
64
  "model.layers.16.mlp.up_proj": 321,
 
 
 
 
65
  "model.layers.17.mlp.down_proj": 303,
66
  "model.layers.17.mlp.gate_proj": 317,
67
  "model.layers.17.mlp.up_proj": 321,
 
 
 
 
68
  "model.layers.18.mlp.down_proj": 306,
69
  "model.layers.18.mlp.gate_proj": 315,
70
  "model.layers.18.mlp.up_proj": 319,
 
 
 
 
71
  "model.layers.19.mlp.down_proj": 306,
72
  "model.layers.19.mlp.gate_proj": 317,
73
  "model.layers.19.mlp.up_proj": 321,
 
 
 
 
74
  "model.layers.2.mlp.down_proj": 317,
75
  "model.layers.2.mlp.gate_proj": 324,
76
  "model.layers.2.mlp.up_proj": 324,
 
 
 
 
77
  "model.layers.20.mlp.down_proj": 309,
78
  "model.layers.20.mlp.gate_proj": 315,
79
  "model.layers.20.mlp.up_proj": 319,
 
 
 
 
80
  "model.layers.21.mlp.down_proj": 308,
81
  "model.layers.21.mlp.gate_proj": 314,
82
  "model.layers.21.mlp.up_proj": 318,
 
 
 
 
83
  "model.layers.22.mlp.down_proj": 306,
84
  "model.layers.22.mlp.gate_proj": 315,
85
  "model.layers.22.mlp.up_proj": 318,
 
 
 
 
86
  "model.layers.23.mlp.down_proj": 309,
87
  "model.layers.23.mlp.gate_proj": 313,
88
  "model.layers.23.mlp.up_proj": 317,
 
 
 
 
89
  "model.layers.24.mlp.down_proj": 311,
90
  "model.layers.24.mlp.gate_proj": 315,
91
  "model.layers.24.mlp.up_proj": 318,
 
 
 
 
92
  "model.layers.25.mlp.down_proj": 311,
93
  "model.layers.25.mlp.gate_proj": 316,
94
  "model.layers.25.mlp.up_proj": 319,
 
 
 
 
95
  "model.layers.26.mlp.down_proj": 311,
96
  "model.layers.26.mlp.gate_proj": 317,
97
  "model.layers.26.mlp.up_proj": 319,
 
 
 
 
98
  "model.layers.27.mlp.down_proj": 310,
99
  "model.layers.27.mlp.gate_proj": 316,
100
  "model.layers.27.mlp.up_proj": 319,
 
 
 
 
101
  "model.layers.28.mlp.down_proj": 311,
102
  "model.layers.28.mlp.gate_proj": 318,
103
  "model.layers.28.mlp.up_proj": 320,
 
 
 
 
104
  "model.layers.29.mlp.down_proj": 306,
105
  "model.layers.29.mlp.gate_proj": 314,
106
  "model.layers.29.mlp.up_proj": 316,
 
 
 
 
107
  "model.layers.3.mlp.down_proj": 315,
108
  "model.layers.3.mlp.gate_proj": 324,
109
  "model.layers.3.mlp.up_proj": 325,
 
 
 
 
110
  "model.layers.4.mlp.down_proj": 312,
111
  "model.layers.4.mlp.gate_proj": 323,
112
  "model.layers.4.mlp.up_proj": 325,
 
 
 
 
113
  "model.layers.5.mlp.down_proj": 312,
114
  "model.layers.5.mlp.gate_proj": 322,
115
  "model.layers.5.mlp.up_proj": 324,
 
 
 
 
116
  "model.layers.6.mlp.down_proj": 313,
117
  "model.layers.6.mlp.gate_proj": 322,
118
  "model.layers.6.mlp.up_proj": 324,
 
 
 
 
119
  "model.layers.7.mlp.down_proj": 313,
120
  "model.layers.7.mlp.gate_proj": 321,
121
  "model.layers.7.mlp.up_proj": 323,
 
 
 
 
122
  "model.layers.8.mlp.down_proj": 314,
123
  "model.layers.8.mlp.gate_proj": 321,
124
  "model.layers.8.mlp.up_proj": 323,
 
 
 
 
125
  "model.layers.9.mlp.down_proj": 316,
126
  "model.layers.9.mlp.gate_proj": 321,
127
+ "model.layers.9.mlp.up_proj": 324
 
 
 
 
128
  }
129
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:39814d463d4dc3c72b4c65a0b55a8222e15d0ab0fb64eb5c139e9c4673975158
3
- size 272375716
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d4e4bcc15f17f4f526542fd5fe3967d211281a8c36ff5fb9b0677fe2b51be8b
3
+ size 286939918