prince-canuma commited on
Commit
4fcbe92
·
verified ·
1 Parent(s): c60f42a

Upload folder using huggingface_hub

Browse files
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a02c9384f0ac15ec3b2c37f82306449308c8f3fe01e7cb966f1adba15cba8ba8
3
- size 5354888445
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a803fa7da14abccba67fd173045831db43d83f4714b058f267e5ac20de47bf9
3
+ size 5153392166
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:85ff58c2be574febac84eff7a5f3168fa5da9a839b0c41df9cf84b4681ead5f0
3
- size 3749723058
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb0162a4d0659df905740556bd562ead0b4a7fd42d55a28089c915837b25a304
3
+ size 1291001325
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 9104484160
4
  },
5
  "weight_map": {
6
  "language_model.lm_head.biases": "model-00002-of-00002.safetensors",
@@ -264,9 +264,9 @@
264
  "language_model.model.blocks.22.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
265
  "language_model.model.blocks.23.attn_norm.weight": "model-00001-of-00002.safetensors",
266
  "language_model.model.blocks.23.ff_norm.weight": "model-00001-of-00002.safetensors",
267
- "language_model.model.blocks.23.mlp.ff_out.biases": "model-00002-of-00002.safetensors",
268
- "language_model.model.blocks.23.mlp.ff_out.scales": "model-00002-of-00002.safetensors",
269
- "language_model.model.blocks.23.mlp.ff_out.weight": "model-00002-of-00002.safetensors",
270
  "language_model.model.blocks.23.mlp.ff_proj.biases": "model-00001-of-00002.safetensors",
271
  "language_model.model.blocks.23.mlp.ff_proj.scales": "model-00001-of-00002.safetensors",
272
  "language_model.model.blocks.23.mlp.ff_proj.weight": "model-00001-of-00002.safetensors",
@@ -278,102 +278,102 @@
278
  "language_model.model.blocks.23.self_attn.attn_out.weight": "model-00001-of-00002.safetensors",
279
  "language_model.model.blocks.23.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
280
  "language_model.model.blocks.23.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
281
- "language_model.model.blocks.24.attn_norm.weight": "model-00002-of-00002.safetensors",
282
- "language_model.model.blocks.24.ff_norm.weight": "model-00002-of-00002.safetensors",
283
- "language_model.model.blocks.24.mlp.ff_out.biases": "model-00002-of-00002.safetensors",
284
- "language_model.model.blocks.24.mlp.ff_out.scales": "model-00002-of-00002.safetensors",
285
- "language_model.model.blocks.24.mlp.ff_out.weight": "model-00002-of-00002.safetensors",
286
- "language_model.model.blocks.24.mlp.ff_proj.biases": "model-00002-of-00002.safetensors",
287
- "language_model.model.blocks.24.mlp.ff_proj.scales": "model-00002-of-00002.safetensors",
288
- "language_model.model.blocks.24.mlp.ff_proj.weight": "model-00002-of-00002.safetensors",
289
- "language_model.model.blocks.24.self_attn.att_proj.biases": "model-00002-of-00002.safetensors",
290
- "language_model.model.blocks.24.self_attn.att_proj.scales": "model-00002-of-00002.safetensors",
291
- "language_model.model.blocks.24.self_attn.att_proj.weight": "model-00002-of-00002.safetensors",
292
- "language_model.model.blocks.24.self_attn.attn_out.biases": "model-00002-of-00002.safetensors",
293
- "language_model.model.blocks.24.self_attn.attn_out.scales": "model-00002-of-00002.safetensors",
294
- "language_model.model.blocks.24.self_attn.attn_out.weight": "model-00002-of-00002.safetensors",
295
- "language_model.model.blocks.24.self_attn.k_norm.weight": "model-00002-of-00002.safetensors",
296
- "language_model.model.blocks.24.self_attn.q_norm.weight": "model-00002-of-00002.safetensors",
297
- "language_model.model.blocks.25.attn_norm.weight": "model-00002-of-00002.safetensors",
298
- "language_model.model.blocks.25.ff_norm.weight": "model-00002-of-00002.safetensors",
299
- "language_model.model.blocks.25.mlp.ff_out.biases": "model-00002-of-00002.safetensors",
300
- "language_model.model.blocks.25.mlp.ff_out.scales": "model-00002-of-00002.safetensors",
301
- "language_model.model.blocks.25.mlp.ff_out.weight": "model-00002-of-00002.safetensors",
302
- "language_model.model.blocks.25.mlp.ff_proj.biases": "model-00002-of-00002.safetensors",
303
- "language_model.model.blocks.25.mlp.ff_proj.scales": "model-00002-of-00002.safetensors",
304
- "language_model.model.blocks.25.mlp.ff_proj.weight": "model-00002-of-00002.safetensors",
305
- "language_model.model.blocks.25.self_attn.att_proj.biases": "model-00002-of-00002.safetensors",
306
- "language_model.model.blocks.25.self_attn.att_proj.scales": "model-00002-of-00002.safetensors",
307
- "language_model.model.blocks.25.self_attn.att_proj.weight": "model-00002-of-00002.safetensors",
308
- "language_model.model.blocks.25.self_attn.attn_out.biases": "model-00002-of-00002.safetensors",
309
- "language_model.model.blocks.25.self_attn.attn_out.scales": "model-00002-of-00002.safetensors",
310
- "language_model.model.blocks.25.self_attn.attn_out.weight": "model-00002-of-00002.safetensors",
311
- "language_model.model.blocks.25.self_attn.k_norm.weight": "model-00002-of-00002.safetensors",
312
- "language_model.model.blocks.25.self_attn.q_norm.weight": "model-00002-of-00002.safetensors",
313
- "language_model.model.blocks.26.attn_norm.weight": "model-00002-of-00002.safetensors",
314
- "language_model.model.blocks.26.ff_norm.weight": "model-00002-of-00002.safetensors",
315
- "language_model.model.blocks.26.mlp.ff_out.biases": "model-00002-of-00002.safetensors",
316
- "language_model.model.blocks.26.mlp.ff_out.scales": "model-00002-of-00002.safetensors",
317
- "language_model.model.blocks.26.mlp.ff_out.weight": "model-00002-of-00002.safetensors",
318
- "language_model.model.blocks.26.mlp.ff_proj.biases": "model-00002-of-00002.safetensors",
319
- "language_model.model.blocks.26.mlp.ff_proj.scales": "model-00002-of-00002.safetensors",
320
- "language_model.model.blocks.26.mlp.ff_proj.weight": "model-00002-of-00002.safetensors",
321
- "language_model.model.blocks.26.self_attn.att_proj.biases": "model-00002-of-00002.safetensors",
322
- "language_model.model.blocks.26.self_attn.att_proj.scales": "model-00002-of-00002.safetensors",
323
- "language_model.model.blocks.26.self_attn.att_proj.weight": "model-00002-of-00002.safetensors",
324
- "language_model.model.blocks.26.self_attn.attn_out.biases": "model-00002-of-00002.safetensors",
325
- "language_model.model.blocks.26.self_attn.attn_out.scales": "model-00002-of-00002.safetensors",
326
- "language_model.model.blocks.26.self_attn.attn_out.weight": "model-00002-of-00002.safetensors",
327
- "language_model.model.blocks.26.self_attn.k_norm.weight": "model-00002-of-00002.safetensors",
328
- "language_model.model.blocks.26.self_attn.q_norm.weight": "model-00002-of-00002.safetensors",
329
- "language_model.model.blocks.27.attn_norm.weight": "model-00002-of-00002.safetensors",
330
- "language_model.model.blocks.27.ff_norm.weight": "model-00002-of-00002.safetensors",
331
- "language_model.model.blocks.27.mlp.ff_out.biases": "model-00002-of-00002.safetensors",
332
- "language_model.model.blocks.27.mlp.ff_out.scales": "model-00002-of-00002.safetensors",
333
- "language_model.model.blocks.27.mlp.ff_out.weight": "model-00002-of-00002.safetensors",
334
- "language_model.model.blocks.27.mlp.ff_proj.biases": "model-00002-of-00002.safetensors",
335
- "language_model.model.blocks.27.mlp.ff_proj.scales": "model-00002-of-00002.safetensors",
336
- "language_model.model.blocks.27.mlp.ff_proj.weight": "model-00002-of-00002.safetensors",
337
- "language_model.model.blocks.27.self_attn.att_proj.biases": "model-00002-of-00002.safetensors",
338
- "language_model.model.blocks.27.self_attn.att_proj.scales": "model-00002-of-00002.safetensors",
339
- "language_model.model.blocks.27.self_attn.att_proj.weight": "model-00002-of-00002.safetensors",
340
- "language_model.model.blocks.27.self_attn.attn_out.biases": "model-00002-of-00002.safetensors",
341
- "language_model.model.blocks.27.self_attn.attn_out.scales": "model-00002-of-00002.safetensors",
342
- "language_model.model.blocks.27.self_attn.attn_out.weight": "model-00002-of-00002.safetensors",
343
- "language_model.model.blocks.27.self_attn.k_norm.weight": "model-00002-of-00002.safetensors",
344
- "language_model.model.blocks.27.self_attn.q_norm.weight": "model-00002-of-00002.safetensors",
345
- "language_model.model.blocks.28.attn_norm.weight": "model-00002-of-00002.safetensors",
346
- "language_model.model.blocks.28.ff_norm.weight": "model-00002-of-00002.safetensors",
347
- "language_model.model.blocks.28.mlp.ff_out.biases": "model-00002-of-00002.safetensors",
348
- "language_model.model.blocks.28.mlp.ff_out.scales": "model-00002-of-00002.safetensors",
349
- "language_model.model.blocks.28.mlp.ff_out.weight": "model-00002-of-00002.safetensors",
350
- "language_model.model.blocks.28.mlp.ff_proj.biases": "model-00002-of-00002.safetensors",
351
- "language_model.model.blocks.28.mlp.ff_proj.scales": "model-00002-of-00002.safetensors",
352
- "language_model.model.blocks.28.mlp.ff_proj.weight": "model-00002-of-00002.safetensors",
353
- "language_model.model.blocks.28.self_attn.att_proj.biases": "model-00002-of-00002.safetensors",
354
- "language_model.model.blocks.28.self_attn.att_proj.scales": "model-00002-of-00002.safetensors",
355
- "language_model.model.blocks.28.self_attn.att_proj.weight": "model-00002-of-00002.safetensors",
356
- "language_model.model.blocks.28.self_attn.attn_out.biases": "model-00002-of-00002.safetensors",
357
- "language_model.model.blocks.28.self_attn.attn_out.scales": "model-00002-of-00002.safetensors",
358
- "language_model.model.blocks.28.self_attn.attn_out.weight": "model-00002-of-00002.safetensors",
359
- "language_model.model.blocks.28.self_attn.k_norm.weight": "model-00002-of-00002.safetensors",
360
- "language_model.model.blocks.28.self_attn.q_norm.weight": "model-00002-of-00002.safetensors",
361
- "language_model.model.blocks.29.attn_norm.weight": "model-00002-of-00002.safetensors",
362
- "language_model.model.blocks.29.ff_norm.weight": "model-00002-of-00002.safetensors",
363
- "language_model.model.blocks.29.mlp.ff_out.biases": "model-00002-of-00002.safetensors",
364
- "language_model.model.blocks.29.mlp.ff_out.scales": "model-00002-of-00002.safetensors",
365
- "language_model.model.blocks.29.mlp.ff_out.weight": "model-00002-of-00002.safetensors",
366
- "language_model.model.blocks.29.mlp.ff_proj.biases": "model-00002-of-00002.safetensors",
367
- "language_model.model.blocks.29.mlp.ff_proj.scales": "model-00002-of-00002.safetensors",
368
- "language_model.model.blocks.29.mlp.ff_proj.weight": "model-00002-of-00002.safetensors",
369
- "language_model.model.blocks.29.self_attn.att_proj.biases": "model-00002-of-00002.safetensors",
370
- "language_model.model.blocks.29.self_attn.att_proj.scales": "model-00002-of-00002.safetensors",
371
- "language_model.model.blocks.29.self_attn.att_proj.weight": "model-00002-of-00002.safetensors",
372
- "language_model.model.blocks.29.self_attn.attn_out.biases": "model-00002-of-00002.safetensors",
373
- "language_model.model.blocks.29.self_attn.attn_out.scales": "model-00002-of-00002.safetensors",
374
- "language_model.model.blocks.29.self_attn.attn_out.weight": "model-00002-of-00002.safetensors",
375
- "language_model.model.blocks.29.self_attn.k_norm.weight": "model-00002-of-00002.safetensors",
376
- "language_model.model.blocks.29.self_attn.q_norm.weight": "model-00002-of-00002.safetensors",
377
  "language_model.model.blocks.3.attn_norm.weight": "model-00001-of-00002.safetensors",
378
  "language_model.model.blocks.3.ff_norm.weight": "model-00001-of-00002.safetensors",
379
  "language_model.model.blocks.3.mlp.ff_out.biases": "model-00001-of-00002.safetensors",
@@ -390,102 +390,102 @@
390
  "language_model.model.blocks.3.self_attn.attn_out.weight": "model-00001-of-00002.safetensors",
391
  "language_model.model.blocks.3.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
392
  "language_model.model.blocks.3.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
393
- "language_model.model.blocks.30.attn_norm.weight": "model-00002-of-00002.safetensors",
394
- "language_model.model.blocks.30.ff_norm.weight": "model-00002-of-00002.safetensors",
395
- "language_model.model.blocks.30.mlp.ff_out.biases": "model-00002-of-00002.safetensors",
396
- "language_model.model.blocks.30.mlp.ff_out.scales": "model-00002-of-00002.safetensors",
397
- "language_model.model.blocks.30.mlp.ff_out.weight": "model-00002-of-00002.safetensors",
398
- "language_model.model.blocks.30.mlp.ff_proj.biases": "model-00002-of-00002.safetensors",
399
- "language_model.model.blocks.30.mlp.ff_proj.scales": "model-00002-of-00002.safetensors",
400
- "language_model.model.blocks.30.mlp.ff_proj.weight": "model-00002-of-00002.safetensors",
401
- "language_model.model.blocks.30.self_attn.att_proj.biases": "model-00002-of-00002.safetensors",
402
- "language_model.model.blocks.30.self_attn.att_proj.scales": "model-00002-of-00002.safetensors",
403
- "language_model.model.blocks.30.self_attn.att_proj.weight": "model-00002-of-00002.safetensors",
404
- "language_model.model.blocks.30.self_attn.attn_out.biases": "model-00002-of-00002.safetensors",
405
- "language_model.model.blocks.30.self_attn.attn_out.scales": "model-00002-of-00002.safetensors",
406
- "language_model.model.blocks.30.self_attn.attn_out.weight": "model-00002-of-00002.safetensors",
407
- "language_model.model.blocks.30.self_attn.k_norm.weight": "model-00002-of-00002.safetensors",
408
- "language_model.model.blocks.30.self_attn.q_norm.weight": "model-00002-of-00002.safetensors",
409
- "language_model.model.blocks.31.attn_norm.weight": "model-00002-of-00002.safetensors",
410
- "language_model.model.blocks.31.ff_norm.weight": "model-00002-of-00002.safetensors",
411
- "language_model.model.blocks.31.mlp.ff_out.biases": "model-00002-of-00002.safetensors",
412
- "language_model.model.blocks.31.mlp.ff_out.scales": "model-00002-of-00002.safetensors",
413
- "language_model.model.blocks.31.mlp.ff_out.weight": "model-00002-of-00002.safetensors",
414
- "language_model.model.blocks.31.mlp.ff_proj.biases": "model-00002-of-00002.safetensors",
415
- "language_model.model.blocks.31.mlp.ff_proj.scales": "model-00002-of-00002.safetensors",
416
- "language_model.model.blocks.31.mlp.ff_proj.weight": "model-00002-of-00002.safetensors",
417
- "language_model.model.blocks.31.self_attn.att_proj.biases": "model-00002-of-00002.safetensors",
418
- "language_model.model.blocks.31.self_attn.att_proj.scales": "model-00002-of-00002.safetensors",
419
- "language_model.model.blocks.31.self_attn.att_proj.weight": "model-00002-of-00002.safetensors",
420
- "language_model.model.blocks.31.self_attn.attn_out.biases": "model-00002-of-00002.safetensors",
421
- "language_model.model.blocks.31.self_attn.attn_out.scales": "model-00002-of-00002.safetensors",
422
- "language_model.model.blocks.31.self_attn.attn_out.weight": "model-00002-of-00002.safetensors",
423
- "language_model.model.blocks.31.self_attn.k_norm.weight": "model-00002-of-00002.safetensors",
424
- "language_model.model.blocks.31.self_attn.q_norm.weight": "model-00002-of-00002.safetensors",
425
- "language_model.model.blocks.32.attn_norm.weight": "model-00002-of-00002.safetensors",
426
- "language_model.model.blocks.32.ff_norm.weight": "model-00002-of-00002.safetensors",
427
- "language_model.model.blocks.32.mlp.ff_out.biases": "model-00002-of-00002.safetensors",
428
- "language_model.model.blocks.32.mlp.ff_out.scales": "model-00002-of-00002.safetensors",
429
- "language_model.model.blocks.32.mlp.ff_out.weight": "model-00002-of-00002.safetensors",
430
- "language_model.model.blocks.32.mlp.ff_proj.biases": "model-00002-of-00002.safetensors",
431
- "language_model.model.blocks.32.mlp.ff_proj.scales": "model-00002-of-00002.safetensors",
432
- "language_model.model.blocks.32.mlp.ff_proj.weight": "model-00002-of-00002.safetensors",
433
- "language_model.model.blocks.32.self_attn.att_proj.biases": "model-00002-of-00002.safetensors",
434
- "language_model.model.blocks.32.self_attn.att_proj.scales": "model-00002-of-00002.safetensors",
435
- "language_model.model.blocks.32.self_attn.att_proj.weight": "model-00002-of-00002.safetensors",
436
- "language_model.model.blocks.32.self_attn.attn_out.biases": "model-00002-of-00002.safetensors",
437
- "language_model.model.blocks.32.self_attn.attn_out.scales": "model-00002-of-00002.safetensors",
438
- "language_model.model.blocks.32.self_attn.attn_out.weight": "model-00002-of-00002.safetensors",
439
- "language_model.model.blocks.32.self_attn.k_norm.weight": "model-00002-of-00002.safetensors",
440
- "language_model.model.blocks.32.self_attn.q_norm.weight": "model-00002-of-00002.safetensors",
441
- "language_model.model.blocks.33.attn_norm.weight": "model-00002-of-00002.safetensors",
442
- "language_model.model.blocks.33.ff_norm.weight": "model-00002-of-00002.safetensors",
443
- "language_model.model.blocks.33.mlp.ff_out.biases": "model-00002-of-00002.safetensors",
444
- "language_model.model.blocks.33.mlp.ff_out.scales": "model-00002-of-00002.safetensors",
445
- "language_model.model.blocks.33.mlp.ff_out.weight": "model-00002-of-00002.safetensors",
446
- "language_model.model.blocks.33.mlp.ff_proj.biases": "model-00002-of-00002.safetensors",
447
- "language_model.model.blocks.33.mlp.ff_proj.scales": "model-00002-of-00002.safetensors",
448
- "language_model.model.blocks.33.mlp.ff_proj.weight": "model-00002-of-00002.safetensors",
449
- "language_model.model.blocks.33.self_attn.att_proj.biases": "model-00002-of-00002.safetensors",
450
- "language_model.model.blocks.33.self_attn.att_proj.scales": "model-00002-of-00002.safetensors",
451
- "language_model.model.blocks.33.self_attn.att_proj.weight": "model-00002-of-00002.safetensors",
452
- "language_model.model.blocks.33.self_attn.attn_out.biases": "model-00002-of-00002.safetensors",
453
- "language_model.model.blocks.33.self_attn.attn_out.scales": "model-00002-of-00002.safetensors",
454
- "language_model.model.blocks.33.self_attn.attn_out.weight": "model-00002-of-00002.safetensors",
455
- "language_model.model.blocks.33.self_attn.k_norm.weight": "model-00002-of-00002.safetensors",
456
- "language_model.model.blocks.33.self_attn.q_norm.weight": "model-00002-of-00002.safetensors",
457
- "language_model.model.blocks.34.attn_norm.weight": "model-00002-of-00002.safetensors",
458
- "language_model.model.blocks.34.ff_norm.weight": "model-00002-of-00002.safetensors",
459
- "language_model.model.blocks.34.mlp.ff_out.biases": "model-00002-of-00002.safetensors",
460
- "language_model.model.blocks.34.mlp.ff_out.scales": "model-00002-of-00002.safetensors",
461
- "language_model.model.blocks.34.mlp.ff_out.weight": "model-00002-of-00002.safetensors",
462
- "language_model.model.blocks.34.mlp.ff_proj.biases": "model-00002-of-00002.safetensors",
463
- "language_model.model.blocks.34.mlp.ff_proj.scales": "model-00002-of-00002.safetensors",
464
- "language_model.model.blocks.34.mlp.ff_proj.weight": "model-00002-of-00002.safetensors",
465
- "language_model.model.blocks.34.self_attn.att_proj.biases": "model-00002-of-00002.safetensors",
466
- "language_model.model.blocks.34.self_attn.att_proj.scales": "model-00002-of-00002.safetensors",
467
- "language_model.model.blocks.34.self_attn.att_proj.weight": "model-00002-of-00002.safetensors",
468
- "language_model.model.blocks.34.self_attn.attn_out.biases": "model-00002-of-00002.safetensors",
469
- "language_model.model.blocks.34.self_attn.attn_out.scales": "model-00002-of-00002.safetensors",
470
- "language_model.model.blocks.34.self_attn.attn_out.weight": "model-00002-of-00002.safetensors",
471
- "language_model.model.blocks.34.self_attn.k_norm.weight": "model-00002-of-00002.safetensors",
472
- "language_model.model.blocks.34.self_attn.q_norm.weight": "model-00002-of-00002.safetensors",
473
- "language_model.model.blocks.35.attn_norm.weight": "model-00002-of-00002.safetensors",
474
- "language_model.model.blocks.35.ff_norm.weight": "model-00002-of-00002.safetensors",
475
- "language_model.model.blocks.35.mlp.ff_out.biases": "model-00002-of-00002.safetensors",
476
- "language_model.model.blocks.35.mlp.ff_out.scales": "model-00002-of-00002.safetensors",
477
- "language_model.model.blocks.35.mlp.ff_out.weight": "model-00002-of-00002.safetensors",
478
- "language_model.model.blocks.35.mlp.ff_proj.biases": "model-00002-of-00002.safetensors",
479
- "language_model.model.blocks.35.mlp.ff_proj.scales": "model-00002-of-00002.safetensors",
480
- "language_model.model.blocks.35.mlp.ff_proj.weight": "model-00002-of-00002.safetensors",
481
- "language_model.model.blocks.35.self_attn.att_proj.biases": "model-00002-of-00002.safetensors",
482
- "language_model.model.blocks.35.self_attn.att_proj.scales": "model-00002-of-00002.safetensors",
483
- "language_model.model.blocks.35.self_attn.att_proj.weight": "model-00002-of-00002.safetensors",
484
- "language_model.model.blocks.35.self_attn.attn_out.biases": "model-00002-of-00002.safetensors",
485
- "language_model.model.blocks.35.self_attn.attn_out.scales": "model-00002-of-00002.safetensors",
486
- "language_model.model.blocks.35.self_attn.attn_out.weight": "model-00002-of-00002.safetensors",
487
- "language_model.model.blocks.35.self_attn.k_norm.weight": "model-00002-of-00002.safetensors",
488
- "language_model.model.blocks.35.self_attn.q_norm.weight": "model-00002-of-00002.safetensors",
489
  "language_model.model.blocks.4.attn_norm.weight": "model-00001-of-00002.safetensors",
490
  "language_model.model.blocks.4.ff_norm.weight": "model-00001-of-00002.safetensors",
491
  "language_model.model.blocks.4.mlp.ff_out.biases": "model-00001-of-00002.safetensors",
@@ -582,7 +582,7 @@
582
  "language_model.model.blocks.9.self_attn.attn_out.weight": "model-00001-of-00002.safetensors",
583
  "language_model.model.blocks.9.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
584
  "language_model.model.blocks.9.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
585
- "language_model.model.ln_f.weight": "model-00002-of-00002.safetensors",
586
  "language_model.model.wte.embedding": "model-00001-of-00002.safetensors",
587
  "language_model.model.wte.new_embedding": "model-00001-of-00002.safetensors",
588
  "vision_tower.image_pooling_2d.wk.bias": "model-00002-of-00002.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 6444266400
4
  },
5
  "weight_map": {
6
  "language_model.lm_head.biases": "model-00002-of-00002.safetensors",
 
264
  "language_model.model.blocks.22.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
265
  "language_model.model.blocks.23.attn_norm.weight": "model-00001-of-00002.safetensors",
266
  "language_model.model.blocks.23.ff_norm.weight": "model-00001-of-00002.safetensors",
267
+ "language_model.model.blocks.23.mlp.ff_out.biases": "model-00001-of-00002.safetensors",
268
+ "language_model.model.blocks.23.mlp.ff_out.scales": "model-00001-of-00002.safetensors",
269
+ "language_model.model.blocks.23.mlp.ff_out.weight": "model-00001-of-00002.safetensors",
270
  "language_model.model.blocks.23.mlp.ff_proj.biases": "model-00001-of-00002.safetensors",
271
  "language_model.model.blocks.23.mlp.ff_proj.scales": "model-00001-of-00002.safetensors",
272
  "language_model.model.blocks.23.mlp.ff_proj.weight": "model-00001-of-00002.safetensors",
 
278
  "language_model.model.blocks.23.self_attn.attn_out.weight": "model-00001-of-00002.safetensors",
279
  "language_model.model.blocks.23.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
280
  "language_model.model.blocks.23.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
281
+ "language_model.model.blocks.24.attn_norm.weight": "model-00001-of-00002.safetensors",
282
+ "language_model.model.blocks.24.ff_norm.weight": "model-00001-of-00002.safetensors",
283
+ "language_model.model.blocks.24.mlp.ff_out.biases": "model-00001-of-00002.safetensors",
284
+ "language_model.model.blocks.24.mlp.ff_out.scales": "model-00001-of-00002.safetensors",
285
+ "language_model.model.blocks.24.mlp.ff_out.weight": "model-00001-of-00002.safetensors",
286
+ "language_model.model.blocks.24.mlp.ff_proj.biases": "model-00001-of-00002.safetensors",
287
+ "language_model.model.blocks.24.mlp.ff_proj.scales": "model-00001-of-00002.safetensors",
288
+ "language_model.model.blocks.24.mlp.ff_proj.weight": "model-00001-of-00002.safetensors",
289
+ "language_model.model.blocks.24.self_attn.att_proj.biases": "model-00001-of-00002.safetensors",
290
+ "language_model.model.blocks.24.self_attn.att_proj.scales": "model-00001-of-00002.safetensors",
291
+ "language_model.model.blocks.24.self_attn.att_proj.weight": "model-00001-of-00002.safetensors",
292
+ "language_model.model.blocks.24.self_attn.attn_out.biases": "model-00001-of-00002.safetensors",
293
+ "language_model.model.blocks.24.self_attn.attn_out.scales": "model-00001-of-00002.safetensors",
294
+ "language_model.model.blocks.24.self_attn.attn_out.weight": "model-00001-of-00002.safetensors",
295
+ "language_model.model.blocks.24.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
296
+ "language_model.model.blocks.24.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
297
+ "language_model.model.blocks.25.attn_norm.weight": "model-00001-of-00002.safetensors",
298
+ "language_model.model.blocks.25.ff_norm.weight": "model-00001-of-00002.safetensors",
299
+ "language_model.model.blocks.25.mlp.ff_out.biases": "model-00001-of-00002.safetensors",
300
+ "language_model.model.blocks.25.mlp.ff_out.scales": "model-00001-of-00002.safetensors",
301
+ "language_model.model.blocks.25.mlp.ff_out.weight": "model-00001-of-00002.safetensors",
302
+ "language_model.model.blocks.25.mlp.ff_proj.biases": "model-00001-of-00002.safetensors",
303
+ "language_model.model.blocks.25.mlp.ff_proj.scales": "model-00001-of-00002.safetensors",
304
+ "language_model.model.blocks.25.mlp.ff_proj.weight": "model-00001-of-00002.safetensors",
305
+ "language_model.model.blocks.25.self_attn.att_proj.biases": "model-00001-of-00002.safetensors",
306
+ "language_model.model.blocks.25.self_attn.att_proj.scales": "model-00001-of-00002.safetensors",
307
+ "language_model.model.blocks.25.self_attn.att_proj.weight": "model-00001-of-00002.safetensors",
308
+ "language_model.model.blocks.25.self_attn.attn_out.biases": "model-00001-of-00002.safetensors",
309
+ "language_model.model.blocks.25.self_attn.attn_out.scales": "model-00001-of-00002.safetensors",
310
+ "language_model.model.blocks.25.self_attn.attn_out.weight": "model-00001-of-00002.safetensors",
311
+ "language_model.model.blocks.25.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
312
+ "language_model.model.blocks.25.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
313
+ "language_model.model.blocks.26.attn_norm.weight": "model-00001-of-00002.safetensors",
314
+ "language_model.model.blocks.26.ff_norm.weight": "model-00001-of-00002.safetensors",
315
+ "language_model.model.blocks.26.mlp.ff_out.biases": "model-00001-of-00002.safetensors",
316
+ "language_model.model.blocks.26.mlp.ff_out.scales": "model-00001-of-00002.safetensors",
317
+ "language_model.model.blocks.26.mlp.ff_out.weight": "model-00001-of-00002.safetensors",
318
+ "language_model.model.blocks.26.mlp.ff_proj.biases": "model-00001-of-00002.safetensors",
319
+ "language_model.model.blocks.26.mlp.ff_proj.scales": "model-00001-of-00002.safetensors",
320
+ "language_model.model.blocks.26.mlp.ff_proj.weight": "model-00001-of-00002.safetensors",
321
+ "language_model.model.blocks.26.self_attn.att_proj.biases": "model-00001-of-00002.safetensors",
322
+ "language_model.model.blocks.26.self_attn.att_proj.scales": "model-00001-of-00002.safetensors",
323
+ "language_model.model.blocks.26.self_attn.att_proj.weight": "model-00001-of-00002.safetensors",
324
+ "language_model.model.blocks.26.self_attn.attn_out.biases": "model-00001-of-00002.safetensors",
325
+ "language_model.model.blocks.26.self_attn.attn_out.scales": "model-00001-of-00002.safetensors",
326
+ "language_model.model.blocks.26.self_attn.attn_out.weight": "model-00001-of-00002.safetensors",
327
+ "language_model.model.blocks.26.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
328
+ "language_model.model.blocks.26.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
329
+ "language_model.model.blocks.27.attn_norm.weight": "model-00001-of-00002.safetensors",
330
+ "language_model.model.blocks.27.ff_norm.weight": "model-00001-of-00002.safetensors",
331
+ "language_model.model.blocks.27.mlp.ff_out.biases": "model-00001-of-00002.safetensors",
332
+ "language_model.model.blocks.27.mlp.ff_out.scales": "model-00001-of-00002.safetensors",
333
+ "language_model.model.blocks.27.mlp.ff_out.weight": "model-00001-of-00002.safetensors",
334
+ "language_model.model.blocks.27.mlp.ff_proj.biases": "model-00001-of-00002.safetensors",
335
+ "language_model.model.blocks.27.mlp.ff_proj.scales": "model-00001-of-00002.safetensors",
336
+ "language_model.model.blocks.27.mlp.ff_proj.weight": "model-00001-of-00002.safetensors",
337
+ "language_model.model.blocks.27.self_attn.att_proj.biases": "model-00001-of-00002.safetensors",
338
+ "language_model.model.blocks.27.self_attn.att_proj.scales": "model-00001-of-00002.safetensors",
339
+ "language_model.model.blocks.27.self_attn.att_proj.weight": "model-00001-of-00002.safetensors",
340
+ "language_model.model.blocks.27.self_attn.attn_out.biases": "model-00001-of-00002.safetensors",
341
+ "language_model.model.blocks.27.self_attn.attn_out.scales": "model-00001-of-00002.safetensors",
342
+ "language_model.model.blocks.27.self_attn.attn_out.weight": "model-00001-of-00002.safetensors",
343
+ "language_model.model.blocks.27.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
344
+ "language_model.model.blocks.27.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
345
+ "language_model.model.blocks.28.attn_norm.weight": "model-00001-of-00002.safetensors",
346
+ "language_model.model.blocks.28.ff_norm.weight": "model-00001-of-00002.safetensors",
347
+ "language_model.model.blocks.28.mlp.ff_out.biases": "model-00001-of-00002.safetensors",
348
+ "language_model.model.blocks.28.mlp.ff_out.scales": "model-00001-of-00002.safetensors",
349
+ "language_model.model.blocks.28.mlp.ff_out.weight": "model-00001-of-00002.safetensors",
350
+ "language_model.model.blocks.28.mlp.ff_proj.biases": "model-00001-of-00002.safetensors",
351
+ "language_model.model.blocks.28.mlp.ff_proj.scales": "model-00001-of-00002.safetensors",
352
+ "language_model.model.blocks.28.mlp.ff_proj.weight": "model-00001-of-00002.safetensors",
353
+ "language_model.model.blocks.28.self_attn.att_proj.biases": "model-00001-of-00002.safetensors",
354
+ "language_model.model.blocks.28.self_attn.att_proj.scales": "model-00001-of-00002.safetensors",
355
+ "language_model.model.blocks.28.self_attn.att_proj.weight": "model-00001-of-00002.safetensors",
356
+ "language_model.model.blocks.28.self_attn.attn_out.biases": "model-00001-of-00002.safetensors",
357
+ "language_model.model.blocks.28.self_attn.attn_out.scales": "model-00001-of-00002.safetensors",
358
+ "language_model.model.blocks.28.self_attn.attn_out.weight": "model-00001-of-00002.safetensors",
359
+ "language_model.model.blocks.28.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
360
+ "language_model.model.blocks.28.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
361
+ "language_model.model.blocks.29.attn_norm.weight": "model-00001-of-00002.safetensors",
362
+ "language_model.model.blocks.29.ff_norm.weight": "model-00001-of-00002.safetensors",
363
+ "language_model.model.blocks.29.mlp.ff_out.biases": "model-00001-of-00002.safetensors",
364
+ "language_model.model.blocks.29.mlp.ff_out.scales": "model-00001-of-00002.safetensors",
365
+ "language_model.model.blocks.29.mlp.ff_out.weight": "model-00001-of-00002.safetensors",
366
+ "language_model.model.blocks.29.mlp.ff_proj.biases": "model-00001-of-00002.safetensors",
367
+ "language_model.model.blocks.29.mlp.ff_proj.scales": "model-00001-of-00002.safetensors",
368
+ "language_model.model.blocks.29.mlp.ff_proj.weight": "model-00001-of-00002.safetensors",
369
+ "language_model.model.blocks.29.self_attn.att_proj.biases": "model-00001-of-00002.safetensors",
370
+ "language_model.model.blocks.29.self_attn.att_proj.scales": "model-00001-of-00002.safetensors",
371
+ "language_model.model.blocks.29.self_attn.att_proj.weight": "model-00001-of-00002.safetensors",
372
+ "language_model.model.blocks.29.self_attn.attn_out.biases": "model-00001-of-00002.safetensors",
373
+ "language_model.model.blocks.29.self_attn.attn_out.scales": "model-00001-of-00002.safetensors",
374
+ "language_model.model.blocks.29.self_attn.attn_out.weight": "model-00001-of-00002.safetensors",
375
+ "language_model.model.blocks.29.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
376
+ "language_model.model.blocks.29.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
377
  "language_model.model.blocks.3.attn_norm.weight": "model-00001-of-00002.safetensors",
378
  "language_model.model.blocks.3.ff_norm.weight": "model-00001-of-00002.safetensors",
379
  "language_model.model.blocks.3.mlp.ff_out.biases": "model-00001-of-00002.safetensors",
 
390
  "language_model.model.blocks.3.self_attn.attn_out.weight": "model-00001-of-00002.safetensors",
391
  "language_model.model.blocks.3.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
392
  "language_model.model.blocks.3.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
393
+ "language_model.model.blocks.30.attn_norm.weight": "model-00001-of-00002.safetensors",
394
+ "language_model.model.blocks.30.ff_norm.weight": "model-00001-of-00002.safetensors",
395
+ "language_model.model.blocks.30.mlp.ff_out.biases": "model-00001-of-00002.safetensors",
396
+ "language_model.model.blocks.30.mlp.ff_out.scales": "model-00001-of-00002.safetensors",
397
+ "language_model.model.blocks.30.mlp.ff_out.weight": "model-00001-of-00002.safetensors",
398
+ "language_model.model.blocks.30.mlp.ff_proj.biases": "model-00001-of-00002.safetensors",
399
+ "language_model.model.blocks.30.mlp.ff_proj.scales": "model-00001-of-00002.safetensors",
400
+ "language_model.model.blocks.30.mlp.ff_proj.weight": "model-00001-of-00002.safetensors",
401
+ "language_model.model.blocks.30.self_attn.att_proj.biases": "model-00001-of-00002.safetensors",
402
+ "language_model.model.blocks.30.self_attn.att_proj.scales": "model-00001-of-00002.safetensors",
403
+ "language_model.model.blocks.30.self_attn.att_proj.weight": "model-00001-of-00002.safetensors",
404
+ "language_model.model.blocks.30.self_attn.attn_out.biases": "model-00001-of-00002.safetensors",
405
+ "language_model.model.blocks.30.self_attn.attn_out.scales": "model-00001-of-00002.safetensors",
406
+ "language_model.model.blocks.30.self_attn.attn_out.weight": "model-00001-of-00002.safetensors",
407
+ "language_model.model.blocks.30.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
408
+ "language_model.model.blocks.30.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
409
+ "language_model.model.blocks.31.attn_norm.weight": "model-00001-of-00002.safetensors",
410
+ "language_model.model.blocks.31.ff_norm.weight": "model-00001-of-00002.safetensors",
411
+ "language_model.model.blocks.31.mlp.ff_out.biases": "model-00001-of-00002.safetensors",
412
+ "language_model.model.blocks.31.mlp.ff_out.scales": "model-00001-of-00002.safetensors",
413
+ "language_model.model.blocks.31.mlp.ff_out.weight": "model-00001-of-00002.safetensors",
414
+ "language_model.model.blocks.31.mlp.ff_proj.biases": "model-00001-of-00002.safetensors",
415
+ "language_model.model.blocks.31.mlp.ff_proj.scales": "model-00001-of-00002.safetensors",
416
+ "language_model.model.blocks.31.mlp.ff_proj.weight": "model-00001-of-00002.safetensors",
417
+ "language_model.model.blocks.31.self_attn.att_proj.biases": "model-00001-of-00002.safetensors",
418
+ "language_model.model.blocks.31.self_attn.att_proj.scales": "model-00001-of-00002.safetensors",
419
+ "language_model.model.blocks.31.self_attn.att_proj.weight": "model-00001-of-00002.safetensors",
420
+ "language_model.model.blocks.31.self_attn.attn_out.biases": "model-00001-of-00002.safetensors",
421
+ "language_model.model.blocks.31.self_attn.attn_out.scales": "model-00001-of-00002.safetensors",
422
+ "language_model.model.blocks.31.self_attn.attn_out.weight": "model-00001-of-00002.safetensors",
423
+ "language_model.model.blocks.31.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
424
+ "language_model.model.blocks.31.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
425
+ "language_model.model.blocks.32.attn_norm.weight": "model-00001-of-00002.safetensors",
426
+ "language_model.model.blocks.32.ff_norm.weight": "model-00001-of-00002.safetensors",
427
+ "language_model.model.blocks.32.mlp.ff_out.biases": "model-00001-of-00002.safetensors",
428
+ "language_model.model.blocks.32.mlp.ff_out.scales": "model-00001-of-00002.safetensors",
429
+ "language_model.model.blocks.32.mlp.ff_out.weight": "model-00001-of-00002.safetensors",
430
+ "language_model.model.blocks.32.mlp.ff_proj.biases": "model-00001-of-00002.safetensors",
431
+ "language_model.model.blocks.32.mlp.ff_proj.scales": "model-00001-of-00002.safetensors",
432
+ "language_model.model.blocks.32.mlp.ff_proj.weight": "model-00001-of-00002.safetensors",
433
+ "language_model.model.blocks.32.self_attn.att_proj.biases": "model-00001-of-00002.safetensors",
434
+ "language_model.model.blocks.32.self_attn.att_proj.scales": "model-00001-of-00002.safetensors",
435
+ "language_model.model.blocks.32.self_attn.att_proj.weight": "model-00001-of-00002.safetensors",
436
+ "language_model.model.blocks.32.self_attn.attn_out.biases": "model-00001-of-00002.safetensors",
437
+ "language_model.model.blocks.32.self_attn.attn_out.scales": "model-00001-of-00002.safetensors",
438
+ "language_model.model.blocks.32.self_attn.attn_out.weight": "model-00001-of-00002.safetensors",
439
+ "language_model.model.blocks.32.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
440
+ "language_model.model.blocks.32.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
441
+ "language_model.model.blocks.33.attn_norm.weight": "model-00001-of-00002.safetensors",
442
+ "language_model.model.blocks.33.ff_norm.weight": "model-00001-of-00002.safetensors",
443
+ "language_model.model.blocks.33.mlp.ff_out.biases": "model-00001-of-00002.safetensors",
444
+ "language_model.model.blocks.33.mlp.ff_out.scales": "model-00001-of-00002.safetensors",
445
+ "language_model.model.blocks.33.mlp.ff_out.weight": "model-00001-of-00002.safetensors",
446
+ "language_model.model.blocks.33.mlp.ff_proj.biases": "model-00001-of-00002.safetensors",
447
+ "language_model.model.blocks.33.mlp.ff_proj.scales": "model-00001-of-00002.safetensors",
448
+ "language_model.model.blocks.33.mlp.ff_proj.weight": "model-00001-of-00002.safetensors",
449
+ "language_model.model.blocks.33.self_attn.att_proj.biases": "model-00001-of-00002.safetensors",
450
+ "language_model.model.blocks.33.self_attn.att_proj.scales": "model-00001-of-00002.safetensors",
451
+ "language_model.model.blocks.33.self_attn.att_proj.weight": "model-00001-of-00002.safetensors",
452
+ "language_model.model.blocks.33.self_attn.attn_out.biases": "model-00001-of-00002.safetensors",
453
+ "language_model.model.blocks.33.self_attn.attn_out.scales": "model-00001-of-00002.safetensors",
454
+ "language_model.model.blocks.33.self_attn.attn_out.weight": "model-00001-of-00002.safetensors",
455
+ "language_model.model.blocks.33.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
456
+ "language_model.model.blocks.33.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
457
+ "language_model.model.blocks.34.attn_norm.weight": "model-00001-of-00002.safetensors",
458
+ "language_model.model.blocks.34.ff_norm.weight": "model-00001-of-00002.safetensors",
459
+ "language_model.model.blocks.34.mlp.ff_out.biases": "model-00001-of-00002.safetensors",
460
+ "language_model.model.blocks.34.mlp.ff_out.scales": "model-00001-of-00002.safetensors",
461
+ "language_model.model.blocks.34.mlp.ff_out.weight": "model-00001-of-00002.safetensors",
462
+ "language_model.model.blocks.34.mlp.ff_proj.biases": "model-00001-of-00002.safetensors",
463
+ "language_model.model.blocks.34.mlp.ff_proj.scales": "model-00001-of-00002.safetensors",
464
+ "language_model.model.blocks.34.mlp.ff_proj.weight": "model-00001-of-00002.safetensors",
465
+ "language_model.model.blocks.34.self_attn.att_proj.biases": "model-00001-of-00002.safetensors",
466
+ "language_model.model.blocks.34.self_attn.att_proj.scales": "model-00001-of-00002.safetensors",
467
+ "language_model.model.blocks.34.self_attn.att_proj.weight": "model-00001-of-00002.safetensors",
468
+ "language_model.model.blocks.34.self_attn.attn_out.biases": "model-00001-of-00002.safetensors",
469
+ "language_model.model.blocks.34.self_attn.attn_out.scales": "model-00001-of-00002.safetensors",
470
+ "language_model.model.blocks.34.self_attn.attn_out.weight": "model-00001-of-00002.safetensors",
471
+ "language_model.model.blocks.34.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
472
+ "language_model.model.blocks.34.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
473
+ "language_model.model.blocks.35.attn_norm.weight": "model-00001-of-00002.safetensors",
474
+ "language_model.model.blocks.35.ff_norm.weight": "model-00001-of-00002.safetensors",
475
+ "language_model.model.blocks.35.mlp.ff_out.biases": "model-00001-of-00002.safetensors",
476
+ "language_model.model.blocks.35.mlp.ff_out.scales": "model-00001-of-00002.safetensors",
477
+ "language_model.model.blocks.35.mlp.ff_out.weight": "model-00001-of-00002.safetensors",
478
+ "language_model.model.blocks.35.mlp.ff_proj.biases": "model-00001-of-00002.safetensors",
479
+ "language_model.model.blocks.35.mlp.ff_proj.scales": "model-00001-of-00002.safetensors",
480
+ "language_model.model.blocks.35.mlp.ff_proj.weight": "model-00001-of-00002.safetensors",
481
+ "language_model.model.blocks.35.self_attn.att_proj.biases": "model-00001-of-00002.safetensors",
482
+ "language_model.model.blocks.35.self_attn.att_proj.scales": "model-00001-of-00002.safetensors",
483
+ "language_model.model.blocks.35.self_attn.att_proj.weight": "model-00001-of-00002.safetensors",
484
+ "language_model.model.blocks.35.self_attn.attn_out.biases": "model-00001-of-00002.safetensors",
485
+ "language_model.model.blocks.35.self_attn.attn_out.scales": "model-00001-of-00002.safetensors",
486
+ "language_model.model.blocks.35.self_attn.attn_out.weight": "model-00001-of-00002.safetensors",
487
+ "language_model.model.blocks.35.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
488
+ "language_model.model.blocks.35.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
489
  "language_model.model.blocks.4.attn_norm.weight": "model-00001-of-00002.safetensors",
490
  "language_model.model.blocks.4.ff_norm.weight": "model-00001-of-00002.safetensors",
491
  "language_model.model.blocks.4.mlp.ff_out.biases": "model-00001-of-00002.safetensors",
 
582
  "language_model.model.blocks.9.self_attn.attn_out.weight": "model-00001-of-00002.safetensors",
583
  "language_model.model.blocks.9.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
584
  "language_model.model.blocks.9.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
585
+ "language_model.model.ln_f.weight": "model-00001-of-00002.safetensors",
586
  "language_model.model.wte.embedding": "model-00001-of-00002.safetensors",
587
  "language_model.model.wte.new_embedding": "model-00001-of-00002.safetensors",
588
  "vision_tower.image_pooling_2d.wk.bias": "model-00002-of-00002.safetensors",