File size: 3,442 Bytes
f93cca0 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 | {
"lm_head.weight": [
[5.5779090, 3.1322198, -404.38358, 62.589108, 0.99358273],
[5.0814748, 2.4687927, -314.44955, 48.671352, 0.7727282],
[3.6916721, 1.7657869, -224.55742, 34.757641, 0.55075526],
[1.4084998, 1.0232025, -134.70717, 20.847967, 0.32766387],
[-1.7680415, 0.24103954, -44.898785, 6.942337, 0.10345399],
[-5.8379521, -0.58070201, 44.867714, -6.9592528, -0.12187435],
[-10.801232, -1.4420221, 134.59233, -20.8568, -0.34832114],
[-16.657881, -2.3429208, 224.27509, -34.750309, -0.57588643],
[-23.4079, -3.2833982, 313.91595, -48.639774, -0.80457014],
[-31.051287, -4.263454, 403.51492, -62.5252, -1.0343723]
],
"embed_tokens.weight": [
[100, 0, 0, 0, 0],
[100, 1, 0, 0, 0],
[100, 2, 0, 0, 0],
[100, 3, 0, 0, 0],
[100, 4, 0, 0, 0],
[100, 5, 0, 0, 0],
[100, 6, 0, 0, 0],
[100, 7, 0, 0, 0],
[100, 8, 0, 0, 0],
[100, 9, 0, 0, 0]
],
"layers.0.input_layernorm.weight": [1, 1, 1, 1, 1],
"layers.0.post_attention_layernorm.weight": [1, 1, 1, 1, 1],
"layers.0.mlp.down_proj.weight": [
[0, 0, 0],
[0, 0, 0],
[0, 0, 0],
[1, -1, 0],
[0, 0, 0]
],
"layers.0.mlp.gate_proj.weight": [
[-0.3353202, -1341.267, 60353.305, 0, 0],
[-13.743691, -1341.8693, 60353.277, 0, 0],
[0, 0, 0, 0, 0]
],
"layers.0.mlp.up_proj.weight": [
[0.014898191, 0.00066922739, 2.9977213e-05, 0, 0],
[0.014898191, 0.00066922739, 2.9977213e-05, 0, 0],
[0, 0, 0, 0, 0]
],
"layers.0.self_attn.q_norm.weight": [16, 16],
"layers.0.self_attn.k_norm.weight": [16, 16],
"layers.0.self_attn.q_proj.weight": [
[0.98502123, 0, 0, 0, 0],
[0.17243294, 0, 0, 0, 0],
[0.96630472, 0, 0, 0, 0],
[-0.25740093, 0, 0, 0, 0]
],
"layers.0.self_attn.k_proj.weight": [
[-0.31672141, 0, 0, 0, 0],
[-0.94851863, 0, 0, 0, 0]
],
"layers.0.self_attn.v_proj.weight": [
[0, 1, 0, 0, 0],
[0, 0, 0, 0, 0]
],
"layers.0.self_attn.o_proj.weight": [
[0, 0, 0, 0],
[0, 0, 0, 0],
[1, 0, 1, 0],
[0, 0, 0, 0],
[0, 0, 0, 0]
],
"layers.1.input_layernorm.weight": [1, 1, 1, 1, 1],
"layers.1.post_attention_layernorm.weight": [1, 1, 1, 1, 1],
"layers.1.mlp.down_proj.weight": [
[0, 0, 0],
[0, 0, 0],
[1, -10, 10],
[0, 0, 0],
[0, 0, 0]
],
"layers.1.mlp.gate_proj.weight": [
[-0.43951669, 5.6323919, 0.4983815, 1343.5575, 60357.68],
[-121.12466, 0.32923722, -5.0313854, 1344.9166, 60357.438],
[-134.53412, -0.2600022, -5.6458039, 1345.0677, 60357.41]
],
"layers.1.mlp.up_proj.weight": [
[0.014899401, 0.00065471046, 0.00068268733, -0.00016779384, 2.9817384e-05],
[0.014899401, 0.00065471046, 0.00068268733, -0.00016779384, 2.9817384e-05],
[0.014899401, 0.00065471046, 0.00068268733, -0.00016779384, 2.9817384e-05]
],
"layers.1.self_attn.q_norm.weight": [16, 16],
"layers.1.self_attn.k_norm.weight": [16, 16],
"layers.1.self_attn.q_proj.weight": [
[-0.25507239, 0, 0, 0, 0],
[0.96692199, 0, 0, 0, 0],
[0.17478994, 0, 0, 0, 0],
[0.98460573, 0, 0, 0, 0]
],
"layers.1.self_attn.k_proj.weight": [
[0.32702553, 0, 0, 0, 0],
[-0.94501549, 0, 0, 0, 0]
],
"layers.1.self_attn.v_proj.weight": [
[0, 1, 0, 0, 0],
[0, 0, 0, 0, 0]
],
"layers.1.self_attn.o_proj.weight": [
[0, 0, 0, 0],
[0, 0, 0, 0],
[0, 0, 0, 0],
[0, 0, 0, 0],
[1, 0, 1, 0]
],
"norm.weight": [1, 1, 1, 1, 1]
} |