{ "lm_head.weight": [ [5.5779090, 3.1322198, -404.38358, 62.589108, 0.99358273], [5.0814748, 2.4687927, -314.44955, 48.671352, 0.7727282], [3.6916721, 1.7657869, -224.55742, 34.757641, 0.55075526], [1.4084998, 1.0232025, -134.70717, 20.847967, 0.32766387], [-1.7680415, 0.24103954, -44.898785, 6.942337, 0.10345399], [-5.8379521, -0.58070201, 44.867714, -6.9592528, -0.12187435], [-10.801232, -1.4420221, 134.59233, -20.8568, -0.34832114], [-16.657881, -2.3429208, 224.27509, -34.750309, -0.57588643], [-23.4079, -3.2833982, 313.91595, -48.639774, -0.80457014], [-31.051287, -4.263454, 403.51492, -62.5252, -1.0343723] ], "embed_tokens.weight": [ [100, 0, 0, 0, 0], [100, 1, 0, 0, 0], [100, 2, 0, 0, 0], [100, 3, 0, 0, 0], [100, 4, 0, 0, 0], [100, 5, 0, 0, 0], [100, 6, 0, 0, 0], [100, 7, 0, 0, 0], [100, 8, 0, 0, 0], [100, 9, 0, 0, 0] ], "layers.0.input_layernorm.weight": [1, 1, 1, 1, 1], "layers.0.post_attention_layernorm.weight": [1, 1, 1, 1, 1], "layers.0.mlp.down_proj.weight": [ [0, 0, 0], [0, 0, 0], [0, 0, 0], [1, -1, 0], [0, 0, 0] ], "layers.0.mlp.gate_proj.weight": [ [-0.3353202, -1341.267, 60353.305, 0, 0], [-13.743691, -1341.8693, 60353.277, 0, 0], [0, 0, 0, 0, 0] ], "layers.0.mlp.up_proj.weight": [ [0.014898191, 0.00066922739, 2.9977213e-05, 0, 0], [0.014898191, 0.00066922739, 2.9977213e-05, 0, 0], [0, 0, 0, 0, 0] ], "layers.0.self_attn.q_norm.weight": [16, 16], "layers.0.self_attn.k_norm.weight": [16, 16], "layers.0.self_attn.q_proj.weight": [ [0.98502123, 0, 0, 0, 0], [0.17243294, 0, 0, 0, 0], [0.96630472, 0, 0, 0, 0], [-0.25740093, 0, 0, 0, 0] ], "layers.0.self_attn.k_proj.weight": [ [-0.31672141, 0, 0, 0, 0], [-0.94851863, 0, 0, 0, 0] ], "layers.0.self_attn.v_proj.weight": [ [0, 1, 0, 0, 0], [0, 0, 0, 0, 0] ], "layers.0.self_attn.o_proj.weight": [ [0, 0, 0, 0], [0, 0, 0, 0], [1, 0, 1, 0], [0, 0, 0, 0], [0, 0, 0, 0] ], "layers.1.input_layernorm.weight": [1, 1, 1, 1, 1], "layers.1.post_attention_layernorm.weight": [1, 1, 1, 1, 1], "layers.1.mlp.down_proj.weight": [ [0, 0, 0], [0, 0, 0], [1, -10, 10], [0, 0, 0], [0, 0, 0] ], "layers.1.mlp.gate_proj.weight": [ [-0.43951669, 5.6323919, 0.4983815, 1343.5575, 60357.68], [-121.12466, 0.32923722, -5.0313854, 1344.9166, 60357.438], [-134.53412, -0.2600022, -5.6458039, 1345.0677, 60357.41] ], "layers.1.mlp.up_proj.weight": [ [0.014899401, 0.00065471046, 0.00068268733, -0.00016779384, 2.9817384e-05], [0.014899401, 0.00065471046, 0.00068268733, -0.00016779384, 2.9817384e-05], [0.014899401, 0.00065471046, 0.00068268733, -0.00016779384, 2.9817384e-05] ], "layers.1.self_attn.q_norm.weight": [16, 16], "layers.1.self_attn.k_norm.weight": [16, 16], "layers.1.self_attn.q_proj.weight": [ [-0.25507239, 0, 0, 0, 0], [0.96692199, 0, 0, 0, 0], [0.17478994, 0, 0, 0, 0], [0.98460573, 0, 0, 0, 0] ], "layers.1.self_attn.k_proj.weight": [ [0.32702553, 0, 0, 0, 0], [-0.94501549, 0, 0, 0, 0] ], "layers.1.self_attn.v_proj.weight": [ [0, 1, 0, 0, 0], [0, 0, 0, 0, 0] ], "layers.1.self_attn.o_proj.weight": [ [0, 0, 0, 0], [0, 0, 0, 0], [0, 0, 0, 0], [0, 0, 0, 0], [1, 0, 1, 0] ], "norm.weight": [1, 1, 1, 1, 1] }