Upload folder using huggingface_hub
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- added_tokens.json +1 -0
- all_results.json +7 -0
- checkpoint-336/added_tokens.json +1 -0
- checkpoint-336/config.json +223 -0
- checkpoint-336/generation_config.json +6 -0
- checkpoint-336/master_weights-00001-of-00001.safetensors +3 -0
- checkpoint-336/master_weights.safetensors.index.json +615 -0
- checkpoint-336/model-00001-of-00001.safetensors +3 -0
- checkpoint-336/model.safetensors.index.json +627 -0
- checkpoint-336/optimizer-00001-of-00001.safetensors +3 -0
- checkpoint-336/optimizer.safetensors.index.json +0 -0
- checkpoint-336/preprocessor_config.json +29 -0
- checkpoint-336/rng_state_0.pth +3 -0
- checkpoint-336/scheduler.pdparams +0 -0
- checkpoint-336/special_tokens_map.json +1 -0
- checkpoint-336/static_name_to_dyg_name.json +1 -0
- checkpoint-336/tokenizer.model +3 -0
- checkpoint-336/tokenizer_config.json +0 -0
- checkpoint-336/trainer_state.json +15 -0
- checkpoint-336/training_args.bin +3 -0
- checkpoint-378/added_tokens.json +1 -0
- checkpoint-378/config.json +223 -0
- checkpoint-378/generation_config.json +6 -0
- checkpoint-378/master_weights-00001-of-00001.safetensors +3 -0
- checkpoint-378/master_weights.safetensors.index.json +615 -0
- checkpoint-378/model-00001-of-00001.safetensors +3 -0
- checkpoint-378/model.safetensors.index.json +627 -0
- checkpoint-378/optimizer-00001-of-00001.safetensors +3 -0
- checkpoint-378/optimizer.safetensors.index.json +0 -0
- checkpoint-378/preprocessor_config.json +29 -0
- checkpoint-378/rng_state_0.pth +3 -0
- checkpoint-378/scheduler.pdparams +0 -0
- checkpoint-378/special_tokens_map.json +1 -0
- checkpoint-378/static_name_to_dyg_name.json +1 -0
- checkpoint-378/tokenizer.model +3 -0
- checkpoint-378/tokenizer_config.json +0 -0
- checkpoint-378/trainer_state.json +15 -0
- checkpoint-378/training_args.bin +3 -0
- checkpoint-420/added_tokens.json +1 -0
- checkpoint-420/config.json +223 -0
- checkpoint-420/generation_config.json +6 -0
- checkpoint-420/master_weights-00001-of-00001.safetensors +3 -0
- checkpoint-420/master_weights.safetensors.index.json +615 -0
- checkpoint-420/model-00001-of-00001.safetensors +3 -0
- checkpoint-420/model.safetensors.index.json +627 -0
- checkpoint-420/optimizer-00001-of-00001.safetensors +3 -0
- checkpoint-420/optimizer.safetensors.index.json +0 -0
- checkpoint-420/preprocessor_config.json +29 -0
- checkpoint-420/rng_state_0.pth +3 -0
- checkpoint-420/scheduler.pdparams +0 -0
added_tokens.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"<unk>": 0, "<s>": 1, "</s>": 2, "0": 3, "1": 4, "2": 5, "3": 6, "4": 7, "5": 8, "6": 9, "7": 10, "8": 11, "9": 12, "<|end_of_sentence|>": 100272, "<|begin_of_sentence|>": 100273, "<mask:1>": 100274, "<mask:4>": 100277, "<mask:5>": 100278, "<mask:6>": 100279, "<mask:7>": 100280, "<|IMAGE_PLACEHOLDER|>": 100295, "<|AUDIO_PLACEHOLDER|>": 100296, "<|LOC_0|>": 100297, "<|LOC_1|>": 100298, "<|LOC_2|>": 100299, "<|LOC_3|>": 100300, "<|LOC_4|>": 100301, "<|LOC_5|>": 100302, "<|LOC_6|>": 100303, "<|LOC_7|>": 100304, "<|LOC_8|>": 100305, "<|LOC_9|>": 100306, "<|LOC_10|>": 100307, "<|LOC_11|>": 100308, "<|LOC_12|>": 100309, "<|LOC_13|>": 100310, "<|LOC_14|>": 100311, "<|LOC_15|>": 100312, "<|LOC_16|>": 100313, "<|LOC_17|>": 100314, "<|LOC_18|>": 100315, "<|LOC_19|>": 100316, "<|LOC_20|>": 100317, "<|LOC_21|>": 100318, "<|LOC_22|>": 100319, "<|LOC_23|>": 100320, "<|LOC_24|>": 100321, "<|LOC_25|>": 100322, "<|LOC_26|>": 100323, "<|LOC_27|>": 100324, "<|LOC_28|>": 100325, "<|LOC_29|>": 100326, "<|LOC_30|>": 100327, "<|LOC_31|>": 100328, "<|LOC_32|>": 100329, "<|LOC_33|>": 100330, "<|LOC_34|>": 100331, "<|LOC_35|>": 100332, "<|LOC_36|>": 100333, "<|LOC_37|>": 100334, "<|LOC_38|>": 100335, "<|LOC_39|>": 100336, "<|LOC_40|>": 100337, "<|LOC_41|>": 100338, "<|LOC_42|>": 100339, "<|LOC_43|>": 100340, "<|LOC_44|>": 100341, "<|LOC_45|>": 100342, "<|LOC_46|>": 100343, "<|LOC_47|>": 100344, "<|LOC_48|>": 100345, "<|LOC_49|>": 100346, "<|LOC_50|>": 100347, "<|LOC_51|>": 100348, "<|LOC_52|>": 100349, "<|LOC_53|>": 100350, "<|LOC_54|>": 100351, "<|LOC_55|>": 100352, "<|LOC_56|>": 100353, "<|LOC_57|>": 100354, "<|LOC_58|>": 100355, "<|LOC_59|>": 100356, "<|LOC_60|>": 100357, "<|LOC_61|>": 100358, "<|LOC_62|>": 100359, "<|LOC_63|>": 100360, "<|LOC_64|>": 100361, "<|LOC_65|>": 100362, "<|LOC_66|>": 100363, "<|LOC_67|>": 100364, "<|LOC_68|>": 100365, "<|LOC_69|>": 100366, "<|LOC_70|>": 100367, "<|LOC_71|>": 100368, "<|LOC_72|>": 100369, "<|LOC_73|>": 100370, "<|LOC_74|>": 100371, "<|LOC_75|>": 100372, "<|LOC_76|>": 100373, "<|LOC_77|>": 100374, "<|LOC_78|>": 100375, "<|LOC_79|>": 100376, "<|LOC_80|>": 100377, "<|LOC_81|>": 100378, "<|LOC_82|>": 100379, "<|LOC_83|>": 100380, "<|LOC_84|>": 100381, "<|LOC_85|>": 100382, "<|LOC_86|>": 100383, "<|LOC_87|>": 100384, "<|LOC_88|>": 100385, "<|LOC_89|>": 100386, "<|LOC_90|>": 100387, "<|LOC_91|>": 100388, "<|LOC_92|>": 100389, "<|LOC_93|>": 100390, "<|LOC_94|>": 100391, "<|LOC_95|>": 100392, "<|LOC_96|>": 100393, "<|LOC_97|>": 100394, "<|LOC_98|>": 100395, "<|LOC_99|>": 100396, "<|LOC_100|>": 100397, "<|LOC_101|>": 100398, "<|LOC_102|>": 100399, "<|LOC_103|>": 100400, "<|LOC_104|>": 100401, "<|LOC_105|>": 100402, "<|LOC_106|>": 100403, "<|LOC_107|>": 100404, "<|LOC_108|>": 100405, "<|LOC_109|>": 100406, "<|LOC_110|>": 100407, "<|LOC_111|>": 100408, "<|LOC_112|>": 100409, "<|LOC_113|>": 100410, "<|LOC_114|>": 100411, "<|LOC_115|>": 100412, "<|LOC_116|>": 100413, "<|LOC_117|>": 100414, "<|LOC_118|>": 100415, "<|LOC_119|>": 100416, "<|LOC_120|>": 100417, "<|LOC_121|>": 100418, "<|LOC_122|>": 100419, "<|LOC_123|>": 100420, "<|LOC_124|>": 100421, "<|LOC_125|>": 100422, "<|LOC_126|>": 100423, "<|LOC_127|>": 100424, "<|LOC_128|>": 100425, "<|LOC_129|>": 100426, "<|LOC_130|>": 100427, "<|LOC_131|>": 100428, "<|LOC_132|>": 100429, "<|LOC_133|>": 100430, "<|LOC_134|>": 100431, "<|LOC_135|>": 100432, "<|LOC_136|>": 100433, "<|LOC_137|>": 100434, "<|LOC_138|>": 100435, "<|LOC_139|>": 100436, "<|LOC_140|>": 100437, "<|LOC_141|>": 100438, "<|LOC_142|>": 100439, "<|LOC_143|>": 100440, "<|LOC_144|>": 100441, "<|LOC_145|>": 100442, "<|LOC_146|>": 100443, "<|LOC_147|>": 100444, "<|LOC_148|>": 100445, "<|LOC_149|>": 100446, "<|LOC_150|>": 100447, "<|LOC_151|>": 100448, "<|LOC_152|>": 100449, "<|LOC_153|>": 100450, "<|LOC_154|>": 100451, "<|LOC_155|>": 100452, "<|LOC_156|>": 100453, "<|LOC_157|>": 100454, "<|LOC_158|>": 100455, "<|LOC_159|>": 100456, "<|LOC_160|>": 100457, "<|LOC_161|>": 100458, "<|LOC_162|>": 100459, "<|LOC_163|>": 100460, "<|LOC_164|>": 100461, "<|LOC_165|>": 100462, "<|LOC_166|>": 100463, "<|LOC_167|>": 100464, "<|LOC_168|>": 100465, "<|LOC_169|>": 100466, "<|LOC_170|>": 100467, "<|LOC_171|>": 100468, "<|LOC_172|>": 100469, "<|LOC_173|>": 100470, "<|LOC_174|>": 100471, "<|LOC_175|>": 100472, "<|LOC_176|>": 100473, "<|LOC_177|>": 100474, "<|LOC_178|>": 100475, "<|LOC_179|>": 100476, "<|LOC_180|>": 100477, "<|LOC_181|>": 100478, "<|LOC_182|>": 100479, "<|LOC_183|>": 100480, "<|LOC_184|>": 100481, "<|LOC_185|>": 100482, "<|LOC_186|>": 100483, "<|LOC_187|>": 100484, "<|LOC_188|>": 100485, "<|LOC_189|>": 100486, "<|LOC_190|>": 100487, "<|LOC_191|>": 100488, "<|LOC_192|>": 100489, "<|LOC_193|>": 100490, "<|LOC_194|>": 100491, "<|LOC_195|>": 100492, "<|LOC_196|>": 100493, "<|LOC_197|>": 100494, "<|LOC_198|>": 100495, "<|LOC_199|>": 100496, "<|LOC_200|>": 100497, "<|LOC_201|>": 100498, "<|LOC_202|>": 100499, "<|LOC_203|>": 100500, "<|LOC_204|>": 100501, "<|LOC_205|>": 100502, "<|LOC_206|>": 100503, "<|LOC_207|>": 100504, "<|LOC_208|>": 100505, "<|LOC_209|>": 100506, "<|LOC_210|>": 100507, "<|LOC_211|>": 100508, "<|LOC_212|>": 100509, "<|LOC_213|>": 100510, "<|LOC_214|>": 100511, "<|LOC_215|>": 100512, "<|LOC_216|>": 100513, "<|LOC_217|>": 100514, "<|LOC_218|>": 100515, "<|LOC_219|>": 100516, "<|LOC_220|>": 100517, "<|LOC_221|>": 100518, "<|LOC_222|>": 100519, "<|LOC_223|>": 100520, "<|LOC_224|>": 100521, "<|LOC_225|>": 100522, "<|LOC_226|>": 100523, "<|LOC_227|>": 100524, "<|LOC_228|>": 100525, "<|LOC_229|>": 100526, "<|LOC_230|>": 100527, "<|LOC_231|>": 100528, "<|LOC_232|>": 100529, "<|LOC_233|>": 100530, "<|LOC_234|>": 100531, "<|LOC_235|>": 100532, "<|LOC_236|>": 100533, "<|LOC_237|>": 100534, "<|LOC_238|>": 100535, "<|LOC_239|>": 100536, "<|LOC_240|>": 100537, "<|LOC_241|>": 100538, "<|LOC_242|>": 100539, "<|LOC_243|>": 100540, "<|LOC_244|>": 100541, "<|LOC_245|>": 100542, "<|LOC_246|>": 100543, "<|LOC_247|>": 100544, "<|LOC_248|>": 100545, "<|LOC_249|>": 100546, "<|LOC_250|>": 100547, "<|LOC_251|>": 100548, "<|LOC_252|>": 100549, "<|LOC_253|>": 100550, "<|LOC_254|>": 100551, "<|LOC_255|>": 100552, "<|LOC_256|>": 100553, "<|LOC_257|>": 100554, "<|LOC_258|>": 100555, "<|LOC_259|>": 100556, "<|LOC_260|>": 100557, "<|LOC_261|>": 100558, "<|LOC_262|>": 100559, "<|LOC_263|>": 100560, "<|LOC_264|>": 100561, "<|LOC_265|>": 100562, "<|LOC_266|>": 100563, "<|LOC_267|>": 100564, "<|LOC_268|>": 100565, "<|LOC_269|>": 100566, "<|LOC_270|>": 100567, "<|LOC_271|>": 100568, "<|LOC_272|>": 100569, "<|LOC_273|>": 100570, "<|LOC_274|>": 100571, "<|LOC_275|>": 100572, "<|LOC_276|>": 100573, "<|LOC_277|>": 100574, "<|LOC_278|>": 100575, "<|LOC_279|>": 100576, "<|LOC_280|>": 100577, "<|LOC_281|>": 100578, "<|LOC_282|>": 100579, "<|LOC_283|>": 100580, "<|LOC_284|>": 100581, "<|LOC_285|>": 100582, "<|LOC_286|>": 100583, "<|LOC_287|>": 100584, "<|LOC_288|>": 100585, "<|LOC_289|>": 100586, "<|LOC_290|>": 100587, "<|LOC_291|>": 100588, "<|LOC_292|>": 100589, "<|LOC_293|>": 100590, "<|LOC_294|>": 100591, "<|LOC_295|>": 100592, "<|LOC_296|>": 100593, "<|LOC_297|>": 100594, "<|LOC_298|>": 100595, "<|LOC_299|>": 100596, "<|LOC_300|>": 100597, "<|LOC_301|>": 100598, "<|LOC_302|>": 100599, "<|LOC_303|>": 100600, "<|LOC_304|>": 100601, "<|LOC_305|>": 100602, "<|LOC_306|>": 100603, "<|LOC_307|>": 100604, "<|LOC_308|>": 100605, "<|LOC_309|>": 100606, "<|LOC_310|>": 100607, "<|LOC_311|>": 100608, "<|LOC_312|>": 100609, "<|LOC_313|>": 100610, "<|LOC_314|>": 100611, "<|LOC_315|>": 100612, "<|LOC_316|>": 100613, "<|LOC_317|>": 100614, "<|LOC_318|>": 100615, "<|LOC_319|>": 100616, "<|LOC_320|>": 100617, "<|LOC_321|>": 100618, "<|LOC_322|>": 100619, "<|LOC_323|>": 100620, "<|LOC_324|>": 100621, "<|LOC_325|>": 100622, "<|LOC_326|>": 100623, "<|LOC_327|>": 100624, "<|LOC_328|>": 100625, "<|LOC_329|>": 100626, "<|LOC_330|>": 100627, "<|LOC_331|>": 100628, "<|LOC_332|>": 100629, "<|LOC_333|>": 100630, "<|LOC_334|>": 100631, "<|LOC_335|>": 100632, "<|LOC_336|>": 100633, "<|LOC_337|>": 100634, "<|LOC_338|>": 100635, "<|LOC_339|>": 100636, "<|LOC_340|>": 100637, "<|LOC_341|>": 100638, "<|LOC_342|>": 100639, "<|LOC_343|>": 100640, "<|LOC_344|>": 100641, "<|LOC_345|>": 100642, "<|LOC_346|>": 100643, "<|LOC_347|>": 100644, "<|LOC_348|>": 100645, "<|LOC_349|>": 100646, "<|LOC_350|>": 100647, "<|LOC_351|>": 100648, "<|LOC_352|>": 100649, "<|LOC_353|>": 100650, "<|LOC_354|>": 100651, "<|LOC_355|>": 100652, "<|LOC_356|>": 100653, "<|LOC_357|>": 100654, "<|LOC_358|>": 100655, "<|LOC_359|>": 100656, "<|LOC_360|>": 100657, "<|LOC_361|>": 100658, "<|LOC_362|>": 100659, "<|LOC_363|>": 100660, "<|LOC_364|>": 100661, "<|LOC_365|>": 100662, "<|LOC_366|>": 100663, "<|LOC_367|>": 100664, "<|LOC_368|>": 100665, "<|LOC_369|>": 100666, "<|LOC_370|>": 100667, "<|LOC_371|>": 100668, "<|LOC_372|>": 100669, "<|LOC_373|>": 100670, "<|LOC_374|>": 100671, "<|LOC_375|>": 100672, "<|LOC_376|>": 100673, "<|LOC_377|>": 100674, "<|LOC_378|>": 100675, "<|LOC_379|>": 100676, "<|LOC_380|>": 100677, "<|LOC_381|>": 100678, "<|LOC_382|>": 100679, "<|LOC_383|>": 100680, "<|LOC_384|>": 100681, "<|LOC_385|>": 100682, "<|LOC_386|>": 100683, "<|LOC_387|>": 100684, "<|LOC_388|>": 100685, "<|LOC_389|>": 100686, "<|LOC_390|>": 100687, "<|LOC_391|>": 100688, "<|LOC_392|>": 100689, "<|LOC_393|>": 100690, "<|LOC_394|>": 100691, "<|LOC_395|>": 100692, "<|LOC_396|>": 100693, "<|LOC_397|>": 100694, "<|LOC_398|>": 100695, "<|LOC_399|>": 100696, "<|LOC_400|>": 100697, "<|LOC_401|>": 100698, "<|LOC_402|>": 100699, "<|LOC_403|>": 100700, "<|LOC_404|>": 100701, "<|LOC_405|>": 100702, "<|LOC_406|>": 100703, "<|LOC_407|>": 100704, "<|LOC_408|>": 100705, "<|LOC_409|>": 100706, "<|LOC_410|>": 100707, "<|LOC_411|>": 100708, "<|LOC_412|>": 100709, "<|LOC_413|>": 100710, "<|LOC_414|>": 100711, "<|LOC_415|>": 100712, "<|LOC_416|>": 100713, "<|LOC_417|>": 100714, "<|LOC_418|>": 100715, "<|LOC_419|>": 100716, "<|LOC_420|>": 100717, "<|LOC_421|>": 100718, "<|LOC_422|>": 100719, "<|LOC_423|>": 100720, "<|LOC_424|>": 100721, "<|LOC_425|>": 100722, "<|LOC_426|>": 100723, "<|LOC_427|>": 100724, "<|LOC_428|>": 100725, "<|LOC_429|>": 100726, "<|LOC_430|>": 100727, "<|LOC_431|>": 100728, "<|LOC_432|>": 100729, "<|LOC_433|>": 100730, "<|LOC_434|>": 100731, "<|LOC_435|>": 100732, "<|LOC_436|>": 100733, "<|LOC_437|>": 100734, "<|LOC_438|>": 100735, "<|LOC_439|>": 100736, "<|LOC_440|>": 100737, "<|LOC_441|>": 100738, "<|LOC_442|>": 100739, "<|LOC_443|>": 100740, "<|LOC_444|>": 100741, "<|LOC_445|>": 100742, "<|LOC_446|>": 100743, "<|LOC_447|>": 100744, "<|LOC_448|>": 100745, "<|LOC_449|>": 100746, "<|LOC_450|>": 100747, "<|LOC_451|>": 100748, "<|LOC_452|>": 100749, "<|LOC_453|>": 100750, "<|LOC_454|>": 100751, "<|LOC_455|>": 100752, "<|LOC_456|>": 100753, "<|LOC_457|>": 100754, "<|LOC_458|>": 100755, "<|LOC_459|>": 100756, "<|LOC_460|>": 100757, "<|LOC_461|>": 100758, "<|LOC_462|>": 100759, "<|LOC_463|>": 100760, "<|LOC_464|>": 100761, "<|LOC_465|>": 100762, "<|LOC_466|>": 100763, "<|LOC_467|>": 100764, "<|LOC_468|>": 100765, "<|LOC_469|>": 100766, "<|LOC_470|>": 100767, "<|LOC_471|>": 100768, "<|LOC_472|>": 100769, "<|LOC_473|>": 100770, "<|LOC_474|>": 100771, "<|LOC_475|>": 100772, "<|LOC_476|>": 100773, "<|LOC_477|>": 100774, "<|LOC_478|>": 100775, "<|LOC_479|>": 100776, "<|LOC_480|>": 100777, "<|LOC_481|>": 100778, "<|LOC_482|>": 100779, "<|LOC_483|>": 100780, "<|LOC_484|>": 100781, "<|LOC_485|>": 100782, "<|LOC_486|>": 100783, "<|LOC_487|>": 100784, "<|LOC_488|>": 100785, "<|LOC_489|>": 100786, "<|LOC_490|>": 100787, "<|LOC_491|>": 100788, "<|LOC_492|>": 100789, "<|LOC_493|>": 100790, "<|LOC_494|>": 100791, "<|LOC_495|>": 100792, "<|LOC_496|>": 100793, "<|LOC_497|>": 100794, "<|LOC_498|>": 100795, "<|LOC_499|>": 100796, "<|LOC_500|>": 100797, "<|LOC_501|>": 100798, "<|LOC_502|>": 100799, "<|LOC_503|>": 100800, "<|LOC_504|>": 100801, "<|LOC_505|>": 100802, "<|LOC_506|>": 100803, "<|LOC_507|>": 100804, "<|LOC_508|>": 100805, "<|LOC_509|>": 100806, "<|LOC_510|>": 100807, "<|LOC_511|>": 100808, "<|LOC_512|>": 100809, "<|LOC_513|>": 100810, "<|LOC_514|>": 100811, "<|LOC_515|>": 100812, "<|LOC_516|>": 100813, "<|LOC_517|>": 100814, "<|LOC_518|>": 100815, "<|LOC_519|>": 100816, "<|LOC_520|>": 100817, "<|LOC_521|>": 100818, "<|LOC_522|>": 100819, "<|LOC_523|>": 100820, "<|LOC_524|>": 100821, "<|LOC_525|>": 100822, "<|LOC_526|>": 100823, "<|LOC_527|>": 100824, "<|LOC_528|>": 100825, "<|LOC_529|>": 100826, "<|LOC_530|>": 100827, "<|LOC_531|>": 100828, "<|LOC_532|>": 100829, "<|LOC_533|>": 100830, "<|LOC_534|>": 100831, "<|LOC_535|>": 100832, "<|LOC_536|>": 100833, "<|LOC_537|>": 100834, "<|LOC_538|>": 100835, "<|LOC_539|>": 100836, "<|LOC_540|>": 100837, "<|LOC_541|>": 100838, "<|LOC_542|>": 100839, "<|LOC_543|>": 100840, "<|LOC_544|>": 100841, "<|LOC_545|>": 100842, "<|LOC_546|>": 100843, "<|LOC_547|>": 100844, "<|LOC_548|>": 100845, "<|LOC_549|>": 100846, "<|LOC_550|>": 100847, "<|LOC_551|>": 100848, "<|LOC_552|>": 100849, "<|LOC_553|>": 100850, "<|LOC_554|>": 100851, "<|LOC_555|>": 100852, "<|LOC_556|>": 100853, "<|LOC_557|>": 100854, "<|LOC_558|>": 100855, "<|LOC_559|>": 100856, "<|LOC_560|>": 100857, "<|LOC_561|>": 100858, "<|LOC_562|>": 100859, "<|LOC_563|>": 100860, "<|LOC_564|>": 100861, "<|LOC_565|>": 100862, "<|LOC_566|>": 100863, "<|LOC_567|>": 100864, "<|LOC_568|>": 100865, "<|LOC_569|>": 100866, "<|LOC_570|>": 100867, "<|LOC_571|>": 100868, "<|LOC_572|>": 100869, "<|LOC_573|>": 100870, "<|LOC_574|>": 100871, "<|LOC_575|>": 100872, "<|LOC_576|>": 100873, "<|LOC_577|>": 100874, "<|LOC_578|>": 100875, "<|LOC_579|>": 100876, "<|LOC_580|>": 100877, "<|LOC_581|>": 100878, "<|LOC_582|>": 100879, "<|LOC_583|>": 100880, "<|LOC_584|>": 100881, "<|LOC_585|>": 100882, "<|LOC_586|>": 100883, "<|LOC_587|>": 100884, "<|LOC_588|>": 100885, "<|LOC_589|>": 100886, "<|LOC_590|>": 100887, "<|LOC_591|>": 100888, "<|LOC_592|>": 100889, "<|LOC_593|>": 100890, "<|LOC_594|>": 100891, "<|LOC_595|>": 100892, "<|LOC_596|>": 100893, "<|LOC_597|>": 100894, "<|LOC_598|>": 100895, "<|LOC_599|>": 100896, "<|LOC_600|>": 100897, "<|LOC_601|>": 100898, "<|LOC_602|>": 100899, "<|LOC_603|>": 100900, "<|LOC_604|>": 100901, "<|LOC_605|>": 100902, "<|LOC_606|>": 100903, "<|LOC_607|>": 100904, "<|LOC_608|>": 100905, "<|LOC_609|>": 100906, "<|LOC_610|>": 100907, "<|LOC_611|>": 100908, "<|LOC_612|>": 100909, "<|LOC_613|>": 100910, "<|LOC_614|>": 100911, "<|LOC_615|>": 100912, "<|LOC_616|>": 100913, "<|LOC_617|>": 100914, "<|LOC_618|>": 100915, "<|LOC_619|>": 100916, "<|LOC_620|>": 100917, "<|LOC_621|>": 100918, "<|LOC_622|>": 100919, "<|LOC_623|>": 100920, "<|LOC_624|>": 100921, "<|LOC_625|>": 100922, "<|LOC_626|>": 100923, "<|LOC_627|>": 100924, "<|LOC_628|>": 100925, "<|LOC_629|>": 100926, "<|LOC_630|>": 100927, "<|LOC_631|>": 100928, "<|LOC_632|>": 100929, "<|LOC_633|>": 100930, "<|LOC_634|>": 100931, "<|LOC_635|>": 100932, "<|LOC_636|>": 100933, "<|LOC_637|>": 100934, "<|LOC_638|>": 100935, "<|LOC_639|>": 100936, "<|LOC_640|>": 100937, "<|LOC_641|>": 100938, "<|LOC_642|>": 100939, "<|LOC_643|>": 100940, "<|LOC_644|>": 100941, "<|LOC_645|>": 100942, "<|LOC_646|>": 100943, "<|LOC_647|>": 100944, "<|LOC_648|>": 100945, "<|LOC_649|>": 100946, "<|LOC_650|>": 100947, "<|LOC_651|>": 100948, "<|LOC_652|>": 100949, "<|LOC_653|>": 100950, "<|LOC_654|>": 100951, "<|LOC_655|>": 100952, "<|LOC_656|>": 100953, "<|LOC_657|>": 100954, "<|LOC_658|>": 100955, "<|LOC_659|>": 100956, "<|LOC_660|>": 100957, "<|LOC_661|>": 100958, "<|LOC_662|>": 100959, "<|LOC_663|>": 100960, "<|LOC_664|>": 100961, "<|LOC_665|>": 100962, "<|LOC_666|>": 100963, "<|LOC_667|>": 100964, "<|LOC_668|>": 100965, "<|LOC_669|>": 100966, "<|LOC_670|>": 100967, "<|LOC_671|>": 100968, "<|LOC_672|>": 100969, "<|LOC_673|>": 100970, "<|LOC_674|>": 100971, "<|LOC_675|>": 100972, "<|LOC_676|>": 100973, "<|LOC_677|>": 100974, "<|LOC_678|>": 100975, "<|LOC_679|>": 100976, "<|LOC_680|>": 100977, "<|LOC_681|>": 100978, "<|LOC_682|>": 100979, "<|LOC_683|>": 100980, "<|LOC_684|>": 100981, "<|LOC_685|>": 100982, "<|LOC_686|>": 100983, "<|LOC_687|>": 100984, "<|LOC_688|>": 100985, "<|LOC_689|>": 100986, "<|LOC_690|>": 100987, "<|LOC_691|>": 100988, "<|LOC_692|>": 100989, "<|LOC_693|>": 100990, "<|LOC_694|>": 100991, "<|LOC_695|>": 100992, "<|LOC_696|>": 100993, "<|LOC_697|>": 100994, "<|LOC_698|>": 100995, "<|LOC_699|>": 100996, "<|LOC_700|>": 100997, "<|LOC_701|>": 100998, "<|LOC_702|>": 100999, "<|LOC_703|>": 101000, "<|LOC_704|>": 101001, "<|LOC_705|>": 101002, "<|LOC_706|>": 101003, "<|LOC_707|>": 101004, "<|LOC_708|>": 101005, "<|LOC_709|>": 101006, "<|LOC_710|>": 101007, "<|LOC_711|>": 101008, "<|LOC_712|>": 101009, "<|LOC_713|>": 101010, "<|LOC_714|>": 101011, "<|LOC_715|>": 101012, "<|LOC_716|>": 101013, "<|LOC_717|>": 101014, "<|LOC_718|>": 101015, "<|LOC_719|>": 101016, "<|LOC_720|>": 101017, "<|LOC_721|>": 101018, "<|LOC_722|>": 101019, "<|LOC_723|>": 101020, "<|LOC_724|>": 101021, "<|LOC_725|>": 101022, "<|LOC_726|>": 101023, "<|LOC_727|>": 101024, "<|LOC_728|>": 101025, "<|LOC_729|>": 101026, "<|LOC_730|>": 101027, "<|LOC_731|>": 101028, "<|LOC_732|>": 101029, "<|LOC_733|>": 101030, "<|LOC_734|>": 101031, "<|LOC_735|>": 101032, "<|LOC_736|>": 101033, "<|LOC_737|>": 101034, "<|LOC_738|>": 101035, "<|LOC_739|>": 101036, "<|LOC_740|>": 101037, "<|LOC_741|>": 101038, "<|LOC_742|>": 101039, "<|LOC_743|>": 101040, "<|LOC_744|>": 101041, "<|LOC_745|>": 101042, "<|LOC_746|>": 101043, "<|LOC_747|>": 101044, "<|LOC_748|>": 101045, "<|LOC_749|>": 101046, "<|LOC_750|>": 101047, "<|LOC_751|>": 101048, "<|LOC_752|>": 101049, "<|LOC_753|>": 101050, "<|LOC_754|>": 101051, "<|LOC_755|>": 101052, "<|LOC_756|>": 101053, "<|LOC_757|>": 101054, "<|LOC_758|>": 101055, "<|LOC_759|>": 101056, "<|LOC_760|>": 101057, "<|LOC_761|>": 101058, "<|LOC_762|>": 101059, "<|LOC_763|>": 101060, "<|LOC_764|>": 101061, "<|LOC_765|>": 101062, "<|LOC_766|>": 101063, "<|LOC_767|>": 101064, "<|LOC_768|>": 101065, "<|LOC_769|>": 101066, "<|LOC_770|>": 101067, "<|LOC_771|>": 101068, "<|LOC_772|>": 101069, "<|LOC_773|>": 101070, "<|LOC_774|>": 101071, "<|LOC_775|>": 101072, "<|LOC_776|>": 101073, "<|LOC_777|>": 101074, "<|LOC_778|>": 101075, "<|LOC_779|>": 101076, "<|LOC_780|>": 101077, "<|LOC_781|>": 101078, "<|LOC_782|>": 101079, "<|LOC_783|>": 101080, "<|LOC_784|>": 101081, "<|LOC_785|>": 101082, "<|LOC_786|>": 101083, "<|LOC_787|>": 101084, "<|LOC_788|>": 101085, "<|LOC_789|>": 101086, "<|LOC_790|>": 101087, "<|LOC_791|>": 101088, "<|LOC_792|>": 101089, "<|LOC_793|>": 101090, "<|LOC_794|>": 101091, "<|LOC_795|>": 101092, "<|LOC_796|>": 101093, "<|LOC_797|>": 101094, "<|LOC_798|>": 101095, "<|LOC_799|>": 101096, "<|LOC_800|>": 101097, "<|LOC_801|>": 101098, "<|LOC_802|>": 101099, "<|LOC_803|>": 101100, "<|LOC_804|>": 101101, "<|LOC_805|>": 101102, "<|LOC_806|>": 101103, "<|LOC_807|>": 101104, "<|LOC_808|>": 101105, "<|LOC_809|>": 101106, "<|LOC_810|>": 101107, "<|LOC_811|>": 101108, "<|LOC_812|>": 101109, "<|LOC_813|>": 101110, "<|LOC_814|>": 101111, "<|LOC_815|>": 101112, "<|LOC_816|>": 101113, "<|LOC_817|>": 101114, "<|LOC_818|>": 101115, "<|LOC_819|>": 101116, "<|LOC_820|>": 101117, "<|LOC_821|>": 101118, "<|LOC_822|>": 101119, "<|LOC_823|>": 101120, "<|LOC_824|>": 101121, "<|LOC_825|>": 101122, "<|LOC_826|>": 101123, "<|LOC_827|>": 101124, "<|LOC_828|>": 101125, "<|LOC_829|>": 101126, "<|LOC_830|>": 101127, "<|LOC_831|>": 101128, "<|LOC_832|>": 101129, "<|LOC_833|>": 101130, "<|LOC_834|>": 101131, "<|LOC_835|>": 101132, "<|LOC_836|>": 101133, "<|LOC_837|>": 101134, "<|LOC_838|>": 101135, "<|LOC_839|>": 101136, "<|LOC_840|>": 101137, "<|LOC_841|>": 101138, "<|LOC_842|>": 101139, "<|LOC_843|>": 101140, "<|LOC_844|>": 101141, "<|LOC_845|>": 101142, "<|LOC_846|>": 101143, "<|LOC_847|>": 101144, "<|LOC_848|>": 101145, "<|LOC_849|>": 101146, "<|LOC_850|>": 101147, "<|LOC_851|>": 101148, "<|LOC_852|>": 101149, "<|LOC_853|>": 101150, "<|LOC_854|>": 101151, "<|LOC_855|>": 101152, "<|LOC_856|>": 101153, "<|LOC_857|>": 101154, "<|LOC_858|>": 101155, "<|LOC_859|>": 101156, "<|LOC_860|>": 101157, "<|LOC_861|>": 101158, "<|LOC_862|>": 101159, "<|LOC_863|>": 101160, "<|LOC_864|>": 101161, "<|LOC_865|>": 101162, "<|LOC_866|>": 101163, "<|LOC_867|>": 101164, "<|LOC_868|>": 101165, "<|LOC_869|>": 101166, "<|LOC_870|>": 101167, "<|LOC_871|>": 101168, "<|LOC_872|>": 101169, "<|LOC_873|>": 101170, "<|LOC_874|>": 101171, "<|LOC_875|>": 101172, "<|LOC_876|>": 101173, "<|LOC_877|>": 101174, "<|LOC_878|>": 101175, "<|LOC_879|>": 101176, "<|LOC_880|>": 101177, "<|LOC_881|>": 101178, "<|LOC_882|>": 101179, "<|LOC_883|>": 101180, "<|LOC_884|>": 101181, "<|LOC_885|>": 101182, "<|LOC_886|>": 101183, "<|LOC_887|>": 101184, "<|LOC_888|>": 101185, "<|LOC_889|>": 101186, "<|LOC_890|>": 101187, "<|LOC_891|>": 101188, "<|LOC_892|>": 101189, "<|LOC_893|>": 101190, "<|LOC_894|>": 101191, "<|LOC_895|>": 101192, "<|LOC_896|>": 101193, "<|LOC_897|>": 101194, "<|LOC_898|>": 101195, "<|LOC_899|>": 101196, "<|LOC_900|>": 101197, "<|LOC_901|>": 101198, "<|LOC_902|>": 101199, "<|LOC_903|>": 101200, "<|LOC_904|>": 101201, "<|LOC_905|>": 101202, "<|LOC_906|>": 101203, "<|LOC_907|>": 101204, "<|LOC_908|>": 101205, "<|LOC_909|>": 101206, "<|LOC_910|>": 101207, "<|LOC_911|>": 101208, "<|LOC_912|>": 101209, "<|LOC_913|>": 101210, "<|LOC_914|>": 101211, "<|LOC_915|>": 101212, "<|LOC_916|>": 101213, "<|LOC_917|>": 101214, "<|LOC_918|>": 101215, "<|LOC_919|>": 101216, "<|LOC_920|>": 101217, "<|LOC_921|>": 101218, "<|LOC_922|>": 101219, "<|LOC_923|>": 101220, "<|LOC_924|>": 101221, "<|LOC_925|>": 101222, "<|LOC_926|>": 101223, "<|LOC_927|>": 101224, "<|LOC_928|>": 101225, "<|LOC_929|>": 101226, "<|LOC_930|>": 101227, "<|LOC_931|>": 101228, "<|LOC_932|>": 101229, "<|LOC_933|>": 101230, "<|LOC_934|>": 101231, "<|LOC_935|>": 101232, "<|LOC_936|>": 101233, "<|LOC_937|>": 101234, "<|LOC_938|>": 101235, "<|LOC_939|>": 101236, "<|LOC_940|>": 101237, "<|LOC_941|>": 101238, "<|LOC_942|>": 101239, "<|LOC_943|>": 101240, "<|LOC_944|>": 101241, "<|LOC_945|>": 101242, "<|LOC_946|>": 101243, "<|LOC_947|>": 101244, "<|LOC_948|>": 101245, "<|LOC_949|>": 101246, "<|LOC_950|>": 101247, "<|LOC_951|>": 101248, "<|LOC_952|>": 101249, "<|LOC_953|>": 101250, "<|LOC_954|>": 101251, "<|LOC_955|>": 101252, "<|LOC_956|>": 101253, "<|LOC_957|>": 101254, "<|LOC_958|>": 101255, "<|LOC_959|>": 101256, "<|LOC_960|>": 101257, "<|LOC_961|>": 101258, "<|LOC_962|>": 101259, "<|LOC_963|>": 101260, "<|LOC_964|>": 101261, "<|LOC_965|>": 101262, "<|LOC_966|>": 101263, "<|LOC_967|>": 101264, "<|LOC_968|>": 101265, "<|LOC_969|>": 101266, "<|LOC_970|>": 101267, "<|LOC_971|>": 101268, "<|LOC_972|>": 101269, "<|LOC_973|>": 101270, "<|LOC_974|>": 101271, "<|LOC_975|>": 101272, "<|LOC_976|>": 101273, "<|LOC_977|>": 101274, "<|LOC_978|>": 101275, "<|LOC_979|>": 101276, "<|LOC_980|>": 101277, "<|LOC_981|>": 101278, "<|LOC_982|>": 101279, "<|LOC_983|>": 101280, "<|LOC_984|>": 101281, "<|LOC_985|>": 101282, "<|LOC_986|>": 101283, "<|LOC_987|>": 101284, "<|LOC_988|>": 101285, "<|LOC_989|>": 101286, "<|LOC_990|>": 101287, "<|LOC_991|>": 101288, "<|LOC_992|>": 101289, "<|LOC_993|>": 101290, "<|LOC_994|>": 101291, "<|LOC_995|>": 101292, "<|LOC_996|>": 101293, "<|LOC_997|>": 101294, "<|LOC_998|>": 101295, "<|LOC_999|>": 101296, "<|LOC_1000|>": 101297, "<|LOC_BEGIN|>": 101298, "<|LOC_END|>": 101299, "<|LOC_SEP|>": 101300, "<|CROP_COL_SEP|>": 101301, "<|CROP_ROW_SEP|>": 101302, "<|IMAGE_SEP|>": 101303, "<|image_pad|>": 101304, "<|IMAGE_START|>": 101305, "<|IMAGE_END|>": 101306, "<|video_pad|>": 101307, "<ecel>": 101308, "<fcel>": 101309, "<xcel>": 101310, "<lcel>": 101311, "<ucel>": 101312, "<nl>": 101313}
|
all_results.json
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"progress_or_epoch": 1.0,
|
| 3 |
+
"train_loss": 0.6455322598193434,
|
| 4 |
+
"train_runtime": 2255.4152,
|
| 5 |
+
"train_samples_per_second": 1.4933,
|
| 6 |
+
"train_steps_per_second": 0.1867
|
| 7 |
+
}
|
checkpoint-336/added_tokens.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"<unk>": 0, "<s>": 1, "</s>": 2, "0": 3, "1": 4, "2": 5, "3": 6, "4": 7, "5": 8, "6": 9, "7": 10, "8": 11, "9": 12, "<|end_of_sentence|>": 100272, "<|begin_of_sentence|>": 100273, "<mask:1>": 100274, "<mask:4>": 100277, "<mask:5>": 100278, "<mask:6>": 100279, "<mask:7>": 100280, "<|IMAGE_PLACEHOLDER|>": 100295, "<|AUDIO_PLACEHOLDER|>": 100296, "<|LOC_0|>": 100297, "<|LOC_1|>": 100298, "<|LOC_2|>": 100299, "<|LOC_3|>": 100300, "<|LOC_4|>": 100301, "<|LOC_5|>": 100302, "<|LOC_6|>": 100303, "<|LOC_7|>": 100304, "<|LOC_8|>": 100305, "<|LOC_9|>": 100306, "<|LOC_10|>": 100307, "<|LOC_11|>": 100308, "<|LOC_12|>": 100309, "<|LOC_13|>": 100310, "<|LOC_14|>": 100311, "<|LOC_15|>": 100312, "<|LOC_16|>": 100313, "<|LOC_17|>": 100314, "<|LOC_18|>": 100315, "<|LOC_19|>": 100316, "<|LOC_20|>": 100317, "<|LOC_21|>": 100318, "<|LOC_22|>": 100319, "<|LOC_23|>": 100320, "<|LOC_24|>": 100321, "<|LOC_25|>": 100322, "<|LOC_26|>": 100323, "<|LOC_27|>": 100324, "<|LOC_28|>": 100325, "<|LOC_29|>": 100326, "<|LOC_30|>": 100327, "<|LOC_31|>": 100328, "<|LOC_32|>": 100329, "<|LOC_33|>": 100330, "<|LOC_34|>": 100331, "<|LOC_35|>": 100332, "<|LOC_36|>": 100333, "<|LOC_37|>": 100334, "<|LOC_38|>": 100335, "<|LOC_39|>": 100336, "<|LOC_40|>": 100337, "<|LOC_41|>": 100338, "<|LOC_42|>": 100339, "<|LOC_43|>": 100340, "<|LOC_44|>": 100341, "<|LOC_45|>": 100342, "<|LOC_46|>": 100343, "<|LOC_47|>": 100344, "<|LOC_48|>": 100345, "<|LOC_49|>": 100346, "<|LOC_50|>": 100347, "<|LOC_51|>": 100348, "<|LOC_52|>": 100349, "<|LOC_53|>": 100350, "<|LOC_54|>": 100351, "<|LOC_55|>": 100352, "<|LOC_56|>": 100353, "<|LOC_57|>": 100354, "<|LOC_58|>": 100355, "<|LOC_59|>": 100356, "<|LOC_60|>": 100357, "<|LOC_61|>": 100358, "<|LOC_62|>": 100359, "<|LOC_63|>": 100360, "<|LOC_64|>": 100361, "<|LOC_65|>": 100362, "<|LOC_66|>": 100363, "<|LOC_67|>": 100364, "<|LOC_68|>": 100365, "<|LOC_69|>": 100366, "<|LOC_70|>": 100367, "<|LOC_71|>": 100368, "<|LOC_72|>": 100369, "<|LOC_73|>": 100370, "<|LOC_74|>": 100371, "<|LOC_75|>": 100372, "<|LOC_76|>": 100373, "<|LOC_77|>": 100374, "<|LOC_78|>": 100375, "<|LOC_79|>": 100376, "<|LOC_80|>": 100377, "<|LOC_81|>": 100378, "<|LOC_82|>": 100379, "<|LOC_83|>": 100380, "<|LOC_84|>": 100381, "<|LOC_85|>": 100382, "<|LOC_86|>": 100383, "<|LOC_87|>": 100384, "<|LOC_88|>": 100385, "<|LOC_89|>": 100386, "<|LOC_90|>": 100387, "<|LOC_91|>": 100388, "<|LOC_92|>": 100389, "<|LOC_93|>": 100390, "<|LOC_94|>": 100391, "<|LOC_95|>": 100392, "<|LOC_96|>": 100393, "<|LOC_97|>": 100394, "<|LOC_98|>": 100395, "<|LOC_99|>": 100396, "<|LOC_100|>": 100397, "<|LOC_101|>": 100398, "<|LOC_102|>": 100399, "<|LOC_103|>": 100400, "<|LOC_104|>": 100401, "<|LOC_105|>": 100402, "<|LOC_106|>": 100403, "<|LOC_107|>": 100404, "<|LOC_108|>": 100405, "<|LOC_109|>": 100406, "<|LOC_110|>": 100407, "<|LOC_111|>": 100408, "<|LOC_112|>": 100409, "<|LOC_113|>": 100410, "<|LOC_114|>": 100411, "<|LOC_115|>": 100412, "<|LOC_116|>": 100413, "<|LOC_117|>": 100414, "<|LOC_118|>": 100415, "<|LOC_119|>": 100416, "<|LOC_120|>": 100417, "<|LOC_121|>": 100418, "<|LOC_122|>": 100419, "<|LOC_123|>": 100420, "<|LOC_124|>": 100421, "<|LOC_125|>": 100422, "<|LOC_126|>": 100423, "<|LOC_127|>": 100424, "<|LOC_128|>": 100425, "<|LOC_129|>": 100426, "<|LOC_130|>": 100427, "<|LOC_131|>": 100428, "<|LOC_132|>": 100429, "<|LOC_133|>": 100430, "<|LOC_134|>": 100431, "<|LOC_135|>": 100432, "<|LOC_136|>": 100433, "<|LOC_137|>": 100434, "<|LOC_138|>": 100435, "<|LOC_139|>": 100436, "<|LOC_140|>": 100437, "<|LOC_141|>": 100438, "<|LOC_142|>": 100439, "<|LOC_143|>": 100440, "<|LOC_144|>": 100441, "<|LOC_145|>": 100442, "<|LOC_146|>": 100443, "<|LOC_147|>": 100444, "<|LOC_148|>": 100445, "<|LOC_149|>": 100446, "<|LOC_150|>": 100447, "<|LOC_151|>": 100448, "<|LOC_152|>": 100449, "<|LOC_153|>": 100450, "<|LOC_154|>": 100451, "<|LOC_155|>": 100452, "<|LOC_156|>": 100453, "<|LOC_157|>": 100454, "<|LOC_158|>": 100455, "<|LOC_159|>": 100456, "<|LOC_160|>": 100457, "<|LOC_161|>": 100458, "<|LOC_162|>": 100459, "<|LOC_163|>": 100460, "<|LOC_164|>": 100461, "<|LOC_165|>": 100462, "<|LOC_166|>": 100463, "<|LOC_167|>": 100464, "<|LOC_168|>": 100465, "<|LOC_169|>": 100466, "<|LOC_170|>": 100467, "<|LOC_171|>": 100468, "<|LOC_172|>": 100469, "<|LOC_173|>": 100470, "<|LOC_174|>": 100471, "<|LOC_175|>": 100472, "<|LOC_176|>": 100473, "<|LOC_177|>": 100474, "<|LOC_178|>": 100475, "<|LOC_179|>": 100476, "<|LOC_180|>": 100477, "<|LOC_181|>": 100478, "<|LOC_182|>": 100479, "<|LOC_183|>": 100480, "<|LOC_184|>": 100481, "<|LOC_185|>": 100482, "<|LOC_186|>": 100483, "<|LOC_187|>": 100484, "<|LOC_188|>": 100485, "<|LOC_189|>": 100486, "<|LOC_190|>": 100487, "<|LOC_191|>": 100488, "<|LOC_192|>": 100489, "<|LOC_193|>": 100490, "<|LOC_194|>": 100491, "<|LOC_195|>": 100492, "<|LOC_196|>": 100493, "<|LOC_197|>": 100494, "<|LOC_198|>": 100495, "<|LOC_199|>": 100496, "<|LOC_200|>": 100497, "<|LOC_201|>": 100498, "<|LOC_202|>": 100499, "<|LOC_203|>": 100500, "<|LOC_204|>": 100501, "<|LOC_205|>": 100502, "<|LOC_206|>": 100503, "<|LOC_207|>": 100504, "<|LOC_208|>": 100505, "<|LOC_209|>": 100506, "<|LOC_210|>": 100507, "<|LOC_211|>": 100508, "<|LOC_212|>": 100509, "<|LOC_213|>": 100510, "<|LOC_214|>": 100511, "<|LOC_215|>": 100512, "<|LOC_216|>": 100513, "<|LOC_217|>": 100514, "<|LOC_218|>": 100515, "<|LOC_219|>": 100516, "<|LOC_220|>": 100517, "<|LOC_221|>": 100518, "<|LOC_222|>": 100519, "<|LOC_223|>": 100520, "<|LOC_224|>": 100521, "<|LOC_225|>": 100522, "<|LOC_226|>": 100523, "<|LOC_227|>": 100524, "<|LOC_228|>": 100525, "<|LOC_229|>": 100526, "<|LOC_230|>": 100527, "<|LOC_231|>": 100528, "<|LOC_232|>": 100529, "<|LOC_233|>": 100530, "<|LOC_234|>": 100531, "<|LOC_235|>": 100532, "<|LOC_236|>": 100533, "<|LOC_237|>": 100534, "<|LOC_238|>": 100535, "<|LOC_239|>": 100536, "<|LOC_240|>": 100537, "<|LOC_241|>": 100538, "<|LOC_242|>": 100539, "<|LOC_243|>": 100540, "<|LOC_244|>": 100541, "<|LOC_245|>": 100542, "<|LOC_246|>": 100543, "<|LOC_247|>": 100544, "<|LOC_248|>": 100545, "<|LOC_249|>": 100546, "<|LOC_250|>": 100547, "<|LOC_251|>": 100548, "<|LOC_252|>": 100549, "<|LOC_253|>": 100550, "<|LOC_254|>": 100551, "<|LOC_255|>": 100552, "<|LOC_256|>": 100553, "<|LOC_257|>": 100554, "<|LOC_258|>": 100555, "<|LOC_259|>": 100556, "<|LOC_260|>": 100557, "<|LOC_261|>": 100558, "<|LOC_262|>": 100559, "<|LOC_263|>": 100560, "<|LOC_264|>": 100561, "<|LOC_265|>": 100562, "<|LOC_266|>": 100563, "<|LOC_267|>": 100564, "<|LOC_268|>": 100565, "<|LOC_269|>": 100566, "<|LOC_270|>": 100567, "<|LOC_271|>": 100568, "<|LOC_272|>": 100569, "<|LOC_273|>": 100570, "<|LOC_274|>": 100571, "<|LOC_275|>": 100572, "<|LOC_276|>": 100573, "<|LOC_277|>": 100574, "<|LOC_278|>": 100575, "<|LOC_279|>": 100576, "<|LOC_280|>": 100577, "<|LOC_281|>": 100578, "<|LOC_282|>": 100579, "<|LOC_283|>": 100580, "<|LOC_284|>": 100581, "<|LOC_285|>": 100582, "<|LOC_286|>": 100583, "<|LOC_287|>": 100584, "<|LOC_288|>": 100585, "<|LOC_289|>": 100586, "<|LOC_290|>": 100587, "<|LOC_291|>": 100588, "<|LOC_292|>": 100589, "<|LOC_293|>": 100590, "<|LOC_294|>": 100591, "<|LOC_295|>": 100592, "<|LOC_296|>": 100593, "<|LOC_297|>": 100594, "<|LOC_298|>": 100595, "<|LOC_299|>": 100596, "<|LOC_300|>": 100597, "<|LOC_301|>": 100598, "<|LOC_302|>": 100599, "<|LOC_303|>": 100600, "<|LOC_304|>": 100601, "<|LOC_305|>": 100602, "<|LOC_306|>": 100603, "<|LOC_307|>": 100604, "<|LOC_308|>": 100605, "<|LOC_309|>": 100606, "<|LOC_310|>": 100607, "<|LOC_311|>": 100608, "<|LOC_312|>": 100609, "<|LOC_313|>": 100610, "<|LOC_314|>": 100611, "<|LOC_315|>": 100612, "<|LOC_316|>": 100613, "<|LOC_317|>": 100614, "<|LOC_318|>": 100615, "<|LOC_319|>": 100616, "<|LOC_320|>": 100617, "<|LOC_321|>": 100618, "<|LOC_322|>": 100619, "<|LOC_323|>": 100620, "<|LOC_324|>": 100621, "<|LOC_325|>": 100622, "<|LOC_326|>": 100623, "<|LOC_327|>": 100624, "<|LOC_328|>": 100625, "<|LOC_329|>": 100626, "<|LOC_330|>": 100627, "<|LOC_331|>": 100628, "<|LOC_332|>": 100629, "<|LOC_333|>": 100630, "<|LOC_334|>": 100631, "<|LOC_335|>": 100632, "<|LOC_336|>": 100633, "<|LOC_337|>": 100634, "<|LOC_338|>": 100635, "<|LOC_339|>": 100636, "<|LOC_340|>": 100637, "<|LOC_341|>": 100638, "<|LOC_342|>": 100639, "<|LOC_343|>": 100640, "<|LOC_344|>": 100641, "<|LOC_345|>": 100642, "<|LOC_346|>": 100643, "<|LOC_347|>": 100644, "<|LOC_348|>": 100645, "<|LOC_349|>": 100646, "<|LOC_350|>": 100647, "<|LOC_351|>": 100648, "<|LOC_352|>": 100649, "<|LOC_353|>": 100650, "<|LOC_354|>": 100651, "<|LOC_355|>": 100652, "<|LOC_356|>": 100653, "<|LOC_357|>": 100654, "<|LOC_358|>": 100655, "<|LOC_359|>": 100656, "<|LOC_360|>": 100657, "<|LOC_361|>": 100658, "<|LOC_362|>": 100659, "<|LOC_363|>": 100660, "<|LOC_364|>": 100661, "<|LOC_365|>": 100662, "<|LOC_366|>": 100663, "<|LOC_367|>": 100664, "<|LOC_368|>": 100665, "<|LOC_369|>": 100666, "<|LOC_370|>": 100667, "<|LOC_371|>": 100668, "<|LOC_372|>": 100669, "<|LOC_373|>": 100670, "<|LOC_374|>": 100671, "<|LOC_375|>": 100672, "<|LOC_376|>": 100673, "<|LOC_377|>": 100674, "<|LOC_378|>": 100675, "<|LOC_379|>": 100676, "<|LOC_380|>": 100677, "<|LOC_381|>": 100678, "<|LOC_382|>": 100679, "<|LOC_383|>": 100680, "<|LOC_384|>": 100681, "<|LOC_385|>": 100682, "<|LOC_386|>": 100683, "<|LOC_387|>": 100684, "<|LOC_388|>": 100685, "<|LOC_389|>": 100686, "<|LOC_390|>": 100687, "<|LOC_391|>": 100688, "<|LOC_392|>": 100689, "<|LOC_393|>": 100690, "<|LOC_394|>": 100691, "<|LOC_395|>": 100692, "<|LOC_396|>": 100693, "<|LOC_397|>": 100694, "<|LOC_398|>": 100695, "<|LOC_399|>": 100696, "<|LOC_400|>": 100697, "<|LOC_401|>": 100698, "<|LOC_402|>": 100699, "<|LOC_403|>": 100700, "<|LOC_404|>": 100701, "<|LOC_405|>": 100702, "<|LOC_406|>": 100703, "<|LOC_407|>": 100704, "<|LOC_408|>": 100705, "<|LOC_409|>": 100706, "<|LOC_410|>": 100707, "<|LOC_411|>": 100708, "<|LOC_412|>": 100709, "<|LOC_413|>": 100710, "<|LOC_414|>": 100711, "<|LOC_415|>": 100712, "<|LOC_416|>": 100713, "<|LOC_417|>": 100714, "<|LOC_418|>": 100715, "<|LOC_419|>": 100716, "<|LOC_420|>": 100717, "<|LOC_421|>": 100718, "<|LOC_422|>": 100719, "<|LOC_423|>": 100720, "<|LOC_424|>": 100721, "<|LOC_425|>": 100722, "<|LOC_426|>": 100723, "<|LOC_427|>": 100724, "<|LOC_428|>": 100725, "<|LOC_429|>": 100726, "<|LOC_430|>": 100727, "<|LOC_431|>": 100728, "<|LOC_432|>": 100729, "<|LOC_433|>": 100730, "<|LOC_434|>": 100731, "<|LOC_435|>": 100732, "<|LOC_436|>": 100733, "<|LOC_437|>": 100734, "<|LOC_438|>": 100735, "<|LOC_439|>": 100736, "<|LOC_440|>": 100737, "<|LOC_441|>": 100738, "<|LOC_442|>": 100739, "<|LOC_443|>": 100740, "<|LOC_444|>": 100741, "<|LOC_445|>": 100742, "<|LOC_446|>": 100743, "<|LOC_447|>": 100744, "<|LOC_448|>": 100745, "<|LOC_449|>": 100746, "<|LOC_450|>": 100747, "<|LOC_451|>": 100748, "<|LOC_452|>": 100749, "<|LOC_453|>": 100750, "<|LOC_454|>": 100751, "<|LOC_455|>": 100752, "<|LOC_456|>": 100753, "<|LOC_457|>": 100754, "<|LOC_458|>": 100755, "<|LOC_459|>": 100756, "<|LOC_460|>": 100757, "<|LOC_461|>": 100758, "<|LOC_462|>": 100759, "<|LOC_463|>": 100760, "<|LOC_464|>": 100761, "<|LOC_465|>": 100762, "<|LOC_466|>": 100763, "<|LOC_467|>": 100764, "<|LOC_468|>": 100765, "<|LOC_469|>": 100766, "<|LOC_470|>": 100767, "<|LOC_471|>": 100768, "<|LOC_472|>": 100769, "<|LOC_473|>": 100770, "<|LOC_474|>": 100771, "<|LOC_475|>": 100772, "<|LOC_476|>": 100773, "<|LOC_477|>": 100774, "<|LOC_478|>": 100775, "<|LOC_479|>": 100776, "<|LOC_480|>": 100777, "<|LOC_481|>": 100778, "<|LOC_482|>": 100779, "<|LOC_483|>": 100780, "<|LOC_484|>": 100781, "<|LOC_485|>": 100782, "<|LOC_486|>": 100783, "<|LOC_487|>": 100784, "<|LOC_488|>": 100785, "<|LOC_489|>": 100786, "<|LOC_490|>": 100787, "<|LOC_491|>": 100788, "<|LOC_492|>": 100789, "<|LOC_493|>": 100790, "<|LOC_494|>": 100791, "<|LOC_495|>": 100792, "<|LOC_496|>": 100793, "<|LOC_497|>": 100794, "<|LOC_498|>": 100795, "<|LOC_499|>": 100796, "<|LOC_500|>": 100797, "<|LOC_501|>": 100798, "<|LOC_502|>": 100799, "<|LOC_503|>": 100800, "<|LOC_504|>": 100801, "<|LOC_505|>": 100802, "<|LOC_506|>": 100803, "<|LOC_507|>": 100804, "<|LOC_508|>": 100805, "<|LOC_509|>": 100806, "<|LOC_510|>": 100807, "<|LOC_511|>": 100808, "<|LOC_512|>": 100809, "<|LOC_513|>": 100810, "<|LOC_514|>": 100811, "<|LOC_515|>": 100812, "<|LOC_516|>": 100813, "<|LOC_517|>": 100814, "<|LOC_518|>": 100815, "<|LOC_519|>": 100816, "<|LOC_520|>": 100817, "<|LOC_521|>": 100818, "<|LOC_522|>": 100819, "<|LOC_523|>": 100820, "<|LOC_524|>": 100821, "<|LOC_525|>": 100822, "<|LOC_526|>": 100823, "<|LOC_527|>": 100824, "<|LOC_528|>": 100825, "<|LOC_529|>": 100826, "<|LOC_530|>": 100827, "<|LOC_531|>": 100828, "<|LOC_532|>": 100829, "<|LOC_533|>": 100830, "<|LOC_534|>": 100831, "<|LOC_535|>": 100832, "<|LOC_536|>": 100833, "<|LOC_537|>": 100834, "<|LOC_538|>": 100835, "<|LOC_539|>": 100836, "<|LOC_540|>": 100837, "<|LOC_541|>": 100838, "<|LOC_542|>": 100839, "<|LOC_543|>": 100840, "<|LOC_544|>": 100841, "<|LOC_545|>": 100842, "<|LOC_546|>": 100843, "<|LOC_547|>": 100844, "<|LOC_548|>": 100845, "<|LOC_549|>": 100846, "<|LOC_550|>": 100847, "<|LOC_551|>": 100848, "<|LOC_552|>": 100849, "<|LOC_553|>": 100850, "<|LOC_554|>": 100851, "<|LOC_555|>": 100852, "<|LOC_556|>": 100853, "<|LOC_557|>": 100854, "<|LOC_558|>": 100855, "<|LOC_559|>": 100856, "<|LOC_560|>": 100857, "<|LOC_561|>": 100858, "<|LOC_562|>": 100859, "<|LOC_563|>": 100860, "<|LOC_564|>": 100861, "<|LOC_565|>": 100862, "<|LOC_566|>": 100863, "<|LOC_567|>": 100864, "<|LOC_568|>": 100865, "<|LOC_569|>": 100866, "<|LOC_570|>": 100867, "<|LOC_571|>": 100868, "<|LOC_572|>": 100869, "<|LOC_573|>": 100870, "<|LOC_574|>": 100871, "<|LOC_575|>": 100872, "<|LOC_576|>": 100873, "<|LOC_577|>": 100874, "<|LOC_578|>": 100875, "<|LOC_579|>": 100876, "<|LOC_580|>": 100877, "<|LOC_581|>": 100878, "<|LOC_582|>": 100879, "<|LOC_583|>": 100880, "<|LOC_584|>": 100881, "<|LOC_585|>": 100882, "<|LOC_586|>": 100883, "<|LOC_587|>": 100884, "<|LOC_588|>": 100885, "<|LOC_589|>": 100886, "<|LOC_590|>": 100887, "<|LOC_591|>": 100888, "<|LOC_592|>": 100889, "<|LOC_593|>": 100890, "<|LOC_594|>": 100891, "<|LOC_595|>": 100892, "<|LOC_596|>": 100893, "<|LOC_597|>": 100894, "<|LOC_598|>": 100895, "<|LOC_599|>": 100896, "<|LOC_600|>": 100897, "<|LOC_601|>": 100898, "<|LOC_602|>": 100899, "<|LOC_603|>": 100900, "<|LOC_604|>": 100901, "<|LOC_605|>": 100902, "<|LOC_606|>": 100903, "<|LOC_607|>": 100904, "<|LOC_608|>": 100905, "<|LOC_609|>": 100906, "<|LOC_610|>": 100907, "<|LOC_611|>": 100908, "<|LOC_612|>": 100909, "<|LOC_613|>": 100910, "<|LOC_614|>": 100911, "<|LOC_615|>": 100912, "<|LOC_616|>": 100913, "<|LOC_617|>": 100914, "<|LOC_618|>": 100915, "<|LOC_619|>": 100916, "<|LOC_620|>": 100917, "<|LOC_621|>": 100918, "<|LOC_622|>": 100919, "<|LOC_623|>": 100920, "<|LOC_624|>": 100921, "<|LOC_625|>": 100922, "<|LOC_626|>": 100923, "<|LOC_627|>": 100924, "<|LOC_628|>": 100925, "<|LOC_629|>": 100926, "<|LOC_630|>": 100927, "<|LOC_631|>": 100928, "<|LOC_632|>": 100929, "<|LOC_633|>": 100930, "<|LOC_634|>": 100931, "<|LOC_635|>": 100932, "<|LOC_636|>": 100933, "<|LOC_637|>": 100934, "<|LOC_638|>": 100935, "<|LOC_639|>": 100936, "<|LOC_640|>": 100937, "<|LOC_641|>": 100938, "<|LOC_642|>": 100939, "<|LOC_643|>": 100940, "<|LOC_644|>": 100941, "<|LOC_645|>": 100942, "<|LOC_646|>": 100943, "<|LOC_647|>": 100944, "<|LOC_648|>": 100945, "<|LOC_649|>": 100946, "<|LOC_650|>": 100947, "<|LOC_651|>": 100948, "<|LOC_652|>": 100949, "<|LOC_653|>": 100950, "<|LOC_654|>": 100951, "<|LOC_655|>": 100952, "<|LOC_656|>": 100953, "<|LOC_657|>": 100954, "<|LOC_658|>": 100955, "<|LOC_659|>": 100956, "<|LOC_660|>": 100957, "<|LOC_661|>": 100958, "<|LOC_662|>": 100959, "<|LOC_663|>": 100960, "<|LOC_664|>": 100961, "<|LOC_665|>": 100962, "<|LOC_666|>": 100963, "<|LOC_667|>": 100964, "<|LOC_668|>": 100965, "<|LOC_669|>": 100966, "<|LOC_670|>": 100967, "<|LOC_671|>": 100968, "<|LOC_672|>": 100969, "<|LOC_673|>": 100970, "<|LOC_674|>": 100971, "<|LOC_675|>": 100972, "<|LOC_676|>": 100973, "<|LOC_677|>": 100974, "<|LOC_678|>": 100975, "<|LOC_679|>": 100976, "<|LOC_680|>": 100977, "<|LOC_681|>": 100978, "<|LOC_682|>": 100979, "<|LOC_683|>": 100980, "<|LOC_684|>": 100981, "<|LOC_685|>": 100982, "<|LOC_686|>": 100983, "<|LOC_687|>": 100984, "<|LOC_688|>": 100985, "<|LOC_689|>": 100986, "<|LOC_690|>": 100987, "<|LOC_691|>": 100988, "<|LOC_692|>": 100989, "<|LOC_693|>": 100990, "<|LOC_694|>": 100991, "<|LOC_695|>": 100992, "<|LOC_696|>": 100993, "<|LOC_697|>": 100994, "<|LOC_698|>": 100995, "<|LOC_699|>": 100996, "<|LOC_700|>": 100997, "<|LOC_701|>": 100998, "<|LOC_702|>": 100999, "<|LOC_703|>": 101000, "<|LOC_704|>": 101001, "<|LOC_705|>": 101002, "<|LOC_706|>": 101003, "<|LOC_707|>": 101004, "<|LOC_708|>": 101005, "<|LOC_709|>": 101006, "<|LOC_710|>": 101007, "<|LOC_711|>": 101008, "<|LOC_712|>": 101009, "<|LOC_713|>": 101010, "<|LOC_714|>": 101011, "<|LOC_715|>": 101012, "<|LOC_716|>": 101013, "<|LOC_717|>": 101014, "<|LOC_718|>": 101015, "<|LOC_719|>": 101016, "<|LOC_720|>": 101017, "<|LOC_721|>": 101018, "<|LOC_722|>": 101019, "<|LOC_723|>": 101020, "<|LOC_724|>": 101021, "<|LOC_725|>": 101022, "<|LOC_726|>": 101023, "<|LOC_727|>": 101024, "<|LOC_728|>": 101025, "<|LOC_729|>": 101026, "<|LOC_730|>": 101027, "<|LOC_731|>": 101028, "<|LOC_732|>": 101029, "<|LOC_733|>": 101030, "<|LOC_734|>": 101031, "<|LOC_735|>": 101032, "<|LOC_736|>": 101033, "<|LOC_737|>": 101034, "<|LOC_738|>": 101035, "<|LOC_739|>": 101036, "<|LOC_740|>": 101037, "<|LOC_741|>": 101038, "<|LOC_742|>": 101039, "<|LOC_743|>": 101040, "<|LOC_744|>": 101041, "<|LOC_745|>": 101042, "<|LOC_746|>": 101043, "<|LOC_747|>": 101044, "<|LOC_748|>": 101045, "<|LOC_749|>": 101046, "<|LOC_750|>": 101047, "<|LOC_751|>": 101048, "<|LOC_752|>": 101049, "<|LOC_753|>": 101050, "<|LOC_754|>": 101051, "<|LOC_755|>": 101052, "<|LOC_756|>": 101053, "<|LOC_757|>": 101054, "<|LOC_758|>": 101055, "<|LOC_759|>": 101056, "<|LOC_760|>": 101057, "<|LOC_761|>": 101058, "<|LOC_762|>": 101059, "<|LOC_763|>": 101060, "<|LOC_764|>": 101061, "<|LOC_765|>": 101062, "<|LOC_766|>": 101063, "<|LOC_767|>": 101064, "<|LOC_768|>": 101065, "<|LOC_769|>": 101066, "<|LOC_770|>": 101067, "<|LOC_771|>": 101068, "<|LOC_772|>": 101069, "<|LOC_773|>": 101070, "<|LOC_774|>": 101071, "<|LOC_775|>": 101072, "<|LOC_776|>": 101073, "<|LOC_777|>": 101074, "<|LOC_778|>": 101075, "<|LOC_779|>": 101076, "<|LOC_780|>": 101077, "<|LOC_781|>": 101078, "<|LOC_782|>": 101079, "<|LOC_783|>": 101080, "<|LOC_784|>": 101081, "<|LOC_785|>": 101082, "<|LOC_786|>": 101083, "<|LOC_787|>": 101084, "<|LOC_788|>": 101085, "<|LOC_789|>": 101086, "<|LOC_790|>": 101087, "<|LOC_791|>": 101088, "<|LOC_792|>": 101089, "<|LOC_793|>": 101090, "<|LOC_794|>": 101091, "<|LOC_795|>": 101092, "<|LOC_796|>": 101093, "<|LOC_797|>": 101094, "<|LOC_798|>": 101095, "<|LOC_799|>": 101096, "<|LOC_800|>": 101097, "<|LOC_801|>": 101098, "<|LOC_802|>": 101099, "<|LOC_803|>": 101100, "<|LOC_804|>": 101101, "<|LOC_805|>": 101102, "<|LOC_806|>": 101103, "<|LOC_807|>": 101104, "<|LOC_808|>": 101105, "<|LOC_809|>": 101106, "<|LOC_810|>": 101107, "<|LOC_811|>": 101108, "<|LOC_812|>": 101109, "<|LOC_813|>": 101110, "<|LOC_814|>": 101111, "<|LOC_815|>": 101112, "<|LOC_816|>": 101113, "<|LOC_817|>": 101114, "<|LOC_818|>": 101115, "<|LOC_819|>": 101116, "<|LOC_820|>": 101117, "<|LOC_821|>": 101118, "<|LOC_822|>": 101119, "<|LOC_823|>": 101120, "<|LOC_824|>": 101121, "<|LOC_825|>": 101122, "<|LOC_826|>": 101123, "<|LOC_827|>": 101124, "<|LOC_828|>": 101125, "<|LOC_829|>": 101126, "<|LOC_830|>": 101127, "<|LOC_831|>": 101128, "<|LOC_832|>": 101129, "<|LOC_833|>": 101130, "<|LOC_834|>": 101131, "<|LOC_835|>": 101132, "<|LOC_836|>": 101133, "<|LOC_837|>": 101134, "<|LOC_838|>": 101135, "<|LOC_839|>": 101136, "<|LOC_840|>": 101137, "<|LOC_841|>": 101138, "<|LOC_842|>": 101139, "<|LOC_843|>": 101140, "<|LOC_844|>": 101141, "<|LOC_845|>": 101142, "<|LOC_846|>": 101143, "<|LOC_847|>": 101144, "<|LOC_848|>": 101145, "<|LOC_849|>": 101146, "<|LOC_850|>": 101147, "<|LOC_851|>": 101148, "<|LOC_852|>": 101149, "<|LOC_853|>": 101150, "<|LOC_854|>": 101151, "<|LOC_855|>": 101152, "<|LOC_856|>": 101153, "<|LOC_857|>": 101154, "<|LOC_858|>": 101155, "<|LOC_859|>": 101156, "<|LOC_860|>": 101157, "<|LOC_861|>": 101158, "<|LOC_862|>": 101159, "<|LOC_863|>": 101160, "<|LOC_864|>": 101161, "<|LOC_865|>": 101162, "<|LOC_866|>": 101163, "<|LOC_867|>": 101164, "<|LOC_868|>": 101165, "<|LOC_869|>": 101166, "<|LOC_870|>": 101167, "<|LOC_871|>": 101168, "<|LOC_872|>": 101169, "<|LOC_873|>": 101170, "<|LOC_874|>": 101171, "<|LOC_875|>": 101172, "<|LOC_876|>": 101173, "<|LOC_877|>": 101174, "<|LOC_878|>": 101175, "<|LOC_879|>": 101176, "<|LOC_880|>": 101177, "<|LOC_881|>": 101178, "<|LOC_882|>": 101179, "<|LOC_883|>": 101180, "<|LOC_884|>": 101181, "<|LOC_885|>": 101182, "<|LOC_886|>": 101183, "<|LOC_887|>": 101184, "<|LOC_888|>": 101185, "<|LOC_889|>": 101186, "<|LOC_890|>": 101187, "<|LOC_891|>": 101188, "<|LOC_892|>": 101189, "<|LOC_893|>": 101190, "<|LOC_894|>": 101191, "<|LOC_895|>": 101192, "<|LOC_896|>": 101193, "<|LOC_897|>": 101194, "<|LOC_898|>": 101195, "<|LOC_899|>": 101196, "<|LOC_900|>": 101197, "<|LOC_901|>": 101198, "<|LOC_902|>": 101199, "<|LOC_903|>": 101200, "<|LOC_904|>": 101201, "<|LOC_905|>": 101202, "<|LOC_906|>": 101203, "<|LOC_907|>": 101204, "<|LOC_908|>": 101205, "<|LOC_909|>": 101206, "<|LOC_910|>": 101207, "<|LOC_911|>": 101208, "<|LOC_912|>": 101209, "<|LOC_913|>": 101210, "<|LOC_914|>": 101211, "<|LOC_915|>": 101212, "<|LOC_916|>": 101213, "<|LOC_917|>": 101214, "<|LOC_918|>": 101215, "<|LOC_919|>": 101216, "<|LOC_920|>": 101217, "<|LOC_921|>": 101218, "<|LOC_922|>": 101219, "<|LOC_923|>": 101220, "<|LOC_924|>": 101221, "<|LOC_925|>": 101222, "<|LOC_926|>": 101223, "<|LOC_927|>": 101224, "<|LOC_928|>": 101225, "<|LOC_929|>": 101226, "<|LOC_930|>": 101227, "<|LOC_931|>": 101228, "<|LOC_932|>": 101229, "<|LOC_933|>": 101230, "<|LOC_934|>": 101231, "<|LOC_935|>": 101232, "<|LOC_936|>": 101233, "<|LOC_937|>": 101234, "<|LOC_938|>": 101235, "<|LOC_939|>": 101236, "<|LOC_940|>": 101237, "<|LOC_941|>": 101238, "<|LOC_942|>": 101239, "<|LOC_943|>": 101240, "<|LOC_944|>": 101241, "<|LOC_945|>": 101242, "<|LOC_946|>": 101243, "<|LOC_947|>": 101244, "<|LOC_948|>": 101245, "<|LOC_949|>": 101246, "<|LOC_950|>": 101247, "<|LOC_951|>": 101248, "<|LOC_952|>": 101249, "<|LOC_953|>": 101250, "<|LOC_954|>": 101251, "<|LOC_955|>": 101252, "<|LOC_956|>": 101253, "<|LOC_957|>": 101254, "<|LOC_958|>": 101255, "<|LOC_959|>": 101256, "<|LOC_960|>": 101257, "<|LOC_961|>": 101258, "<|LOC_962|>": 101259, "<|LOC_963|>": 101260, "<|LOC_964|>": 101261, "<|LOC_965|>": 101262, "<|LOC_966|>": 101263, "<|LOC_967|>": 101264, "<|LOC_968|>": 101265, "<|LOC_969|>": 101266, "<|LOC_970|>": 101267, "<|LOC_971|>": 101268, "<|LOC_972|>": 101269, "<|LOC_973|>": 101270, "<|LOC_974|>": 101271, "<|LOC_975|>": 101272, "<|LOC_976|>": 101273, "<|LOC_977|>": 101274, "<|LOC_978|>": 101275, "<|LOC_979|>": 101276, "<|LOC_980|>": 101277, "<|LOC_981|>": 101278, "<|LOC_982|>": 101279, "<|LOC_983|>": 101280, "<|LOC_984|>": 101281, "<|LOC_985|>": 101282, "<|LOC_986|>": 101283, "<|LOC_987|>": 101284, "<|LOC_988|>": 101285, "<|LOC_989|>": 101286, "<|LOC_990|>": 101287, "<|LOC_991|>": 101288, "<|LOC_992|>": 101289, "<|LOC_993|>": 101290, "<|LOC_994|>": 101291, "<|LOC_995|>": 101292, "<|LOC_996|>": 101293, "<|LOC_997|>": 101294, "<|LOC_998|>": 101295, "<|LOC_999|>": 101296, "<|LOC_1000|>": 101297, "<|LOC_BEGIN|>": 101298, "<|LOC_END|>": 101299, "<|LOC_SEP|>": 101300, "<|CROP_COL_SEP|>": 101301, "<|CROP_ROW_SEP|>": 101302, "<|IMAGE_SEP|>": 101303, "<|image_pad|>": 101304, "<|IMAGE_START|>": 101305, "<|IMAGE_END|>": 101306, "<|video_pad|>": 101307, "<ecel>": 101308, "<fcel>": 101309, "<xcel>": 101310, "<lcel>": 101311, "<ucel>": 101312, "<nl>": 101313}
|
checkpoint-336/config.json
ADDED
|
@@ -0,0 +1,223 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"architectures": [
|
| 3 |
+
"PaddleOCRVLForConditionalGeneration"
|
| 4 |
+
],
|
| 5 |
+
"attention_probs_dropout_prob": 0.0,
|
| 6 |
+
"auto_map": {
|
| 7 |
+
"AutoConfig": "configuration_paddleocr_vl.PaddleOCRVLConfig",
|
| 8 |
+
"AutoModel": "modeling_paddleocr_vl.PaddleOCRVLForConditionalGeneration",
|
| 9 |
+
"AutoModelForCausalLM": "modeling_paddleocr_vl.PaddleOCRVLForConditionalGeneration"
|
| 10 |
+
},
|
| 11 |
+
"compression_ratio": 1.0,
|
| 12 |
+
"disable_pipeline_warmup": false,
|
| 13 |
+
"enable_mtp_magic_send": false,
|
| 14 |
+
"fp16_opt_level": "O2",
|
| 15 |
+
"freq_allocation": 20,
|
| 16 |
+
"fuse_ln": false,
|
| 17 |
+
"fuse_rms_norm": true,
|
| 18 |
+
"head_dim": 128,
|
| 19 |
+
"hidden_act": "silu",
|
| 20 |
+
"hidden_dropout_prob": 0.0,
|
| 21 |
+
"hidden_size": 1024,
|
| 22 |
+
"ignored_index": -100,
|
| 23 |
+
"im_patch_id": 100295,
|
| 24 |
+
"image_token_id": 100295,
|
| 25 |
+
"intermediate_size": 3072,
|
| 26 |
+
"max_position_embeddings": 131072,
|
| 27 |
+
"max_text_id": 100295,
|
| 28 |
+
"model_type": "paddleocr_vl",
|
| 29 |
+
"moe_dropout_prob": 0.0,
|
| 30 |
+
"moe_multimodal_dispatch_use_allgather": "v2-alltoall-unpad",
|
| 31 |
+
"num_attention_heads": 16,
|
| 32 |
+
"num_hidden_layers": 18,
|
| 33 |
+
"num_key_value_heads": 2,
|
| 34 |
+
"paddleformers_version": "0.4.0",
|
| 35 |
+
"pixel_hidden_size": 1152,
|
| 36 |
+
"rms_norm_eps": 1e-05,
|
| 37 |
+
"rope_is_neox_style": true,
|
| 38 |
+
"rope_scaling": {
|
| 39 |
+
"mrope_section": [
|
| 40 |
+
16,
|
| 41 |
+
24,
|
| 42 |
+
24
|
| 43 |
+
],
|
| 44 |
+
"rope_type": "default",
|
| 45 |
+
"type": "default"
|
| 46 |
+
},
|
| 47 |
+
"rope_theta": 500000,
|
| 48 |
+
"scale_qk_coeff": 1.0,
|
| 49 |
+
"seqlen": 16384,
|
| 50 |
+
"sliding_window": null,
|
| 51 |
+
"tie_word_embeddings": false,
|
| 52 |
+
"token_balance_loss": false,
|
| 53 |
+
"token_balance_seqlen": 16384,
|
| 54 |
+
"torch_dtype": "bfloat16",
|
| 55 |
+
"use_3d_rope": true,
|
| 56 |
+
"use_bias": false,
|
| 57 |
+
"use_flash_attn_with_mask": true,
|
| 58 |
+
"use_fp8": false,
|
| 59 |
+
"use_mem_eff_attn": true,
|
| 60 |
+
"use_recompute_moe": false,
|
| 61 |
+
"use_rmsnorm": true,
|
| 62 |
+
"video_token_id": 101307,
|
| 63 |
+
"vision_config": {
|
| 64 |
+
"_attn_implementation": "eager",
|
| 65 |
+
"_name_or_path": "",
|
| 66 |
+
"_save_to_hf": false,
|
| 67 |
+
"add_cross_attention": false,
|
| 68 |
+
"add_tail_layers": 0,
|
| 69 |
+
"architectures": [
|
| 70 |
+
"PaddleOCRVisionModel"
|
| 71 |
+
],
|
| 72 |
+
"attention_dropout": 0.0,
|
| 73 |
+
"auto_map": {
|
| 74 |
+
"AutoConfig": "configuration_paddleocr_vl.PaddleOCRVLConfig",
|
| 75 |
+
"AutoModel": "modeling_paddleocr_vl.PaddleOCRVisionModel"
|
| 76 |
+
},
|
| 77 |
+
"bad_words_ids": null,
|
| 78 |
+
"begin_suppress_tokens": null,
|
| 79 |
+
"bos_token_id": null,
|
| 80 |
+
"chunk_size_feed_forward": 0,
|
| 81 |
+
"classifier_dropout": null,
|
| 82 |
+
"context_parallel_degree": 1,
|
| 83 |
+
"cross_attention_hidden_size": null,
|
| 84 |
+
"decoder_start_token_id": null,
|
| 85 |
+
"diversity_penalty": 0.0,
|
| 86 |
+
"do_sample": false,
|
| 87 |
+
"dpo_config": null,
|
| 88 |
+
"dtype": "bfloat16",
|
| 89 |
+
"early_stopping": false,
|
| 90 |
+
"encoder_no_repeat_ngram_size": 0,
|
| 91 |
+
"eos_token_id": null,
|
| 92 |
+
"ep_communication_type": "deepep",
|
| 93 |
+
"exponential_decay_length_penalty": null,
|
| 94 |
+
"finetuning_task": null,
|
| 95 |
+
"forced_bos_token_id": null,
|
| 96 |
+
"forced_eos_token_id": null,
|
| 97 |
+
"fuse_attention_ffn": false,
|
| 98 |
+
"fuse_attention_qkv": false,
|
| 99 |
+
"fuse_linear": false,
|
| 100 |
+
"fuse_rope": false,
|
| 101 |
+
"fuse_sequence_parallel_allreduce": false,
|
| 102 |
+
"fuse_swiglu": false,
|
| 103 |
+
"hidden_act": "gelu_new",
|
| 104 |
+
"hidden_size": 1152,
|
| 105 |
+
"id2label": {
|
| 106 |
+
"0": "LABEL_0",
|
| 107 |
+
"1": "LABEL_1"
|
| 108 |
+
},
|
| 109 |
+
"image_size": 384,
|
| 110 |
+
"intermediate_size": 4304,
|
| 111 |
+
"is_decoder": false,
|
| 112 |
+
"is_encoder_decoder": false,
|
| 113 |
+
"kto_config": null,
|
| 114 |
+
"label2id": {
|
| 115 |
+
"LABEL_0": 0,
|
| 116 |
+
"LABEL_1": 1
|
| 117 |
+
},
|
| 118 |
+
"layer_norm_eps": 1e-06,
|
| 119 |
+
"length_penalty": 1.0,
|
| 120 |
+
"loss_subbatch_sequence_length": -1,
|
| 121 |
+
"max_length": 20,
|
| 122 |
+
"min_length": 0,
|
| 123 |
+
"model_type": "paddleocr_vl",
|
| 124 |
+
"moe_subbatch_token_num": 0,
|
| 125 |
+
"no_recompute_layers": null,
|
| 126 |
+
"no_repeat_ngram_size": 0,
|
| 127 |
+
"num_attention_heads": 16,
|
| 128 |
+
"num_beam_groups": 1,
|
| 129 |
+
"num_beams": 1,
|
| 130 |
+
"num_channels": 3,
|
| 131 |
+
"num_choices": null,
|
| 132 |
+
"num_hidden_layers": 27,
|
| 133 |
+
"num_nextn_predict_layers": 0,
|
| 134 |
+
"num_return_sequences": 1,
|
| 135 |
+
"offload_recompute_inputs": false,
|
| 136 |
+
"output_attentions": false,
|
| 137 |
+
"output_hidden_states": false,
|
| 138 |
+
"output_scores": false,
|
| 139 |
+
"pad_token_id": 0,
|
| 140 |
+
"patch_size": 14,
|
| 141 |
+
"pipeline_parallel_degree": 1,
|
| 142 |
+
"pp_recompute_interval": 1,
|
| 143 |
+
"prefix": null,
|
| 144 |
+
"problem_type": null,
|
| 145 |
+
"pruned_heads": {},
|
| 146 |
+
"quantization_config": {
|
| 147 |
+
"act_quant_method": "abs_max",
|
| 148 |
+
"activation_scheme": null,
|
| 149 |
+
"actscale_moving_rate": 0.01,
|
| 150 |
+
"apply_hadamard": false,
|
| 151 |
+
"apply_online_actscale_step": 200,
|
| 152 |
+
"dense_quant_type": "",
|
| 153 |
+
"dtype": null,
|
| 154 |
+
"fmt": null,
|
| 155 |
+
"fp8_format_type": "hybrid",
|
| 156 |
+
"group_size": -1,
|
| 157 |
+
"hadamard_block_size": 32,
|
| 158 |
+
"ignore_modules": null,
|
| 159 |
+
"llm_int8_threshold": 6.0,
|
| 160 |
+
"moe_quant_type": "",
|
| 161 |
+
"qlora_weight_blocksize": 64,
|
| 162 |
+
"qlora_weight_double_quant": false,
|
| 163 |
+
"qlora_weight_double_quant_block_size": 256,
|
| 164 |
+
"quant_input_grad": false,
|
| 165 |
+
"quant_method": null,
|
| 166 |
+
"quant_round_type": 0,
|
| 167 |
+
"quant_type": null,
|
| 168 |
+
"quant_weight_grad": false,
|
| 169 |
+
"quantization": "",
|
| 170 |
+
"scale_epsilon": 1e-08,
|
| 171 |
+
"shift": false,
|
| 172 |
+
"shift_smooth_all_linears": false,
|
| 173 |
+
"smooth": false,
|
| 174 |
+
"weight_block_size": null,
|
| 175 |
+
"weight_quant_method": "abs_max_channel_wise",
|
| 176 |
+
"weight_quantize_algo": null
|
| 177 |
+
},
|
| 178 |
+
"recompute": true,
|
| 179 |
+
"recompute_granularity": "full",
|
| 180 |
+
"recompute_use_reentrant": false,
|
| 181 |
+
"refined_recompute": "",
|
| 182 |
+
"remove_invalid_values": false,
|
| 183 |
+
"repetition_penalty": 1.0,
|
| 184 |
+
"return_dict": false,
|
| 185 |
+
"return_dict_in_generate": false,
|
| 186 |
+
"sep_parallel_degree": 1,
|
| 187 |
+
"sep_token_id": null,
|
| 188 |
+
"sequence_parallel": false,
|
| 189 |
+
"spatial_merge_size": 2,
|
| 190 |
+
"suppress_tokens": null,
|
| 191 |
+
"task_specific_params": null,
|
| 192 |
+
"temperature": 1.0,
|
| 193 |
+
"temporal_patch_size": 2,
|
| 194 |
+
"tensor_parallel_degree": 1,
|
| 195 |
+
"tensor_parallel_output": true,
|
| 196 |
+
"tensor_parallel_rank": 0,
|
| 197 |
+
"tie_encoder_decoder": false,
|
| 198 |
+
"tie_word_embeddings": true,
|
| 199 |
+
"tokenizer_class": null,
|
| 200 |
+
"tokens_per_second": 2,
|
| 201 |
+
"top_k": 50,
|
| 202 |
+
"top_p": 1.0,
|
| 203 |
+
"typical_p": 1.0,
|
| 204 |
+
"use_cache": false,
|
| 205 |
+
"use_filtered_label_loss": false,
|
| 206 |
+
"use_flash_attention": true,
|
| 207 |
+
"use_fused_dropout_add": false,
|
| 208 |
+
"use_fused_head_and_loss_fn": false,
|
| 209 |
+
"use_fused_linear": false,
|
| 210 |
+
"use_fused_linear_cross_entropy": false,
|
| 211 |
+
"use_fused_rms_norm": false,
|
| 212 |
+
"use_fused_rope": false,
|
| 213 |
+
"use_sparse_flash_attn": true,
|
| 214 |
+
"use_sparse_head_and_loss_fn": false,
|
| 215 |
+
"use_unified_moe": false,
|
| 216 |
+
"using_fake_gate": false,
|
| 217 |
+
"virtual_pp_degree": 1
|
| 218 |
+
},
|
| 219 |
+
"vision_end_token_id": 101306,
|
| 220 |
+
"vision_start_token_id": 101305,
|
| 221 |
+
"vocab_size": 103424,
|
| 222 |
+
"weight_share_add_bias": true
|
| 223 |
+
}
|
checkpoint-336/generation_config.json
ADDED
|
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_from_model_config": true,
|
| 3 |
+
"eos_token_id": 2,
|
| 4 |
+
"pad_token_id": 0,
|
| 5 |
+
"use_cache": false
|
| 6 |
+
}
|
checkpoint-336/master_weights-00001-of-00001.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3a0c4828027d40f6474016cae12335498d8c612eb2531e2fdbedb34699f503ce
|
| 3 |
+
size 3622483352
|
checkpoint-336/master_weights.safetensors.index.json
ADDED
|
@@ -0,0 +1,615 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"metadata": {
|
| 3 |
+
"total_size": 3622406592
|
| 4 |
+
},
|
| 5 |
+
"weight_map": {
|
| 6 |
+
"mlp_AR.pre_norm.weight": "master_weights-00001-of-00001.safetensors",
|
| 7 |
+
"mlp_AR.pre_norm.bias": "master_weights-00001-of-00001.safetensors",
|
| 8 |
+
"mlp_AR.linear_1.weight": "master_weights-00001-of-00001.safetensors",
|
| 9 |
+
"mlp_AR.linear_1.bias": "master_weights-00001-of-00001.safetensors",
|
| 10 |
+
"mlp_AR.linear_2.weight": "master_weights-00001-of-00001.safetensors",
|
| 11 |
+
"mlp_AR.linear_2.bias": "master_weights-00001-of-00001.safetensors",
|
| 12 |
+
"visual.vision_model.embeddings.patch_embedding.weight": "master_weights-00001-of-00001.safetensors",
|
| 13 |
+
"visual.vision_model.embeddings.patch_embedding.bias": "master_weights-00001-of-00001.safetensors",
|
| 14 |
+
"visual.vision_model.embeddings.position_embedding.weight": "master_weights-00001-of-00001.safetensors",
|
| 15 |
+
"visual.vision_model.encoder.layers.0.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 16 |
+
"visual.vision_model.encoder.layers.0.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 17 |
+
"visual.vision_model.encoder.layers.0.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 18 |
+
"visual.vision_model.encoder.layers.0.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 19 |
+
"visual.vision_model.encoder.layers.0.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 20 |
+
"visual.vision_model.encoder.layers.0.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 21 |
+
"visual.vision_model.encoder.layers.0.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 22 |
+
"visual.vision_model.encoder.layers.0.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 23 |
+
"visual.vision_model.encoder.layers.0.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 24 |
+
"visual.vision_model.encoder.layers.0.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 25 |
+
"visual.vision_model.encoder.layers.0.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 26 |
+
"visual.vision_model.encoder.layers.0.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 27 |
+
"visual.vision_model.encoder.layers.0.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 28 |
+
"visual.vision_model.encoder.layers.0.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 29 |
+
"visual.vision_model.encoder.layers.0.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 30 |
+
"visual.vision_model.encoder.layers.0.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 31 |
+
"visual.vision_model.encoder.layers.1.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 32 |
+
"visual.vision_model.encoder.layers.1.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 33 |
+
"visual.vision_model.encoder.layers.1.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 34 |
+
"visual.vision_model.encoder.layers.1.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 35 |
+
"visual.vision_model.encoder.layers.1.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 36 |
+
"visual.vision_model.encoder.layers.1.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 37 |
+
"visual.vision_model.encoder.layers.1.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 38 |
+
"visual.vision_model.encoder.layers.1.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 39 |
+
"visual.vision_model.encoder.layers.1.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 40 |
+
"visual.vision_model.encoder.layers.1.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 41 |
+
"visual.vision_model.encoder.layers.1.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 42 |
+
"visual.vision_model.encoder.layers.1.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 43 |
+
"visual.vision_model.encoder.layers.1.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 44 |
+
"visual.vision_model.encoder.layers.1.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 45 |
+
"visual.vision_model.encoder.layers.1.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 46 |
+
"visual.vision_model.encoder.layers.1.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 47 |
+
"visual.vision_model.encoder.layers.2.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 48 |
+
"visual.vision_model.encoder.layers.2.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 49 |
+
"visual.vision_model.encoder.layers.2.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 50 |
+
"visual.vision_model.encoder.layers.2.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 51 |
+
"visual.vision_model.encoder.layers.2.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 52 |
+
"visual.vision_model.encoder.layers.2.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 53 |
+
"visual.vision_model.encoder.layers.2.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 54 |
+
"visual.vision_model.encoder.layers.2.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 55 |
+
"visual.vision_model.encoder.layers.2.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 56 |
+
"visual.vision_model.encoder.layers.2.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 57 |
+
"visual.vision_model.encoder.layers.2.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 58 |
+
"visual.vision_model.encoder.layers.2.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 59 |
+
"visual.vision_model.encoder.layers.2.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 60 |
+
"visual.vision_model.encoder.layers.2.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 61 |
+
"visual.vision_model.encoder.layers.2.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 62 |
+
"visual.vision_model.encoder.layers.2.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 63 |
+
"visual.vision_model.encoder.layers.3.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 64 |
+
"visual.vision_model.encoder.layers.3.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 65 |
+
"visual.vision_model.encoder.layers.3.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 66 |
+
"visual.vision_model.encoder.layers.3.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 67 |
+
"visual.vision_model.encoder.layers.3.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 68 |
+
"visual.vision_model.encoder.layers.3.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 69 |
+
"visual.vision_model.encoder.layers.3.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 70 |
+
"visual.vision_model.encoder.layers.3.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 71 |
+
"visual.vision_model.encoder.layers.3.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 72 |
+
"visual.vision_model.encoder.layers.3.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 73 |
+
"visual.vision_model.encoder.layers.3.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 74 |
+
"visual.vision_model.encoder.layers.3.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 75 |
+
"visual.vision_model.encoder.layers.3.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 76 |
+
"visual.vision_model.encoder.layers.3.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 77 |
+
"visual.vision_model.encoder.layers.3.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 78 |
+
"visual.vision_model.encoder.layers.3.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 79 |
+
"visual.vision_model.encoder.layers.4.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 80 |
+
"visual.vision_model.encoder.layers.4.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 81 |
+
"visual.vision_model.encoder.layers.4.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 82 |
+
"visual.vision_model.encoder.layers.4.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 83 |
+
"visual.vision_model.encoder.layers.4.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 84 |
+
"visual.vision_model.encoder.layers.4.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 85 |
+
"visual.vision_model.encoder.layers.4.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 86 |
+
"visual.vision_model.encoder.layers.4.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 87 |
+
"visual.vision_model.encoder.layers.4.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 88 |
+
"visual.vision_model.encoder.layers.4.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 89 |
+
"visual.vision_model.encoder.layers.4.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 90 |
+
"visual.vision_model.encoder.layers.4.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 91 |
+
"visual.vision_model.encoder.layers.4.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 92 |
+
"visual.vision_model.encoder.layers.4.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 93 |
+
"visual.vision_model.encoder.layers.4.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 94 |
+
"visual.vision_model.encoder.layers.4.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 95 |
+
"visual.vision_model.encoder.layers.5.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 96 |
+
"visual.vision_model.encoder.layers.5.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 97 |
+
"visual.vision_model.encoder.layers.5.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 98 |
+
"visual.vision_model.encoder.layers.5.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 99 |
+
"visual.vision_model.encoder.layers.5.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 100 |
+
"visual.vision_model.encoder.layers.5.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 101 |
+
"visual.vision_model.encoder.layers.5.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 102 |
+
"visual.vision_model.encoder.layers.5.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 103 |
+
"visual.vision_model.encoder.layers.5.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 104 |
+
"visual.vision_model.encoder.layers.5.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 105 |
+
"visual.vision_model.encoder.layers.5.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 106 |
+
"visual.vision_model.encoder.layers.5.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 107 |
+
"visual.vision_model.encoder.layers.5.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 108 |
+
"visual.vision_model.encoder.layers.5.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 109 |
+
"visual.vision_model.encoder.layers.5.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 110 |
+
"visual.vision_model.encoder.layers.5.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 111 |
+
"visual.vision_model.encoder.layers.6.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 112 |
+
"visual.vision_model.encoder.layers.6.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 113 |
+
"visual.vision_model.encoder.layers.6.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 114 |
+
"visual.vision_model.encoder.layers.6.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 115 |
+
"visual.vision_model.encoder.layers.6.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 116 |
+
"visual.vision_model.encoder.layers.6.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 117 |
+
"visual.vision_model.encoder.layers.6.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 118 |
+
"visual.vision_model.encoder.layers.6.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 119 |
+
"visual.vision_model.encoder.layers.6.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 120 |
+
"visual.vision_model.encoder.layers.6.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 121 |
+
"visual.vision_model.encoder.layers.6.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 122 |
+
"visual.vision_model.encoder.layers.6.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 123 |
+
"visual.vision_model.encoder.layers.6.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 124 |
+
"visual.vision_model.encoder.layers.6.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 125 |
+
"visual.vision_model.encoder.layers.6.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 126 |
+
"visual.vision_model.encoder.layers.6.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 127 |
+
"visual.vision_model.encoder.layers.7.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 128 |
+
"visual.vision_model.encoder.layers.7.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 129 |
+
"visual.vision_model.encoder.layers.7.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 130 |
+
"visual.vision_model.encoder.layers.7.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 131 |
+
"visual.vision_model.encoder.layers.7.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 132 |
+
"visual.vision_model.encoder.layers.7.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 133 |
+
"visual.vision_model.encoder.layers.7.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 134 |
+
"visual.vision_model.encoder.layers.7.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 135 |
+
"visual.vision_model.encoder.layers.7.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 136 |
+
"visual.vision_model.encoder.layers.7.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 137 |
+
"visual.vision_model.encoder.layers.7.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 138 |
+
"visual.vision_model.encoder.layers.7.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 139 |
+
"visual.vision_model.encoder.layers.7.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 140 |
+
"visual.vision_model.encoder.layers.7.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 141 |
+
"visual.vision_model.encoder.layers.7.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 142 |
+
"visual.vision_model.encoder.layers.7.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 143 |
+
"visual.vision_model.encoder.layers.8.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 144 |
+
"visual.vision_model.encoder.layers.8.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 145 |
+
"visual.vision_model.encoder.layers.8.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 146 |
+
"visual.vision_model.encoder.layers.8.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 147 |
+
"visual.vision_model.encoder.layers.8.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 148 |
+
"visual.vision_model.encoder.layers.8.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 149 |
+
"visual.vision_model.encoder.layers.8.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 150 |
+
"visual.vision_model.encoder.layers.8.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 151 |
+
"visual.vision_model.encoder.layers.8.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 152 |
+
"visual.vision_model.encoder.layers.8.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 153 |
+
"visual.vision_model.encoder.layers.8.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 154 |
+
"visual.vision_model.encoder.layers.8.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 155 |
+
"visual.vision_model.encoder.layers.8.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 156 |
+
"visual.vision_model.encoder.layers.8.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 157 |
+
"visual.vision_model.encoder.layers.8.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 158 |
+
"visual.vision_model.encoder.layers.8.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 159 |
+
"visual.vision_model.encoder.layers.9.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 160 |
+
"visual.vision_model.encoder.layers.9.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 161 |
+
"visual.vision_model.encoder.layers.9.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 162 |
+
"visual.vision_model.encoder.layers.9.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 163 |
+
"visual.vision_model.encoder.layers.9.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 164 |
+
"visual.vision_model.encoder.layers.9.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 165 |
+
"visual.vision_model.encoder.layers.9.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 166 |
+
"visual.vision_model.encoder.layers.9.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 167 |
+
"visual.vision_model.encoder.layers.9.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 168 |
+
"visual.vision_model.encoder.layers.9.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 169 |
+
"visual.vision_model.encoder.layers.9.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 170 |
+
"visual.vision_model.encoder.layers.9.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 171 |
+
"visual.vision_model.encoder.layers.9.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 172 |
+
"visual.vision_model.encoder.layers.9.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 173 |
+
"visual.vision_model.encoder.layers.9.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 174 |
+
"visual.vision_model.encoder.layers.9.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 175 |
+
"visual.vision_model.encoder.layers.10.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 176 |
+
"visual.vision_model.encoder.layers.10.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 177 |
+
"visual.vision_model.encoder.layers.10.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 178 |
+
"visual.vision_model.encoder.layers.10.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 179 |
+
"visual.vision_model.encoder.layers.10.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 180 |
+
"visual.vision_model.encoder.layers.10.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 181 |
+
"visual.vision_model.encoder.layers.10.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 182 |
+
"visual.vision_model.encoder.layers.10.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 183 |
+
"visual.vision_model.encoder.layers.10.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 184 |
+
"visual.vision_model.encoder.layers.10.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 185 |
+
"visual.vision_model.encoder.layers.10.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 186 |
+
"visual.vision_model.encoder.layers.10.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 187 |
+
"visual.vision_model.encoder.layers.10.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 188 |
+
"visual.vision_model.encoder.layers.10.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 189 |
+
"visual.vision_model.encoder.layers.10.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 190 |
+
"visual.vision_model.encoder.layers.10.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 191 |
+
"visual.vision_model.encoder.layers.11.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 192 |
+
"visual.vision_model.encoder.layers.11.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 193 |
+
"visual.vision_model.encoder.layers.11.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 194 |
+
"visual.vision_model.encoder.layers.11.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 195 |
+
"visual.vision_model.encoder.layers.11.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 196 |
+
"visual.vision_model.encoder.layers.11.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 197 |
+
"visual.vision_model.encoder.layers.11.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 198 |
+
"visual.vision_model.encoder.layers.11.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 199 |
+
"visual.vision_model.encoder.layers.11.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 200 |
+
"visual.vision_model.encoder.layers.11.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 201 |
+
"visual.vision_model.encoder.layers.11.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 202 |
+
"visual.vision_model.encoder.layers.11.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 203 |
+
"visual.vision_model.encoder.layers.11.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 204 |
+
"visual.vision_model.encoder.layers.11.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 205 |
+
"visual.vision_model.encoder.layers.11.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 206 |
+
"visual.vision_model.encoder.layers.11.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 207 |
+
"visual.vision_model.encoder.layers.12.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 208 |
+
"visual.vision_model.encoder.layers.12.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 209 |
+
"visual.vision_model.encoder.layers.12.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 210 |
+
"visual.vision_model.encoder.layers.12.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 211 |
+
"visual.vision_model.encoder.layers.12.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 212 |
+
"visual.vision_model.encoder.layers.12.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 213 |
+
"visual.vision_model.encoder.layers.12.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 214 |
+
"visual.vision_model.encoder.layers.12.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 215 |
+
"visual.vision_model.encoder.layers.12.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 216 |
+
"visual.vision_model.encoder.layers.12.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 217 |
+
"visual.vision_model.encoder.layers.12.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 218 |
+
"visual.vision_model.encoder.layers.12.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 219 |
+
"visual.vision_model.encoder.layers.12.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 220 |
+
"visual.vision_model.encoder.layers.12.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 221 |
+
"visual.vision_model.encoder.layers.12.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 222 |
+
"visual.vision_model.encoder.layers.12.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 223 |
+
"visual.vision_model.encoder.layers.13.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 224 |
+
"visual.vision_model.encoder.layers.13.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 225 |
+
"visual.vision_model.encoder.layers.13.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 226 |
+
"visual.vision_model.encoder.layers.13.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 227 |
+
"visual.vision_model.encoder.layers.13.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 228 |
+
"visual.vision_model.encoder.layers.13.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 229 |
+
"visual.vision_model.encoder.layers.13.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 230 |
+
"visual.vision_model.encoder.layers.13.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 231 |
+
"visual.vision_model.encoder.layers.13.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 232 |
+
"visual.vision_model.encoder.layers.13.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 233 |
+
"visual.vision_model.encoder.layers.13.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 234 |
+
"visual.vision_model.encoder.layers.13.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 235 |
+
"visual.vision_model.encoder.layers.13.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 236 |
+
"visual.vision_model.encoder.layers.13.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 237 |
+
"visual.vision_model.encoder.layers.13.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 238 |
+
"visual.vision_model.encoder.layers.13.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 239 |
+
"visual.vision_model.encoder.layers.14.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 240 |
+
"visual.vision_model.encoder.layers.14.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 241 |
+
"visual.vision_model.encoder.layers.14.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 242 |
+
"visual.vision_model.encoder.layers.14.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 243 |
+
"visual.vision_model.encoder.layers.14.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 244 |
+
"visual.vision_model.encoder.layers.14.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 245 |
+
"visual.vision_model.encoder.layers.14.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 246 |
+
"visual.vision_model.encoder.layers.14.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 247 |
+
"visual.vision_model.encoder.layers.14.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 248 |
+
"visual.vision_model.encoder.layers.14.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 249 |
+
"visual.vision_model.encoder.layers.14.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 250 |
+
"visual.vision_model.encoder.layers.14.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 251 |
+
"visual.vision_model.encoder.layers.14.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 252 |
+
"visual.vision_model.encoder.layers.14.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 253 |
+
"visual.vision_model.encoder.layers.14.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 254 |
+
"visual.vision_model.encoder.layers.14.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 255 |
+
"visual.vision_model.encoder.layers.15.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 256 |
+
"visual.vision_model.encoder.layers.15.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 257 |
+
"visual.vision_model.encoder.layers.15.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 258 |
+
"visual.vision_model.encoder.layers.15.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 259 |
+
"visual.vision_model.encoder.layers.15.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 260 |
+
"visual.vision_model.encoder.layers.15.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 261 |
+
"visual.vision_model.encoder.layers.15.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 262 |
+
"visual.vision_model.encoder.layers.15.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 263 |
+
"visual.vision_model.encoder.layers.15.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 264 |
+
"visual.vision_model.encoder.layers.15.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 265 |
+
"visual.vision_model.encoder.layers.15.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 266 |
+
"visual.vision_model.encoder.layers.15.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 267 |
+
"visual.vision_model.encoder.layers.15.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 268 |
+
"visual.vision_model.encoder.layers.15.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 269 |
+
"visual.vision_model.encoder.layers.15.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 270 |
+
"visual.vision_model.encoder.layers.15.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 271 |
+
"visual.vision_model.encoder.layers.16.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 272 |
+
"visual.vision_model.encoder.layers.16.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 273 |
+
"visual.vision_model.encoder.layers.16.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 274 |
+
"visual.vision_model.encoder.layers.16.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 275 |
+
"visual.vision_model.encoder.layers.16.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 276 |
+
"visual.vision_model.encoder.layers.16.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 277 |
+
"visual.vision_model.encoder.layers.16.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 278 |
+
"visual.vision_model.encoder.layers.16.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 279 |
+
"visual.vision_model.encoder.layers.16.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 280 |
+
"visual.vision_model.encoder.layers.16.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 281 |
+
"visual.vision_model.encoder.layers.16.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 282 |
+
"visual.vision_model.encoder.layers.16.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 283 |
+
"visual.vision_model.encoder.layers.16.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 284 |
+
"visual.vision_model.encoder.layers.16.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 285 |
+
"visual.vision_model.encoder.layers.16.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 286 |
+
"visual.vision_model.encoder.layers.16.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 287 |
+
"visual.vision_model.encoder.layers.17.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 288 |
+
"visual.vision_model.encoder.layers.17.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 289 |
+
"visual.vision_model.encoder.layers.17.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 290 |
+
"visual.vision_model.encoder.layers.17.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 291 |
+
"visual.vision_model.encoder.layers.17.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 292 |
+
"visual.vision_model.encoder.layers.17.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 293 |
+
"visual.vision_model.encoder.layers.17.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 294 |
+
"visual.vision_model.encoder.layers.17.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 295 |
+
"visual.vision_model.encoder.layers.17.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 296 |
+
"visual.vision_model.encoder.layers.17.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 297 |
+
"visual.vision_model.encoder.layers.17.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 298 |
+
"visual.vision_model.encoder.layers.17.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 299 |
+
"visual.vision_model.encoder.layers.17.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 300 |
+
"visual.vision_model.encoder.layers.17.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 301 |
+
"visual.vision_model.encoder.layers.17.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 302 |
+
"visual.vision_model.encoder.layers.17.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 303 |
+
"visual.vision_model.encoder.layers.18.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 304 |
+
"visual.vision_model.encoder.layers.18.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 305 |
+
"visual.vision_model.encoder.layers.18.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 306 |
+
"visual.vision_model.encoder.layers.18.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 307 |
+
"visual.vision_model.encoder.layers.18.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 308 |
+
"visual.vision_model.encoder.layers.18.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 309 |
+
"visual.vision_model.encoder.layers.18.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 310 |
+
"visual.vision_model.encoder.layers.18.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 311 |
+
"visual.vision_model.encoder.layers.18.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 312 |
+
"visual.vision_model.encoder.layers.18.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 313 |
+
"visual.vision_model.encoder.layers.18.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 314 |
+
"visual.vision_model.encoder.layers.18.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 315 |
+
"visual.vision_model.encoder.layers.18.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 316 |
+
"visual.vision_model.encoder.layers.18.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 317 |
+
"visual.vision_model.encoder.layers.18.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 318 |
+
"visual.vision_model.encoder.layers.18.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 319 |
+
"visual.vision_model.encoder.layers.19.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 320 |
+
"visual.vision_model.encoder.layers.19.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 321 |
+
"visual.vision_model.encoder.layers.19.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 322 |
+
"visual.vision_model.encoder.layers.19.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 323 |
+
"visual.vision_model.encoder.layers.19.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 324 |
+
"visual.vision_model.encoder.layers.19.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 325 |
+
"visual.vision_model.encoder.layers.19.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 326 |
+
"visual.vision_model.encoder.layers.19.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 327 |
+
"visual.vision_model.encoder.layers.19.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 328 |
+
"visual.vision_model.encoder.layers.19.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 329 |
+
"visual.vision_model.encoder.layers.19.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 330 |
+
"visual.vision_model.encoder.layers.19.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 331 |
+
"visual.vision_model.encoder.layers.19.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 332 |
+
"visual.vision_model.encoder.layers.19.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 333 |
+
"visual.vision_model.encoder.layers.19.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 334 |
+
"visual.vision_model.encoder.layers.19.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 335 |
+
"visual.vision_model.encoder.layers.20.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 336 |
+
"visual.vision_model.encoder.layers.20.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 337 |
+
"visual.vision_model.encoder.layers.20.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 338 |
+
"visual.vision_model.encoder.layers.20.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 339 |
+
"visual.vision_model.encoder.layers.20.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 340 |
+
"visual.vision_model.encoder.layers.20.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 341 |
+
"visual.vision_model.encoder.layers.20.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 342 |
+
"visual.vision_model.encoder.layers.20.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 343 |
+
"visual.vision_model.encoder.layers.20.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 344 |
+
"visual.vision_model.encoder.layers.20.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 345 |
+
"visual.vision_model.encoder.layers.20.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 346 |
+
"visual.vision_model.encoder.layers.20.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 347 |
+
"visual.vision_model.encoder.layers.20.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 348 |
+
"visual.vision_model.encoder.layers.20.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 349 |
+
"visual.vision_model.encoder.layers.20.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 350 |
+
"visual.vision_model.encoder.layers.20.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 351 |
+
"visual.vision_model.encoder.layers.21.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 352 |
+
"visual.vision_model.encoder.layers.21.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 353 |
+
"visual.vision_model.encoder.layers.21.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 354 |
+
"visual.vision_model.encoder.layers.21.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 355 |
+
"visual.vision_model.encoder.layers.21.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 356 |
+
"visual.vision_model.encoder.layers.21.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 357 |
+
"visual.vision_model.encoder.layers.21.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 358 |
+
"visual.vision_model.encoder.layers.21.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 359 |
+
"visual.vision_model.encoder.layers.21.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 360 |
+
"visual.vision_model.encoder.layers.21.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 361 |
+
"visual.vision_model.encoder.layers.21.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 362 |
+
"visual.vision_model.encoder.layers.21.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 363 |
+
"visual.vision_model.encoder.layers.21.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 364 |
+
"visual.vision_model.encoder.layers.21.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 365 |
+
"visual.vision_model.encoder.layers.21.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 366 |
+
"visual.vision_model.encoder.layers.21.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 367 |
+
"visual.vision_model.encoder.layers.22.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 368 |
+
"visual.vision_model.encoder.layers.22.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 369 |
+
"visual.vision_model.encoder.layers.22.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 370 |
+
"visual.vision_model.encoder.layers.22.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 371 |
+
"visual.vision_model.encoder.layers.22.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 372 |
+
"visual.vision_model.encoder.layers.22.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 373 |
+
"visual.vision_model.encoder.layers.22.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 374 |
+
"visual.vision_model.encoder.layers.22.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 375 |
+
"visual.vision_model.encoder.layers.22.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 376 |
+
"visual.vision_model.encoder.layers.22.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 377 |
+
"visual.vision_model.encoder.layers.22.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 378 |
+
"visual.vision_model.encoder.layers.22.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 379 |
+
"visual.vision_model.encoder.layers.22.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 380 |
+
"visual.vision_model.encoder.layers.22.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 381 |
+
"visual.vision_model.encoder.layers.22.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 382 |
+
"visual.vision_model.encoder.layers.22.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 383 |
+
"visual.vision_model.encoder.layers.23.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 384 |
+
"visual.vision_model.encoder.layers.23.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 385 |
+
"visual.vision_model.encoder.layers.23.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 386 |
+
"visual.vision_model.encoder.layers.23.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 387 |
+
"visual.vision_model.encoder.layers.23.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 388 |
+
"visual.vision_model.encoder.layers.23.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 389 |
+
"visual.vision_model.encoder.layers.23.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 390 |
+
"visual.vision_model.encoder.layers.23.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 391 |
+
"visual.vision_model.encoder.layers.23.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 392 |
+
"visual.vision_model.encoder.layers.23.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 393 |
+
"visual.vision_model.encoder.layers.23.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 394 |
+
"visual.vision_model.encoder.layers.23.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 395 |
+
"visual.vision_model.encoder.layers.23.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 396 |
+
"visual.vision_model.encoder.layers.23.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 397 |
+
"visual.vision_model.encoder.layers.23.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 398 |
+
"visual.vision_model.encoder.layers.23.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 399 |
+
"visual.vision_model.encoder.layers.24.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 400 |
+
"visual.vision_model.encoder.layers.24.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 401 |
+
"visual.vision_model.encoder.layers.24.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 402 |
+
"visual.vision_model.encoder.layers.24.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 403 |
+
"visual.vision_model.encoder.layers.24.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 404 |
+
"visual.vision_model.encoder.layers.24.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 405 |
+
"visual.vision_model.encoder.layers.24.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 406 |
+
"visual.vision_model.encoder.layers.24.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 407 |
+
"visual.vision_model.encoder.layers.24.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 408 |
+
"visual.vision_model.encoder.layers.24.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 409 |
+
"visual.vision_model.encoder.layers.24.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 410 |
+
"visual.vision_model.encoder.layers.24.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 411 |
+
"visual.vision_model.encoder.layers.24.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 412 |
+
"visual.vision_model.encoder.layers.24.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 413 |
+
"visual.vision_model.encoder.layers.24.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 414 |
+
"visual.vision_model.encoder.layers.24.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 415 |
+
"visual.vision_model.encoder.layers.25.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 416 |
+
"visual.vision_model.encoder.layers.25.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 417 |
+
"visual.vision_model.encoder.layers.25.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 418 |
+
"visual.vision_model.encoder.layers.25.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 419 |
+
"visual.vision_model.encoder.layers.25.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 420 |
+
"visual.vision_model.encoder.layers.25.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 421 |
+
"visual.vision_model.encoder.layers.25.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 422 |
+
"visual.vision_model.encoder.layers.25.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 423 |
+
"visual.vision_model.encoder.layers.25.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 424 |
+
"visual.vision_model.encoder.layers.25.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 425 |
+
"visual.vision_model.encoder.layers.25.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 426 |
+
"visual.vision_model.encoder.layers.25.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 427 |
+
"visual.vision_model.encoder.layers.25.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 428 |
+
"visual.vision_model.encoder.layers.25.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 429 |
+
"visual.vision_model.encoder.layers.25.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 430 |
+
"visual.vision_model.encoder.layers.25.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 431 |
+
"visual.vision_model.encoder.layers.26.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 432 |
+
"visual.vision_model.encoder.layers.26.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 433 |
+
"visual.vision_model.encoder.layers.26.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 434 |
+
"visual.vision_model.encoder.layers.26.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 435 |
+
"visual.vision_model.encoder.layers.26.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 436 |
+
"visual.vision_model.encoder.layers.26.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 437 |
+
"visual.vision_model.encoder.layers.26.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 438 |
+
"visual.vision_model.encoder.layers.26.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 439 |
+
"visual.vision_model.encoder.layers.26.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 440 |
+
"visual.vision_model.encoder.layers.26.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 441 |
+
"visual.vision_model.encoder.layers.26.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 442 |
+
"visual.vision_model.encoder.layers.26.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 443 |
+
"visual.vision_model.encoder.layers.26.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 444 |
+
"visual.vision_model.encoder.layers.26.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 445 |
+
"visual.vision_model.encoder.layers.26.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 446 |
+
"visual.vision_model.encoder.layers.26.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 447 |
+
"visual.vision_model.post_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 448 |
+
"visual.vision_model.post_layernorm.bias": "master_weights-00001-of-00001.safetensors",
|
| 449 |
+
"model.embed_tokens.weight": "master_weights-00001-of-00001.safetensors",
|
| 450 |
+
"model.layers.0.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 451 |
+
"model.layers.0.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 452 |
+
"model.layers.0.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 453 |
+
"model.layers.0.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 454 |
+
"model.layers.0.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 455 |
+
"model.layers.0.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 456 |
+
"model.layers.0.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 457 |
+
"model.layers.0.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 458 |
+
"model.layers.0.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 459 |
+
"model.layers.1.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 460 |
+
"model.layers.1.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 461 |
+
"model.layers.1.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 462 |
+
"model.layers.1.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 463 |
+
"model.layers.1.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 464 |
+
"model.layers.1.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 465 |
+
"model.layers.1.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 466 |
+
"model.layers.1.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 467 |
+
"model.layers.1.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 468 |
+
"model.layers.2.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 469 |
+
"model.layers.2.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 470 |
+
"model.layers.2.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 471 |
+
"model.layers.2.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 472 |
+
"model.layers.2.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 473 |
+
"model.layers.2.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 474 |
+
"model.layers.2.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 475 |
+
"model.layers.2.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 476 |
+
"model.layers.2.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 477 |
+
"model.layers.3.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 478 |
+
"model.layers.3.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 479 |
+
"model.layers.3.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 480 |
+
"model.layers.3.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 481 |
+
"model.layers.3.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 482 |
+
"model.layers.3.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 483 |
+
"model.layers.3.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 484 |
+
"model.layers.3.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 485 |
+
"model.layers.3.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 486 |
+
"model.layers.4.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 487 |
+
"model.layers.4.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 488 |
+
"model.layers.4.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 489 |
+
"model.layers.4.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 490 |
+
"model.layers.4.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 491 |
+
"model.layers.4.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 492 |
+
"model.layers.4.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 493 |
+
"model.layers.4.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 494 |
+
"model.layers.4.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 495 |
+
"model.layers.5.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 496 |
+
"model.layers.5.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 497 |
+
"model.layers.5.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 498 |
+
"model.layers.5.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 499 |
+
"model.layers.5.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 500 |
+
"model.layers.5.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 501 |
+
"model.layers.5.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 502 |
+
"model.layers.5.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 503 |
+
"model.layers.5.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 504 |
+
"model.layers.6.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 505 |
+
"model.layers.6.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 506 |
+
"model.layers.6.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 507 |
+
"model.layers.6.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 508 |
+
"model.layers.6.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 509 |
+
"model.layers.6.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 510 |
+
"model.layers.6.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 511 |
+
"model.layers.6.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 512 |
+
"model.layers.6.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 513 |
+
"model.layers.7.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 514 |
+
"model.layers.7.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 515 |
+
"model.layers.7.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 516 |
+
"model.layers.7.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 517 |
+
"model.layers.7.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 518 |
+
"model.layers.7.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 519 |
+
"model.layers.7.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 520 |
+
"model.layers.7.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 521 |
+
"model.layers.7.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 522 |
+
"model.layers.8.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 523 |
+
"model.layers.8.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 524 |
+
"model.layers.8.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 525 |
+
"model.layers.8.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 526 |
+
"model.layers.8.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 527 |
+
"model.layers.8.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 528 |
+
"model.layers.8.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 529 |
+
"model.layers.8.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 530 |
+
"model.layers.8.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 531 |
+
"model.layers.9.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 532 |
+
"model.layers.9.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 533 |
+
"model.layers.9.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 534 |
+
"model.layers.9.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 535 |
+
"model.layers.9.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 536 |
+
"model.layers.9.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 537 |
+
"model.layers.9.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 538 |
+
"model.layers.9.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 539 |
+
"model.layers.9.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 540 |
+
"model.layers.10.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 541 |
+
"model.layers.10.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 542 |
+
"model.layers.10.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 543 |
+
"model.layers.10.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 544 |
+
"model.layers.10.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 545 |
+
"model.layers.10.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 546 |
+
"model.layers.10.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 547 |
+
"model.layers.10.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 548 |
+
"model.layers.10.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 549 |
+
"model.layers.11.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 550 |
+
"model.layers.11.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 551 |
+
"model.layers.11.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 552 |
+
"model.layers.11.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 553 |
+
"model.layers.11.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 554 |
+
"model.layers.11.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 555 |
+
"model.layers.11.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 556 |
+
"model.layers.11.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 557 |
+
"model.layers.11.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 558 |
+
"model.layers.12.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 559 |
+
"model.layers.12.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 560 |
+
"model.layers.12.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 561 |
+
"model.layers.12.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 562 |
+
"model.layers.12.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 563 |
+
"model.layers.12.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 564 |
+
"model.layers.12.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 565 |
+
"model.layers.12.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 566 |
+
"model.layers.12.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 567 |
+
"model.layers.13.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 568 |
+
"model.layers.13.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 569 |
+
"model.layers.13.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 570 |
+
"model.layers.13.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 571 |
+
"model.layers.13.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 572 |
+
"model.layers.13.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 573 |
+
"model.layers.13.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 574 |
+
"model.layers.13.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 575 |
+
"model.layers.13.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 576 |
+
"model.layers.14.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 577 |
+
"model.layers.14.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 578 |
+
"model.layers.14.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 579 |
+
"model.layers.14.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 580 |
+
"model.layers.14.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 581 |
+
"model.layers.14.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 582 |
+
"model.layers.14.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 583 |
+
"model.layers.14.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 584 |
+
"model.layers.14.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 585 |
+
"model.layers.15.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 586 |
+
"model.layers.15.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 587 |
+
"model.layers.15.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 588 |
+
"model.layers.15.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 589 |
+
"model.layers.15.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 590 |
+
"model.layers.15.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 591 |
+
"model.layers.15.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 592 |
+
"model.layers.15.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 593 |
+
"model.layers.15.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 594 |
+
"model.layers.16.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 595 |
+
"model.layers.16.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 596 |
+
"model.layers.16.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 597 |
+
"model.layers.16.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 598 |
+
"model.layers.16.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 599 |
+
"model.layers.16.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 600 |
+
"model.layers.16.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 601 |
+
"model.layers.16.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 602 |
+
"model.layers.16.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 603 |
+
"model.layers.17.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 604 |
+
"model.layers.17.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 605 |
+
"model.layers.17.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 606 |
+
"model.layers.17.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 607 |
+
"model.layers.17.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 608 |
+
"model.layers.17.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 609 |
+
"model.layers.17.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 610 |
+
"model.layers.17.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 611 |
+
"model.layers.17.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 612 |
+
"model.norm.weight": "master_weights-00001-of-00001.safetensors",
|
| 613 |
+
"lm_head.weight": "master_weights-00001-of-00001.safetensors"
|
| 614 |
+
}
|
| 615 |
+
}
|
checkpoint-336/model-00001-of-00001.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8765aee223483d0bf21fd28e690f6de5a1e3668de260b7c197b99beac936adfd
|
| 3 |
+
size 1917255968
|
checkpoint-336/model.safetensors.index.json
ADDED
|
@@ -0,0 +1,627 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"metadata": {
|
| 3 |
+
"total_size": 1917177472
|
| 4 |
+
},
|
| 5 |
+
"weight_map": {
|
| 6 |
+
"mlp_AR.pre_norm.weight": "model-00001-of-00001.safetensors",
|
| 7 |
+
"mlp_AR.pre_norm.bias": "model-00001-of-00001.safetensors",
|
| 8 |
+
"mlp_AR.linear_1.weight": "model-00001-of-00001.safetensors",
|
| 9 |
+
"mlp_AR.linear_1.bias": "model-00001-of-00001.safetensors",
|
| 10 |
+
"mlp_AR.linear_2.weight": "model-00001-of-00001.safetensors",
|
| 11 |
+
"mlp_AR.linear_2.bias": "model-00001-of-00001.safetensors",
|
| 12 |
+
"visual.vision_model.embeddings.patch_embedding.weight": "model-00001-of-00001.safetensors",
|
| 13 |
+
"visual.vision_model.embeddings.patch_embedding.bias": "model-00001-of-00001.safetensors",
|
| 14 |
+
"visual.vision_model.embeddings.position_embedding.weight": "model-00001-of-00001.safetensors",
|
| 15 |
+
"visual.vision_model.embeddings.packing_position_embedding.weight": "model-00001-of-00001.safetensors",
|
| 16 |
+
"visual.vision_model.encoder.layers.0.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 17 |
+
"visual.vision_model.encoder.layers.0.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 18 |
+
"visual.vision_model.encoder.layers.0.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 19 |
+
"visual.vision_model.encoder.layers.0.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 20 |
+
"visual.vision_model.encoder.layers.0.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 21 |
+
"visual.vision_model.encoder.layers.0.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 22 |
+
"visual.vision_model.encoder.layers.0.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 23 |
+
"visual.vision_model.encoder.layers.0.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 24 |
+
"visual.vision_model.encoder.layers.0.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 25 |
+
"visual.vision_model.encoder.layers.0.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 26 |
+
"visual.vision_model.encoder.layers.0.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 27 |
+
"visual.vision_model.encoder.layers.0.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 28 |
+
"visual.vision_model.encoder.layers.0.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 29 |
+
"visual.vision_model.encoder.layers.0.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 30 |
+
"visual.vision_model.encoder.layers.0.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 31 |
+
"visual.vision_model.encoder.layers.0.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 32 |
+
"visual.vision_model.encoder.layers.1.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 33 |
+
"visual.vision_model.encoder.layers.1.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 34 |
+
"visual.vision_model.encoder.layers.1.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 35 |
+
"visual.vision_model.encoder.layers.1.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 36 |
+
"visual.vision_model.encoder.layers.1.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 37 |
+
"visual.vision_model.encoder.layers.1.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 38 |
+
"visual.vision_model.encoder.layers.1.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 39 |
+
"visual.vision_model.encoder.layers.1.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 40 |
+
"visual.vision_model.encoder.layers.1.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 41 |
+
"visual.vision_model.encoder.layers.1.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 42 |
+
"visual.vision_model.encoder.layers.1.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 43 |
+
"visual.vision_model.encoder.layers.1.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 44 |
+
"visual.vision_model.encoder.layers.1.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 45 |
+
"visual.vision_model.encoder.layers.1.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 46 |
+
"visual.vision_model.encoder.layers.1.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 47 |
+
"visual.vision_model.encoder.layers.1.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 48 |
+
"visual.vision_model.encoder.layers.2.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 49 |
+
"visual.vision_model.encoder.layers.2.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 50 |
+
"visual.vision_model.encoder.layers.2.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 51 |
+
"visual.vision_model.encoder.layers.2.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 52 |
+
"visual.vision_model.encoder.layers.2.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 53 |
+
"visual.vision_model.encoder.layers.2.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 54 |
+
"visual.vision_model.encoder.layers.2.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 55 |
+
"visual.vision_model.encoder.layers.2.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 56 |
+
"visual.vision_model.encoder.layers.2.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 57 |
+
"visual.vision_model.encoder.layers.2.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 58 |
+
"visual.vision_model.encoder.layers.2.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 59 |
+
"visual.vision_model.encoder.layers.2.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 60 |
+
"visual.vision_model.encoder.layers.2.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 61 |
+
"visual.vision_model.encoder.layers.2.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 62 |
+
"visual.vision_model.encoder.layers.2.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 63 |
+
"visual.vision_model.encoder.layers.2.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 64 |
+
"visual.vision_model.encoder.layers.3.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 65 |
+
"visual.vision_model.encoder.layers.3.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 66 |
+
"visual.vision_model.encoder.layers.3.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 67 |
+
"visual.vision_model.encoder.layers.3.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 68 |
+
"visual.vision_model.encoder.layers.3.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 69 |
+
"visual.vision_model.encoder.layers.3.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 70 |
+
"visual.vision_model.encoder.layers.3.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 71 |
+
"visual.vision_model.encoder.layers.3.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 72 |
+
"visual.vision_model.encoder.layers.3.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 73 |
+
"visual.vision_model.encoder.layers.3.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 74 |
+
"visual.vision_model.encoder.layers.3.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 75 |
+
"visual.vision_model.encoder.layers.3.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 76 |
+
"visual.vision_model.encoder.layers.3.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 77 |
+
"visual.vision_model.encoder.layers.3.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 78 |
+
"visual.vision_model.encoder.layers.3.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 79 |
+
"visual.vision_model.encoder.layers.3.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 80 |
+
"visual.vision_model.encoder.layers.4.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 81 |
+
"visual.vision_model.encoder.layers.4.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 82 |
+
"visual.vision_model.encoder.layers.4.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 83 |
+
"visual.vision_model.encoder.layers.4.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 84 |
+
"visual.vision_model.encoder.layers.4.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 85 |
+
"visual.vision_model.encoder.layers.4.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 86 |
+
"visual.vision_model.encoder.layers.4.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 87 |
+
"visual.vision_model.encoder.layers.4.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 88 |
+
"visual.vision_model.encoder.layers.4.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 89 |
+
"visual.vision_model.encoder.layers.4.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 90 |
+
"visual.vision_model.encoder.layers.4.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 91 |
+
"visual.vision_model.encoder.layers.4.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 92 |
+
"visual.vision_model.encoder.layers.4.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 93 |
+
"visual.vision_model.encoder.layers.4.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 94 |
+
"visual.vision_model.encoder.layers.4.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 95 |
+
"visual.vision_model.encoder.layers.4.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 96 |
+
"visual.vision_model.encoder.layers.5.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 97 |
+
"visual.vision_model.encoder.layers.5.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 98 |
+
"visual.vision_model.encoder.layers.5.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 99 |
+
"visual.vision_model.encoder.layers.5.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 100 |
+
"visual.vision_model.encoder.layers.5.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 101 |
+
"visual.vision_model.encoder.layers.5.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 102 |
+
"visual.vision_model.encoder.layers.5.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 103 |
+
"visual.vision_model.encoder.layers.5.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 104 |
+
"visual.vision_model.encoder.layers.5.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 105 |
+
"visual.vision_model.encoder.layers.5.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 106 |
+
"visual.vision_model.encoder.layers.5.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 107 |
+
"visual.vision_model.encoder.layers.5.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 108 |
+
"visual.vision_model.encoder.layers.5.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 109 |
+
"visual.vision_model.encoder.layers.5.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 110 |
+
"visual.vision_model.encoder.layers.5.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 111 |
+
"visual.vision_model.encoder.layers.5.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 112 |
+
"visual.vision_model.encoder.layers.6.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 113 |
+
"visual.vision_model.encoder.layers.6.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 114 |
+
"visual.vision_model.encoder.layers.6.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 115 |
+
"visual.vision_model.encoder.layers.6.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 116 |
+
"visual.vision_model.encoder.layers.6.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 117 |
+
"visual.vision_model.encoder.layers.6.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 118 |
+
"visual.vision_model.encoder.layers.6.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 119 |
+
"visual.vision_model.encoder.layers.6.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 120 |
+
"visual.vision_model.encoder.layers.6.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 121 |
+
"visual.vision_model.encoder.layers.6.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 122 |
+
"visual.vision_model.encoder.layers.6.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 123 |
+
"visual.vision_model.encoder.layers.6.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 124 |
+
"visual.vision_model.encoder.layers.6.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 125 |
+
"visual.vision_model.encoder.layers.6.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 126 |
+
"visual.vision_model.encoder.layers.6.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 127 |
+
"visual.vision_model.encoder.layers.6.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 128 |
+
"visual.vision_model.encoder.layers.7.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 129 |
+
"visual.vision_model.encoder.layers.7.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 130 |
+
"visual.vision_model.encoder.layers.7.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 131 |
+
"visual.vision_model.encoder.layers.7.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 132 |
+
"visual.vision_model.encoder.layers.7.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 133 |
+
"visual.vision_model.encoder.layers.7.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 134 |
+
"visual.vision_model.encoder.layers.7.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 135 |
+
"visual.vision_model.encoder.layers.7.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 136 |
+
"visual.vision_model.encoder.layers.7.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 137 |
+
"visual.vision_model.encoder.layers.7.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 138 |
+
"visual.vision_model.encoder.layers.7.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 139 |
+
"visual.vision_model.encoder.layers.7.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 140 |
+
"visual.vision_model.encoder.layers.7.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 141 |
+
"visual.vision_model.encoder.layers.7.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 142 |
+
"visual.vision_model.encoder.layers.7.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 143 |
+
"visual.vision_model.encoder.layers.7.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 144 |
+
"visual.vision_model.encoder.layers.8.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 145 |
+
"visual.vision_model.encoder.layers.8.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 146 |
+
"visual.vision_model.encoder.layers.8.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 147 |
+
"visual.vision_model.encoder.layers.8.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 148 |
+
"visual.vision_model.encoder.layers.8.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 149 |
+
"visual.vision_model.encoder.layers.8.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 150 |
+
"visual.vision_model.encoder.layers.8.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 151 |
+
"visual.vision_model.encoder.layers.8.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 152 |
+
"visual.vision_model.encoder.layers.8.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 153 |
+
"visual.vision_model.encoder.layers.8.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 154 |
+
"visual.vision_model.encoder.layers.8.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 155 |
+
"visual.vision_model.encoder.layers.8.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 156 |
+
"visual.vision_model.encoder.layers.8.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 157 |
+
"visual.vision_model.encoder.layers.8.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 158 |
+
"visual.vision_model.encoder.layers.8.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 159 |
+
"visual.vision_model.encoder.layers.8.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 160 |
+
"visual.vision_model.encoder.layers.9.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 161 |
+
"visual.vision_model.encoder.layers.9.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 162 |
+
"visual.vision_model.encoder.layers.9.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 163 |
+
"visual.vision_model.encoder.layers.9.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 164 |
+
"visual.vision_model.encoder.layers.9.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 165 |
+
"visual.vision_model.encoder.layers.9.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 166 |
+
"visual.vision_model.encoder.layers.9.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 167 |
+
"visual.vision_model.encoder.layers.9.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 168 |
+
"visual.vision_model.encoder.layers.9.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 169 |
+
"visual.vision_model.encoder.layers.9.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 170 |
+
"visual.vision_model.encoder.layers.9.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 171 |
+
"visual.vision_model.encoder.layers.9.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 172 |
+
"visual.vision_model.encoder.layers.9.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 173 |
+
"visual.vision_model.encoder.layers.9.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 174 |
+
"visual.vision_model.encoder.layers.9.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 175 |
+
"visual.vision_model.encoder.layers.9.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 176 |
+
"visual.vision_model.encoder.layers.10.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 177 |
+
"visual.vision_model.encoder.layers.10.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 178 |
+
"visual.vision_model.encoder.layers.10.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 179 |
+
"visual.vision_model.encoder.layers.10.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 180 |
+
"visual.vision_model.encoder.layers.10.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 181 |
+
"visual.vision_model.encoder.layers.10.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 182 |
+
"visual.vision_model.encoder.layers.10.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 183 |
+
"visual.vision_model.encoder.layers.10.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 184 |
+
"visual.vision_model.encoder.layers.10.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 185 |
+
"visual.vision_model.encoder.layers.10.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 186 |
+
"visual.vision_model.encoder.layers.10.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 187 |
+
"visual.vision_model.encoder.layers.10.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 188 |
+
"visual.vision_model.encoder.layers.10.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 189 |
+
"visual.vision_model.encoder.layers.10.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 190 |
+
"visual.vision_model.encoder.layers.10.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 191 |
+
"visual.vision_model.encoder.layers.10.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 192 |
+
"visual.vision_model.encoder.layers.11.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 193 |
+
"visual.vision_model.encoder.layers.11.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 194 |
+
"visual.vision_model.encoder.layers.11.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 195 |
+
"visual.vision_model.encoder.layers.11.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 196 |
+
"visual.vision_model.encoder.layers.11.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 197 |
+
"visual.vision_model.encoder.layers.11.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 198 |
+
"visual.vision_model.encoder.layers.11.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 199 |
+
"visual.vision_model.encoder.layers.11.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 200 |
+
"visual.vision_model.encoder.layers.11.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 201 |
+
"visual.vision_model.encoder.layers.11.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 202 |
+
"visual.vision_model.encoder.layers.11.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 203 |
+
"visual.vision_model.encoder.layers.11.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 204 |
+
"visual.vision_model.encoder.layers.11.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 205 |
+
"visual.vision_model.encoder.layers.11.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 206 |
+
"visual.vision_model.encoder.layers.11.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 207 |
+
"visual.vision_model.encoder.layers.11.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 208 |
+
"visual.vision_model.encoder.layers.12.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 209 |
+
"visual.vision_model.encoder.layers.12.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 210 |
+
"visual.vision_model.encoder.layers.12.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 211 |
+
"visual.vision_model.encoder.layers.12.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 212 |
+
"visual.vision_model.encoder.layers.12.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 213 |
+
"visual.vision_model.encoder.layers.12.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 214 |
+
"visual.vision_model.encoder.layers.12.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 215 |
+
"visual.vision_model.encoder.layers.12.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 216 |
+
"visual.vision_model.encoder.layers.12.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 217 |
+
"visual.vision_model.encoder.layers.12.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 218 |
+
"visual.vision_model.encoder.layers.12.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 219 |
+
"visual.vision_model.encoder.layers.12.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 220 |
+
"visual.vision_model.encoder.layers.12.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 221 |
+
"visual.vision_model.encoder.layers.12.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 222 |
+
"visual.vision_model.encoder.layers.12.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 223 |
+
"visual.vision_model.encoder.layers.12.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 224 |
+
"visual.vision_model.encoder.layers.13.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 225 |
+
"visual.vision_model.encoder.layers.13.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 226 |
+
"visual.vision_model.encoder.layers.13.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 227 |
+
"visual.vision_model.encoder.layers.13.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 228 |
+
"visual.vision_model.encoder.layers.13.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 229 |
+
"visual.vision_model.encoder.layers.13.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 230 |
+
"visual.vision_model.encoder.layers.13.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 231 |
+
"visual.vision_model.encoder.layers.13.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 232 |
+
"visual.vision_model.encoder.layers.13.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 233 |
+
"visual.vision_model.encoder.layers.13.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 234 |
+
"visual.vision_model.encoder.layers.13.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 235 |
+
"visual.vision_model.encoder.layers.13.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 236 |
+
"visual.vision_model.encoder.layers.13.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 237 |
+
"visual.vision_model.encoder.layers.13.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 238 |
+
"visual.vision_model.encoder.layers.13.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 239 |
+
"visual.vision_model.encoder.layers.13.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 240 |
+
"visual.vision_model.encoder.layers.14.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 241 |
+
"visual.vision_model.encoder.layers.14.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 242 |
+
"visual.vision_model.encoder.layers.14.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 243 |
+
"visual.vision_model.encoder.layers.14.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 244 |
+
"visual.vision_model.encoder.layers.14.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 245 |
+
"visual.vision_model.encoder.layers.14.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 246 |
+
"visual.vision_model.encoder.layers.14.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 247 |
+
"visual.vision_model.encoder.layers.14.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 248 |
+
"visual.vision_model.encoder.layers.14.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 249 |
+
"visual.vision_model.encoder.layers.14.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 250 |
+
"visual.vision_model.encoder.layers.14.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 251 |
+
"visual.vision_model.encoder.layers.14.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 252 |
+
"visual.vision_model.encoder.layers.14.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 253 |
+
"visual.vision_model.encoder.layers.14.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 254 |
+
"visual.vision_model.encoder.layers.14.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 255 |
+
"visual.vision_model.encoder.layers.14.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 256 |
+
"visual.vision_model.encoder.layers.15.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 257 |
+
"visual.vision_model.encoder.layers.15.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 258 |
+
"visual.vision_model.encoder.layers.15.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 259 |
+
"visual.vision_model.encoder.layers.15.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 260 |
+
"visual.vision_model.encoder.layers.15.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 261 |
+
"visual.vision_model.encoder.layers.15.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 262 |
+
"visual.vision_model.encoder.layers.15.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 263 |
+
"visual.vision_model.encoder.layers.15.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 264 |
+
"visual.vision_model.encoder.layers.15.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 265 |
+
"visual.vision_model.encoder.layers.15.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 266 |
+
"visual.vision_model.encoder.layers.15.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 267 |
+
"visual.vision_model.encoder.layers.15.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 268 |
+
"visual.vision_model.encoder.layers.15.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 269 |
+
"visual.vision_model.encoder.layers.15.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 270 |
+
"visual.vision_model.encoder.layers.15.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 271 |
+
"visual.vision_model.encoder.layers.15.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 272 |
+
"visual.vision_model.encoder.layers.16.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 273 |
+
"visual.vision_model.encoder.layers.16.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 274 |
+
"visual.vision_model.encoder.layers.16.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 275 |
+
"visual.vision_model.encoder.layers.16.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 276 |
+
"visual.vision_model.encoder.layers.16.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 277 |
+
"visual.vision_model.encoder.layers.16.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 278 |
+
"visual.vision_model.encoder.layers.16.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 279 |
+
"visual.vision_model.encoder.layers.16.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 280 |
+
"visual.vision_model.encoder.layers.16.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 281 |
+
"visual.vision_model.encoder.layers.16.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 282 |
+
"visual.vision_model.encoder.layers.16.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 283 |
+
"visual.vision_model.encoder.layers.16.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 284 |
+
"visual.vision_model.encoder.layers.16.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 285 |
+
"visual.vision_model.encoder.layers.16.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 286 |
+
"visual.vision_model.encoder.layers.16.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 287 |
+
"visual.vision_model.encoder.layers.16.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 288 |
+
"visual.vision_model.encoder.layers.17.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 289 |
+
"visual.vision_model.encoder.layers.17.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 290 |
+
"visual.vision_model.encoder.layers.17.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 291 |
+
"visual.vision_model.encoder.layers.17.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 292 |
+
"visual.vision_model.encoder.layers.17.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 293 |
+
"visual.vision_model.encoder.layers.17.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 294 |
+
"visual.vision_model.encoder.layers.17.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 295 |
+
"visual.vision_model.encoder.layers.17.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 296 |
+
"visual.vision_model.encoder.layers.17.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 297 |
+
"visual.vision_model.encoder.layers.17.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 298 |
+
"visual.vision_model.encoder.layers.17.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 299 |
+
"visual.vision_model.encoder.layers.17.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 300 |
+
"visual.vision_model.encoder.layers.17.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 301 |
+
"visual.vision_model.encoder.layers.17.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 302 |
+
"visual.vision_model.encoder.layers.17.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 303 |
+
"visual.vision_model.encoder.layers.17.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 304 |
+
"visual.vision_model.encoder.layers.18.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 305 |
+
"visual.vision_model.encoder.layers.18.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 306 |
+
"visual.vision_model.encoder.layers.18.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 307 |
+
"visual.vision_model.encoder.layers.18.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 308 |
+
"visual.vision_model.encoder.layers.18.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 309 |
+
"visual.vision_model.encoder.layers.18.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 310 |
+
"visual.vision_model.encoder.layers.18.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 311 |
+
"visual.vision_model.encoder.layers.18.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 312 |
+
"visual.vision_model.encoder.layers.18.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 313 |
+
"visual.vision_model.encoder.layers.18.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 314 |
+
"visual.vision_model.encoder.layers.18.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 315 |
+
"visual.vision_model.encoder.layers.18.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 316 |
+
"visual.vision_model.encoder.layers.18.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 317 |
+
"visual.vision_model.encoder.layers.18.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 318 |
+
"visual.vision_model.encoder.layers.18.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 319 |
+
"visual.vision_model.encoder.layers.18.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 320 |
+
"visual.vision_model.encoder.layers.19.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 321 |
+
"visual.vision_model.encoder.layers.19.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 322 |
+
"visual.vision_model.encoder.layers.19.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 323 |
+
"visual.vision_model.encoder.layers.19.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 324 |
+
"visual.vision_model.encoder.layers.19.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 325 |
+
"visual.vision_model.encoder.layers.19.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 326 |
+
"visual.vision_model.encoder.layers.19.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 327 |
+
"visual.vision_model.encoder.layers.19.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 328 |
+
"visual.vision_model.encoder.layers.19.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 329 |
+
"visual.vision_model.encoder.layers.19.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 330 |
+
"visual.vision_model.encoder.layers.19.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 331 |
+
"visual.vision_model.encoder.layers.19.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 332 |
+
"visual.vision_model.encoder.layers.19.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 333 |
+
"visual.vision_model.encoder.layers.19.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 334 |
+
"visual.vision_model.encoder.layers.19.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 335 |
+
"visual.vision_model.encoder.layers.19.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 336 |
+
"visual.vision_model.encoder.layers.20.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 337 |
+
"visual.vision_model.encoder.layers.20.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 338 |
+
"visual.vision_model.encoder.layers.20.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 339 |
+
"visual.vision_model.encoder.layers.20.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 340 |
+
"visual.vision_model.encoder.layers.20.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 341 |
+
"visual.vision_model.encoder.layers.20.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 342 |
+
"visual.vision_model.encoder.layers.20.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 343 |
+
"visual.vision_model.encoder.layers.20.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 344 |
+
"visual.vision_model.encoder.layers.20.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 345 |
+
"visual.vision_model.encoder.layers.20.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 346 |
+
"visual.vision_model.encoder.layers.20.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 347 |
+
"visual.vision_model.encoder.layers.20.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 348 |
+
"visual.vision_model.encoder.layers.20.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 349 |
+
"visual.vision_model.encoder.layers.20.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 350 |
+
"visual.vision_model.encoder.layers.20.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 351 |
+
"visual.vision_model.encoder.layers.20.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 352 |
+
"visual.vision_model.encoder.layers.21.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 353 |
+
"visual.vision_model.encoder.layers.21.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 354 |
+
"visual.vision_model.encoder.layers.21.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 355 |
+
"visual.vision_model.encoder.layers.21.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 356 |
+
"visual.vision_model.encoder.layers.21.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 357 |
+
"visual.vision_model.encoder.layers.21.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 358 |
+
"visual.vision_model.encoder.layers.21.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 359 |
+
"visual.vision_model.encoder.layers.21.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 360 |
+
"visual.vision_model.encoder.layers.21.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 361 |
+
"visual.vision_model.encoder.layers.21.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 362 |
+
"visual.vision_model.encoder.layers.21.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 363 |
+
"visual.vision_model.encoder.layers.21.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 364 |
+
"visual.vision_model.encoder.layers.21.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 365 |
+
"visual.vision_model.encoder.layers.21.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 366 |
+
"visual.vision_model.encoder.layers.21.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 367 |
+
"visual.vision_model.encoder.layers.21.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 368 |
+
"visual.vision_model.encoder.layers.22.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 369 |
+
"visual.vision_model.encoder.layers.22.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 370 |
+
"visual.vision_model.encoder.layers.22.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 371 |
+
"visual.vision_model.encoder.layers.22.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 372 |
+
"visual.vision_model.encoder.layers.22.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 373 |
+
"visual.vision_model.encoder.layers.22.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 374 |
+
"visual.vision_model.encoder.layers.22.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 375 |
+
"visual.vision_model.encoder.layers.22.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 376 |
+
"visual.vision_model.encoder.layers.22.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 377 |
+
"visual.vision_model.encoder.layers.22.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 378 |
+
"visual.vision_model.encoder.layers.22.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 379 |
+
"visual.vision_model.encoder.layers.22.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 380 |
+
"visual.vision_model.encoder.layers.22.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 381 |
+
"visual.vision_model.encoder.layers.22.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 382 |
+
"visual.vision_model.encoder.layers.22.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 383 |
+
"visual.vision_model.encoder.layers.22.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 384 |
+
"visual.vision_model.encoder.layers.23.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 385 |
+
"visual.vision_model.encoder.layers.23.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 386 |
+
"visual.vision_model.encoder.layers.23.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 387 |
+
"visual.vision_model.encoder.layers.23.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 388 |
+
"visual.vision_model.encoder.layers.23.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 389 |
+
"visual.vision_model.encoder.layers.23.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 390 |
+
"visual.vision_model.encoder.layers.23.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 391 |
+
"visual.vision_model.encoder.layers.23.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 392 |
+
"visual.vision_model.encoder.layers.23.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 393 |
+
"visual.vision_model.encoder.layers.23.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 394 |
+
"visual.vision_model.encoder.layers.23.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 395 |
+
"visual.vision_model.encoder.layers.23.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 396 |
+
"visual.vision_model.encoder.layers.23.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 397 |
+
"visual.vision_model.encoder.layers.23.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 398 |
+
"visual.vision_model.encoder.layers.23.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 399 |
+
"visual.vision_model.encoder.layers.23.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 400 |
+
"visual.vision_model.encoder.layers.24.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 401 |
+
"visual.vision_model.encoder.layers.24.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 402 |
+
"visual.vision_model.encoder.layers.24.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 403 |
+
"visual.vision_model.encoder.layers.24.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 404 |
+
"visual.vision_model.encoder.layers.24.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 405 |
+
"visual.vision_model.encoder.layers.24.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 406 |
+
"visual.vision_model.encoder.layers.24.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 407 |
+
"visual.vision_model.encoder.layers.24.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 408 |
+
"visual.vision_model.encoder.layers.24.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 409 |
+
"visual.vision_model.encoder.layers.24.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 410 |
+
"visual.vision_model.encoder.layers.24.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 411 |
+
"visual.vision_model.encoder.layers.24.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 412 |
+
"visual.vision_model.encoder.layers.24.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 413 |
+
"visual.vision_model.encoder.layers.24.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 414 |
+
"visual.vision_model.encoder.layers.24.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 415 |
+
"visual.vision_model.encoder.layers.24.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 416 |
+
"visual.vision_model.encoder.layers.25.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 417 |
+
"visual.vision_model.encoder.layers.25.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 418 |
+
"visual.vision_model.encoder.layers.25.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 419 |
+
"visual.vision_model.encoder.layers.25.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 420 |
+
"visual.vision_model.encoder.layers.25.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 421 |
+
"visual.vision_model.encoder.layers.25.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 422 |
+
"visual.vision_model.encoder.layers.25.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 423 |
+
"visual.vision_model.encoder.layers.25.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 424 |
+
"visual.vision_model.encoder.layers.25.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 425 |
+
"visual.vision_model.encoder.layers.25.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 426 |
+
"visual.vision_model.encoder.layers.25.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 427 |
+
"visual.vision_model.encoder.layers.25.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 428 |
+
"visual.vision_model.encoder.layers.25.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 429 |
+
"visual.vision_model.encoder.layers.25.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 430 |
+
"visual.vision_model.encoder.layers.25.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 431 |
+
"visual.vision_model.encoder.layers.25.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 432 |
+
"visual.vision_model.encoder.layers.26.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 433 |
+
"visual.vision_model.encoder.layers.26.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 434 |
+
"visual.vision_model.encoder.layers.26.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 435 |
+
"visual.vision_model.encoder.layers.26.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 436 |
+
"visual.vision_model.encoder.layers.26.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 437 |
+
"visual.vision_model.encoder.layers.26.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 438 |
+
"visual.vision_model.encoder.layers.26.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 439 |
+
"visual.vision_model.encoder.layers.26.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 440 |
+
"visual.vision_model.encoder.layers.26.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 441 |
+
"visual.vision_model.encoder.layers.26.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 442 |
+
"visual.vision_model.encoder.layers.26.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 443 |
+
"visual.vision_model.encoder.layers.26.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 444 |
+
"visual.vision_model.encoder.layers.26.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 445 |
+
"visual.vision_model.encoder.layers.26.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 446 |
+
"visual.vision_model.encoder.layers.26.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 447 |
+
"visual.vision_model.encoder.layers.26.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 448 |
+
"visual.vision_model.post_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 449 |
+
"visual.vision_model.post_layernorm.bias": "model-00001-of-00001.safetensors",
|
| 450 |
+
"visual.vision_model.head.probe": "model-00001-of-00001.safetensors",
|
| 451 |
+
"visual.vision_model.head.attention.in_proj_weight": "model-00001-of-00001.safetensors",
|
| 452 |
+
"visual.vision_model.head.attention.in_proj_bias": "model-00001-of-00001.safetensors",
|
| 453 |
+
"visual.vision_model.head.attention.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 454 |
+
"visual.vision_model.head.attention.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 455 |
+
"visual.vision_model.head.layernorm.weight": "model-00001-of-00001.safetensors",
|
| 456 |
+
"visual.vision_model.head.layernorm.bias": "model-00001-of-00001.safetensors",
|
| 457 |
+
"visual.vision_model.head.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 458 |
+
"visual.vision_model.head.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 459 |
+
"visual.vision_model.head.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 460 |
+
"visual.vision_model.head.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 461 |
+
"model.embed_tokens.weight": "model-00001-of-00001.safetensors",
|
| 462 |
+
"model.layers.0.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 463 |
+
"model.layers.0.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 464 |
+
"model.layers.0.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 465 |
+
"model.layers.0.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 466 |
+
"model.layers.0.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 467 |
+
"model.layers.0.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 468 |
+
"model.layers.0.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 469 |
+
"model.layers.0.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 470 |
+
"model.layers.0.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 471 |
+
"model.layers.1.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 472 |
+
"model.layers.1.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 473 |
+
"model.layers.1.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 474 |
+
"model.layers.1.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 475 |
+
"model.layers.1.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 476 |
+
"model.layers.1.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 477 |
+
"model.layers.1.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 478 |
+
"model.layers.1.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 479 |
+
"model.layers.1.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 480 |
+
"model.layers.2.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 481 |
+
"model.layers.2.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 482 |
+
"model.layers.2.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 483 |
+
"model.layers.2.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 484 |
+
"model.layers.2.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 485 |
+
"model.layers.2.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 486 |
+
"model.layers.2.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 487 |
+
"model.layers.2.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 488 |
+
"model.layers.2.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 489 |
+
"model.layers.3.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 490 |
+
"model.layers.3.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 491 |
+
"model.layers.3.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 492 |
+
"model.layers.3.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 493 |
+
"model.layers.3.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 494 |
+
"model.layers.3.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 495 |
+
"model.layers.3.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 496 |
+
"model.layers.3.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 497 |
+
"model.layers.3.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 498 |
+
"model.layers.4.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 499 |
+
"model.layers.4.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 500 |
+
"model.layers.4.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 501 |
+
"model.layers.4.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 502 |
+
"model.layers.4.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 503 |
+
"model.layers.4.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 504 |
+
"model.layers.4.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 505 |
+
"model.layers.4.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 506 |
+
"model.layers.4.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 507 |
+
"model.layers.5.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 508 |
+
"model.layers.5.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 509 |
+
"model.layers.5.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 510 |
+
"model.layers.5.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 511 |
+
"model.layers.5.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 512 |
+
"model.layers.5.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 513 |
+
"model.layers.5.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 514 |
+
"model.layers.5.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 515 |
+
"model.layers.5.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 516 |
+
"model.layers.6.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 517 |
+
"model.layers.6.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 518 |
+
"model.layers.6.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 519 |
+
"model.layers.6.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 520 |
+
"model.layers.6.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 521 |
+
"model.layers.6.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 522 |
+
"model.layers.6.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 523 |
+
"model.layers.6.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 524 |
+
"model.layers.6.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 525 |
+
"model.layers.7.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 526 |
+
"model.layers.7.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 527 |
+
"model.layers.7.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 528 |
+
"model.layers.7.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 529 |
+
"model.layers.7.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 530 |
+
"model.layers.7.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 531 |
+
"model.layers.7.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 532 |
+
"model.layers.7.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 533 |
+
"model.layers.7.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 534 |
+
"model.layers.8.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 535 |
+
"model.layers.8.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 536 |
+
"model.layers.8.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 537 |
+
"model.layers.8.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 538 |
+
"model.layers.8.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 539 |
+
"model.layers.8.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 540 |
+
"model.layers.8.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 541 |
+
"model.layers.8.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 542 |
+
"model.layers.8.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 543 |
+
"model.layers.9.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 544 |
+
"model.layers.9.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 545 |
+
"model.layers.9.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 546 |
+
"model.layers.9.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 547 |
+
"model.layers.9.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 548 |
+
"model.layers.9.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 549 |
+
"model.layers.9.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 550 |
+
"model.layers.9.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 551 |
+
"model.layers.9.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 552 |
+
"model.layers.10.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 553 |
+
"model.layers.10.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 554 |
+
"model.layers.10.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 555 |
+
"model.layers.10.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 556 |
+
"model.layers.10.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 557 |
+
"model.layers.10.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 558 |
+
"model.layers.10.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 559 |
+
"model.layers.10.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 560 |
+
"model.layers.10.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 561 |
+
"model.layers.11.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 562 |
+
"model.layers.11.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 563 |
+
"model.layers.11.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 564 |
+
"model.layers.11.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 565 |
+
"model.layers.11.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 566 |
+
"model.layers.11.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 567 |
+
"model.layers.11.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 568 |
+
"model.layers.11.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 569 |
+
"model.layers.11.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 570 |
+
"model.layers.12.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 571 |
+
"model.layers.12.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 572 |
+
"model.layers.12.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 573 |
+
"model.layers.12.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 574 |
+
"model.layers.12.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 575 |
+
"model.layers.12.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 576 |
+
"model.layers.12.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 577 |
+
"model.layers.12.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 578 |
+
"model.layers.12.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 579 |
+
"model.layers.13.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 580 |
+
"model.layers.13.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 581 |
+
"model.layers.13.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 582 |
+
"model.layers.13.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 583 |
+
"model.layers.13.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 584 |
+
"model.layers.13.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 585 |
+
"model.layers.13.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 586 |
+
"model.layers.13.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 587 |
+
"model.layers.13.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 588 |
+
"model.layers.14.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 589 |
+
"model.layers.14.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 590 |
+
"model.layers.14.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 591 |
+
"model.layers.14.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 592 |
+
"model.layers.14.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 593 |
+
"model.layers.14.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 594 |
+
"model.layers.14.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 595 |
+
"model.layers.14.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 596 |
+
"model.layers.14.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 597 |
+
"model.layers.15.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 598 |
+
"model.layers.15.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 599 |
+
"model.layers.15.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 600 |
+
"model.layers.15.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 601 |
+
"model.layers.15.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 602 |
+
"model.layers.15.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 603 |
+
"model.layers.15.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 604 |
+
"model.layers.15.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 605 |
+
"model.layers.15.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 606 |
+
"model.layers.16.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 607 |
+
"model.layers.16.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 608 |
+
"model.layers.16.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 609 |
+
"model.layers.16.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 610 |
+
"model.layers.16.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 611 |
+
"model.layers.16.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 612 |
+
"model.layers.16.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 613 |
+
"model.layers.16.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 614 |
+
"model.layers.16.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 615 |
+
"model.layers.17.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 616 |
+
"model.layers.17.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 617 |
+
"model.layers.17.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 618 |
+
"model.layers.17.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 619 |
+
"model.layers.17.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 620 |
+
"model.layers.17.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 621 |
+
"model.layers.17.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 622 |
+
"model.layers.17.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 623 |
+
"model.layers.17.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 624 |
+
"model.norm.weight": "model-00001-of-00001.safetensors",
|
| 625 |
+
"lm_head.weight": "model-00001-of-00001.safetensors"
|
| 626 |
+
}
|
| 627 |
+
}
|
checkpoint-336/optimizer-00001-of-00001.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:09a0535d83eb6a76d79e74bd6c9f2b380e6d5b1d4f8499f0a666f0d82260720a
|
| 3 |
+
size 7245150152
|
checkpoint-336/optimizer.safetensors.index.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
checkpoint-336/preprocessor_config.json
ADDED
|
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"do_convert_rgb": true,
|
| 3 |
+
"do_normalize": true,
|
| 4 |
+
"do_rescale": true,
|
| 5 |
+
"do_resize": true,
|
| 6 |
+
"image_mean": [
|
| 7 |
+
0.5,
|
| 8 |
+
0.5,
|
| 9 |
+
0.5
|
| 10 |
+
],
|
| 11 |
+
"image_processor_type": "SiglipImageProcessor",
|
| 12 |
+
"image_std": [
|
| 13 |
+
0.5,
|
| 14 |
+
0.5,
|
| 15 |
+
0.5
|
| 16 |
+
],
|
| 17 |
+
"max_pixels": 2822400,
|
| 18 |
+
"merge_size": 2,
|
| 19 |
+
"min_pixels": 147384,
|
| 20 |
+
"patch_size": 14,
|
| 21 |
+
"resample": 3,
|
| 22 |
+
"rescale_factor": 0.00392156862745098,
|
| 23 |
+
"size": {
|
| 24 |
+
"max_pixels": 2822400,
|
| 25 |
+
"min_pixels": 147384
|
| 26 |
+
},
|
| 27 |
+
"temporal_conv_size": 1,
|
| 28 |
+
"temporal_patch_size": 1
|
| 29 |
+
}
|
checkpoint-336/rng_state_0.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:29cde6478b49d2682dfa8c033ed7da4f33ab76ccd1941cfe950453d3923dfec3
|
| 3 |
+
size 32008
|
checkpoint-336/scheduler.pdparams
ADDED
|
Binary file (51 Bytes). View file
|
|
|
checkpoint-336/special_tokens_map.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"bos_token": {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": false, "special": false}, "eos_token": "</s>", "unk_token": {"content": "<unk>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": false, "special": false}, "sep_token": "<|end_of_sentence|>", "pad_token": {"content": "<unk>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": false, "special": false}, "cls_token": "<|begin_of_sentence|>", "mask_token": {"content": "<mask:1>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": false, "special": false}, "additional_special_tokens": ["<|IMAGE_PLACEHOLDER|>", "<|image_pad|>", "<|IMAGE_START|>", "<|IMAGE_END|>", "<|video_pad|>"]}
|
checkpoint-336/static_name_to_dyg_name.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"layer_norm_0.w_0": "mlp_AR.pre_norm.weight", "layer_norm_0.b_0": "mlp_AR.pre_norm.bias", "linear_0.w_0": "mlp_AR.linear_1.weight", "linear_0.b_0": "mlp_AR.linear_1.bias", "linear_1.w_0": "mlp_AR.linear_2.weight", "linear_1.b_0": "mlp_AR.linear_2.bias", "conv2d_0.w_0": "visual.vision_model.embeddings.patch_embedding.weight", "conv2d_0.b_0": "visual.vision_model.embeddings.patch_embedding.bias", "embedding_0.w_0": "visual.vision_model.embeddings.position_embedding.weight", "embedding_1.w_0": "visual.vision_model.embeddings.packing_position_embedding.weight", "layer_norm_1.w_0": "visual.vision_model.encoder.layers.0.layer_norm1.weight", "layer_norm_1.b_0": "visual.vision_model.encoder.layers.0.layer_norm1.bias", "linear_2.w_0": "visual.vision_model.encoder.layers.0.self_attn.k_proj.weight", "linear_2.b_0": "visual.vision_model.encoder.layers.0.self_attn.k_proj.bias", "linear_3.w_0": "visual.vision_model.encoder.layers.0.self_attn.v_proj.weight", "linear_3.b_0": "visual.vision_model.encoder.layers.0.self_attn.v_proj.bias", "linear_4.w_0": "visual.vision_model.encoder.layers.0.self_attn.q_proj.weight", "linear_4.b_0": "visual.vision_model.encoder.layers.0.self_attn.q_proj.bias", "linear_5.w_0": "visual.vision_model.encoder.layers.0.self_attn.out_proj.weight", "linear_5.b_0": "visual.vision_model.encoder.layers.0.self_attn.out_proj.bias", "layer_norm_2.w_0": "visual.vision_model.encoder.layers.0.layer_norm2.weight", "layer_norm_2.b_0": "visual.vision_model.encoder.layers.0.layer_norm2.bias", "linear_6.w_0": "visual.vision_model.encoder.layers.0.mlp.fc1.weight", "linear_6.b_0": "visual.vision_model.encoder.layers.0.mlp.fc1.bias", "linear_7.w_0": "visual.vision_model.encoder.layers.0.mlp.fc2.weight", "linear_7.b_0": "visual.vision_model.encoder.layers.0.mlp.fc2.bias", "layer_norm_3.w_0": "visual.vision_model.encoder.layers.1.layer_norm1.weight", "layer_norm_3.b_0": "visual.vision_model.encoder.layers.1.layer_norm1.bias", "linear_8.w_0": "visual.vision_model.encoder.layers.1.self_attn.k_proj.weight", "linear_8.b_0": "visual.vision_model.encoder.layers.1.self_attn.k_proj.bias", "linear_9.w_0": "visual.vision_model.encoder.layers.1.self_attn.v_proj.weight", "linear_9.b_0": "visual.vision_model.encoder.layers.1.self_attn.v_proj.bias", "linear_10.w_0": "visual.vision_model.encoder.layers.1.self_attn.q_proj.weight", "linear_10.b_0": "visual.vision_model.encoder.layers.1.self_attn.q_proj.bias", "linear_11.w_0": "visual.vision_model.encoder.layers.1.self_attn.out_proj.weight", "linear_11.b_0": "visual.vision_model.encoder.layers.1.self_attn.out_proj.bias", "layer_norm_4.w_0": "visual.vision_model.encoder.layers.1.layer_norm2.weight", "layer_norm_4.b_0": "visual.vision_model.encoder.layers.1.layer_norm2.bias", "linear_12.w_0": "visual.vision_model.encoder.layers.1.mlp.fc1.weight", "linear_12.b_0": "visual.vision_model.encoder.layers.1.mlp.fc1.bias", "linear_13.w_0": "visual.vision_model.encoder.layers.1.mlp.fc2.weight", "linear_13.b_0": "visual.vision_model.encoder.layers.1.mlp.fc2.bias", "layer_norm_5.w_0": "visual.vision_model.encoder.layers.2.layer_norm1.weight", "layer_norm_5.b_0": "visual.vision_model.encoder.layers.2.layer_norm1.bias", "linear_14.w_0": "visual.vision_model.encoder.layers.2.self_attn.k_proj.weight", "linear_14.b_0": "visual.vision_model.encoder.layers.2.self_attn.k_proj.bias", "linear_15.w_0": "visual.vision_model.encoder.layers.2.self_attn.v_proj.weight", "linear_15.b_0": "visual.vision_model.encoder.layers.2.self_attn.v_proj.bias", "linear_16.w_0": "visual.vision_model.encoder.layers.2.self_attn.q_proj.weight", "linear_16.b_0": "visual.vision_model.encoder.layers.2.self_attn.q_proj.bias", "linear_17.w_0": "visual.vision_model.encoder.layers.2.self_attn.out_proj.weight", "linear_17.b_0": "visual.vision_model.encoder.layers.2.self_attn.out_proj.bias", "layer_norm_6.w_0": "visual.vision_model.encoder.layers.2.layer_norm2.weight", "layer_norm_6.b_0": "visual.vision_model.encoder.layers.2.layer_norm2.bias", "linear_18.w_0": "visual.vision_model.encoder.layers.2.mlp.fc1.weight", "linear_18.b_0": "visual.vision_model.encoder.layers.2.mlp.fc1.bias", "linear_19.w_0": "visual.vision_model.encoder.layers.2.mlp.fc2.weight", "linear_19.b_0": "visual.vision_model.encoder.layers.2.mlp.fc2.bias", "layer_norm_7.w_0": "visual.vision_model.encoder.layers.3.layer_norm1.weight", "layer_norm_7.b_0": "visual.vision_model.encoder.layers.3.layer_norm1.bias", "linear_20.w_0": "visual.vision_model.encoder.layers.3.self_attn.k_proj.weight", "linear_20.b_0": "visual.vision_model.encoder.layers.3.self_attn.k_proj.bias", "linear_21.w_0": "visual.vision_model.encoder.layers.3.self_attn.v_proj.weight", "linear_21.b_0": "visual.vision_model.encoder.layers.3.self_attn.v_proj.bias", "linear_22.w_0": "visual.vision_model.encoder.layers.3.self_attn.q_proj.weight", "linear_22.b_0": "visual.vision_model.encoder.layers.3.self_attn.q_proj.bias", "linear_23.w_0": "visual.vision_model.encoder.layers.3.self_attn.out_proj.weight", "linear_23.b_0": "visual.vision_model.encoder.layers.3.self_attn.out_proj.bias", "layer_norm_8.w_0": "visual.vision_model.encoder.layers.3.layer_norm2.weight", "layer_norm_8.b_0": "visual.vision_model.encoder.layers.3.layer_norm2.bias", "linear_24.w_0": "visual.vision_model.encoder.layers.3.mlp.fc1.weight", "linear_24.b_0": "visual.vision_model.encoder.layers.3.mlp.fc1.bias", "linear_25.w_0": "visual.vision_model.encoder.layers.3.mlp.fc2.weight", "linear_25.b_0": "visual.vision_model.encoder.layers.3.mlp.fc2.bias", "layer_norm_9.w_0": "visual.vision_model.encoder.layers.4.layer_norm1.weight", "layer_norm_9.b_0": "visual.vision_model.encoder.layers.4.layer_norm1.bias", "linear_26.w_0": "visual.vision_model.encoder.layers.4.self_attn.k_proj.weight", "linear_26.b_0": "visual.vision_model.encoder.layers.4.self_attn.k_proj.bias", "linear_27.w_0": "visual.vision_model.encoder.layers.4.self_attn.v_proj.weight", "linear_27.b_0": "visual.vision_model.encoder.layers.4.self_attn.v_proj.bias", "linear_28.w_0": "visual.vision_model.encoder.layers.4.self_attn.q_proj.weight", "linear_28.b_0": "visual.vision_model.encoder.layers.4.self_attn.q_proj.bias", "linear_29.w_0": "visual.vision_model.encoder.layers.4.self_attn.out_proj.weight", "linear_29.b_0": "visual.vision_model.encoder.layers.4.self_attn.out_proj.bias", "layer_norm_10.w_0": "visual.vision_model.encoder.layers.4.layer_norm2.weight", "layer_norm_10.b_0": "visual.vision_model.encoder.layers.4.layer_norm2.bias", "linear_30.w_0": "visual.vision_model.encoder.layers.4.mlp.fc1.weight", "linear_30.b_0": "visual.vision_model.encoder.layers.4.mlp.fc1.bias", "linear_31.w_0": "visual.vision_model.encoder.layers.4.mlp.fc2.weight", "linear_31.b_0": "visual.vision_model.encoder.layers.4.mlp.fc2.bias", "layer_norm_11.w_0": "visual.vision_model.encoder.layers.5.layer_norm1.weight", "layer_norm_11.b_0": "visual.vision_model.encoder.layers.5.layer_norm1.bias", "linear_32.w_0": "visual.vision_model.encoder.layers.5.self_attn.k_proj.weight", "linear_32.b_0": "visual.vision_model.encoder.layers.5.self_attn.k_proj.bias", "linear_33.w_0": "visual.vision_model.encoder.layers.5.self_attn.v_proj.weight", "linear_33.b_0": "visual.vision_model.encoder.layers.5.self_attn.v_proj.bias", "linear_34.w_0": "visual.vision_model.encoder.layers.5.self_attn.q_proj.weight", "linear_34.b_0": "visual.vision_model.encoder.layers.5.self_attn.q_proj.bias", "linear_35.w_0": "visual.vision_model.encoder.layers.5.self_attn.out_proj.weight", "linear_35.b_0": "visual.vision_model.encoder.layers.5.self_attn.out_proj.bias", "layer_norm_12.w_0": "visual.vision_model.encoder.layers.5.layer_norm2.weight", "layer_norm_12.b_0": "visual.vision_model.encoder.layers.5.layer_norm2.bias", "linear_36.w_0": "visual.vision_model.encoder.layers.5.mlp.fc1.weight", "linear_36.b_0": "visual.vision_model.encoder.layers.5.mlp.fc1.bias", "linear_37.w_0": "visual.vision_model.encoder.layers.5.mlp.fc2.weight", "linear_37.b_0": "visual.vision_model.encoder.layers.5.mlp.fc2.bias", "layer_norm_13.w_0": "visual.vision_model.encoder.layers.6.layer_norm1.weight", "layer_norm_13.b_0": "visual.vision_model.encoder.layers.6.layer_norm1.bias", "linear_38.w_0": "visual.vision_model.encoder.layers.6.self_attn.k_proj.weight", "linear_38.b_0": "visual.vision_model.encoder.layers.6.self_attn.k_proj.bias", "linear_39.w_0": "visual.vision_model.encoder.layers.6.self_attn.v_proj.weight", "linear_39.b_0": "visual.vision_model.encoder.layers.6.self_attn.v_proj.bias", "linear_40.w_0": "visual.vision_model.encoder.layers.6.self_attn.q_proj.weight", "linear_40.b_0": "visual.vision_model.encoder.layers.6.self_attn.q_proj.bias", "linear_41.w_0": "visual.vision_model.encoder.layers.6.self_attn.out_proj.weight", "linear_41.b_0": "visual.vision_model.encoder.layers.6.self_attn.out_proj.bias", "layer_norm_14.w_0": "visual.vision_model.encoder.layers.6.layer_norm2.weight", "layer_norm_14.b_0": "visual.vision_model.encoder.layers.6.layer_norm2.bias", "linear_42.w_0": "visual.vision_model.encoder.layers.6.mlp.fc1.weight", "linear_42.b_0": "visual.vision_model.encoder.layers.6.mlp.fc1.bias", "linear_43.w_0": "visual.vision_model.encoder.layers.6.mlp.fc2.weight", "linear_43.b_0": "visual.vision_model.encoder.layers.6.mlp.fc2.bias", "layer_norm_15.w_0": "visual.vision_model.encoder.layers.7.layer_norm1.weight", "layer_norm_15.b_0": "visual.vision_model.encoder.layers.7.layer_norm1.bias", "linear_44.w_0": "visual.vision_model.encoder.layers.7.self_attn.k_proj.weight", "linear_44.b_0": "visual.vision_model.encoder.layers.7.self_attn.k_proj.bias", "linear_45.w_0": "visual.vision_model.encoder.layers.7.self_attn.v_proj.weight", "linear_45.b_0": "visual.vision_model.encoder.layers.7.self_attn.v_proj.bias", "linear_46.w_0": "visual.vision_model.encoder.layers.7.self_attn.q_proj.weight", "linear_46.b_0": "visual.vision_model.encoder.layers.7.self_attn.q_proj.bias", "linear_47.w_0": "visual.vision_model.encoder.layers.7.self_attn.out_proj.weight", "linear_47.b_0": "visual.vision_model.encoder.layers.7.self_attn.out_proj.bias", "layer_norm_16.w_0": "visual.vision_model.encoder.layers.7.layer_norm2.weight", "layer_norm_16.b_0": "visual.vision_model.encoder.layers.7.layer_norm2.bias", "linear_48.w_0": "visual.vision_model.encoder.layers.7.mlp.fc1.weight", "linear_48.b_0": "visual.vision_model.encoder.layers.7.mlp.fc1.bias", "linear_49.w_0": "visual.vision_model.encoder.layers.7.mlp.fc2.weight", "linear_49.b_0": "visual.vision_model.encoder.layers.7.mlp.fc2.bias", "layer_norm_17.w_0": "visual.vision_model.encoder.layers.8.layer_norm1.weight", "layer_norm_17.b_0": "visual.vision_model.encoder.layers.8.layer_norm1.bias", "linear_50.w_0": "visual.vision_model.encoder.layers.8.self_attn.k_proj.weight", "linear_50.b_0": "visual.vision_model.encoder.layers.8.self_attn.k_proj.bias", "linear_51.w_0": "visual.vision_model.encoder.layers.8.self_attn.v_proj.weight", "linear_51.b_0": "visual.vision_model.encoder.layers.8.self_attn.v_proj.bias", "linear_52.w_0": "visual.vision_model.encoder.layers.8.self_attn.q_proj.weight", "linear_52.b_0": "visual.vision_model.encoder.layers.8.self_attn.q_proj.bias", "linear_53.w_0": "visual.vision_model.encoder.layers.8.self_attn.out_proj.weight", "linear_53.b_0": "visual.vision_model.encoder.layers.8.self_attn.out_proj.bias", "layer_norm_18.w_0": "visual.vision_model.encoder.layers.8.layer_norm2.weight", "layer_norm_18.b_0": "visual.vision_model.encoder.layers.8.layer_norm2.bias", "linear_54.w_0": "visual.vision_model.encoder.layers.8.mlp.fc1.weight", "linear_54.b_0": "visual.vision_model.encoder.layers.8.mlp.fc1.bias", "linear_55.w_0": "visual.vision_model.encoder.layers.8.mlp.fc2.weight", "linear_55.b_0": "visual.vision_model.encoder.layers.8.mlp.fc2.bias", "layer_norm_19.w_0": "visual.vision_model.encoder.layers.9.layer_norm1.weight", "layer_norm_19.b_0": "visual.vision_model.encoder.layers.9.layer_norm1.bias", "linear_56.w_0": "visual.vision_model.encoder.layers.9.self_attn.k_proj.weight", "linear_56.b_0": "visual.vision_model.encoder.layers.9.self_attn.k_proj.bias", "linear_57.w_0": "visual.vision_model.encoder.layers.9.self_attn.v_proj.weight", "linear_57.b_0": "visual.vision_model.encoder.layers.9.self_attn.v_proj.bias", "linear_58.w_0": "visual.vision_model.encoder.layers.9.self_attn.q_proj.weight", "linear_58.b_0": "visual.vision_model.encoder.layers.9.self_attn.q_proj.bias", "linear_59.w_0": "visual.vision_model.encoder.layers.9.self_attn.out_proj.weight", "linear_59.b_0": "visual.vision_model.encoder.layers.9.self_attn.out_proj.bias", "layer_norm_20.w_0": "visual.vision_model.encoder.layers.9.layer_norm2.weight", "layer_norm_20.b_0": "visual.vision_model.encoder.layers.9.layer_norm2.bias", "linear_60.w_0": "visual.vision_model.encoder.layers.9.mlp.fc1.weight", "linear_60.b_0": "visual.vision_model.encoder.layers.9.mlp.fc1.bias", "linear_61.w_0": "visual.vision_model.encoder.layers.9.mlp.fc2.weight", "linear_61.b_0": "visual.vision_model.encoder.layers.9.mlp.fc2.bias", "layer_norm_21.w_0": "visual.vision_model.encoder.layers.10.layer_norm1.weight", "layer_norm_21.b_0": "visual.vision_model.encoder.layers.10.layer_norm1.bias", "linear_62.w_0": "visual.vision_model.encoder.layers.10.self_attn.k_proj.weight", "linear_62.b_0": "visual.vision_model.encoder.layers.10.self_attn.k_proj.bias", "linear_63.w_0": "visual.vision_model.encoder.layers.10.self_attn.v_proj.weight", "linear_63.b_0": "visual.vision_model.encoder.layers.10.self_attn.v_proj.bias", "linear_64.w_0": "visual.vision_model.encoder.layers.10.self_attn.q_proj.weight", "linear_64.b_0": "visual.vision_model.encoder.layers.10.self_attn.q_proj.bias", "linear_65.w_0": "visual.vision_model.encoder.layers.10.self_attn.out_proj.weight", "linear_65.b_0": "visual.vision_model.encoder.layers.10.self_attn.out_proj.bias", "layer_norm_22.w_0": "visual.vision_model.encoder.layers.10.layer_norm2.weight", "layer_norm_22.b_0": "visual.vision_model.encoder.layers.10.layer_norm2.bias", "linear_66.w_0": "visual.vision_model.encoder.layers.10.mlp.fc1.weight", "linear_66.b_0": "visual.vision_model.encoder.layers.10.mlp.fc1.bias", "linear_67.w_0": "visual.vision_model.encoder.layers.10.mlp.fc2.weight", "linear_67.b_0": "visual.vision_model.encoder.layers.10.mlp.fc2.bias", "layer_norm_23.w_0": "visual.vision_model.encoder.layers.11.layer_norm1.weight", "layer_norm_23.b_0": "visual.vision_model.encoder.layers.11.layer_norm1.bias", "linear_68.w_0": "visual.vision_model.encoder.layers.11.self_attn.k_proj.weight", "linear_68.b_0": "visual.vision_model.encoder.layers.11.self_attn.k_proj.bias", "linear_69.w_0": "visual.vision_model.encoder.layers.11.self_attn.v_proj.weight", "linear_69.b_0": "visual.vision_model.encoder.layers.11.self_attn.v_proj.bias", "linear_70.w_0": "visual.vision_model.encoder.layers.11.self_attn.q_proj.weight", "linear_70.b_0": "visual.vision_model.encoder.layers.11.self_attn.q_proj.bias", "linear_71.w_0": "visual.vision_model.encoder.layers.11.self_attn.out_proj.weight", "linear_71.b_0": "visual.vision_model.encoder.layers.11.self_attn.out_proj.bias", "layer_norm_24.w_0": "visual.vision_model.encoder.layers.11.layer_norm2.weight", "layer_norm_24.b_0": "visual.vision_model.encoder.layers.11.layer_norm2.bias", "linear_72.w_0": "visual.vision_model.encoder.layers.11.mlp.fc1.weight", "linear_72.b_0": "visual.vision_model.encoder.layers.11.mlp.fc1.bias", "linear_73.w_0": "visual.vision_model.encoder.layers.11.mlp.fc2.weight", "linear_73.b_0": "visual.vision_model.encoder.layers.11.mlp.fc2.bias", "layer_norm_25.w_0": "visual.vision_model.encoder.layers.12.layer_norm1.weight", "layer_norm_25.b_0": "visual.vision_model.encoder.layers.12.layer_norm1.bias", "linear_74.w_0": "visual.vision_model.encoder.layers.12.self_attn.k_proj.weight", "linear_74.b_0": "visual.vision_model.encoder.layers.12.self_attn.k_proj.bias", "linear_75.w_0": "visual.vision_model.encoder.layers.12.self_attn.v_proj.weight", "linear_75.b_0": "visual.vision_model.encoder.layers.12.self_attn.v_proj.bias", "linear_76.w_0": "visual.vision_model.encoder.layers.12.self_attn.q_proj.weight", "linear_76.b_0": "visual.vision_model.encoder.layers.12.self_attn.q_proj.bias", "linear_77.w_0": "visual.vision_model.encoder.layers.12.self_attn.out_proj.weight", "linear_77.b_0": "visual.vision_model.encoder.layers.12.self_attn.out_proj.bias", "layer_norm_26.w_0": "visual.vision_model.encoder.layers.12.layer_norm2.weight", "layer_norm_26.b_0": "visual.vision_model.encoder.layers.12.layer_norm2.bias", "linear_78.w_0": "visual.vision_model.encoder.layers.12.mlp.fc1.weight", "linear_78.b_0": "visual.vision_model.encoder.layers.12.mlp.fc1.bias", "linear_79.w_0": "visual.vision_model.encoder.layers.12.mlp.fc2.weight", "linear_79.b_0": "visual.vision_model.encoder.layers.12.mlp.fc2.bias", "layer_norm_27.w_0": "visual.vision_model.encoder.layers.13.layer_norm1.weight", "layer_norm_27.b_0": "visual.vision_model.encoder.layers.13.layer_norm1.bias", "linear_80.w_0": "visual.vision_model.encoder.layers.13.self_attn.k_proj.weight", "linear_80.b_0": "visual.vision_model.encoder.layers.13.self_attn.k_proj.bias", "linear_81.w_0": "visual.vision_model.encoder.layers.13.self_attn.v_proj.weight", "linear_81.b_0": "visual.vision_model.encoder.layers.13.self_attn.v_proj.bias", "linear_82.w_0": "visual.vision_model.encoder.layers.13.self_attn.q_proj.weight", "linear_82.b_0": "visual.vision_model.encoder.layers.13.self_attn.q_proj.bias", "linear_83.w_0": "visual.vision_model.encoder.layers.13.self_attn.out_proj.weight", "linear_83.b_0": "visual.vision_model.encoder.layers.13.self_attn.out_proj.bias", "layer_norm_28.w_0": "visual.vision_model.encoder.layers.13.layer_norm2.weight", "layer_norm_28.b_0": "visual.vision_model.encoder.layers.13.layer_norm2.bias", "linear_84.w_0": "visual.vision_model.encoder.layers.13.mlp.fc1.weight", "linear_84.b_0": "visual.vision_model.encoder.layers.13.mlp.fc1.bias", "linear_85.w_0": "visual.vision_model.encoder.layers.13.mlp.fc2.weight", "linear_85.b_0": "visual.vision_model.encoder.layers.13.mlp.fc2.bias", "layer_norm_29.w_0": "visual.vision_model.encoder.layers.14.layer_norm1.weight", "layer_norm_29.b_0": "visual.vision_model.encoder.layers.14.layer_norm1.bias", "linear_86.w_0": "visual.vision_model.encoder.layers.14.self_attn.k_proj.weight", "linear_86.b_0": "visual.vision_model.encoder.layers.14.self_attn.k_proj.bias", "linear_87.w_0": "visual.vision_model.encoder.layers.14.self_attn.v_proj.weight", "linear_87.b_0": "visual.vision_model.encoder.layers.14.self_attn.v_proj.bias", "linear_88.w_0": "visual.vision_model.encoder.layers.14.self_attn.q_proj.weight", "linear_88.b_0": "visual.vision_model.encoder.layers.14.self_attn.q_proj.bias", "linear_89.w_0": "visual.vision_model.encoder.layers.14.self_attn.out_proj.weight", "linear_89.b_0": "visual.vision_model.encoder.layers.14.self_attn.out_proj.bias", "layer_norm_30.w_0": "visual.vision_model.encoder.layers.14.layer_norm2.weight", "layer_norm_30.b_0": "visual.vision_model.encoder.layers.14.layer_norm2.bias", "linear_90.w_0": "visual.vision_model.encoder.layers.14.mlp.fc1.weight", "linear_90.b_0": "visual.vision_model.encoder.layers.14.mlp.fc1.bias", "linear_91.w_0": "visual.vision_model.encoder.layers.14.mlp.fc2.weight", "linear_91.b_0": "visual.vision_model.encoder.layers.14.mlp.fc2.bias", "layer_norm_31.w_0": "visual.vision_model.encoder.layers.15.layer_norm1.weight", "layer_norm_31.b_0": "visual.vision_model.encoder.layers.15.layer_norm1.bias", "linear_92.w_0": "visual.vision_model.encoder.layers.15.self_attn.k_proj.weight", "linear_92.b_0": "visual.vision_model.encoder.layers.15.self_attn.k_proj.bias", "linear_93.w_0": "visual.vision_model.encoder.layers.15.self_attn.v_proj.weight", "linear_93.b_0": "visual.vision_model.encoder.layers.15.self_attn.v_proj.bias", "linear_94.w_0": "visual.vision_model.encoder.layers.15.self_attn.q_proj.weight", "linear_94.b_0": "visual.vision_model.encoder.layers.15.self_attn.q_proj.bias", "linear_95.w_0": "visual.vision_model.encoder.layers.15.self_attn.out_proj.weight", "linear_95.b_0": "visual.vision_model.encoder.layers.15.self_attn.out_proj.bias", "layer_norm_32.w_0": "visual.vision_model.encoder.layers.15.layer_norm2.weight", "layer_norm_32.b_0": "visual.vision_model.encoder.layers.15.layer_norm2.bias", "linear_96.w_0": "visual.vision_model.encoder.layers.15.mlp.fc1.weight", "linear_96.b_0": "visual.vision_model.encoder.layers.15.mlp.fc1.bias", "linear_97.w_0": "visual.vision_model.encoder.layers.15.mlp.fc2.weight", "linear_97.b_0": "visual.vision_model.encoder.layers.15.mlp.fc2.bias", "layer_norm_33.w_0": "visual.vision_model.encoder.layers.16.layer_norm1.weight", "layer_norm_33.b_0": "visual.vision_model.encoder.layers.16.layer_norm1.bias", "linear_98.w_0": "visual.vision_model.encoder.layers.16.self_attn.k_proj.weight", "linear_98.b_0": "visual.vision_model.encoder.layers.16.self_attn.k_proj.bias", "linear_99.w_0": "visual.vision_model.encoder.layers.16.self_attn.v_proj.weight", "linear_99.b_0": "visual.vision_model.encoder.layers.16.self_attn.v_proj.bias", "linear_100.w_0": "visual.vision_model.encoder.layers.16.self_attn.q_proj.weight", "linear_100.b_0": "visual.vision_model.encoder.layers.16.self_attn.q_proj.bias", "linear_101.w_0": "visual.vision_model.encoder.layers.16.self_attn.out_proj.weight", "linear_101.b_0": "visual.vision_model.encoder.layers.16.self_attn.out_proj.bias", "layer_norm_34.w_0": "visual.vision_model.encoder.layers.16.layer_norm2.weight", "layer_norm_34.b_0": "visual.vision_model.encoder.layers.16.layer_norm2.bias", "linear_102.w_0": "visual.vision_model.encoder.layers.16.mlp.fc1.weight", "linear_102.b_0": "visual.vision_model.encoder.layers.16.mlp.fc1.bias", "linear_103.w_0": "visual.vision_model.encoder.layers.16.mlp.fc2.weight", "linear_103.b_0": "visual.vision_model.encoder.layers.16.mlp.fc2.bias", "layer_norm_35.w_0": "visual.vision_model.encoder.layers.17.layer_norm1.weight", "layer_norm_35.b_0": "visual.vision_model.encoder.layers.17.layer_norm1.bias", "linear_104.w_0": "visual.vision_model.encoder.layers.17.self_attn.k_proj.weight", "linear_104.b_0": "visual.vision_model.encoder.layers.17.self_attn.k_proj.bias", "linear_105.w_0": "visual.vision_model.encoder.layers.17.self_attn.v_proj.weight", "linear_105.b_0": "visual.vision_model.encoder.layers.17.self_attn.v_proj.bias", "linear_106.w_0": "visual.vision_model.encoder.layers.17.self_attn.q_proj.weight", "linear_106.b_0": "visual.vision_model.encoder.layers.17.self_attn.q_proj.bias", "linear_107.w_0": "visual.vision_model.encoder.layers.17.self_attn.out_proj.weight", "linear_107.b_0": "visual.vision_model.encoder.layers.17.self_attn.out_proj.bias", "layer_norm_36.w_0": "visual.vision_model.encoder.layers.17.layer_norm2.weight", "layer_norm_36.b_0": "visual.vision_model.encoder.layers.17.layer_norm2.bias", "linear_108.w_0": "visual.vision_model.encoder.layers.17.mlp.fc1.weight", "linear_108.b_0": "visual.vision_model.encoder.layers.17.mlp.fc1.bias", "linear_109.w_0": "visual.vision_model.encoder.layers.17.mlp.fc2.weight", "linear_109.b_0": "visual.vision_model.encoder.layers.17.mlp.fc2.bias", "layer_norm_37.w_0": "visual.vision_model.encoder.layers.18.layer_norm1.weight", "layer_norm_37.b_0": "visual.vision_model.encoder.layers.18.layer_norm1.bias", "linear_110.w_0": "visual.vision_model.encoder.layers.18.self_attn.k_proj.weight", "linear_110.b_0": "visual.vision_model.encoder.layers.18.self_attn.k_proj.bias", "linear_111.w_0": "visual.vision_model.encoder.layers.18.self_attn.v_proj.weight", "linear_111.b_0": "visual.vision_model.encoder.layers.18.self_attn.v_proj.bias", "linear_112.w_0": "visual.vision_model.encoder.layers.18.self_attn.q_proj.weight", "linear_112.b_0": "visual.vision_model.encoder.layers.18.self_attn.q_proj.bias", "linear_113.w_0": "visual.vision_model.encoder.layers.18.self_attn.out_proj.weight", "linear_113.b_0": "visual.vision_model.encoder.layers.18.self_attn.out_proj.bias", "layer_norm_38.w_0": "visual.vision_model.encoder.layers.18.layer_norm2.weight", "layer_norm_38.b_0": "visual.vision_model.encoder.layers.18.layer_norm2.bias", "linear_114.w_0": "visual.vision_model.encoder.layers.18.mlp.fc1.weight", "linear_114.b_0": "visual.vision_model.encoder.layers.18.mlp.fc1.bias", "linear_115.w_0": "visual.vision_model.encoder.layers.18.mlp.fc2.weight", "linear_115.b_0": "visual.vision_model.encoder.layers.18.mlp.fc2.bias", "layer_norm_39.w_0": "visual.vision_model.encoder.layers.19.layer_norm1.weight", "layer_norm_39.b_0": "visual.vision_model.encoder.layers.19.layer_norm1.bias", "linear_116.w_0": "visual.vision_model.encoder.layers.19.self_attn.k_proj.weight", "linear_116.b_0": "visual.vision_model.encoder.layers.19.self_attn.k_proj.bias", "linear_117.w_0": "visual.vision_model.encoder.layers.19.self_attn.v_proj.weight", "linear_117.b_0": "visual.vision_model.encoder.layers.19.self_attn.v_proj.bias", "linear_118.w_0": "visual.vision_model.encoder.layers.19.self_attn.q_proj.weight", "linear_118.b_0": "visual.vision_model.encoder.layers.19.self_attn.q_proj.bias", "linear_119.w_0": "visual.vision_model.encoder.layers.19.self_attn.out_proj.weight", "linear_119.b_0": "visual.vision_model.encoder.layers.19.self_attn.out_proj.bias", "layer_norm_40.w_0": "visual.vision_model.encoder.layers.19.layer_norm2.weight", "layer_norm_40.b_0": "visual.vision_model.encoder.layers.19.layer_norm2.bias", "linear_120.w_0": "visual.vision_model.encoder.layers.19.mlp.fc1.weight", "linear_120.b_0": "visual.vision_model.encoder.layers.19.mlp.fc1.bias", "linear_121.w_0": "visual.vision_model.encoder.layers.19.mlp.fc2.weight", "linear_121.b_0": "visual.vision_model.encoder.layers.19.mlp.fc2.bias", "layer_norm_41.w_0": "visual.vision_model.encoder.layers.20.layer_norm1.weight", "layer_norm_41.b_0": "visual.vision_model.encoder.layers.20.layer_norm1.bias", "linear_122.w_0": "visual.vision_model.encoder.layers.20.self_attn.k_proj.weight", "linear_122.b_0": "visual.vision_model.encoder.layers.20.self_attn.k_proj.bias", "linear_123.w_0": "visual.vision_model.encoder.layers.20.self_attn.v_proj.weight", "linear_123.b_0": "visual.vision_model.encoder.layers.20.self_attn.v_proj.bias", "linear_124.w_0": "visual.vision_model.encoder.layers.20.self_attn.q_proj.weight", "linear_124.b_0": "visual.vision_model.encoder.layers.20.self_attn.q_proj.bias", "linear_125.w_0": "visual.vision_model.encoder.layers.20.self_attn.out_proj.weight", "linear_125.b_0": "visual.vision_model.encoder.layers.20.self_attn.out_proj.bias", "layer_norm_42.w_0": "visual.vision_model.encoder.layers.20.layer_norm2.weight", "layer_norm_42.b_0": "visual.vision_model.encoder.layers.20.layer_norm2.bias", "linear_126.w_0": "visual.vision_model.encoder.layers.20.mlp.fc1.weight", "linear_126.b_0": "visual.vision_model.encoder.layers.20.mlp.fc1.bias", "linear_127.w_0": "visual.vision_model.encoder.layers.20.mlp.fc2.weight", "linear_127.b_0": "visual.vision_model.encoder.layers.20.mlp.fc2.bias", "layer_norm_43.w_0": "visual.vision_model.encoder.layers.21.layer_norm1.weight", "layer_norm_43.b_0": "visual.vision_model.encoder.layers.21.layer_norm1.bias", "linear_128.w_0": "visual.vision_model.encoder.layers.21.self_attn.k_proj.weight", "linear_128.b_0": "visual.vision_model.encoder.layers.21.self_attn.k_proj.bias", "linear_129.w_0": "visual.vision_model.encoder.layers.21.self_attn.v_proj.weight", "linear_129.b_0": "visual.vision_model.encoder.layers.21.self_attn.v_proj.bias", "linear_130.w_0": "visual.vision_model.encoder.layers.21.self_attn.q_proj.weight", "linear_130.b_0": "visual.vision_model.encoder.layers.21.self_attn.q_proj.bias", "linear_131.w_0": "visual.vision_model.encoder.layers.21.self_attn.out_proj.weight", "linear_131.b_0": "visual.vision_model.encoder.layers.21.self_attn.out_proj.bias", "layer_norm_44.w_0": "visual.vision_model.encoder.layers.21.layer_norm2.weight", "layer_norm_44.b_0": "visual.vision_model.encoder.layers.21.layer_norm2.bias", "linear_132.w_0": "visual.vision_model.encoder.layers.21.mlp.fc1.weight", "linear_132.b_0": "visual.vision_model.encoder.layers.21.mlp.fc1.bias", "linear_133.w_0": "visual.vision_model.encoder.layers.21.mlp.fc2.weight", "linear_133.b_0": "visual.vision_model.encoder.layers.21.mlp.fc2.bias", "layer_norm_45.w_0": "visual.vision_model.encoder.layers.22.layer_norm1.weight", "layer_norm_45.b_0": "visual.vision_model.encoder.layers.22.layer_norm1.bias", "linear_134.w_0": "visual.vision_model.encoder.layers.22.self_attn.k_proj.weight", "linear_134.b_0": "visual.vision_model.encoder.layers.22.self_attn.k_proj.bias", "linear_135.w_0": "visual.vision_model.encoder.layers.22.self_attn.v_proj.weight", "linear_135.b_0": "visual.vision_model.encoder.layers.22.self_attn.v_proj.bias", "linear_136.w_0": "visual.vision_model.encoder.layers.22.self_attn.q_proj.weight", "linear_136.b_0": "visual.vision_model.encoder.layers.22.self_attn.q_proj.bias", "linear_137.w_0": "visual.vision_model.encoder.layers.22.self_attn.out_proj.weight", "linear_137.b_0": "visual.vision_model.encoder.layers.22.self_attn.out_proj.bias", "layer_norm_46.w_0": "visual.vision_model.encoder.layers.22.layer_norm2.weight", "layer_norm_46.b_0": "visual.vision_model.encoder.layers.22.layer_norm2.bias", "linear_138.w_0": "visual.vision_model.encoder.layers.22.mlp.fc1.weight", "linear_138.b_0": "visual.vision_model.encoder.layers.22.mlp.fc1.bias", "linear_139.w_0": "visual.vision_model.encoder.layers.22.mlp.fc2.weight", "linear_139.b_0": "visual.vision_model.encoder.layers.22.mlp.fc2.bias", "layer_norm_47.w_0": "visual.vision_model.encoder.layers.23.layer_norm1.weight", "layer_norm_47.b_0": "visual.vision_model.encoder.layers.23.layer_norm1.bias", "linear_140.w_0": "visual.vision_model.encoder.layers.23.self_attn.k_proj.weight", "linear_140.b_0": "visual.vision_model.encoder.layers.23.self_attn.k_proj.bias", "linear_141.w_0": "visual.vision_model.encoder.layers.23.self_attn.v_proj.weight", "linear_141.b_0": "visual.vision_model.encoder.layers.23.self_attn.v_proj.bias", "linear_142.w_0": "visual.vision_model.encoder.layers.23.self_attn.q_proj.weight", "linear_142.b_0": "visual.vision_model.encoder.layers.23.self_attn.q_proj.bias", "linear_143.w_0": "visual.vision_model.encoder.layers.23.self_attn.out_proj.weight", "linear_143.b_0": "visual.vision_model.encoder.layers.23.self_attn.out_proj.bias", "layer_norm_48.w_0": "visual.vision_model.encoder.layers.23.layer_norm2.weight", "layer_norm_48.b_0": "visual.vision_model.encoder.layers.23.layer_norm2.bias", "linear_144.w_0": "visual.vision_model.encoder.layers.23.mlp.fc1.weight", "linear_144.b_0": "visual.vision_model.encoder.layers.23.mlp.fc1.bias", "linear_145.w_0": "visual.vision_model.encoder.layers.23.mlp.fc2.weight", "linear_145.b_0": "visual.vision_model.encoder.layers.23.mlp.fc2.bias", "layer_norm_49.w_0": "visual.vision_model.encoder.layers.24.layer_norm1.weight", "layer_norm_49.b_0": "visual.vision_model.encoder.layers.24.layer_norm1.bias", "linear_146.w_0": "visual.vision_model.encoder.layers.24.self_attn.k_proj.weight", "linear_146.b_0": "visual.vision_model.encoder.layers.24.self_attn.k_proj.bias", "linear_147.w_0": "visual.vision_model.encoder.layers.24.self_attn.v_proj.weight", "linear_147.b_0": "visual.vision_model.encoder.layers.24.self_attn.v_proj.bias", "linear_148.w_0": "visual.vision_model.encoder.layers.24.self_attn.q_proj.weight", "linear_148.b_0": "visual.vision_model.encoder.layers.24.self_attn.q_proj.bias", "linear_149.w_0": "visual.vision_model.encoder.layers.24.self_attn.out_proj.weight", "linear_149.b_0": "visual.vision_model.encoder.layers.24.self_attn.out_proj.bias", "layer_norm_50.w_0": "visual.vision_model.encoder.layers.24.layer_norm2.weight", "layer_norm_50.b_0": "visual.vision_model.encoder.layers.24.layer_norm2.bias", "linear_150.w_0": "visual.vision_model.encoder.layers.24.mlp.fc1.weight", "linear_150.b_0": "visual.vision_model.encoder.layers.24.mlp.fc1.bias", "linear_151.w_0": "visual.vision_model.encoder.layers.24.mlp.fc2.weight", "linear_151.b_0": "visual.vision_model.encoder.layers.24.mlp.fc2.bias", "layer_norm_51.w_0": "visual.vision_model.encoder.layers.25.layer_norm1.weight", "layer_norm_51.b_0": "visual.vision_model.encoder.layers.25.layer_norm1.bias", "linear_152.w_0": "visual.vision_model.encoder.layers.25.self_attn.k_proj.weight", "linear_152.b_0": "visual.vision_model.encoder.layers.25.self_attn.k_proj.bias", "linear_153.w_0": "visual.vision_model.encoder.layers.25.self_attn.v_proj.weight", "linear_153.b_0": "visual.vision_model.encoder.layers.25.self_attn.v_proj.bias", "linear_154.w_0": "visual.vision_model.encoder.layers.25.self_attn.q_proj.weight", "linear_154.b_0": "visual.vision_model.encoder.layers.25.self_attn.q_proj.bias", "linear_155.w_0": "visual.vision_model.encoder.layers.25.self_attn.out_proj.weight", "linear_155.b_0": "visual.vision_model.encoder.layers.25.self_attn.out_proj.bias", "layer_norm_52.w_0": "visual.vision_model.encoder.layers.25.layer_norm2.weight", "layer_norm_52.b_0": "visual.vision_model.encoder.layers.25.layer_norm2.bias", "linear_156.w_0": "visual.vision_model.encoder.layers.25.mlp.fc1.weight", "linear_156.b_0": "visual.vision_model.encoder.layers.25.mlp.fc1.bias", "linear_157.w_0": "visual.vision_model.encoder.layers.25.mlp.fc2.weight", "linear_157.b_0": "visual.vision_model.encoder.layers.25.mlp.fc2.bias", "layer_norm_53.w_0": "visual.vision_model.encoder.layers.26.layer_norm1.weight", "layer_norm_53.b_0": "visual.vision_model.encoder.layers.26.layer_norm1.bias", "linear_158.w_0": "visual.vision_model.encoder.layers.26.self_attn.k_proj.weight", "linear_158.b_0": "visual.vision_model.encoder.layers.26.self_attn.k_proj.bias", "linear_159.w_0": "visual.vision_model.encoder.layers.26.self_attn.v_proj.weight", "linear_159.b_0": "visual.vision_model.encoder.layers.26.self_attn.v_proj.bias", "linear_160.w_0": "visual.vision_model.encoder.layers.26.self_attn.q_proj.weight", "linear_160.b_0": "visual.vision_model.encoder.layers.26.self_attn.q_proj.bias", "linear_161.w_0": "visual.vision_model.encoder.layers.26.self_attn.out_proj.weight", "linear_161.b_0": "visual.vision_model.encoder.layers.26.self_attn.out_proj.bias", "layer_norm_54.w_0": "visual.vision_model.encoder.layers.26.layer_norm2.weight", "layer_norm_54.b_0": "visual.vision_model.encoder.layers.26.layer_norm2.bias", "linear_162.w_0": "visual.vision_model.encoder.layers.26.mlp.fc1.weight", "linear_162.b_0": "visual.vision_model.encoder.layers.26.mlp.fc1.bias", "linear_163.w_0": "visual.vision_model.encoder.layers.26.mlp.fc2.weight", "linear_163.b_0": "visual.vision_model.encoder.layers.26.mlp.fc2.bias", "layer_norm_55.w_0": "visual.vision_model.post_layernorm.weight", "layer_norm_55.b_0": "visual.vision_model.post_layernorm.bias", "siglip_multihead_attention_pooling_head_0.w_0": "visual.vision_model.head.probe", "multi_head_attention_0.w_0": "visual.vision_model.head.attention.in_proj_weight", "multi_head_attention_0.w_1": "visual.vision_model.head.attention.in_proj_bias", "linear_164.w_0": "visual.vision_model.head.attention.out_proj.weight", "linear_164.b_0": "visual.vision_model.head.attention.out_proj.bias", "layer_norm_56.w_0": "visual.vision_model.head.layernorm.weight", "layer_norm_56.b_0": "visual.vision_model.head.layernorm.bias", "linear_165.w_0": "visual.vision_model.head.mlp.fc1.weight", "linear_165.b_0": "visual.vision_model.head.mlp.fc1.bias", "linear_166.w_0": "visual.vision_model.head.mlp.fc2.weight", "linear_166.b_0": "visual.vision_model.head.mlp.fc2.bias", "embedding_2.w_0": "model.embed_tokens.weight", "linear_167.w_0": "model.layers.0.self_attn.q_proj.weight", "linear_168.w_0": "model.layers.0.self_attn.k_proj.weight", "linear_169.w_0": "model.layers.0.self_attn.v_proj.weight", "linear_170.w_0": "model.layers.0.self_attn.o_proj.weight", "linear_171.w_0": "model.layers.0.mlp.gate_proj.weight", "linear_172.w_0": "model.layers.0.mlp.up_proj.weight", "linear_173.w_0": "model.layers.0.mlp.down_proj.weight", "create_parameter_0.w_0": "model.layers.0.input_layernorm.weight", "create_parameter_1.w_0": "model.layers.0.post_attention_layernorm.weight", "linear_174.w_0": "model.layers.1.self_attn.q_proj.weight", "linear_175.w_0": "model.layers.1.self_attn.k_proj.weight", "linear_176.w_0": "model.layers.1.self_attn.v_proj.weight", "linear_177.w_0": "model.layers.1.self_attn.o_proj.weight", "linear_178.w_0": "model.layers.1.mlp.gate_proj.weight", "linear_179.w_0": "model.layers.1.mlp.up_proj.weight", "linear_180.w_0": "model.layers.1.mlp.down_proj.weight", "create_parameter_2.w_0": "model.layers.1.input_layernorm.weight", "create_parameter_3.w_0": "model.layers.1.post_attention_layernorm.weight", "linear_181.w_0": "model.layers.2.self_attn.q_proj.weight", "linear_182.w_0": "model.layers.2.self_attn.k_proj.weight", "linear_183.w_0": "model.layers.2.self_attn.v_proj.weight", "linear_184.w_0": "model.layers.2.self_attn.o_proj.weight", "linear_185.w_0": "model.layers.2.mlp.gate_proj.weight", "linear_186.w_0": "model.layers.2.mlp.up_proj.weight", "linear_187.w_0": "model.layers.2.mlp.down_proj.weight", "create_parameter_4.w_0": "model.layers.2.input_layernorm.weight", "create_parameter_5.w_0": "model.layers.2.post_attention_layernorm.weight", "linear_188.w_0": "model.layers.3.self_attn.q_proj.weight", "linear_189.w_0": "model.layers.3.self_attn.k_proj.weight", "linear_190.w_0": "model.layers.3.self_attn.v_proj.weight", "linear_191.w_0": "model.layers.3.self_attn.o_proj.weight", "linear_192.w_0": "model.layers.3.mlp.gate_proj.weight", "linear_193.w_0": "model.layers.3.mlp.up_proj.weight", "linear_194.w_0": "model.layers.3.mlp.down_proj.weight", "create_parameter_6.w_0": "model.layers.3.input_layernorm.weight", "create_parameter_7.w_0": "model.layers.3.post_attention_layernorm.weight", "linear_195.w_0": "model.layers.4.self_attn.q_proj.weight", "linear_196.w_0": "model.layers.4.self_attn.k_proj.weight", "linear_197.w_0": "model.layers.4.self_attn.v_proj.weight", "linear_198.w_0": "model.layers.4.self_attn.o_proj.weight", "linear_199.w_0": "model.layers.4.mlp.gate_proj.weight", "linear_200.w_0": "model.layers.4.mlp.up_proj.weight", "linear_201.w_0": "model.layers.4.mlp.down_proj.weight", "create_parameter_8.w_0": "model.layers.4.input_layernorm.weight", "create_parameter_9.w_0": "model.layers.4.post_attention_layernorm.weight", "linear_202.w_0": "model.layers.5.self_attn.q_proj.weight", "linear_203.w_0": "model.layers.5.self_attn.k_proj.weight", "linear_204.w_0": "model.layers.5.self_attn.v_proj.weight", "linear_205.w_0": "model.layers.5.self_attn.o_proj.weight", "linear_206.w_0": "model.layers.5.mlp.gate_proj.weight", "linear_207.w_0": "model.layers.5.mlp.up_proj.weight", "linear_208.w_0": "model.layers.5.mlp.down_proj.weight", "create_parameter_10.w_0": "model.layers.5.input_layernorm.weight", "create_parameter_11.w_0": "model.layers.5.post_attention_layernorm.weight", "linear_209.w_0": "model.layers.6.self_attn.q_proj.weight", "linear_210.w_0": "model.layers.6.self_attn.k_proj.weight", "linear_211.w_0": "model.layers.6.self_attn.v_proj.weight", "linear_212.w_0": "model.layers.6.self_attn.o_proj.weight", "linear_213.w_0": "model.layers.6.mlp.gate_proj.weight", "linear_214.w_0": "model.layers.6.mlp.up_proj.weight", "linear_215.w_0": "model.layers.6.mlp.down_proj.weight", "create_parameter_12.w_0": "model.layers.6.input_layernorm.weight", "create_parameter_13.w_0": "model.layers.6.post_attention_layernorm.weight", "linear_216.w_0": "model.layers.7.self_attn.q_proj.weight", "linear_217.w_0": "model.layers.7.self_attn.k_proj.weight", "linear_218.w_0": "model.layers.7.self_attn.v_proj.weight", "linear_219.w_0": "model.layers.7.self_attn.o_proj.weight", "linear_220.w_0": "model.layers.7.mlp.gate_proj.weight", "linear_221.w_0": "model.layers.7.mlp.up_proj.weight", "linear_222.w_0": "model.layers.7.mlp.down_proj.weight", "create_parameter_14.w_0": "model.layers.7.input_layernorm.weight", "create_parameter_15.w_0": "model.layers.7.post_attention_layernorm.weight", "linear_223.w_0": "model.layers.8.self_attn.q_proj.weight", "linear_224.w_0": "model.layers.8.self_attn.k_proj.weight", "linear_225.w_0": "model.layers.8.self_attn.v_proj.weight", "linear_226.w_0": "model.layers.8.self_attn.o_proj.weight", "linear_227.w_0": "model.layers.8.mlp.gate_proj.weight", "linear_228.w_0": "model.layers.8.mlp.up_proj.weight", "linear_229.w_0": "model.layers.8.mlp.down_proj.weight", "create_parameter_16.w_0": "model.layers.8.input_layernorm.weight", "create_parameter_17.w_0": "model.layers.8.post_attention_layernorm.weight", "linear_230.w_0": "model.layers.9.self_attn.q_proj.weight", "linear_231.w_0": "model.layers.9.self_attn.k_proj.weight", "linear_232.w_0": "model.layers.9.self_attn.v_proj.weight", "linear_233.w_0": "model.layers.9.self_attn.o_proj.weight", "linear_234.w_0": "model.layers.9.mlp.gate_proj.weight", "linear_235.w_0": "model.layers.9.mlp.up_proj.weight", "linear_236.w_0": "model.layers.9.mlp.down_proj.weight", "create_parameter_18.w_0": "model.layers.9.input_layernorm.weight", "create_parameter_19.w_0": "model.layers.9.post_attention_layernorm.weight", "linear_237.w_0": "model.layers.10.self_attn.q_proj.weight", "linear_238.w_0": "model.layers.10.self_attn.k_proj.weight", "linear_239.w_0": "model.layers.10.self_attn.v_proj.weight", "linear_240.w_0": "model.layers.10.self_attn.o_proj.weight", "linear_241.w_0": "model.layers.10.mlp.gate_proj.weight", "linear_242.w_0": "model.layers.10.mlp.up_proj.weight", "linear_243.w_0": "model.layers.10.mlp.down_proj.weight", "create_parameter_20.w_0": "model.layers.10.input_layernorm.weight", "create_parameter_21.w_0": "model.layers.10.post_attention_layernorm.weight", "linear_244.w_0": "model.layers.11.self_attn.q_proj.weight", "linear_245.w_0": "model.layers.11.self_attn.k_proj.weight", "linear_246.w_0": "model.layers.11.self_attn.v_proj.weight", "linear_247.w_0": "model.layers.11.self_attn.o_proj.weight", "linear_248.w_0": "model.layers.11.mlp.gate_proj.weight", "linear_249.w_0": "model.layers.11.mlp.up_proj.weight", "linear_250.w_0": "model.layers.11.mlp.down_proj.weight", "create_parameter_22.w_0": "model.layers.11.input_layernorm.weight", "create_parameter_23.w_0": "model.layers.11.post_attention_layernorm.weight", "linear_251.w_0": "model.layers.12.self_attn.q_proj.weight", "linear_252.w_0": "model.layers.12.self_attn.k_proj.weight", "linear_253.w_0": "model.layers.12.self_attn.v_proj.weight", "linear_254.w_0": "model.layers.12.self_attn.o_proj.weight", "linear_255.w_0": "model.layers.12.mlp.gate_proj.weight", "linear_256.w_0": "model.layers.12.mlp.up_proj.weight", "linear_257.w_0": "model.layers.12.mlp.down_proj.weight", "create_parameter_24.w_0": "model.layers.12.input_layernorm.weight", "create_parameter_25.w_0": "model.layers.12.post_attention_layernorm.weight", "linear_258.w_0": "model.layers.13.self_attn.q_proj.weight", "linear_259.w_0": "model.layers.13.self_attn.k_proj.weight", "linear_260.w_0": "model.layers.13.self_attn.v_proj.weight", "linear_261.w_0": "model.layers.13.self_attn.o_proj.weight", "linear_262.w_0": "model.layers.13.mlp.gate_proj.weight", "linear_263.w_0": "model.layers.13.mlp.up_proj.weight", "linear_264.w_0": "model.layers.13.mlp.down_proj.weight", "create_parameter_26.w_0": "model.layers.13.input_layernorm.weight", "create_parameter_27.w_0": "model.layers.13.post_attention_layernorm.weight", "linear_265.w_0": "model.layers.14.self_attn.q_proj.weight", "linear_266.w_0": "model.layers.14.self_attn.k_proj.weight", "linear_267.w_0": "model.layers.14.self_attn.v_proj.weight", "linear_268.w_0": "model.layers.14.self_attn.o_proj.weight", "linear_269.w_0": "model.layers.14.mlp.gate_proj.weight", "linear_270.w_0": "model.layers.14.mlp.up_proj.weight", "linear_271.w_0": "model.layers.14.mlp.down_proj.weight", "create_parameter_28.w_0": "model.layers.14.input_layernorm.weight", "create_parameter_29.w_0": "model.layers.14.post_attention_layernorm.weight", "linear_272.w_0": "model.layers.15.self_attn.q_proj.weight", "linear_273.w_0": "model.layers.15.self_attn.k_proj.weight", "linear_274.w_0": "model.layers.15.self_attn.v_proj.weight", "linear_275.w_0": "model.layers.15.self_attn.o_proj.weight", "linear_276.w_0": "model.layers.15.mlp.gate_proj.weight", "linear_277.w_0": "model.layers.15.mlp.up_proj.weight", "linear_278.w_0": "model.layers.15.mlp.down_proj.weight", "create_parameter_30.w_0": "model.layers.15.input_layernorm.weight", "create_parameter_31.w_0": "model.layers.15.post_attention_layernorm.weight", "linear_279.w_0": "model.layers.16.self_attn.q_proj.weight", "linear_280.w_0": "model.layers.16.self_attn.k_proj.weight", "linear_281.w_0": "model.layers.16.self_attn.v_proj.weight", "linear_282.w_0": "model.layers.16.self_attn.o_proj.weight", "linear_283.w_0": "model.layers.16.mlp.gate_proj.weight", "linear_284.w_0": "model.layers.16.mlp.up_proj.weight", "linear_285.w_0": "model.layers.16.mlp.down_proj.weight", "create_parameter_32.w_0": "model.layers.16.input_layernorm.weight", "create_parameter_33.w_0": "model.layers.16.post_attention_layernorm.weight", "linear_286.w_0": "model.layers.17.self_attn.q_proj.weight", "linear_287.w_0": "model.layers.17.self_attn.k_proj.weight", "linear_288.w_0": "model.layers.17.self_attn.v_proj.weight", "linear_289.w_0": "model.layers.17.self_attn.o_proj.weight", "linear_290.w_0": "model.layers.17.mlp.gate_proj.weight", "linear_291.w_0": "model.layers.17.mlp.up_proj.weight", "linear_292.w_0": "model.layers.17.mlp.down_proj.weight", "create_parameter_34.w_0": "model.layers.17.input_layernorm.weight", "create_parameter_35.w_0": "model.layers.17.post_attention_layernorm.weight", "create_parameter_36.w_0": "model.norm.weight", "linear_293.w_0": "lm_head.weight"}
|
checkpoint-336/tokenizer.model
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:34ef7db83df785924fb83d7b887b6e822a031c56e15cff40aaf9b982988180df
|
| 3 |
+
size 1614363
|
checkpoint-336/tokenizer_config.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
checkpoint-336/trainer_state.json
ADDED
|
@@ -0,0 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"best_metric": null,
|
| 3 |
+
"best_model_checkpoint": null,
|
| 4 |
+
"consumed_samples": 0,
|
| 5 |
+
"epoch": 0.7980997624703088,
|
| 6 |
+
"global_step": 336,
|
| 7 |
+
"is_local_process_zero": true,
|
| 8 |
+
"is_world_process_zero": true,
|
| 9 |
+
"log_history": [],
|
| 10 |
+
"max_steps": 421,
|
| 11 |
+
"num_train_epochs": 9223372036854775807,
|
| 12 |
+
"total_flos": 0,
|
| 13 |
+
"trial_name": null,
|
| 14 |
+
"trial_params": null
|
| 15 |
+
}
|
checkpoint-336/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0888f0d3f65facdf04c57f4bd4b2036e3fec9976c17b7eb281a443129536c8e7
|
| 3 |
+
size 6425
|
checkpoint-378/added_tokens.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"<unk>": 0, "<s>": 1, "</s>": 2, "0": 3, "1": 4, "2": 5, "3": 6, "4": 7, "5": 8, "6": 9, "7": 10, "8": 11, "9": 12, "<|end_of_sentence|>": 100272, "<|begin_of_sentence|>": 100273, "<mask:1>": 100274, "<mask:4>": 100277, "<mask:5>": 100278, "<mask:6>": 100279, "<mask:7>": 100280, "<|IMAGE_PLACEHOLDER|>": 100295, "<|AUDIO_PLACEHOLDER|>": 100296, "<|LOC_0|>": 100297, "<|LOC_1|>": 100298, "<|LOC_2|>": 100299, "<|LOC_3|>": 100300, "<|LOC_4|>": 100301, "<|LOC_5|>": 100302, "<|LOC_6|>": 100303, "<|LOC_7|>": 100304, "<|LOC_8|>": 100305, "<|LOC_9|>": 100306, "<|LOC_10|>": 100307, "<|LOC_11|>": 100308, "<|LOC_12|>": 100309, "<|LOC_13|>": 100310, "<|LOC_14|>": 100311, "<|LOC_15|>": 100312, "<|LOC_16|>": 100313, "<|LOC_17|>": 100314, "<|LOC_18|>": 100315, "<|LOC_19|>": 100316, "<|LOC_20|>": 100317, "<|LOC_21|>": 100318, "<|LOC_22|>": 100319, "<|LOC_23|>": 100320, "<|LOC_24|>": 100321, "<|LOC_25|>": 100322, "<|LOC_26|>": 100323, "<|LOC_27|>": 100324, "<|LOC_28|>": 100325, "<|LOC_29|>": 100326, "<|LOC_30|>": 100327, "<|LOC_31|>": 100328, "<|LOC_32|>": 100329, "<|LOC_33|>": 100330, "<|LOC_34|>": 100331, "<|LOC_35|>": 100332, "<|LOC_36|>": 100333, "<|LOC_37|>": 100334, "<|LOC_38|>": 100335, "<|LOC_39|>": 100336, "<|LOC_40|>": 100337, "<|LOC_41|>": 100338, "<|LOC_42|>": 100339, "<|LOC_43|>": 100340, "<|LOC_44|>": 100341, "<|LOC_45|>": 100342, "<|LOC_46|>": 100343, "<|LOC_47|>": 100344, "<|LOC_48|>": 100345, "<|LOC_49|>": 100346, "<|LOC_50|>": 100347, "<|LOC_51|>": 100348, "<|LOC_52|>": 100349, "<|LOC_53|>": 100350, "<|LOC_54|>": 100351, "<|LOC_55|>": 100352, "<|LOC_56|>": 100353, "<|LOC_57|>": 100354, "<|LOC_58|>": 100355, "<|LOC_59|>": 100356, "<|LOC_60|>": 100357, "<|LOC_61|>": 100358, "<|LOC_62|>": 100359, "<|LOC_63|>": 100360, "<|LOC_64|>": 100361, "<|LOC_65|>": 100362, "<|LOC_66|>": 100363, "<|LOC_67|>": 100364, "<|LOC_68|>": 100365, "<|LOC_69|>": 100366, "<|LOC_70|>": 100367, "<|LOC_71|>": 100368, "<|LOC_72|>": 100369, "<|LOC_73|>": 100370, "<|LOC_74|>": 100371, "<|LOC_75|>": 100372, "<|LOC_76|>": 100373, "<|LOC_77|>": 100374, "<|LOC_78|>": 100375, "<|LOC_79|>": 100376, "<|LOC_80|>": 100377, "<|LOC_81|>": 100378, "<|LOC_82|>": 100379, "<|LOC_83|>": 100380, "<|LOC_84|>": 100381, "<|LOC_85|>": 100382, "<|LOC_86|>": 100383, "<|LOC_87|>": 100384, "<|LOC_88|>": 100385, "<|LOC_89|>": 100386, "<|LOC_90|>": 100387, "<|LOC_91|>": 100388, "<|LOC_92|>": 100389, "<|LOC_93|>": 100390, "<|LOC_94|>": 100391, "<|LOC_95|>": 100392, "<|LOC_96|>": 100393, "<|LOC_97|>": 100394, "<|LOC_98|>": 100395, "<|LOC_99|>": 100396, "<|LOC_100|>": 100397, "<|LOC_101|>": 100398, "<|LOC_102|>": 100399, "<|LOC_103|>": 100400, "<|LOC_104|>": 100401, "<|LOC_105|>": 100402, "<|LOC_106|>": 100403, "<|LOC_107|>": 100404, "<|LOC_108|>": 100405, "<|LOC_109|>": 100406, "<|LOC_110|>": 100407, "<|LOC_111|>": 100408, "<|LOC_112|>": 100409, "<|LOC_113|>": 100410, "<|LOC_114|>": 100411, "<|LOC_115|>": 100412, "<|LOC_116|>": 100413, "<|LOC_117|>": 100414, "<|LOC_118|>": 100415, "<|LOC_119|>": 100416, "<|LOC_120|>": 100417, "<|LOC_121|>": 100418, "<|LOC_122|>": 100419, "<|LOC_123|>": 100420, "<|LOC_124|>": 100421, "<|LOC_125|>": 100422, "<|LOC_126|>": 100423, "<|LOC_127|>": 100424, "<|LOC_128|>": 100425, "<|LOC_129|>": 100426, "<|LOC_130|>": 100427, "<|LOC_131|>": 100428, "<|LOC_132|>": 100429, "<|LOC_133|>": 100430, "<|LOC_134|>": 100431, "<|LOC_135|>": 100432, "<|LOC_136|>": 100433, "<|LOC_137|>": 100434, "<|LOC_138|>": 100435, "<|LOC_139|>": 100436, "<|LOC_140|>": 100437, "<|LOC_141|>": 100438, "<|LOC_142|>": 100439, "<|LOC_143|>": 100440, "<|LOC_144|>": 100441, "<|LOC_145|>": 100442, "<|LOC_146|>": 100443, "<|LOC_147|>": 100444, "<|LOC_148|>": 100445, "<|LOC_149|>": 100446, "<|LOC_150|>": 100447, "<|LOC_151|>": 100448, "<|LOC_152|>": 100449, "<|LOC_153|>": 100450, "<|LOC_154|>": 100451, "<|LOC_155|>": 100452, "<|LOC_156|>": 100453, "<|LOC_157|>": 100454, "<|LOC_158|>": 100455, "<|LOC_159|>": 100456, "<|LOC_160|>": 100457, "<|LOC_161|>": 100458, "<|LOC_162|>": 100459, "<|LOC_163|>": 100460, "<|LOC_164|>": 100461, "<|LOC_165|>": 100462, "<|LOC_166|>": 100463, "<|LOC_167|>": 100464, "<|LOC_168|>": 100465, "<|LOC_169|>": 100466, "<|LOC_170|>": 100467, "<|LOC_171|>": 100468, "<|LOC_172|>": 100469, "<|LOC_173|>": 100470, "<|LOC_174|>": 100471, "<|LOC_175|>": 100472, "<|LOC_176|>": 100473, "<|LOC_177|>": 100474, "<|LOC_178|>": 100475, "<|LOC_179|>": 100476, "<|LOC_180|>": 100477, "<|LOC_181|>": 100478, "<|LOC_182|>": 100479, "<|LOC_183|>": 100480, "<|LOC_184|>": 100481, "<|LOC_185|>": 100482, "<|LOC_186|>": 100483, "<|LOC_187|>": 100484, "<|LOC_188|>": 100485, "<|LOC_189|>": 100486, "<|LOC_190|>": 100487, "<|LOC_191|>": 100488, "<|LOC_192|>": 100489, "<|LOC_193|>": 100490, "<|LOC_194|>": 100491, "<|LOC_195|>": 100492, "<|LOC_196|>": 100493, "<|LOC_197|>": 100494, "<|LOC_198|>": 100495, "<|LOC_199|>": 100496, "<|LOC_200|>": 100497, "<|LOC_201|>": 100498, "<|LOC_202|>": 100499, "<|LOC_203|>": 100500, "<|LOC_204|>": 100501, "<|LOC_205|>": 100502, "<|LOC_206|>": 100503, "<|LOC_207|>": 100504, "<|LOC_208|>": 100505, "<|LOC_209|>": 100506, "<|LOC_210|>": 100507, "<|LOC_211|>": 100508, "<|LOC_212|>": 100509, "<|LOC_213|>": 100510, "<|LOC_214|>": 100511, "<|LOC_215|>": 100512, "<|LOC_216|>": 100513, "<|LOC_217|>": 100514, "<|LOC_218|>": 100515, "<|LOC_219|>": 100516, "<|LOC_220|>": 100517, "<|LOC_221|>": 100518, "<|LOC_222|>": 100519, "<|LOC_223|>": 100520, "<|LOC_224|>": 100521, "<|LOC_225|>": 100522, "<|LOC_226|>": 100523, "<|LOC_227|>": 100524, "<|LOC_228|>": 100525, "<|LOC_229|>": 100526, "<|LOC_230|>": 100527, "<|LOC_231|>": 100528, "<|LOC_232|>": 100529, "<|LOC_233|>": 100530, "<|LOC_234|>": 100531, "<|LOC_235|>": 100532, "<|LOC_236|>": 100533, "<|LOC_237|>": 100534, "<|LOC_238|>": 100535, "<|LOC_239|>": 100536, "<|LOC_240|>": 100537, "<|LOC_241|>": 100538, "<|LOC_242|>": 100539, "<|LOC_243|>": 100540, "<|LOC_244|>": 100541, "<|LOC_245|>": 100542, "<|LOC_246|>": 100543, "<|LOC_247|>": 100544, "<|LOC_248|>": 100545, "<|LOC_249|>": 100546, "<|LOC_250|>": 100547, "<|LOC_251|>": 100548, "<|LOC_252|>": 100549, "<|LOC_253|>": 100550, "<|LOC_254|>": 100551, "<|LOC_255|>": 100552, "<|LOC_256|>": 100553, "<|LOC_257|>": 100554, "<|LOC_258|>": 100555, "<|LOC_259|>": 100556, "<|LOC_260|>": 100557, "<|LOC_261|>": 100558, "<|LOC_262|>": 100559, "<|LOC_263|>": 100560, "<|LOC_264|>": 100561, "<|LOC_265|>": 100562, "<|LOC_266|>": 100563, "<|LOC_267|>": 100564, "<|LOC_268|>": 100565, "<|LOC_269|>": 100566, "<|LOC_270|>": 100567, "<|LOC_271|>": 100568, "<|LOC_272|>": 100569, "<|LOC_273|>": 100570, "<|LOC_274|>": 100571, "<|LOC_275|>": 100572, "<|LOC_276|>": 100573, "<|LOC_277|>": 100574, "<|LOC_278|>": 100575, "<|LOC_279|>": 100576, "<|LOC_280|>": 100577, "<|LOC_281|>": 100578, "<|LOC_282|>": 100579, "<|LOC_283|>": 100580, "<|LOC_284|>": 100581, "<|LOC_285|>": 100582, "<|LOC_286|>": 100583, "<|LOC_287|>": 100584, "<|LOC_288|>": 100585, "<|LOC_289|>": 100586, "<|LOC_290|>": 100587, "<|LOC_291|>": 100588, "<|LOC_292|>": 100589, "<|LOC_293|>": 100590, "<|LOC_294|>": 100591, "<|LOC_295|>": 100592, "<|LOC_296|>": 100593, "<|LOC_297|>": 100594, "<|LOC_298|>": 100595, "<|LOC_299|>": 100596, "<|LOC_300|>": 100597, "<|LOC_301|>": 100598, "<|LOC_302|>": 100599, "<|LOC_303|>": 100600, "<|LOC_304|>": 100601, "<|LOC_305|>": 100602, "<|LOC_306|>": 100603, "<|LOC_307|>": 100604, "<|LOC_308|>": 100605, "<|LOC_309|>": 100606, "<|LOC_310|>": 100607, "<|LOC_311|>": 100608, "<|LOC_312|>": 100609, "<|LOC_313|>": 100610, "<|LOC_314|>": 100611, "<|LOC_315|>": 100612, "<|LOC_316|>": 100613, "<|LOC_317|>": 100614, "<|LOC_318|>": 100615, "<|LOC_319|>": 100616, "<|LOC_320|>": 100617, "<|LOC_321|>": 100618, "<|LOC_322|>": 100619, "<|LOC_323|>": 100620, "<|LOC_324|>": 100621, "<|LOC_325|>": 100622, "<|LOC_326|>": 100623, "<|LOC_327|>": 100624, "<|LOC_328|>": 100625, "<|LOC_329|>": 100626, "<|LOC_330|>": 100627, "<|LOC_331|>": 100628, "<|LOC_332|>": 100629, "<|LOC_333|>": 100630, "<|LOC_334|>": 100631, "<|LOC_335|>": 100632, "<|LOC_336|>": 100633, "<|LOC_337|>": 100634, "<|LOC_338|>": 100635, "<|LOC_339|>": 100636, "<|LOC_340|>": 100637, "<|LOC_341|>": 100638, "<|LOC_342|>": 100639, "<|LOC_343|>": 100640, "<|LOC_344|>": 100641, "<|LOC_345|>": 100642, "<|LOC_346|>": 100643, "<|LOC_347|>": 100644, "<|LOC_348|>": 100645, "<|LOC_349|>": 100646, "<|LOC_350|>": 100647, "<|LOC_351|>": 100648, "<|LOC_352|>": 100649, "<|LOC_353|>": 100650, "<|LOC_354|>": 100651, "<|LOC_355|>": 100652, "<|LOC_356|>": 100653, "<|LOC_357|>": 100654, "<|LOC_358|>": 100655, "<|LOC_359|>": 100656, "<|LOC_360|>": 100657, "<|LOC_361|>": 100658, "<|LOC_362|>": 100659, "<|LOC_363|>": 100660, "<|LOC_364|>": 100661, "<|LOC_365|>": 100662, "<|LOC_366|>": 100663, "<|LOC_367|>": 100664, "<|LOC_368|>": 100665, "<|LOC_369|>": 100666, "<|LOC_370|>": 100667, "<|LOC_371|>": 100668, "<|LOC_372|>": 100669, "<|LOC_373|>": 100670, "<|LOC_374|>": 100671, "<|LOC_375|>": 100672, "<|LOC_376|>": 100673, "<|LOC_377|>": 100674, "<|LOC_378|>": 100675, "<|LOC_379|>": 100676, "<|LOC_380|>": 100677, "<|LOC_381|>": 100678, "<|LOC_382|>": 100679, "<|LOC_383|>": 100680, "<|LOC_384|>": 100681, "<|LOC_385|>": 100682, "<|LOC_386|>": 100683, "<|LOC_387|>": 100684, "<|LOC_388|>": 100685, "<|LOC_389|>": 100686, "<|LOC_390|>": 100687, "<|LOC_391|>": 100688, "<|LOC_392|>": 100689, "<|LOC_393|>": 100690, "<|LOC_394|>": 100691, "<|LOC_395|>": 100692, "<|LOC_396|>": 100693, "<|LOC_397|>": 100694, "<|LOC_398|>": 100695, "<|LOC_399|>": 100696, "<|LOC_400|>": 100697, "<|LOC_401|>": 100698, "<|LOC_402|>": 100699, "<|LOC_403|>": 100700, "<|LOC_404|>": 100701, "<|LOC_405|>": 100702, "<|LOC_406|>": 100703, "<|LOC_407|>": 100704, "<|LOC_408|>": 100705, "<|LOC_409|>": 100706, "<|LOC_410|>": 100707, "<|LOC_411|>": 100708, "<|LOC_412|>": 100709, "<|LOC_413|>": 100710, "<|LOC_414|>": 100711, "<|LOC_415|>": 100712, "<|LOC_416|>": 100713, "<|LOC_417|>": 100714, "<|LOC_418|>": 100715, "<|LOC_419|>": 100716, "<|LOC_420|>": 100717, "<|LOC_421|>": 100718, "<|LOC_422|>": 100719, "<|LOC_423|>": 100720, "<|LOC_424|>": 100721, "<|LOC_425|>": 100722, "<|LOC_426|>": 100723, "<|LOC_427|>": 100724, "<|LOC_428|>": 100725, "<|LOC_429|>": 100726, "<|LOC_430|>": 100727, "<|LOC_431|>": 100728, "<|LOC_432|>": 100729, "<|LOC_433|>": 100730, "<|LOC_434|>": 100731, "<|LOC_435|>": 100732, "<|LOC_436|>": 100733, "<|LOC_437|>": 100734, "<|LOC_438|>": 100735, "<|LOC_439|>": 100736, "<|LOC_440|>": 100737, "<|LOC_441|>": 100738, "<|LOC_442|>": 100739, "<|LOC_443|>": 100740, "<|LOC_444|>": 100741, "<|LOC_445|>": 100742, "<|LOC_446|>": 100743, "<|LOC_447|>": 100744, "<|LOC_448|>": 100745, "<|LOC_449|>": 100746, "<|LOC_450|>": 100747, "<|LOC_451|>": 100748, "<|LOC_452|>": 100749, "<|LOC_453|>": 100750, "<|LOC_454|>": 100751, "<|LOC_455|>": 100752, "<|LOC_456|>": 100753, "<|LOC_457|>": 100754, "<|LOC_458|>": 100755, "<|LOC_459|>": 100756, "<|LOC_460|>": 100757, "<|LOC_461|>": 100758, "<|LOC_462|>": 100759, "<|LOC_463|>": 100760, "<|LOC_464|>": 100761, "<|LOC_465|>": 100762, "<|LOC_466|>": 100763, "<|LOC_467|>": 100764, "<|LOC_468|>": 100765, "<|LOC_469|>": 100766, "<|LOC_470|>": 100767, "<|LOC_471|>": 100768, "<|LOC_472|>": 100769, "<|LOC_473|>": 100770, "<|LOC_474|>": 100771, "<|LOC_475|>": 100772, "<|LOC_476|>": 100773, "<|LOC_477|>": 100774, "<|LOC_478|>": 100775, "<|LOC_479|>": 100776, "<|LOC_480|>": 100777, "<|LOC_481|>": 100778, "<|LOC_482|>": 100779, "<|LOC_483|>": 100780, "<|LOC_484|>": 100781, "<|LOC_485|>": 100782, "<|LOC_486|>": 100783, "<|LOC_487|>": 100784, "<|LOC_488|>": 100785, "<|LOC_489|>": 100786, "<|LOC_490|>": 100787, "<|LOC_491|>": 100788, "<|LOC_492|>": 100789, "<|LOC_493|>": 100790, "<|LOC_494|>": 100791, "<|LOC_495|>": 100792, "<|LOC_496|>": 100793, "<|LOC_497|>": 100794, "<|LOC_498|>": 100795, "<|LOC_499|>": 100796, "<|LOC_500|>": 100797, "<|LOC_501|>": 100798, "<|LOC_502|>": 100799, "<|LOC_503|>": 100800, "<|LOC_504|>": 100801, "<|LOC_505|>": 100802, "<|LOC_506|>": 100803, "<|LOC_507|>": 100804, "<|LOC_508|>": 100805, "<|LOC_509|>": 100806, "<|LOC_510|>": 100807, "<|LOC_511|>": 100808, "<|LOC_512|>": 100809, "<|LOC_513|>": 100810, "<|LOC_514|>": 100811, "<|LOC_515|>": 100812, "<|LOC_516|>": 100813, "<|LOC_517|>": 100814, "<|LOC_518|>": 100815, "<|LOC_519|>": 100816, "<|LOC_520|>": 100817, "<|LOC_521|>": 100818, "<|LOC_522|>": 100819, "<|LOC_523|>": 100820, "<|LOC_524|>": 100821, "<|LOC_525|>": 100822, "<|LOC_526|>": 100823, "<|LOC_527|>": 100824, "<|LOC_528|>": 100825, "<|LOC_529|>": 100826, "<|LOC_530|>": 100827, "<|LOC_531|>": 100828, "<|LOC_532|>": 100829, "<|LOC_533|>": 100830, "<|LOC_534|>": 100831, "<|LOC_535|>": 100832, "<|LOC_536|>": 100833, "<|LOC_537|>": 100834, "<|LOC_538|>": 100835, "<|LOC_539|>": 100836, "<|LOC_540|>": 100837, "<|LOC_541|>": 100838, "<|LOC_542|>": 100839, "<|LOC_543|>": 100840, "<|LOC_544|>": 100841, "<|LOC_545|>": 100842, "<|LOC_546|>": 100843, "<|LOC_547|>": 100844, "<|LOC_548|>": 100845, "<|LOC_549|>": 100846, "<|LOC_550|>": 100847, "<|LOC_551|>": 100848, "<|LOC_552|>": 100849, "<|LOC_553|>": 100850, "<|LOC_554|>": 100851, "<|LOC_555|>": 100852, "<|LOC_556|>": 100853, "<|LOC_557|>": 100854, "<|LOC_558|>": 100855, "<|LOC_559|>": 100856, "<|LOC_560|>": 100857, "<|LOC_561|>": 100858, "<|LOC_562|>": 100859, "<|LOC_563|>": 100860, "<|LOC_564|>": 100861, "<|LOC_565|>": 100862, "<|LOC_566|>": 100863, "<|LOC_567|>": 100864, "<|LOC_568|>": 100865, "<|LOC_569|>": 100866, "<|LOC_570|>": 100867, "<|LOC_571|>": 100868, "<|LOC_572|>": 100869, "<|LOC_573|>": 100870, "<|LOC_574|>": 100871, "<|LOC_575|>": 100872, "<|LOC_576|>": 100873, "<|LOC_577|>": 100874, "<|LOC_578|>": 100875, "<|LOC_579|>": 100876, "<|LOC_580|>": 100877, "<|LOC_581|>": 100878, "<|LOC_582|>": 100879, "<|LOC_583|>": 100880, "<|LOC_584|>": 100881, "<|LOC_585|>": 100882, "<|LOC_586|>": 100883, "<|LOC_587|>": 100884, "<|LOC_588|>": 100885, "<|LOC_589|>": 100886, "<|LOC_590|>": 100887, "<|LOC_591|>": 100888, "<|LOC_592|>": 100889, "<|LOC_593|>": 100890, "<|LOC_594|>": 100891, "<|LOC_595|>": 100892, "<|LOC_596|>": 100893, "<|LOC_597|>": 100894, "<|LOC_598|>": 100895, "<|LOC_599|>": 100896, "<|LOC_600|>": 100897, "<|LOC_601|>": 100898, "<|LOC_602|>": 100899, "<|LOC_603|>": 100900, "<|LOC_604|>": 100901, "<|LOC_605|>": 100902, "<|LOC_606|>": 100903, "<|LOC_607|>": 100904, "<|LOC_608|>": 100905, "<|LOC_609|>": 100906, "<|LOC_610|>": 100907, "<|LOC_611|>": 100908, "<|LOC_612|>": 100909, "<|LOC_613|>": 100910, "<|LOC_614|>": 100911, "<|LOC_615|>": 100912, "<|LOC_616|>": 100913, "<|LOC_617|>": 100914, "<|LOC_618|>": 100915, "<|LOC_619|>": 100916, "<|LOC_620|>": 100917, "<|LOC_621|>": 100918, "<|LOC_622|>": 100919, "<|LOC_623|>": 100920, "<|LOC_624|>": 100921, "<|LOC_625|>": 100922, "<|LOC_626|>": 100923, "<|LOC_627|>": 100924, "<|LOC_628|>": 100925, "<|LOC_629|>": 100926, "<|LOC_630|>": 100927, "<|LOC_631|>": 100928, "<|LOC_632|>": 100929, "<|LOC_633|>": 100930, "<|LOC_634|>": 100931, "<|LOC_635|>": 100932, "<|LOC_636|>": 100933, "<|LOC_637|>": 100934, "<|LOC_638|>": 100935, "<|LOC_639|>": 100936, "<|LOC_640|>": 100937, "<|LOC_641|>": 100938, "<|LOC_642|>": 100939, "<|LOC_643|>": 100940, "<|LOC_644|>": 100941, "<|LOC_645|>": 100942, "<|LOC_646|>": 100943, "<|LOC_647|>": 100944, "<|LOC_648|>": 100945, "<|LOC_649|>": 100946, "<|LOC_650|>": 100947, "<|LOC_651|>": 100948, "<|LOC_652|>": 100949, "<|LOC_653|>": 100950, "<|LOC_654|>": 100951, "<|LOC_655|>": 100952, "<|LOC_656|>": 100953, "<|LOC_657|>": 100954, "<|LOC_658|>": 100955, "<|LOC_659|>": 100956, "<|LOC_660|>": 100957, "<|LOC_661|>": 100958, "<|LOC_662|>": 100959, "<|LOC_663|>": 100960, "<|LOC_664|>": 100961, "<|LOC_665|>": 100962, "<|LOC_666|>": 100963, "<|LOC_667|>": 100964, "<|LOC_668|>": 100965, "<|LOC_669|>": 100966, "<|LOC_670|>": 100967, "<|LOC_671|>": 100968, "<|LOC_672|>": 100969, "<|LOC_673|>": 100970, "<|LOC_674|>": 100971, "<|LOC_675|>": 100972, "<|LOC_676|>": 100973, "<|LOC_677|>": 100974, "<|LOC_678|>": 100975, "<|LOC_679|>": 100976, "<|LOC_680|>": 100977, "<|LOC_681|>": 100978, "<|LOC_682|>": 100979, "<|LOC_683|>": 100980, "<|LOC_684|>": 100981, "<|LOC_685|>": 100982, "<|LOC_686|>": 100983, "<|LOC_687|>": 100984, "<|LOC_688|>": 100985, "<|LOC_689|>": 100986, "<|LOC_690|>": 100987, "<|LOC_691|>": 100988, "<|LOC_692|>": 100989, "<|LOC_693|>": 100990, "<|LOC_694|>": 100991, "<|LOC_695|>": 100992, "<|LOC_696|>": 100993, "<|LOC_697|>": 100994, "<|LOC_698|>": 100995, "<|LOC_699|>": 100996, "<|LOC_700|>": 100997, "<|LOC_701|>": 100998, "<|LOC_702|>": 100999, "<|LOC_703|>": 101000, "<|LOC_704|>": 101001, "<|LOC_705|>": 101002, "<|LOC_706|>": 101003, "<|LOC_707|>": 101004, "<|LOC_708|>": 101005, "<|LOC_709|>": 101006, "<|LOC_710|>": 101007, "<|LOC_711|>": 101008, "<|LOC_712|>": 101009, "<|LOC_713|>": 101010, "<|LOC_714|>": 101011, "<|LOC_715|>": 101012, "<|LOC_716|>": 101013, "<|LOC_717|>": 101014, "<|LOC_718|>": 101015, "<|LOC_719|>": 101016, "<|LOC_720|>": 101017, "<|LOC_721|>": 101018, "<|LOC_722|>": 101019, "<|LOC_723|>": 101020, "<|LOC_724|>": 101021, "<|LOC_725|>": 101022, "<|LOC_726|>": 101023, "<|LOC_727|>": 101024, "<|LOC_728|>": 101025, "<|LOC_729|>": 101026, "<|LOC_730|>": 101027, "<|LOC_731|>": 101028, "<|LOC_732|>": 101029, "<|LOC_733|>": 101030, "<|LOC_734|>": 101031, "<|LOC_735|>": 101032, "<|LOC_736|>": 101033, "<|LOC_737|>": 101034, "<|LOC_738|>": 101035, "<|LOC_739|>": 101036, "<|LOC_740|>": 101037, "<|LOC_741|>": 101038, "<|LOC_742|>": 101039, "<|LOC_743|>": 101040, "<|LOC_744|>": 101041, "<|LOC_745|>": 101042, "<|LOC_746|>": 101043, "<|LOC_747|>": 101044, "<|LOC_748|>": 101045, "<|LOC_749|>": 101046, "<|LOC_750|>": 101047, "<|LOC_751|>": 101048, "<|LOC_752|>": 101049, "<|LOC_753|>": 101050, "<|LOC_754|>": 101051, "<|LOC_755|>": 101052, "<|LOC_756|>": 101053, "<|LOC_757|>": 101054, "<|LOC_758|>": 101055, "<|LOC_759|>": 101056, "<|LOC_760|>": 101057, "<|LOC_761|>": 101058, "<|LOC_762|>": 101059, "<|LOC_763|>": 101060, "<|LOC_764|>": 101061, "<|LOC_765|>": 101062, "<|LOC_766|>": 101063, "<|LOC_767|>": 101064, "<|LOC_768|>": 101065, "<|LOC_769|>": 101066, "<|LOC_770|>": 101067, "<|LOC_771|>": 101068, "<|LOC_772|>": 101069, "<|LOC_773|>": 101070, "<|LOC_774|>": 101071, "<|LOC_775|>": 101072, "<|LOC_776|>": 101073, "<|LOC_777|>": 101074, "<|LOC_778|>": 101075, "<|LOC_779|>": 101076, "<|LOC_780|>": 101077, "<|LOC_781|>": 101078, "<|LOC_782|>": 101079, "<|LOC_783|>": 101080, "<|LOC_784|>": 101081, "<|LOC_785|>": 101082, "<|LOC_786|>": 101083, "<|LOC_787|>": 101084, "<|LOC_788|>": 101085, "<|LOC_789|>": 101086, "<|LOC_790|>": 101087, "<|LOC_791|>": 101088, "<|LOC_792|>": 101089, "<|LOC_793|>": 101090, "<|LOC_794|>": 101091, "<|LOC_795|>": 101092, "<|LOC_796|>": 101093, "<|LOC_797|>": 101094, "<|LOC_798|>": 101095, "<|LOC_799|>": 101096, "<|LOC_800|>": 101097, "<|LOC_801|>": 101098, "<|LOC_802|>": 101099, "<|LOC_803|>": 101100, "<|LOC_804|>": 101101, "<|LOC_805|>": 101102, "<|LOC_806|>": 101103, "<|LOC_807|>": 101104, "<|LOC_808|>": 101105, "<|LOC_809|>": 101106, "<|LOC_810|>": 101107, "<|LOC_811|>": 101108, "<|LOC_812|>": 101109, "<|LOC_813|>": 101110, "<|LOC_814|>": 101111, "<|LOC_815|>": 101112, "<|LOC_816|>": 101113, "<|LOC_817|>": 101114, "<|LOC_818|>": 101115, "<|LOC_819|>": 101116, "<|LOC_820|>": 101117, "<|LOC_821|>": 101118, "<|LOC_822|>": 101119, "<|LOC_823|>": 101120, "<|LOC_824|>": 101121, "<|LOC_825|>": 101122, "<|LOC_826|>": 101123, "<|LOC_827|>": 101124, "<|LOC_828|>": 101125, "<|LOC_829|>": 101126, "<|LOC_830|>": 101127, "<|LOC_831|>": 101128, "<|LOC_832|>": 101129, "<|LOC_833|>": 101130, "<|LOC_834|>": 101131, "<|LOC_835|>": 101132, "<|LOC_836|>": 101133, "<|LOC_837|>": 101134, "<|LOC_838|>": 101135, "<|LOC_839|>": 101136, "<|LOC_840|>": 101137, "<|LOC_841|>": 101138, "<|LOC_842|>": 101139, "<|LOC_843|>": 101140, "<|LOC_844|>": 101141, "<|LOC_845|>": 101142, "<|LOC_846|>": 101143, "<|LOC_847|>": 101144, "<|LOC_848|>": 101145, "<|LOC_849|>": 101146, "<|LOC_850|>": 101147, "<|LOC_851|>": 101148, "<|LOC_852|>": 101149, "<|LOC_853|>": 101150, "<|LOC_854|>": 101151, "<|LOC_855|>": 101152, "<|LOC_856|>": 101153, "<|LOC_857|>": 101154, "<|LOC_858|>": 101155, "<|LOC_859|>": 101156, "<|LOC_860|>": 101157, "<|LOC_861|>": 101158, "<|LOC_862|>": 101159, "<|LOC_863|>": 101160, "<|LOC_864|>": 101161, "<|LOC_865|>": 101162, "<|LOC_866|>": 101163, "<|LOC_867|>": 101164, "<|LOC_868|>": 101165, "<|LOC_869|>": 101166, "<|LOC_870|>": 101167, "<|LOC_871|>": 101168, "<|LOC_872|>": 101169, "<|LOC_873|>": 101170, "<|LOC_874|>": 101171, "<|LOC_875|>": 101172, "<|LOC_876|>": 101173, "<|LOC_877|>": 101174, "<|LOC_878|>": 101175, "<|LOC_879|>": 101176, "<|LOC_880|>": 101177, "<|LOC_881|>": 101178, "<|LOC_882|>": 101179, "<|LOC_883|>": 101180, "<|LOC_884|>": 101181, "<|LOC_885|>": 101182, "<|LOC_886|>": 101183, "<|LOC_887|>": 101184, "<|LOC_888|>": 101185, "<|LOC_889|>": 101186, "<|LOC_890|>": 101187, "<|LOC_891|>": 101188, "<|LOC_892|>": 101189, "<|LOC_893|>": 101190, "<|LOC_894|>": 101191, "<|LOC_895|>": 101192, "<|LOC_896|>": 101193, "<|LOC_897|>": 101194, "<|LOC_898|>": 101195, "<|LOC_899|>": 101196, "<|LOC_900|>": 101197, "<|LOC_901|>": 101198, "<|LOC_902|>": 101199, "<|LOC_903|>": 101200, "<|LOC_904|>": 101201, "<|LOC_905|>": 101202, "<|LOC_906|>": 101203, "<|LOC_907|>": 101204, "<|LOC_908|>": 101205, "<|LOC_909|>": 101206, "<|LOC_910|>": 101207, "<|LOC_911|>": 101208, "<|LOC_912|>": 101209, "<|LOC_913|>": 101210, "<|LOC_914|>": 101211, "<|LOC_915|>": 101212, "<|LOC_916|>": 101213, "<|LOC_917|>": 101214, "<|LOC_918|>": 101215, "<|LOC_919|>": 101216, "<|LOC_920|>": 101217, "<|LOC_921|>": 101218, "<|LOC_922|>": 101219, "<|LOC_923|>": 101220, "<|LOC_924|>": 101221, "<|LOC_925|>": 101222, "<|LOC_926|>": 101223, "<|LOC_927|>": 101224, "<|LOC_928|>": 101225, "<|LOC_929|>": 101226, "<|LOC_930|>": 101227, "<|LOC_931|>": 101228, "<|LOC_932|>": 101229, "<|LOC_933|>": 101230, "<|LOC_934|>": 101231, "<|LOC_935|>": 101232, "<|LOC_936|>": 101233, "<|LOC_937|>": 101234, "<|LOC_938|>": 101235, "<|LOC_939|>": 101236, "<|LOC_940|>": 101237, "<|LOC_941|>": 101238, "<|LOC_942|>": 101239, "<|LOC_943|>": 101240, "<|LOC_944|>": 101241, "<|LOC_945|>": 101242, "<|LOC_946|>": 101243, "<|LOC_947|>": 101244, "<|LOC_948|>": 101245, "<|LOC_949|>": 101246, "<|LOC_950|>": 101247, "<|LOC_951|>": 101248, "<|LOC_952|>": 101249, "<|LOC_953|>": 101250, "<|LOC_954|>": 101251, "<|LOC_955|>": 101252, "<|LOC_956|>": 101253, "<|LOC_957|>": 101254, "<|LOC_958|>": 101255, "<|LOC_959|>": 101256, "<|LOC_960|>": 101257, "<|LOC_961|>": 101258, "<|LOC_962|>": 101259, "<|LOC_963|>": 101260, "<|LOC_964|>": 101261, "<|LOC_965|>": 101262, "<|LOC_966|>": 101263, "<|LOC_967|>": 101264, "<|LOC_968|>": 101265, "<|LOC_969|>": 101266, "<|LOC_970|>": 101267, "<|LOC_971|>": 101268, "<|LOC_972|>": 101269, "<|LOC_973|>": 101270, "<|LOC_974|>": 101271, "<|LOC_975|>": 101272, "<|LOC_976|>": 101273, "<|LOC_977|>": 101274, "<|LOC_978|>": 101275, "<|LOC_979|>": 101276, "<|LOC_980|>": 101277, "<|LOC_981|>": 101278, "<|LOC_982|>": 101279, "<|LOC_983|>": 101280, "<|LOC_984|>": 101281, "<|LOC_985|>": 101282, "<|LOC_986|>": 101283, "<|LOC_987|>": 101284, "<|LOC_988|>": 101285, "<|LOC_989|>": 101286, "<|LOC_990|>": 101287, "<|LOC_991|>": 101288, "<|LOC_992|>": 101289, "<|LOC_993|>": 101290, "<|LOC_994|>": 101291, "<|LOC_995|>": 101292, "<|LOC_996|>": 101293, "<|LOC_997|>": 101294, "<|LOC_998|>": 101295, "<|LOC_999|>": 101296, "<|LOC_1000|>": 101297, "<|LOC_BEGIN|>": 101298, "<|LOC_END|>": 101299, "<|LOC_SEP|>": 101300, "<|CROP_COL_SEP|>": 101301, "<|CROP_ROW_SEP|>": 101302, "<|IMAGE_SEP|>": 101303, "<|image_pad|>": 101304, "<|IMAGE_START|>": 101305, "<|IMAGE_END|>": 101306, "<|video_pad|>": 101307, "<ecel>": 101308, "<fcel>": 101309, "<xcel>": 101310, "<lcel>": 101311, "<ucel>": 101312, "<nl>": 101313}
|
checkpoint-378/config.json
ADDED
|
@@ -0,0 +1,223 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"architectures": [
|
| 3 |
+
"PaddleOCRVLForConditionalGeneration"
|
| 4 |
+
],
|
| 5 |
+
"attention_probs_dropout_prob": 0.0,
|
| 6 |
+
"auto_map": {
|
| 7 |
+
"AutoConfig": "configuration_paddleocr_vl.PaddleOCRVLConfig",
|
| 8 |
+
"AutoModel": "modeling_paddleocr_vl.PaddleOCRVLForConditionalGeneration",
|
| 9 |
+
"AutoModelForCausalLM": "modeling_paddleocr_vl.PaddleOCRVLForConditionalGeneration"
|
| 10 |
+
},
|
| 11 |
+
"compression_ratio": 1.0,
|
| 12 |
+
"disable_pipeline_warmup": false,
|
| 13 |
+
"enable_mtp_magic_send": false,
|
| 14 |
+
"fp16_opt_level": "O2",
|
| 15 |
+
"freq_allocation": 20,
|
| 16 |
+
"fuse_ln": false,
|
| 17 |
+
"fuse_rms_norm": true,
|
| 18 |
+
"head_dim": 128,
|
| 19 |
+
"hidden_act": "silu",
|
| 20 |
+
"hidden_dropout_prob": 0.0,
|
| 21 |
+
"hidden_size": 1024,
|
| 22 |
+
"ignored_index": -100,
|
| 23 |
+
"im_patch_id": 100295,
|
| 24 |
+
"image_token_id": 100295,
|
| 25 |
+
"intermediate_size": 3072,
|
| 26 |
+
"max_position_embeddings": 131072,
|
| 27 |
+
"max_text_id": 100295,
|
| 28 |
+
"model_type": "paddleocr_vl",
|
| 29 |
+
"moe_dropout_prob": 0.0,
|
| 30 |
+
"moe_multimodal_dispatch_use_allgather": "v2-alltoall-unpad",
|
| 31 |
+
"num_attention_heads": 16,
|
| 32 |
+
"num_hidden_layers": 18,
|
| 33 |
+
"num_key_value_heads": 2,
|
| 34 |
+
"paddleformers_version": "0.4.0",
|
| 35 |
+
"pixel_hidden_size": 1152,
|
| 36 |
+
"rms_norm_eps": 1e-05,
|
| 37 |
+
"rope_is_neox_style": true,
|
| 38 |
+
"rope_scaling": {
|
| 39 |
+
"mrope_section": [
|
| 40 |
+
16,
|
| 41 |
+
24,
|
| 42 |
+
24
|
| 43 |
+
],
|
| 44 |
+
"rope_type": "default",
|
| 45 |
+
"type": "default"
|
| 46 |
+
},
|
| 47 |
+
"rope_theta": 500000,
|
| 48 |
+
"scale_qk_coeff": 1.0,
|
| 49 |
+
"seqlen": 16384,
|
| 50 |
+
"sliding_window": null,
|
| 51 |
+
"tie_word_embeddings": false,
|
| 52 |
+
"token_balance_loss": false,
|
| 53 |
+
"token_balance_seqlen": 16384,
|
| 54 |
+
"torch_dtype": "bfloat16",
|
| 55 |
+
"use_3d_rope": true,
|
| 56 |
+
"use_bias": false,
|
| 57 |
+
"use_flash_attn_with_mask": true,
|
| 58 |
+
"use_fp8": false,
|
| 59 |
+
"use_mem_eff_attn": true,
|
| 60 |
+
"use_recompute_moe": false,
|
| 61 |
+
"use_rmsnorm": true,
|
| 62 |
+
"video_token_id": 101307,
|
| 63 |
+
"vision_config": {
|
| 64 |
+
"_attn_implementation": "eager",
|
| 65 |
+
"_name_or_path": "",
|
| 66 |
+
"_save_to_hf": false,
|
| 67 |
+
"add_cross_attention": false,
|
| 68 |
+
"add_tail_layers": 0,
|
| 69 |
+
"architectures": [
|
| 70 |
+
"PaddleOCRVisionModel"
|
| 71 |
+
],
|
| 72 |
+
"attention_dropout": 0.0,
|
| 73 |
+
"auto_map": {
|
| 74 |
+
"AutoConfig": "configuration_paddleocr_vl.PaddleOCRVLConfig",
|
| 75 |
+
"AutoModel": "modeling_paddleocr_vl.PaddleOCRVisionModel"
|
| 76 |
+
},
|
| 77 |
+
"bad_words_ids": null,
|
| 78 |
+
"begin_suppress_tokens": null,
|
| 79 |
+
"bos_token_id": null,
|
| 80 |
+
"chunk_size_feed_forward": 0,
|
| 81 |
+
"classifier_dropout": null,
|
| 82 |
+
"context_parallel_degree": 1,
|
| 83 |
+
"cross_attention_hidden_size": null,
|
| 84 |
+
"decoder_start_token_id": null,
|
| 85 |
+
"diversity_penalty": 0.0,
|
| 86 |
+
"do_sample": false,
|
| 87 |
+
"dpo_config": null,
|
| 88 |
+
"dtype": "bfloat16",
|
| 89 |
+
"early_stopping": false,
|
| 90 |
+
"encoder_no_repeat_ngram_size": 0,
|
| 91 |
+
"eos_token_id": null,
|
| 92 |
+
"ep_communication_type": "deepep",
|
| 93 |
+
"exponential_decay_length_penalty": null,
|
| 94 |
+
"finetuning_task": null,
|
| 95 |
+
"forced_bos_token_id": null,
|
| 96 |
+
"forced_eos_token_id": null,
|
| 97 |
+
"fuse_attention_ffn": false,
|
| 98 |
+
"fuse_attention_qkv": false,
|
| 99 |
+
"fuse_linear": false,
|
| 100 |
+
"fuse_rope": false,
|
| 101 |
+
"fuse_sequence_parallel_allreduce": false,
|
| 102 |
+
"fuse_swiglu": false,
|
| 103 |
+
"hidden_act": "gelu_new",
|
| 104 |
+
"hidden_size": 1152,
|
| 105 |
+
"id2label": {
|
| 106 |
+
"0": "LABEL_0",
|
| 107 |
+
"1": "LABEL_1"
|
| 108 |
+
},
|
| 109 |
+
"image_size": 384,
|
| 110 |
+
"intermediate_size": 4304,
|
| 111 |
+
"is_decoder": false,
|
| 112 |
+
"is_encoder_decoder": false,
|
| 113 |
+
"kto_config": null,
|
| 114 |
+
"label2id": {
|
| 115 |
+
"LABEL_0": 0,
|
| 116 |
+
"LABEL_1": 1
|
| 117 |
+
},
|
| 118 |
+
"layer_norm_eps": 1e-06,
|
| 119 |
+
"length_penalty": 1.0,
|
| 120 |
+
"loss_subbatch_sequence_length": -1,
|
| 121 |
+
"max_length": 20,
|
| 122 |
+
"min_length": 0,
|
| 123 |
+
"model_type": "paddleocr_vl",
|
| 124 |
+
"moe_subbatch_token_num": 0,
|
| 125 |
+
"no_recompute_layers": null,
|
| 126 |
+
"no_repeat_ngram_size": 0,
|
| 127 |
+
"num_attention_heads": 16,
|
| 128 |
+
"num_beam_groups": 1,
|
| 129 |
+
"num_beams": 1,
|
| 130 |
+
"num_channels": 3,
|
| 131 |
+
"num_choices": null,
|
| 132 |
+
"num_hidden_layers": 27,
|
| 133 |
+
"num_nextn_predict_layers": 0,
|
| 134 |
+
"num_return_sequences": 1,
|
| 135 |
+
"offload_recompute_inputs": false,
|
| 136 |
+
"output_attentions": false,
|
| 137 |
+
"output_hidden_states": false,
|
| 138 |
+
"output_scores": false,
|
| 139 |
+
"pad_token_id": 0,
|
| 140 |
+
"patch_size": 14,
|
| 141 |
+
"pipeline_parallel_degree": 1,
|
| 142 |
+
"pp_recompute_interval": 1,
|
| 143 |
+
"prefix": null,
|
| 144 |
+
"problem_type": null,
|
| 145 |
+
"pruned_heads": {},
|
| 146 |
+
"quantization_config": {
|
| 147 |
+
"act_quant_method": "abs_max",
|
| 148 |
+
"activation_scheme": null,
|
| 149 |
+
"actscale_moving_rate": 0.01,
|
| 150 |
+
"apply_hadamard": false,
|
| 151 |
+
"apply_online_actscale_step": 200,
|
| 152 |
+
"dense_quant_type": "",
|
| 153 |
+
"dtype": null,
|
| 154 |
+
"fmt": null,
|
| 155 |
+
"fp8_format_type": "hybrid",
|
| 156 |
+
"group_size": -1,
|
| 157 |
+
"hadamard_block_size": 32,
|
| 158 |
+
"ignore_modules": null,
|
| 159 |
+
"llm_int8_threshold": 6.0,
|
| 160 |
+
"moe_quant_type": "",
|
| 161 |
+
"qlora_weight_blocksize": 64,
|
| 162 |
+
"qlora_weight_double_quant": false,
|
| 163 |
+
"qlora_weight_double_quant_block_size": 256,
|
| 164 |
+
"quant_input_grad": false,
|
| 165 |
+
"quant_method": null,
|
| 166 |
+
"quant_round_type": 0,
|
| 167 |
+
"quant_type": null,
|
| 168 |
+
"quant_weight_grad": false,
|
| 169 |
+
"quantization": "",
|
| 170 |
+
"scale_epsilon": 1e-08,
|
| 171 |
+
"shift": false,
|
| 172 |
+
"shift_smooth_all_linears": false,
|
| 173 |
+
"smooth": false,
|
| 174 |
+
"weight_block_size": null,
|
| 175 |
+
"weight_quant_method": "abs_max_channel_wise",
|
| 176 |
+
"weight_quantize_algo": null
|
| 177 |
+
},
|
| 178 |
+
"recompute": true,
|
| 179 |
+
"recompute_granularity": "full",
|
| 180 |
+
"recompute_use_reentrant": false,
|
| 181 |
+
"refined_recompute": "",
|
| 182 |
+
"remove_invalid_values": false,
|
| 183 |
+
"repetition_penalty": 1.0,
|
| 184 |
+
"return_dict": false,
|
| 185 |
+
"return_dict_in_generate": false,
|
| 186 |
+
"sep_parallel_degree": 1,
|
| 187 |
+
"sep_token_id": null,
|
| 188 |
+
"sequence_parallel": false,
|
| 189 |
+
"spatial_merge_size": 2,
|
| 190 |
+
"suppress_tokens": null,
|
| 191 |
+
"task_specific_params": null,
|
| 192 |
+
"temperature": 1.0,
|
| 193 |
+
"temporal_patch_size": 2,
|
| 194 |
+
"tensor_parallel_degree": 1,
|
| 195 |
+
"tensor_parallel_output": true,
|
| 196 |
+
"tensor_parallel_rank": 0,
|
| 197 |
+
"tie_encoder_decoder": false,
|
| 198 |
+
"tie_word_embeddings": true,
|
| 199 |
+
"tokenizer_class": null,
|
| 200 |
+
"tokens_per_second": 2,
|
| 201 |
+
"top_k": 50,
|
| 202 |
+
"top_p": 1.0,
|
| 203 |
+
"typical_p": 1.0,
|
| 204 |
+
"use_cache": false,
|
| 205 |
+
"use_filtered_label_loss": false,
|
| 206 |
+
"use_flash_attention": true,
|
| 207 |
+
"use_fused_dropout_add": false,
|
| 208 |
+
"use_fused_head_and_loss_fn": false,
|
| 209 |
+
"use_fused_linear": false,
|
| 210 |
+
"use_fused_linear_cross_entropy": false,
|
| 211 |
+
"use_fused_rms_norm": false,
|
| 212 |
+
"use_fused_rope": false,
|
| 213 |
+
"use_sparse_flash_attn": true,
|
| 214 |
+
"use_sparse_head_and_loss_fn": false,
|
| 215 |
+
"use_unified_moe": false,
|
| 216 |
+
"using_fake_gate": false,
|
| 217 |
+
"virtual_pp_degree": 1
|
| 218 |
+
},
|
| 219 |
+
"vision_end_token_id": 101306,
|
| 220 |
+
"vision_start_token_id": 101305,
|
| 221 |
+
"vocab_size": 103424,
|
| 222 |
+
"weight_share_add_bias": true
|
| 223 |
+
}
|
checkpoint-378/generation_config.json
ADDED
|
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_from_model_config": true,
|
| 3 |
+
"eos_token_id": 2,
|
| 4 |
+
"pad_token_id": 0,
|
| 5 |
+
"use_cache": false
|
| 6 |
+
}
|
checkpoint-378/master_weights-00001-of-00001.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:cd1c110b76954659e6dac8e032e963ae04742fa40d976974604898de01f72a71
|
| 3 |
+
size 3622483352
|
checkpoint-378/master_weights.safetensors.index.json
ADDED
|
@@ -0,0 +1,615 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"metadata": {
|
| 3 |
+
"total_size": 3622406592
|
| 4 |
+
},
|
| 5 |
+
"weight_map": {
|
| 6 |
+
"mlp_AR.pre_norm.weight": "master_weights-00001-of-00001.safetensors",
|
| 7 |
+
"mlp_AR.pre_norm.bias": "master_weights-00001-of-00001.safetensors",
|
| 8 |
+
"mlp_AR.linear_1.weight": "master_weights-00001-of-00001.safetensors",
|
| 9 |
+
"mlp_AR.linear_1.bias": "master_weights-00001-of-00001.safetensors",
|
| 10 |
+
"mlp_AR.linear_2.weight": "master_weights-00001-of-00001.safetensors",
|
| 11 |
+
"mlp_AR.linear_2.bias": "master_weights-00001-of-00001.safetensors",
|
| 12 |
+
"visual.vision_model.embeddings.patch_embedding.weight": "master_weights-00001-of-00001.safetensors",
|
| 13 |
+
"visual.vision_model.embeddings.patch_embedding.bias": "master_weights-00001-of-00001.safetensors",
|
| 14 |
+
"visual.vision_model.embeddings.position_embedding.weight": "master_weights-00001-of-00001.safetensors",
|
| 15 |
+
"visual.vision_model.encoder.layers.0.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 16 |
+
"visual.vision_model.encoder.layers.0.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 17 |
+
"visual.vision_model.encoder.layers.0.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 18 |
+
"visual.vision_model.encoder.layers.0.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 19 |
+
"visual.vision_model.encoder.layers.0.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 20 |
+
"visual.vision_model.encoder.layers.0.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 21 |
+
"visual.vision_model.encoder.layers.0.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 22 |
+
"visual.vision_model.encoder.layers.0.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 23 |
+
"visual.vision_model.encoder.layers.0.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 24 |
+
"visual.vision_model.encoder.layers.0.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 25 |
+
"visual.vision_model.encoder.layers.0.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 26 |
+
"visual.vision_model.encoder.layers.0.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 27 |
+
"visual.vision_model.encoder.layers.0.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 28 |
+
"visual.vision_model.encoder.layers.0.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 29 |
+
"visual.vision_model.encoder.layers.0.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 30 |
+
"visual.vision_model.encoder.layers.0.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 31 |
+
"visual.vision_model.encoder.layers.1.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 32 |
+
"visual.vision_model.encoder.layers.1.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 33 |
+
"visual.vision_model.encoder.layers.1.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 34 |
+
"visual.vision_model.encoder.layers.1.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 35 |
+
"visual.vision_model.encoder.layers.1.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 36 |
+
"visual.vision_model.encoder.layers.1.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 37 |
+
"visual.vision_model.encoder.layers.1.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 38 |
+
"visual.vision_model.encoder.layers.1.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 39 |
+
"visual.vision_model.encoder.layers.1.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 40 |
+
"visual.vision_model.encoder.layers.1.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 41 |
+
"visual.vision_model.encoder.layers.1.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 42 |
+
"visual.vision_model.encoder.layers.1.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 43 |
+
"visual.vision_model.encoder.layers.1.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 44 |
+
"visual.vision_model.encoder.layers.1.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 45 |
+
"visual.vision_model.encoder.layers.1.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 46 |
+
"visual.vision_model.encoder.layers.1.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 47 |
+
"visual.vision_model.encoder.layers.2.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 48 |
+
"visual.vision_model.encoder.layers.2.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 49 |
+
"visual.vision_model.encoder.layers.2.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 50 |
+
"visual.vision_model.encoder.layers.2.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 51 |
+
"visual.vision_model.encoder.layers.2.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 52 |
+
"visual.vision_model.encoder.layers.2.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 53 |
+
"visual.vision_model.encoder.layers.2.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 54 |
+
"visual.vision_model.encoder.layers.2.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 55 |
+
"visual.vision_model.encoder.layers.2.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 56 |
+
"visual.vision_model.encoder.layers.2.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 57 |
+
"visual.vision_model.encoder.layers.2.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 58 |
+
"visual.vision_model.encoder.layers.2.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 59 |
+
"visual.vision_model.encoder.layers.2.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 60 |
+
"visual.vision_model.encoder.layers.2.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 61 |
+
"visual.vision_model.encoder.layers.2.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 62 |
+
"visual.vision_model.encoder.layers.2.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 63 |
+
"visual.vision_model.encoder.layers.3.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 64 |
+
"visual.vision_model.encoder.layers.3.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 65 |
+
"visual.vision_model.encoder.layers.3.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 66 |
+
"visual.vision_model.encoder.layers.3.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 67 |
+
"visual.vision_model.encoder.layers.3.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 68 |
+
"visual.vision_model.encoder.layers.3.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 69 |
+
"visual.vision_model.encoder.layers.3.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 70 |
+
"visual.vision_model.encoder.layers.3.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 71 |
+
"visual.vision_model.encoder.layers.3.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 72 |
+
"visual.vision_model.encoder.layers.3.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 73 |
+
"visual.vision_model.encoder.layers.3.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 74 |
+
"visual.vision_model.encoder.layers.3.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 75 |
+
"visual.vision_model.encoder.layers.3.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 76 |
+
"visual.vision_model.encoder.layers.3.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 77 |
+
"visual.vision_model.encoder.layers.3.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 78 |
+
"visual.vision_model.encoder.layers.3.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 79 |
+
"visual.vision_model.encoder.layers.4.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 80 |
+
"visual.vision_model.encoder.layers.4.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 81 |
+
"visual.vision_model.encoder.layers.4.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 82 |
+
"visual.vision_model.encoder.layers.4.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 83 |
+
"visual.vision_model.encoder.layers.4.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 84 |
+
"visual.vision_model.encoder.layers.4.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 85 |
+
"visual.vision_model.encoder.layers.4.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 86 |
+
"visual.vision_model.encoder.layers.4.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 87 |
+
"visual.vision_model.encoder.layers.4.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 88 |
+
"visual.vision_model.encoder.layers.4.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 89 |
+
"visual.vision_model.encoder.layers.4.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 90 |
+
"visual.vision_model.encoder.layers.4.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 91 |
+
"visual.vision_model.encoder.layers.4.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 92 |
+
"visual.vision_model.encoder.layers.4.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 93 |
+
"visual.vision_model.encoder.layers.4.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 94 |
+
"visual.vision_model.encoder.layers.4.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 95 |
+
"visual.vision_model.encoder.layers.5.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 96 |
+
"visual.vision_model.encoder.layers.5.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 97 |
+
"visual.vision_model.encoder.layers.5.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 98 |
+
"visual.vision_model.encoder.layers.5.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 99 |
+
"visual.vision_model.encoder.layers.5.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 100 |
+
"visual.vision_model.encoder.layers.5.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 101 |
+
"visual.vision_model.encoder.layers.5.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 102 |
+
"visual.vision_model.encoder.layers.5.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 103 |
+
"visual.vision_model.encoder.layers.5.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 104 |
+
"visual.vision_model.encoder.layers.5.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 105 |
+
"visual.vision_model.encoder.layers.5.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 106 |
+
"visual.vision_model.encoder.layers.5.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 107 |
+
"visual.vision_model.encoder.layers.5.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 108 |
+
"visual.vision_model.encoder.layers.5.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 109 |
+
"visual.vision_model.encoder.layers.5.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 110 |
+
"visual.vision_model.encoder.layers.5.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 111 |
+
"visual.vision_model.encoder.layers.6.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 112 |
+
"visual.vision_model.encoder.layers.6.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 113 |
+
"visual.vision_model.encoder.layers.6.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 114 |
+
"visual.vision_model.encoder.layers.6.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 115 |
+
"visual.vision_model.encoder.layers.6.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 116 |
+
"visual.vision_model.encoder.layers.6.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 117 |
+
"visual.vision_model.encoder.layers.6.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 118 |
+
"visual.vision_model.encoder.layers.6.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 119 |
+
"visual.vision_model.encoder.layers.6.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 120 |
+
"visual.vision_model.encoder.layers.6.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 121 |
+
"visual.vision_model.encoder.layers.6.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 122 |
+
"visual.vision_model.encoder.layers.6.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 123 |
+
"visual.vision_model.encoder.layers.6.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 124 |
+
"visual.vision_model.encoder.layers.6.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 125 |
+
"visual.vision_model.encoder.layers.6.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 126 |
+
"visual.vision_model.encoder.layers.6.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 127 |
+
"visual.vision_model.encoder.layers.7.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 128 |
+
"visual.vision_model.encoder.layers.7.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 129 |
+
"visual.vision_model.encoder.layers.7.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 130 |
+
"visual.vision_model.encoder.layers.7.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 131 |
+
"visual.vision_model.encoder.layers.7.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 132 |
+
"visual.vision_model.encoder.layers.7.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 133 |
+
"visual.vision_model.encoder.layers.7.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 134 |
+
"visual.vision_model.encoder.layers.7.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 135 |
+
"visual.vision_model.encoder.layers.7.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 136 |
+
"visual.vision_model.encoder.layers.7.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 137 |
+
"visual.vision_model.encoder.layers.7.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 138 |
+
"visual.vision_model.encoder.layers.7.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 139 |
+
"visual.vision_model.encoder.layers.7.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 140 |
+
"visual.vision_model.encoder.layers.7.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 141 |
+
"visual.vision_model.encoder.layers.7.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 142 |
+
"visual.vision_model.encoder.layers.7.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 143 |
+
"visual.vision_model.encoder.layers.8.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 144 |
+
"visual.vision_model.encoder.layers.8.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 145 |
+
"visual.vision_model.encoder.layers.8.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 146 |
+
"visual.vision_model.encoder.layers.8.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 147 |
+
"visual.vision_model.encoder.layers.8.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 148 |
+
"visual.vision_model.encoder.layers.8.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 149 |
+
"visual.vision_model.encoder.layers.8.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 150 |
+
"visual.vision_model.encoder.layers.8.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 151 |
+
"visual.vision_model.encoder.layers.8.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 152 |
+
"visual.vision_model.encoder.layers.8.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 153 |
+
"visual.vision_model.encoder.layers.8.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 154 |
+
"visual.vision_model.encoder.layers.8.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 155 |
+
"visual.vision_model.encoder.layers.8.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 156 |
+
"visual.vision_model.encoder.layers.8.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 157 |
+
"visual.vision_model.encoder.layers.8.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 158 |
+
"visual.vision_model.encoder.layers.8.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 159 |
+
"visual.vision_model.encoder.layers.9.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 160 |
+
"visual.vision_model.encoder.layers.9.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 161 |
+
"visual.vision_model.encoder.layers.9.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 162 |
+
"visual.vision_model.encoder.layers.9.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 163 |
+
"visual.vision_model.encoder.layers.9.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 164 |
+
"visual.vision_model.encoder.layers.9.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 165 |
+
"visual.vision_model.encoder.layers.9.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 166 |
+
"visual.vision_model.encoder.layers.9.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 167 |
+
"visual.vision_model.encoder.layers.9.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 168 |
+
"visual.vision_model.encoder.layers.9.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 169 |
+
"visual.vision_model.encoder.layers.9.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 170 |
+
"visual.vision_model.encoder.layers.9.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 171 |
+
"visual.vision_model.encoder.layers.9.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 172 |
+
"visual.vision_model.encoder.layers.9.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 173 |
+
"visual.vision_model.encoder.layers.9.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 174 |
+
"visual.vision_model.encoder.layers.9.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 175 |
+
"visual.vision_model.encoder.layers.10.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 176 |
+
"visual.vision_model.encoder.layers.10.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 177 |
+
"visual.vision_model.encoder.layers.10.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 178 |
+
"visual.vision_model.encoder.layers.10.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 179 |
+
"visual.vision_model.encoder.layers.10.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 180 |
+
"visual.vision_model.encoder.layers.10.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 181 |
+
"visual.vision_model.encoder.layers.10.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 182 |
+
"visual.vision_model.encoder.layers.10.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 183 |
+
"visual.vision_model.encoder.layers.10.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 184 |
+
"visual.vision_model.encoder.layers.10.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 185 |
+
"visual.vision_model.encoder.layers.10.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 186 |
+
"visual.vision_model.encoder.layers.10.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 187 |
+
"visual.vision_model.encoder.layers.10.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 188 |
+
"visual.vision_model.encoder.layers.10.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 189 |
+
"visual.vision_model.encoder.layers.10.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 190 |
+
"visual.vision_model.encoder.layers.10.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 191 |
+
"visual.vision_model.encoder.layers.11.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 192 |
+
"visual.vision_model.encoder.layers.11.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 193 |
+
"visual.vision_model.encoder.layers.11.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 194 |
+
"visual.vision_model.encoder.layers.11.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 195 |
+
"visual.vision_model.encoder.layers.11.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 196 |
+
"visual.vision_model.encoder.layers.11.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 197 |
+
"visual.vision_model.encoder.layers.11.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 198 |
+
"visual.vision_model.encoder.layers.11.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 199 |
+
"visual.vision_model.encoder.layers.11.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 200 |
+
"visual.vision_model.encoder.layers.11.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 201 |
+
"visual.vision_model.encoder.layers.11.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 202 |
+
"visual.vision_model.encoder.layers.11.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 203 |
+
"visual.vision_model.encoder.layers.11.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 204 |
+
"visual.vision_model.encoder.layers.11.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 205 |
+
"visual.vision_model.encoder.layers.11.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 206 |
+
"visual.vision_model.encoder.layers.11.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 207 |
+
"visual.vision_model.encoder.layers.12.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 208 |
+
"visual.vision_model.encoder.layers.12.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 209 |
+
"visual.vision_model.encoder.layers.12.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 210 |
+
"visual.vision_model.encoder.layers.12.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 211 |
+
"visual.vision_model.encoder.layers.12.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 212 |
+
"visual.vision_model.encoder.layers.12.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 213 |
+
"visual.vision_model.encoder.layers.12.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 214 |
+
"visual.vision_model.encoder.layers.12.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 215 |
+
"visual.vision_model.encoder.layers.12.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 216 |
+
"visual.vision_model.encoder.layers.12.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 217 |
+
"visual.vision_model.encoder.layers.12.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 218 |
+
"visual.vision_model.encoder.layers.12.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 219 |
+
"visual.vision_model.encoder.layers.12.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 220 |
+
"visual.vision_model.encoder.layers.12.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 221 |
+
"visual.vision_model.encoder.layers.12.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 222 |
+
"visual.vision_model.encoder.layers.12.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 223 |
+
"visual.vision_model.encoder.layers.13.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 224 |
+
"visual.vision_model.encoder.layers.13.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 225 |
+
"visual.vision_model.encoder.layers.13.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 226 |
+
"visual.vision_model.encoder.layers.13.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 227 |
+
"visual.vision_model.encoder.layers.13.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 228 |
+
"visual.vision_model.encoder.layers.13.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 229 |
+
"visual.vision_model.encoder.layers.13.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 230 |
+
"visual.vision_model.encoder.layers.13.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 231 |
+
"visual.vision_model.encoder.layers.13.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 232 |
+
"visual.vision_model.encoder.layers.13.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 233 |
+
"visual.vision_model.encoder.layers.13.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 234 |
+
"visual.vision_model.encoder.layers.13.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 235 |
+
"visual.vision_model.encoder.layers.13.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 236 |
+
"visual.vision_model.encoder.layers.13.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 237 |
+
"visual.vision_model.encoder.layers.13.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 238 |
+
"visual.vision_model.encoder.layers.13.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 239 |
+
"visual.vision_model.encoder.layers.14.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 240 |
+
"visual.vision_model.encoder.layers.14.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 241 |
+
"visual.vision_model.encoder.layers.14.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 242 |
+
"visual.vision_model.encoder.layers.14.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 243 |
+
"visual.vision_model.encoder.layers.14.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 244 |
+
"visual.vision_model.encoder.layers.14.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 245 |
+
"visual.vision_model.encoder.layers.14.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 246 |
+
"visual.vision_model.encoder.layers.14.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 247 |
+
"visual.vision_model.encoder.layers.14.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 248 |
+
"visual.vision_model.encoder.layers.14.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 249 |
+
"visual.vision_model.encoder.layers.14.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 250 |
+
"visual.vision_model.encoder.layers.14.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 251 |
+
"visual.vision_model.encoder.layers.14.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 252 |
+
"visual.vision_model.encoder.layers.14.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 253 |
+
"visual.vision_model.encoder.layers.14.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 254 |
+
"visual.vision_model.encoder.layers.14.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 255 |
+
"visual.vision_model.encoder.layers.15.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 256 |
+
"visual.vision_model.encoder.layers.15.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 257 |
+
"visual.vision_model.encoder.layers.15.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 258 |
+
"visual.vision_model.encoder.layers.15.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 259 |
+
"visual.vision_model.encoder.layers.15.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 260 |
+
"visual.vision_model.encoder.layers.15.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 261 |
+
"visual.vision_model.encoder.layers.15.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 262 |
+
"visual.vision_model.encoder.layers.15.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 263 |
+
"visual.vision_model.encoder.layers.15.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 264 |
+
"visual.vision_model.encoder.layers.15.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 265 |
+
"visual.vision_model.encoder.layers.15.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 266 |
+
"visual.vision_model.encoder.layers.15.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 267 |
+
"visual.vision_model.encoder.layers.15.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 268 |
+
"visual.vision_model.encoder.layers.15.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 269 |
+
"visual.vision_model.encoder.layers.15.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 270 |
+
"visual.vision_model.encoder.layers.15.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 271 |
+
"visual.vision_model.encoder.layers.16.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 272 |
+
"visual.vision_model.encoder.layers.16.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 273 |
+
"visual.vision_model.encoder.layers.16.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 274 |
+
"visual.vision_model.encoder.layers.16.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 275 |
+
"visual.vision_model.encoder.layers.16.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 276 |
+
"visual.vision_model.encoder.layers.16.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 277 |
+
"visual.vision_model.encoder.layers.16.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 278 |
+
"visual.vision_model.encoder.layers.16.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 279 |
+
"visual.vision_model.encoder.layers.16.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 280 |
+
"visual.vision_model.encoder.layers.16.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 281 |
+
"visual.vision_model.encoder.layers.16.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 282 |
+
"visual.vision_model.encoder.layers.16.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 283 |
+
"visual.vision_model.encoder.layers.16.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 284 |
+
"visual.vision_model.encoder.layers.16.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 285 |
+
"visual.vision_model.encoder.layers.16.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 286 |
+
"visual.vision_model.encoder.layers.16.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 287 |
+
"visual.vision_model.encoder.layers.17.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 288 |
+
"visual.vision_model.encoder.layers.17.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 289 |
+
"visual.vision_model.encoder.layers.17.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 290 |
+
"visual.vision_model.encoder.layers.17.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 291 |
+
"visual.vision_model.encoder.layers.17.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 292 |
+
"visual.vision_model.encoder.layers.17.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 293 |
+
"visual.vision_model.encoder.layers.17.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 294 |
+
"visual.vision_model.encoder.layers.17.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 295 |
+
"visual.vision_model.encoder.layers.17.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 296 |
+
"visual.vision_model.encoder.layers.17.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 297 |
+
"visual.vision_model.encoder.layers.17.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 298 |
+
"visual.vision_model.encoder.layers.17.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 299 |
+
"visual.vision_model.encoder.layers.17.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 300 |
+
"visual.vision_model.encoder.layers.17.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 301 |
+
"visual.vision_model.encoder.layers.17.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 302 |
+
"visual.vision_model.encoder.layers.17.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 303 |
+
"visual.vision_model.encoder.layers.18.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 304 |
+
"visual.vision_model.encoder.layers.18.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 305 |
+
"visual.vision_model.encoder.layers.18.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 306 |
+
"visual.vision_model.encoder.layers.18.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 307 |
+
"visual.vision_model.encoder.layers.18.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 308 |
+
"visual.vision_model.encoder.layers.18.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 309 |
+
"visual.vision_model.encoder.layers.18.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 310 |
+
"visual.vision_model.encoder.layers.18.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 311 |
+
"visual.vision_model.encoder.layers.18.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 312 |
+
"visual.vision_model.encoder.layers.18.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 313 |
+
"visual.vision_model.encoder.layers.18.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 314 |
+
"visual.vision_model.encoder.layers.18.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 315 |
+
"visual.vision_model.encoder.layers.18.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 316 |
+
"visual.vision_model.encoder.layers.18.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 317 |
+
"visual.vision_model.encoder.layers.18.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 318 |
+
"visual.vision_model.encoder.layers.18.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 319 |
+
"visual.vision_model.encoder.layers.19.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 320 |
+
"visual.vision_model.encoder.layers.19.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 321 |
+
"visual.vision_model.encoder.layers.19.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 322 |
+
"visual.vision_model.encoder.layers.19.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 323 |
+
"visual.vision_model.encoder.layers.19.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 324 |
+
"visual.vision_model.encoder.layers.19.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 325 |
+
"visual.vision_model.encoder.layers.19.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 326 |
+
"visual.vision_model.encoder.layers.19.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 327 |
+
"visual.vision_model.encoder.layers.19.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 328 |
+
"visual.vision_model.encoder.layers.19.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 329 |
+
"visual.vision_model.encoder.layers.19.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 330 |
+
"visual.vision_model.encoder.layers.19.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 331 |
+
"visual.vision_model.encoder.layers.19.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 332 |
+
"visual.vision_model.encoder.layers.19.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 333 |
+
"visual.vision_model.encoder.layers.19.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 334 |
+
"visual.vision_model.encoder.layers.19.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 335 |
+
"visual.vision_model.encoder.layers.20.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 336 |
+
"visual.vision_model.encoder.layers.20.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 337 |
+
"visual.vision_model.encoder.layers.20.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 338 |
+
"visual.vision_model.encoder.layers.20.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 339 |
+
"visual.vision_model.encoder.layers.20.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 340 |
+
"visual.vision_model.encoder.layers.20.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 341 |
+
"visual.vision_model.encoder.layers.20.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 342 |
+
"visual.vision_model.encoder.layers.20.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 343 |
+
"visual.vision_model.encoder.layers.20.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 344 |
+
"visual.vision_model.encoder.layers.20.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 345 |
+
"visual.vision_model.encoder.layers.20.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 346 |
+
"visual.vision_model.encoder.layers.20.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 347 |
+
"visual.vision_model.encoder.layers.20.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 348 |
+
"visual.vision_model.encoder.layers.20.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 349 |
+
"visual.vision_model.encoder.layers.20.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 350 |
+
"visual.vision_model.encoder.layers.20.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 351 |
+
"visual.vision_model.encoder.layers.21.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 352 |
+
"visual.vision_model.encoder.layers.21.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 353 |
+
"visual.vision_model.encoder.layers.21.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 354 |
+
"visual.vision_model.encoder.layers.21.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 355 |
+
"visual.vision_model.encoder.layers.21.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 356 |
+
"visual.vision_model.encoder.layers.21.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 357 |
+
"visual.vision_model.encoder.layers.21.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 358 |
+
"visual.vision_model.encoder.layers.21.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 359 |
+
"visual.vision_model.encoder.layers.21.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 360 |
+
"visual.vision_model.encoder.layers.21.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 361 |
+
"visual.vision_model.encoder.layers.21.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 362 |
+
"visual.vision_model.encoder.layers.21.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 363 |
+
"visual.vision_model.encoder.layers.21.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 364 |
+
"visual.vision_model.encoder.layers.21.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 365 |
+
"visual.vision_model.encoder.layers.21.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 366 |
+
"visual.vision_model.encoder.layers.21.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 367 |
+
"visual.vision_model.encoder.layers.22.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 368 |
+
"visual.vision_model.encoder.layers.22.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 369 |
+
"visual.vision_model.encoder.layers.22.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 370 |
+
"visual.vision_model.encoder.layers.22.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 371 |
+
"visual.vision_model.encoder.layers.22.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 372 |
+
"visual.vision_model.encoder.layers.22.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 373 |
+
"visual.vision_model.encoder.layers.22.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 374 |
+
"visual.vision_model.encoder.layers.22.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 375 |
+
"visual.vision_model.encoder.layers.22.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 376 |
+
"visual.vision_model.encoder.layers.22.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 377 |
+
"visual.vision_model.encoder.layers.22.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 378 |
+
"visual.vision_model.encoder.layers.22.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 379 |
+
"visual.vision_model.encoder.layers.22.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 380 |
+
"visual.vision_model.encoder.layers.22.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 381 |
+
"visual.vision_model.encoder.layers.22.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 382 |
+
"visual.vision_model.encoder.layers.22.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 383 |
+
"visual.vision_model.encoder.layers.23.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 384 |
+
"visual.vision_model.encoder.layers.23.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 385 |
+
"visual.vision_model.encoder.layers.23.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 386 |
+
"visual.vision_model.encoder.layers.23.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 387 |
+
"visual.vision_model.encoder.layers.23.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 388 |
+
"visual.vision_model.encoder.layers.23.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 389 |
+
"visual.vision_model.encoder.layers.23.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 390 |
+
"visual.vision_model.encoder.layers.23.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 391 |
+
"visual.vision_model.encoder.layers.23.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 392 |
+
"visual.vision_model.encoder.layers.23.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 393 |
+
"visual.vision_model.encoder.layers.23.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 394 |
+
"visual.vision_model.encoder.layers.23.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 395 |
+
"visual.vision_model.encoder.layers.23.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 396 |
+
"visual.vision_model.encoder.layers.23.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 397 |
+
"visual.vision_model.encoder.layers.23.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 398 |
+
"visual.vision_model.encoder.layers.23.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 399 |
+
"visual.vision_model.encoder.layers.24.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 400 |
+
"visual.vision_model.encoder.layers.24.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 401 |
+
"visual.vision_model.encoder.layers.24.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 402 |
+
"visual.vision_model.encoder.layers.24.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 403 |
+
"visual.vision_model.encoder.layers.24.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 404 |
+
"visual.vision_model.encoder.layers.24.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 405 |
+
"visual.vision_model.encoder.layers.24.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 406 |
+
"visual.vision_model.encoder.layers.24.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 407 |
+
"visual.vision_model.encoder.layers.24.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 408 |
+
"visual.vision_model.encoder.layers.24.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 409 |
+
"visual.vision_model.encoder.layers.24.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 410 |
+
"visual.vision_model.encoder.layers.24.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 411 |
+
"visual.vision_model.encoder.layers.24.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 412 |
+
"visual.vision_model.encoder.layers.24.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 413 |
+
"visual.vision_model.encoder.layers.24.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 414 |
+
"visual.vision_model.encoder.layers.24.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 415 |
+
"visual.vision_model.encoder.layers.25.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 416 |
+
"visual.vision_model.encoder.layers.25.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 417 |
+
"visual.vision_model.encoder.layers.25.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 418 |
+
"visual.vision_model.encoder.layers.25.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 419 |
+
"visual.vision_model.encoder.layers.25.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 420 |
+
"visual.vision_model.encoder.layers.25.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 421 |
+
"visual.vision_model.encoder.layers.25.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 422 |
+
"visual.vision_model.encoder.layers.25.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 423 |
+
"visual.vision_model.encoder.layers.25.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 424 |
+
"visual.vision_model.encoder.layers.25.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 425 |
+
"visual.vision_model.encoder.layers.25.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 426 |
+
"visual.vision_model.encoder.layers.25.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 427 |
+
"visual.vision_model.encoder.layers.25.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 428 |
+
"visual.vision_model.encoder.layers.25.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 429 |
+
"visual.vision_model.encoder.layers.25.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 430 |
+
"visual.vision_model.encoder.layers.25.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 431 |
+
"visual.vision_model.encoder.layers.26.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 432 |
+
"visual.vision_model.encoder.layers.26.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 433 |
+
"visual.vision_model.encoder.layers.26.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 434 |
+
"visual.vision_model.encoder.layers.26.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 435 |
+
"visual.vision_model.encoder.layers.26.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 436 |
+
"visual.vision_model.encoder.layers.26.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 437 |
+
"visual.vision_model.encoder.layers.26.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 438 |
+
"visual.vision_model.encoder.layers.26.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 439 |
+
"visual.vision_model.encoder.layers.26.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 440 |
+
"visual.vision_model.encoder.layers.26.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 441 |
+
"visual.vision_model.encoder.layers.26.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 442 |
+
"visual.vision_model.encoder.layers.26.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 443 |
+
"visual.vision_model.encoder.layers.26.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 444 |
+
"visual.vision_model.encoder.layers.26.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 445 |
+
"visual.vision_model.encoder.layers.26.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 446 |
+
"visual.vision_model.encoder.layers.26.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 447 |
+
"visual.vision_model.post_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 448 |
+
"visual.vision_model.post_layernorm.bias": "master_weights-00001-of-00001.safetensors",
|
| 449 |
+
"model.embed_tokens.weight": "master_weights-00001-of-00001.safetensors",
|
| 450 |
+
"model.layers.0.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 451 |
+
"model.layers.0.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 452 |
+
"model.layers.0.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 453 |
+
"model.layers.0.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 454 |
+
"model.layers.0.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 455 |
+
"model.layers.0.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 456 |
+
"model.layers.0.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 457 |
+
"model.layers.0.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 458 |
+
"model.layers.0.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 459 |
+
"model.layers.1.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 460 |
+
"model.layers.1.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 461 |
+
"model.layers.1.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 462 |
+
"model.layers.1.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 463 |
+
"model.layers.1.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 464 |
+
"model.layers.1.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 465 |
+
"model.layers.1.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 466 |
+
"model.layers.1.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 467 |
+
"model.layers.1.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 468 |
+
"model.layers.2.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 469 |
+
"model.layers.2.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 470 |
+
"model.layers.2.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 471 |
+
"model.layers.2.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 472 |
+
"model.layers.2.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 473 |
+
"model.layers.2.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 474 |
+
"model.layers.2.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 475 |
+
"model.layers.2.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 476 |
+
"model.layers.2.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 477 |
+
"model.layers.3.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 478 |
+
"model.layers.3.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 479 |
+
"model.layers.3.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 480 |
+
"model.layers.3.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 481 |
+
"model.layers.3.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 482 |
+
"model.layers.3.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 483 |
+
"model.layers.3.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 484 |
+
"model.layers.3.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 485 |
+
"model.layers.3.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 486 |
+
"model.layers.4.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 487 |
+
"model.layers.4.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 488 |
+
"model.layers.4.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 489 |
+
"model.layers.4.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 490 |
+
"model.layers.4.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 491 |
+
"model.layers.4.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 492 |
+
"model.layers.4.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 493 |
+
"model.layers.4.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 494 |
+
"model.layers.4.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 495 |
+
"model.layers.5.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 496 |
+
"model.layers.5.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 497 |
+
"model.layers.5.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 498 |
+
"model.layers.5.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 499 |
+
"model.layers.5.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 500 |
+
"model.layers.5.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 501 |
+
"model.layers.5.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 502 |
+
"model.layers.5.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 503 |
+
"model.layers.5.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 504 |
+
"model.layers.6.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 505 |
+
"model.layers.6.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 506 |
+
"model.layers.6.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 507 |
+
"model.layers.6.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 508 |
+
"model.layers.6.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 509 |
+
"model.layers.6.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 510 |
+
"model.layers.6.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 511 |
+
"model.layers.6.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 512 |
+
"model.layers.6.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 513 |
+
"model.layers.7.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 514 |
+
"model.layers.7.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 515 |
+
"model.layers.7.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 516 |
+
"model.layers.7.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 517 |
+
"model.layers.7.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 518 |
+
"model.layers.7.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 519 |
+
"model.layers.7.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 520 |
+
"model.layers.7.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 521 |
+
"model.layers.7.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 522 |
+
"model.layers.8.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 523 |
+
"model.layers.8.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 524 |
+
"model.layers.8.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 525 |
+
"model.layers.8.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 526 |
+
"model.layers.8.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 527 |
+
"model.layers.8.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 528 |
+
"model.layers.8.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 529 |
+
"model.layers.8.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 530 |
+
"model.layers.8.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 531 |
+
"model.layers.9.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 532 |
+
"model.layers.9.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 533 |
+
"model.layers.9.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 534 |
+
"model.layers.9.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 535 |
+
"model.layers.9.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 536 |
+
"model.layers.9.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 537 |
+
"model.layers.9.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 538 |
+
"model.layers.9.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 539 |
+
"model.layers.9.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 540 |
+
"model.layers.10.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 541 |
+
"model.layers.10.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 542 |
+
"model.layers.10.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 543 |
+
"model.layers.10.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 544 |
+
"model.layers.10.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 545 |
+
"model.layers.10.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 546 |
+
"model.layers.10.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 547 |
+
"model.layers.10.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 548 |
+
"model.layers.10.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 549 |
+
"model.layers.11.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 550 |
+
"model.layers.11.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 551 |
+
"model.layers.11.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 552 |
+
"model.layers.11.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 553 |
+
"model.layers.11.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 554 |
+
"model.layers.11.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 555 |
+
"model.layers.11.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 556 |
+
"model.layers.11.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 557 |
+
"model.layers.11.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 558 |
+
"model.layers.12.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 559 |
+
"model.layers.12.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 560 |
+
"model.layers.12.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 561 |
+
"model.layers.12.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 562 |
+
"model.layers.12.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 563 |
+
"model.layers.12.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 564 |
+
"model.layers.12.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 565 |
+
"model.layers.12.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 566 |
+
"model.layers.12.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 567 |
+
"model.layers.13.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 568 |
+
"model.layers.13.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 569 |
+
"model.layers.13.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 570 |
+
"model.layers.13.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 571 |
+
"model.layers.13.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 572 |
+
"model.layers.13.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 573 |
+
"model.layers.13.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 574 |
+
"model.layers.13.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 575 |
+
"model.layers.13.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 576 |
+
"model.layers.14.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 577 |
+
"model.layers.14.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 578 |
+
"model.layers.14.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 579 |
+
"model.layers.14.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 580 |
+
"model.layers.14.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 581 |
+
"model.layers.14.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 582 |
+
"model.layers.14.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 583 |
+
"model.layers.14.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 584 |
+
"model.layers.14.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 585 |
+
"model.layers.15.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 586 |
+
"model.layers.15.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 587 |
+
"model.layers.15.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 588 |
+
"model.layers.15.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 589 |
+
"model.layers.15.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 590 |
+
"model.layers.15.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 591 |
+
"model.layers.15.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 592 |
+
"model.layers.15.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 593 |
+
"model.layers.15.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 594 |
+
"model.layers.16.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 595 |
+
"model.layers.16.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 596 |
+
"model.layers.16.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 597 |
+
"model.layers.16.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 598 |
+
"model.layers.16.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 599 |
+
"model.layers.16.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 600 |
+
"model.layers.16.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 601 |
+
"model.layers.16.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 602 |
+
"model.layers.16.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 603 |
+
"model.layers.17.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 604 |
+
"model.layers.17.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 605 |
+
"model.layers.17.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 606 |
+
"model.layers.17.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 607 |
+
"model.layers.17.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 608 |
+
"model.layers.17.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 609 |
+
"model.layers.17.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 610 |
+
"model.layers.17.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 611 |
+
"model.layers.17.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 612 |
+
"model.norm.weight": "master_weights-00001-of-00001.safetensors",
|
| 613 |
+
"lm_head.weight": "master_weights-00001-of-00001.safetensors"
|
| 614 |
+
}
|
| 615 |
+
}
|
checkpoint-378/model-00001-of-00001.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b1daec2a5420a9168e38e1ee07bc43febae885c45dbb49f81d30ab572eb25686
|
| 3 |
+
size 1917255968
|
checkpoint-378/model.safetensors.index.json
ADDED
|
@@ -0,0 +1,627 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"metadata": {
|
| 3 |
+
"total_size": 1917177472
|
| 4 |
+
},
|
| 5 |
+
"weight_map": {
|
| 6 |
+
"mlp_AR.pre_norm.weight": "model-00001-of-00001.safetensors",
|
| 7 |
+
"mlp_AR.pre_norm.bias": "model-00001-of-00001.safetensors",
|
| 8 |
+
"mlp_AR.linear_1.weight": "model-00001-of-00001.safetensors",
|
| 9 |
+
"mlp_AR.linear_1.bias": "model-00001-of-00001.safetensors",
|
| 10 |
+
"mlp_AR.linear_2.weight": "model-00001-of-00001.safetensors",
|
| 11 |
+
"mlp_AR.linear_2.bias": "model-00001-of-00001.safetensors",
|
| 12 |
+
"visual.vision_model.embeddings.patch_embedding.weight": "model-00001-of-00001.safetensors",
|
| 13 |
+
"visual.vision_model.embeddings.patch_embedding.bias": "model-00001-of-00001.safetensors",
|
| 14 |
+
"visual.vision_model.embeddings.position_embedding.weight": "model-00001-of-00001.safetensors",
|
| 15 |
+
"visual.vision_model.embeddings.packing_position_embedding.weight": "model-00001-of-00001.safetensors",
|
| 16 |
+
"visual.vision_model.encoder.layers.0.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 17 |
+
"visual.vision_model.encoder.layers.0.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 18 |
+
"visual.vision_model.encoder.layers.0.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 19 |
+
"visual.vision_model.encoder.layers.0.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 20 |
+
"visual.vision_model.encoder.layers.0.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 21 |
+
"visual.vision_model.encoder.layers.0.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 22 |
+
"visual.vision_model.encoder.layers.0.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 23 |
+
"visual.vision_model.encoder.layers.0.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 24 |
+
"visual.vision_model.encoder.layers.0.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 25 |
+
"visual.vision_model.encoder.layers.0.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 26 |
+
"visual.vision_model.encoder.layers.0.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 27 |
+
"visual.vision_model.encoder.layers.0.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 28 |
+
"visual.vision_model.encoder.layers.0.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 29 |
+
"visual.vision_model.encoder.layers.0.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 30 |
+
"visual.vision_model.encoder.layers.0.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 31 |
+
"visual.vision_model.encoder.layers.0.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 32 |
+
"visual.vision_model.encoder.layers.1.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 33 |
+
"visual.vision_model.encoder.layers.1.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 34 |
+
"visual.vision_model.encoder.layers.1.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 35 |
+
"visual.vision_model.encoder.layers.1.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 36 |
+
"visual.vision_model.encoder.layers.1.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 37 |
+
"visual.vision_model.encoder.layers.1.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 38 |
+
"visual.vision_model.encoder.layers.1.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 39 |
+
"visual.vision_model.encoder.layers.1.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 40 |
+
"visual.vision_model.encoder.layers.1.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 41 |
+
"visual.vision_model.encoder.layers.1.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 42 |
+
"visual.vision_model.encoder.layers.1.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 43 |
+
"visual.vision_model.encoder.layers.1.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 44 |
+
"visual.vision_model.encoder.layers.1.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 45 |
+
"visual.vision_model.encoder.layers.1.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 46 |
+
"visual.vision_model.encoder.layers.1.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 47 |
+
"visual.vision_model.encoder.layers.1.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 48 |
+
"visual.vision_model.encoder.layers.2.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 49 |
+
"visual.vision_model.encoder.layers.2.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 50 |
+
"visual.vision_model.encoder.layers.2.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 51 |
+
"visual.vision_model.encoder.layers.2.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 52 |
+
"visual.vision_model.encoder.layers.2.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 53 |
+
"visual.vision_model.encoder.layers.2.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 54 |
+
"visual.vision_model.encoder.layers.2.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 55 |
+
"visual.vision_model.encoder.layers.2.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 56 |
+
"visual.vision_model.encoder.layers.2.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 57 |
+
"visual.vision_model.encoder.layers.2.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 58 |
+
"visual.vision_model.encoder.layers.2.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 59 |
+
"visual.vision_model.encoder.layers.2.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 60 |
+
"visual.vision_model.encoder.layers.2.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 61 |
+
"visual.vision_model.encoder.layers.2.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 62 |
+
"visual.vision_model.encoder.layers.2.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 63 |
+
"visual.vision_model.encoder.layers.2.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 64 |
+
"visual.vision_model.encoder.layers.3.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 65 |
+
"visual.vision_model.encoder.layers.3.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 66 |
+
"visual.vision_model.encoder.layers.3.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 67 |
+
"visual.vision_model.encoder.layers.3.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 68 |
+
"visual.vision_model.encoder.layers.3.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 69 |
+
"visual.vision_model.encoder.layers.3.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 70 |
+
"visual.vision_model.encoder.layers.3.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 71 |
+
"visual.vision_model.encoder.layers.3.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 72 |
+
"visual.vision_model.encoder.layers.3.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 73 |
+
"visual.vision_model.encoder.layers.3.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 74 |
+
"visual.vision_model.encoder.layers.3.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 75 |
+
"visual.vision_model.encoder.layers.3.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 76 |
+
"visual.vision_model.encoder.layers.3.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 77 |
+
"visual.vision_model.encoder.layers.3.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 78 |
+
"visual.vision_model.encoder.layers.3.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 79 |
+
"visual.vision_model.encoder.layers.3.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 80 |
+
"visual.vision_model.encoder.layers.4.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 81 |
+
"visual.vision_model.encoder.layers.4.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 82 |
+
"visual.vision_model.encoder.layers.4.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 83 |
+
"visual.vision_model.encoder.layers.4.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 84 |
+
"visual.vision_model.encoder.layers.4.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 85 |
+
"visual.vision_model.encoder.layers.4.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 86 |
+
"visual.vision_model.encoder.layers.4.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 87 |
+
"visual.vision_model.encoder.layers.4.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 88 |
+
"visual.vision_model.encoder.layers.4.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 89 |
+
"visual.vision_model.encoder.layers.4.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 90 |
+
"visual.vision_model.encoder.layers.4.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 91 |
+
"visual.vision_model.encoder.layers.4.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 92 |
+
"visual.vision_model.encoder.layers.4.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 93 |
+
"visual.vision_model.encoder.layers.4.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 94 |
+
"visual.vision_model.encoder.layers.4.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 95 |
+
"visual.vision_model.encoder.layers.4.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 96 |
+
"visual.vision_model.encoder.layers.5.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 97 |
+
"visual.vision_model.encoder.layers.5.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 98 |
+
"visual.vision_model.encoder.layers.5.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 99 |
+
"visual.vision_model.encoder.layers.5.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 100 |
+
"visual.vision_model.encoder.layers.5.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 101 |
+
"visual.vision_model.encoder.layers.5.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 102 |
+
"visual.vision_model.encoder.layers.5.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 103 |
+
"visual.vision_model.encoder.layers.5.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 104 |
+
"visual.vision_model.encoder.layers.5.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 105 |
+
"visual.vision_model.encoder.layers.5.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 106 |
+
"visual.vision_model.encoder.layers.5.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 107 |
+
"visual.vision_model.encoder.layers.5.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 108 |
+
"visual.vision_model.encoder.layers.5.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 109 |
+
"visual.vision_model.encoder.layers.5.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 110 |
+
"visual.vision_model.encoder.layers.5.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 111 |
+
"visual.vision_model.encoder.layers.5.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 112 |
+
"visual.vision_model.encoder.layers.6.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 113 |
+
"visual.vision_model.encoder.layers.6.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 114 |
+
"visual.vision_model.encoder.layers.6.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 115 |
+
"visual.vision_model.encoder.layers.6.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 116 |
+
"visual.vision_model.encoder.layers.6.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 117 |
+
"visual.vision_model.encoder.layers.6.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 118 |
+
"visual.vision_model.encoder.layers.6.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 119 |
+
"visual.vision_model.encoder.layers.6.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 120 |
+
"visual.vision_model.encoder.layers.6.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 121 |
+
"visual.vision_model.encoder.layers.6.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 122 |
+
"visual.vision_model.encoder.layers.6.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 123 |
+
"visual.vision_model.encoder.layers.6.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 124 |
+
"visual.vision_model.encoder.layers.6.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 125 |
+
"visual.vision_model.encoder.layers.6.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 126 |
+
"visual.vision_model.encoder.layers.6.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 127 |
+
"visual.vision_model.encoder.layers.6.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 128 |
+
"visual.vision_model.encoder.layers.7.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 129 |
+
"visual.vision_model.encoder.layers.7.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 130 |
+
"visual.vision_model.encoder.layers.7.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 131 |
+
"visual.vision_model.encoder.layers.7.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 132 |
+
"visual.vision_model.encoder.layers.7.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 133 |
+
"visual.vision_model.encoder.layers.7.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 134 |
+
"visual.vision_model.encoder.layers.7.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 135 |
+
"visual.vision_model.encoder.layers.7.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 136 |
+
"visual.vision_model.encoder.layers.7.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 137 |
+
"visual.vision_model.encoder.layers.7.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 138 |
+
"visual.vision_model.encoder.layers.7.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 139 |
+
"visual.vision_model.encoder.layers.7.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 140 |
+
"visual.vision_model.encoder.layers.7.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 141 |
+
"visual.vision_model.encoder.layers.7.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 142 |
+
"visual.vision_model.encoder.layers.7.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 143 |
+
"visual.vision_model.encoder.layers.7.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 144 |
+
"visual.vision_model.encoder.layers.8.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 145 |
+
"visual.vision_model.encoder.layers.8.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 146 |
+
"visual.vision_model.encoder.layers.8.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 147 |
+
"visual.vision_model.encoder.layers.8.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 148 |
+
"visual.vision_model.encoder.layers.8.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 149 |
+
"visual.vision_model.encoder.layers.8.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 150 |
+
"visual.vision_model.encoder.layers.8.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 151 |
+
"visual.vision_model.encoder.layers.8.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 152 |
+
"visual.vision_model.encoder.layers.8.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 153 |
+
"visual.vision_model.encoder.layers.8.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 154 |
+
"visual.vision_model.encoder.layers.8.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 155 |
+
"visual.vision_model.encoder.layers.8.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 156 |
+
"visual.vision_model.encoder.layers.8.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 157 |
+
"visual.vision_model.encoder.layers.8.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 158 |
+
"visual.vision_model.encoder.layers.8.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 159 |
+
"visual.vision_model.encoder.layers.8.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 160 |
+
"visual.vision_model.encoder.layers.9.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 161 |
+
"visual.vision_model.encoder.layers.9.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 162 |
+
"visual.vision_model.encoder.layers.9.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 163 |
+
"visual.vision_model.encoder.layers.9.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 164 |
+
"visual.vision_model.encoder.layers.9.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 165 |
+
"visual.vision_model.encoder.layers.9.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 166 |
+
"visual.vision_model.encoder.layers.9.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 167 |
+
"visual.vision_model.encoder.layers.9.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 168 |
+
"visual.vision_model.encoder.layers.9.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 169 |
+
"visual.vision_model.encoder.layers.9.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 170 |
+
"visual.vision_model.encoder.layers.9.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 171 |
+
"visual.vision_model.encoder.layers.9.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 172 |
+
"visual.vision_model.encoder.layers.9.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 173 |
+
"visual.vision_model.encoder.layers.9.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 174 |
+
"visual.vision_model.encoder.layers.9.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 175 |
+
"visual.vision_model.encoder.layers.9.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 176 |
+
"visual.vision_model.encoder.layers.10.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 177 |
+
"visual.vision_model.encoder.layers.10.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 178 |
+
"visual.vision_model.encoder.layers.10.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 179 |
+
"visual.vision_model.encoder.layers.10.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 180 |
+
"visual.vision_model.encoder.layers.10.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 181 |
+
"visual.vision_model.encoder.layers.10.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 182 |
+
"visual.vision_model.encoder.layers.10.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 183 |
+
"visual.vision_model.encoder.layers.10.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 184 |
+
"visual.vision_model.encoder.layers.10.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 185 |
+
"visual.vision_model.encoder.layers.10.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 186 |
+
"visual.vision_model.encoder.layers.10.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 187 |
+
"visual.vision_model.encoder.layers.10.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 188 |
+
"visual.vision_model.encoder.layers.10.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 189 |
+
"visual.vision_model.encoder.layers.10.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 190 |
+
"visual.vision_model.encoder.layers.10.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 191 |
+
"visual.vision_model.encoder.layers.10.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 192 |
+
"visual.vision_model.encoder.layers.11.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 193 |
+
"visual.vision_model.encoder.layers.11.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 194 |
+
"visual.vision_model.encoder.layers.11.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 195 |
+
"visual.vision_model.encoder.layers.11.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 196 |
+
"visual.vision_model.encoder.layers.11.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 197 |
+
"visual.vision_model.encoder.layers.11.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 198 |
+
"visual.vision_model.encoder.layers.11.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 199 |
+
"visual.vision_model.encoder.layers.11.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 200 |
+
"visual.vision_model.encoder.layers.11.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 201 |
+
"visual.vision_model.encoder.layers.11.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 202 |
+
"visual.vision_model.encoder.layers.11.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 203 |
+
"visual.vision_model.encoder.layers.11.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 204 |
+
"visual.vision_model.encoder.layers.11.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 205 |
+
"visual.vision_model.encoder.layers.11.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 206 |
+
"visual.vision_model.encoder.layers.11.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 207 |
+
"visual.vision_model.encoder.layers.11.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 208 |
+
"visual.vision_model.encoder.layers.12.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 209 |
+
"visual.vision_model.encoder.layers.12.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 210 |
+
"visual.vision_model.encoder.layers.12.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 211 |
+
"visual.vision_model.encoder.layers.12.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 212 |
+
"visual.vision_model.encoder.layers.12.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 213 |
+
"visual.vision_model.encoder.layers.12.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 214 |
+
"visual.vision_model.encoder.layers.12.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 215 |
+
"visual.vision_model.encoder.layers.12.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 216 |
+
"visual.vision_model.encoder.layers.12.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 217 |
+
"visual.vision_model.encoder.layers.12.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 218 |
+
"visual.vision_model.encoder.layers.12.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 219 |
+
"visual.vision_model.encoder.layers.12.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 220 |
+
"visual.vision_model.encoder.layers.12.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 221 |
+
"visual.vision_model.encoder.layers.12.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 222 |
+
"visual.vision_model.encoder.layers.12.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 223 |
+
"visual.vision_model.encoder.layers.12.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 224 |
+
"visual.vision_model.encoder.layers.13.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 225 |
+
"visual.vision_model.encoder.layers.13.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 226 |
+
"visual.vision_model.encoder.layers.13.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 227 |
+
"visual.vision_model.encoder.layers.13.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 228 |
+
"visual.vision_model.encoder.layers.13.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 229 |
+
"visual.vision_model.encoder.layers.13.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 230 |
+
"visual.vision_model.encoder.layers.13.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 231 |
+
"visual.vision_model.encoder.layers.13.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 232 |
+
"visual.vision_model.encoder.layers.13.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 233 |
+
"visual.vision_model.encoder.layers.13.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 234 |
+
"visual.vision_model.encoder.layers.13.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 235 |
+
"visual.vision_model.encoder.layers.13.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 236 |
+
"visual.vision_model.encoder.layers.13.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 237 |
+
"visual.vision_model.encoder.layers.13.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 238 |
+
"visual.vision_model.encoder.layers.13.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 239 |
+
"visual.vision_model.encoder.layers.13.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 240 |
+
"visual.vision_model.encoder.layers.14.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 241 |
+
"visual.vision_model.encoder.layers.14.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 242 |
+
"visual.vision_model.encoder.layers.14.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 243 |
+
"visual.vision_model.encoder.layers.14.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 244 |
+
"visual.vision_model.encoder.layers.14.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 245 |
+
"visual.vision_model.encoder.layers.14.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 246 |
+
"visual.vision_model.encoder.layers.14.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 247 |
+
"visual.vision_model.encoder.layers.14.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 248 |
+
"visual.vision_model.encoder.layers.14.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 249 |
+
"visual.vision_model.encoder.layers.14.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 250 |
+
"visual.vision_model.encoder.layers.14.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 251 |
+
"visual.vision_model.encoder.layers.14.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 252 |
+
"visual.vision_model.encoder.layers.14.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 253 |
+
"visual.vision_model.encoder.layers.14.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 254 |
+
"visual.vision_model.encoder.layers.14.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 255 |
+
"visual.vision_model.encoder.layers.14.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 256 |
+
"visual.vision_model.encoder.layers.15.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 257 |
+
"visual.vision_model.encoder.layers.15.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 258 |
+
"visual.vision_model.encoder.layers.15.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 259 |
+
"visual.vision_model.encoder.layers.15.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 260 |
+
"visual.vision_model.encoder.layers.15.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 261 |
+
"visual.vision_model.encoder.layers.15.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 262 |
+
"visual.vision_model.encoder.layers.15.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 263 |
+
"visual.vision_model.encoder.layers.15.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 264 |
+
"visual.vision_model.encoder.layers.15.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 265 |
+
"visual.vision_model.encoder.layers.15.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 266 |
+
"visual.vision_model.encoder.layers.15.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 267 |
+
"visual.vision_model.encoder.layers.15.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 268 |
+
"visual.vision_model.encoder.layers.15.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 269 |
+
"visual.vision_model.encoder.layers.15.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 270 |
+
"visual.vision_model.encoder.layers.15.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 271 |
+
"visual.vision_model.encoder.layers.15.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 272 |
+
"visual.vision_model.encoder.layers.16.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 273 |
+
"visual.vision_model.encoder.layers.16.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 274 |
+
"visual.vision_model.encoder.layers.16.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 275 |
+
"visual.vision_model.encoder.layers.16.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 276 |
+
"visual.vision_model.encoder.layers.16.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 277 |
+
"visual.vision_model.encoder.layers.16.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 278 |
+
"visual.vision_model.encoder.layers.16.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 279 |
+
"visual.vision_model.encoder.layers.16.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 280 |
+
"visual.vision_model.encoder.layers.16.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 281 |
+
"visual.vision_model.encoder.layers.16.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 282 |
+
"visual.vision_model.encoder.layers.16.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 283 |
+
"visual.vision_model.encoder.layers.16.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 284 |
+
"visual.vision_model.encoder.layers.16.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 285 |
+
"visual.vision_model.encoder.layers.16.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 286 |
+
"visual.vision_model.encoder.layers.16.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 287 |
+
"visual.vision_model.encoder.layers.16.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 288 |
+
"visual.vision_model.encoder.layers.17.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 289 |
+
"visual.vision_model.encoder.layers.17.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 290 |
+
"visual.vision_model.encoder.layers.17.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 291 |
+
"visual.vision_model.encoder.layers.17.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 292 |
+
"visual.vision_model.encoder.layers.17.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 293 |
+
"visual.vision_model.encoder.layers.17.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 294 |
+
"visual.vision_model.encoder.layers.17.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 295 |
+
"visual.vision_model.encoder.layers.17.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 296 |
+
"visual.vision_model.encoder.layers.17.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 297 |
+
"visual.vision_model.encoder.layers.17.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 298 |
+
"visual.vision_model.encoder.layers.17.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 299 |
+
"visual.vision_model.encoder.layers.17.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 300 |
+
"visual.vision_model.encoder.layers.17.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 301 |
+
"visual.vision_model.encoder.layers.17.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 302 |
+
"visual.vision_model.encoder.layers.17.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 303 |
+
"visual.vision_model.encoder.layers.17.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 304 |
+
"visual.vision_model.encoder.layers.18.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 305 |
+
"visual.vision_model.encoder.layers.18.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 306 |
+
"visual.vision_model.encoder.layers.18.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 307 |
+
"visual.vision_model.encoder.layers.18.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 308 |
+
"visual.vision_model.encoder.layers.18.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 309 |
+
"visual.vision_model.encoder.layers.18.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 310 |
+
"visual.vision_model.encoder.layers.18.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 311 |
+
"visual.vision_model.encoder.layers.18.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 312 |
+
"visual.vision_model.encoder.layers.18.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 313 |
+
"visual.vision_model.encoder.layers.18.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 314 |
+
"visual.vision_model.encoder.layers.18.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 315 |
+
"visual.vision_model.encoder.layers.18.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 316 |
+
"visual.vision_model.encoder.layers.18.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 317 |
+
"visual.vision_model.encoder.layers.18.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 318 |
+
"visual.vision_model.encoder.layers.18.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 319 |
+
"visual.vision_model.encoder.layers.18.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 320 |
+
"visual.vision_model.encoder.layers.19.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 321 |
+
"visual.vision_model.encoder.layers.19.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 322 |
+
"visual.vision_model.encoder.layers.19.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 323 |
+
"visual.vision_model.encoder.layers.19.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 324 |
+
"visual.vision_model.encoder.layers.19.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 325 |
+
"visual.vision_model.encoder.layers.19.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 326 |
+
"visual.vision_model.encoder.layers.19.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 327 |
+
"visual.vision_model.encoder.layers.19.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 328 |
+
"visual.vision_model.encoder.layers.19.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 329 |
+
"visual.vision_model.encoder.layers.19.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 330 |
+
"visual.vision_model.encoder.layers.19.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 331 |
+
"visual.vision_model.encoder.layers.19.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 332 |
+
"visual.vision_model.encoder.layers.19.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 333 |
+
"visual.vision_model.encoder.layers.19.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 334 |
+
"visual.vision_model.encoder.layers.19.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 335 |
+
"visual.vision_model.encoder.layers.19.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 336 |
+
"visual.vision_model.encoder.layers.20.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 337 |
+
"visual.vision_model.encoder.layers.20.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 338 |
+
"visual.vision_model.encoder.layers.20.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 339 |
+
"visual.vision_model.encoder.layers.20.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 340 |
+
"visual.vision_model.encoder.layers.20.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 341 |
+
"visual.vision_model.encoder.layers.20.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 342 |
+
"visual.vision_model.encoder.layers.20.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 343 |
+
"visual.vision_model.encoder.layers.20.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 344 |
+
"visual.vision_model.encoder.layers.20.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 345 |
+
"visual.vision_model.encoder.layers.20.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 346 |
+
"visual.vision_model.encoder.layers.20.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 347 |
+
"visual.vision_model.encoder.layers.20.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 348 |
+
"visual.vision_model.encoder.layers.20.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 349 |
+
"visual.vision_model.encoder.layers.20.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 350 |
+
"visual.vision_model.encoder.layers.20.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 351 |
+
"visual.vision_model.encoder.layers.20.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 352 |
+
"visual.vision_model.encoder.layers.21.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 353 |
+
"visual.vision_model.encoder.layers.21.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 354 |
+
"visual.vision_model.encoder.layers.21.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 355 |
+
"visual.vision_model.encoder.layers.21.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 356 |
+
"visual.vision_model.encoder.layers.21.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 357 |
+
"visual.vision_model.encoder.layers.21.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 358 |
+
"visual.vision_model.encoder.layers.21.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 359 |
+
"visual.vision_model.encoder.layers.21.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 360 |
+
"visual.vision_model.encoder.layers.21.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 361 |
+
"visual.vision_model.encoder.layers.21.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 362 |
+
"visual.vision_model.encoder.layers.21.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 363 |
+
"visual.vision_model.encoder.layers.21.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 364 |
+
"visual.vision_model.encoder.layers.21.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 365 |
+
"visual.vision_model.encoder.layers.21.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 366 |
+
"visual.vision_model.encoder.layers.21.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 367 |
+
"visual.vision_model.encoder.layers.21.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 368 |
+
"visual.vision_model.encoder.layers.22.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 369 |
+
"visual.vision_model.encoder.layers.22.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 370 |
+
"visual.vision_model.encoder.layers.22.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 371 |
+
"visual.vision_model.encoder.layers.22.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 372 |
+
"visual.vision_model.encoder.layers.22.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 373 |
+
"visual.vision_model.encoder.layers.22.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 374 |
+
"visual.vision_model.encoder.layers.22.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 375 |
+
"visual.vision_model.encoder.layers.22.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 376 |
+
"visual.vision_model.encoder.layers.22.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 377 |
+
"visual.vision_model.encoder.layers.22.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 378 |
+
"visual.vision_model.encoder.layers.22.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 379 |
+
"visual.vision_model.encoder.layers.22.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 380 |
+
"visual.vision_model.encoder.layers.22.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 381 |
+
"visual.vision_model.encoder.layers.22.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 382 |
+
"visual.vision_model.encoder.layers.22.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 383 |
+
"visual.vision_model.encoder.layers.22.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 384 |
+
"visual.vision_model.encoder.layers.23.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 385 |
+
"visual.vision_model.encoder.layers.23.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 386 |
+
"visual.vision_model.encoder.layers.23.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 387 |
+
"visual.vision_model.encoder.layers.23.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 388 |
+
"visual.vision_model.encoder.layers.23.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 389 |
+
"visual.vision_model.encoder.layers.23.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 390 |
+
"visual.vision_model.encoder.layers.23.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 391 |
+
"visual.vision_model.encoder.layers.23.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 392 |
+
"visual.vision_model.encoder.layers.23.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 393 |
+
"visual.vision_model.encoder.layers.23.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 394 |
+
"visual.vision_model.encoder.layers.23.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 395 |
+
"visual.vision_model.encoder.layers.23.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 396 |
+
"visual.vision_model.encoder.layers.23.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 397 |
+
"visual.vision_model.encoder.layers.23.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 398 |
+
"visual.vision_model.encoder.layers.23.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 399 |
+
"visual.vision_model.encoder.layers.23.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 400 |
+
"visual.vision_model.encoder.layers.24.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 401 |
+
"visual.vision_model.encoder.layers.24.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 402 |
+
"visual.vision_model.encoder.layers.24.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 403 |
+
"visual.vision_model.encoder.layers.24.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 404 |
+
"visual.vision_model.encoder.layers.24.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 405 |
+
"visual.vision_model.encoder.layers.24.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 406 |
+
"visual.vision_model.encoder.layers.24.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 407 |
+
"visual.vision_model.encoder.layers.24.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 408 |
+
"visual.vision_model.encoder.layers.24.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 409 |
+
"visual.vision_model.encoder.layers.24.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 410 |
+
"visual.vision_model.encoder.layers.24.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 411 |
+
"visual.vision_model.encoder.layers.24.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 412 |
+
"visual.vision_model.encoder.layers.24.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 413 |
+
"visual.vision_model.encoder.layers.24.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 414 |
+
"visual.vision_model.encoder.layers.24.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 415 |
+
"visual.vision_model.encoder.layers.24.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 416 |
+
"visual.vision_model.encoder.layers.25.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 417 |
+
"visual.vision_model.encoder.layers.25.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 418 |
+
"visual.vision_model.encoder.layers.25.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 419 |
+
"visual.vision_model.encoder.layers.25.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 420 |
+
"visual.vision_model.encoder.layers.25.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 421 |
+
"visual.vision_model.encoder.layers.25.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 422 |
+
"visual.vision_model.encoder.layers.25.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 423 |
+
"visual.vision_model.encoder.layers.25.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 424 |
+
"visual.vision_model.encoder.layers.25.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 425 |
+
"visual.vision_model.encoder.layers.25.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 426 |
+
"visual.vision_model.encoder.layers.25.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 427 |
+
"visual.vision_model.encoder.layers.25.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 428 |
+
"visual.vision_model.encoder.layers.25.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 429 |
+
"visual.vision_model.encoder.layers.25.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 430 |
+
"visual.vision_model.encoder.layers.25.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 431 |
+
"visual.vision_model.encoder.layers.25.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 432 |
+
"visual.vision_model.encoder.layers.26.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 433 |
+
"visual.vision_model.encoder.layers.26.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 434 |
+
"visual.vision_model.encoder.layers.26.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 435 |
+
"visual.vision_model.encoder.layers.26.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 436 |
+
"visual.vision_model.encoder.layers.26.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 437 |
+
"visual.vision_model.encoder.layers.26.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 438 |
+
"visual.vision_model.encoder.layers.26.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 439 |
+
"visual.vision_model.encoder.layers.26.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 440 |
+
"visual.vision_model.encoder.layers.26.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 441 |
+
"visual.vision_model.encoder.layers.26.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 442 |
+
"visual.vision_model.encoder.layers.26.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 443 |
+
"visual.vision_model.encoder.layers.26.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 444 |
+
"visual.vision_model.encoder.layers.26.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 445 |
+
"visual.vision_model.encoder.layers.26.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 446 |
+
"visual.vision_model.encoder.layers.26.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 447 |
+
"visual.vision_model.encoder.layers.26.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 448 |
+
"visual.vision_model.post_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 449 |
+
"visual.vision_model.post_layernorm.bias": "model-00001-of-00001.safetensors",
|
| 450 |
+
"visual.vision_model.head.probe": "model-00001-of-00001.safetensors",
|
| 451 |
+
"visual.vision_model.head.attention.in_proj_weight": "model-00001-of-00001.safetensors",
|
| 452 |
+
"visual.vision_model.head.attention.in_proj_bias": "model-00001-of-00001.safetensors",
|
| 453 |
+
"visual.vision_model.head.attention.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 454 |
+
"visual.vision_model.head.attention.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 455 |
+
"visual.vision_model.head.layernorm.weight": "model-00001-of-00001.safetensors",
|
| 456 |
+
"visual.vision_model.head.layernorm.bias": "model-00001-of-00001.safetensors",
|
| 457 |
+
"visual.vision_model.head.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 458 |
+
"visual.vision_model.head.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 459 |
+
"visual.vision_model.head.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 460 |
+
"visual.vision_model.head.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 461 |
+
"model.embed_tokens.weight": "model-00001-of-00001.safetensors",
|
| 462 |
+
"model.layers.0.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 463 |
+
"model.layers.0.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 464 |
+
"model.layers.0.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 465 |
+
"model.layers.0.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 466 |
+
"model.layers.0.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 467 |
+
"model.layers.0.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 468 |
+
"model.layers.0.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 469 |
+
"model.layers.0.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 470 |
+
"model.layers.0.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 471 |
+
"model.layers.1.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 472 |
+
"model.layers.1.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 473 |
+
"model.layers.1.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 474 |
+
"model.layers.1.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 475 |
+
"model.layers.1.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 476 |
+
"model.layers.1.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 477 |
+
"model.layers.1.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 478 |
+
"model.layers.1.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 479 |
+
"model.layers.1.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 480 |
+
"model.layers.2.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 481 |
+
"model.layers.2.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 482 |
+
"model.layers.2.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 483 |
+
"model.layers.2.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 484 |
+
"model.layers.2.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 485 |
+
"model.layers.2.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 486 |
+
"model.layers.2.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 487 |
+
"model.layers.2.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 488 |
+
"model.layers.2.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 489 |
+
"model.layers.3.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 490 |
+
"model.layers.3.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 491 |
+
"model.layers.3.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 492 |
+
"model.layers.3.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 493 |
+
"model.layers.3.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 494 |
+
"model.layers.3.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 495 |
+
"model.layers.3.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 496 |
+
"model.layers.3.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 497 |
+
"model.layers.3.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 498 |
+
"model.layers.4.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 499 |
+
"model.layers.4.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 500 |
+
"model.layers.4.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 501 |
+
"model.layers.4.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 502 |
+
"model.layers.4.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 503 |
+
"model.layers.4.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 504 |
+
"model.layers.4.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 505 |
+
"model.layers.4.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 506 |
+
"model.layers.4.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 507 |
+
"model.layers.5.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 508 |
+
"model.layers.5.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 509 |
+
"model.layers.5.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 510 |
+
"model.layers.5.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 511 |
+
"model.layers.5.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 512 |
+
"model.layers.5.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 513 |
+
"model.layers.5.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 514 |
+
"model.layers.5.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 515 |
+
"model.layers.5.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 516 |
+
"model.layers.6.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 517 |
+
"model.layers.6.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 518 |
+
"model.layers.6.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 519 |
+
"model.layers.6.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 520 |
+
"model.layers.6.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 521 |
+
"model.layers.6.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 522 |
+
"model.layers.6.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 523 |
+
"model.layers.6.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 524 |
+
"model.layers.6.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 525 |
+
"model.layers.7.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 526 |
+
"model.layers.7.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 527 |
+
"model.layers.7.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 528 |
+
"model.layers.7.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 529 |
+
"model.layers.7.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 530 |
+
"model.layers.7.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 531 |
+
"model.layers.7.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 532 |
+
"model.layers.7.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 533 |
+
"model.layers.7.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 534 |
+
"model.layers.8.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 535 |
+
"model.layers.8.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 536 |
+
"model.layers.8.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 537 |
+
"model.layers.8.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 538 |
+
"model.layers.8.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 539 |
+
"model.layers.8.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 540 |
+
"model.layers.8.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 541 |
+
"model.layers.8.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 542 |
+
"model.layers.8.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 543 |
+
"model.layers.9.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 544 |
+
"model.layers.9.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 545 |
+
"model.layers.9.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 546 |
+
"model.layers.9.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 547 |
+
"model.layers.9.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 548 |
+
"model.layers.9.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 549 |
+
"model.layers.9.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 550 |
+
"model.layers.9.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 551 |
+
"model.layers.9.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 552 |
+
"model.layers.10.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 553 |
+
"model.layers.10.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 554 |
+
"model.layers.10.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 555 |
+
"model.layers.10.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 556 |
+
"model.layers.10.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 557 |
+
"model.layers.10.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 558 |
+
"model.layers.10.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 559 |
+
"model.layers.10.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 560 |
+
"model.layers.10.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 561 |
+
"model.layers.11.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 562 |
+
"model.layers.11.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 563 |
+
"model.layers.11.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 564 |
+
"model.layers.11.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 565 |
+
"model.layers.11.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 566 |
+
"model.layers.11.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 567 |
+
"model.layers.11.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 568 |
+
"model.layers.11.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 569 |
+
"model.layers.11.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 570 |
+
"model.layers.12.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 571 |
+
"model.layers.12.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 572 |
+
"model.layers.12.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 573 |
+
"model.layers.12.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 574 |
+
"model.layers.12.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 575 |
+
"model.layers.12.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 576 |
+
"model.layers.12.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 577 |
+
"model.layers.12.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 578 |
+
"model.layers.12.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 579 |
+
"model.layers.13.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 580 |
+
"model.layers.13.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 581 |
+
"model.layers.13.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 582 |
+
"model.layers.13.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 583 |
+
"model.layers.13.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 584 |
+
"model.layers.13.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 585 |
+
"model.layers.13.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 586 |
+
"model.layers.13.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 587 |
+
"model.layers.13.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 588 |
+
"model.layers.14.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 589 |
+
"model.layers.14.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 590 |
+
"model.layers.14.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 591 |
+
"model.layers.14.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 592 |
+
"model.layers.14.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 593 |
+
"model.layers.14.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 594 |
+
"model.layers.14.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 595 |
+
"model.layers.14.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 596 |
+
"model.layers.14.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 597 |
+
"model.layers.15.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 598 |
+
"model.layers.15.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 599 |
+
"model.layers.15.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 600 |
+
"model.layers.15.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 601 |
+
"model.layers.15.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 602 |
+
"model.layers.15.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 603 |
+
"model.layers.15.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 604 |
+
"model.layers.15.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 605 |
+
"model.layers.15.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 606 |
+
"model.layers.16.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 607 |
+
"model.layers.16.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 608 |
+
"model.layers.16.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 609 |
+
"model.layers.16.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 610 |
+
"model.layers.16.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 611 |
+
"model.layers.16.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 612 |
+
"model.layers.16.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 613 |
+
"model.layers.16.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 614 |
+
"model.layers.16.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 615 |
+
"model.layers.17.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 616 |
+
"model.layers.17.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 617 |
+
"model.layers.17.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 618 |
+
"model.layers.17.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 619 |
+
"model.layers.17.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 620 |
+
"model.layers.17.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 621 |
+
"model.layers.17.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 622 |
+
"model.layers.17.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 623 |
+
"model.layers.17.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 624 |
+
"model.norm.weight": "model-00001-of-00001.safetensors",
|
| 625 |
+
"lm_head.weight": "model-00001-of-00001.safetensors"
|
| 626 |
+
}
|
| 627 |
+
}
|
checkpoint-378/optimizer-00001-of-00001.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d15e7372d8ff5e0050feda539c5dee7463c8bcc5ec1b4e3bc4d2f3e64b83df6b
|
| 3 |
+
size 7245150152
|
checkpoint-378/optimizer.safetensors.index.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
checkpoint-378/preprocessor_config.json
ADDED
|
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"do_convert_rgb": true,
|
| 3 |
+
"do_normalize": true,
|
| 4 |
+
"do_rescale": true,
|
| 5 |
+
"do_resize": true,
|
| 6 |
+
"image_mean": [
|
| 7 |
+
0.5,
|
| 8 |
+
0.5,
|
| 9 |
+
0.5
|
| 10 |
+
],
|
| 11 |
+
"image_processor_type": "SiglipImageProcessor",
|
| 12 |
+
"image_std": [
|
| 13 |
+
0.5,
|
| 14 |
+
0.5,
|
| 15 |
+
0.5
|
| 16 |
+
],
|
| 17 |
+
"max_pixels": 2822400,
|
| 18 |
+
"merge_size": 2,
|
| 19 |
+
"min_pixels": 147384,
|
| 20 |
+
"patch_size": 14,
|
| 21 |
+
"resample": 3,
|
| 22 |
+
"rescale_factor": 0.00392156862745098,
|
| 23 |
+
"size": {
|
| 24 |
+
"max_pixels": 2822400,
|
| 25 |
+
"min_pixels": 147384
|
| 26 |
+
},
|
| 27 |
+
"temporal_conv_size": 1,
|
| 28 |
+
"temporal_patch_size": 1
|
| 29 |
+
}
|
checkpoint-378/rng_state_0.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:29cde6478b49d2682dfa8c033ed7da4f33ab76ccd1941cfe950453d3923dfec3
|
| 3 |
+
size 32008
|
checkpoint-378/scheduler.pdparams
ADDED
|
Binary file (51 Bytes). View file
|
|
|
checkpoint-378/special_tokens_map.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"bos_token": {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": false, "special": false}, "eos_token": "</s>", "unk_token": {"content": "<unk>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": false, "special": false}, "sep_token": "<|end_of_sentence|>", "pad_token": {"content": "<unk>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": false, "special": false}, "cls_token": "<|begin_of_sentence|>", "mask_token": {"content": "<mask:1>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": false, "special": false}, "additional_special_tokens": ["<|IMAGE_PLACEHOLDER|>", "<|image_pad|>", "<|IMAGE_START|>", "<|IMAGE_END|>", "<|video_pad|>"]}
|
checkpoint-378/static_name_to_dyg_name.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"layer_norm_0.w_0": "mlp_AR.pre_norm.weight", "layer_norm_0.b_0": "mlp_AR.pre_norm.bias", "linear_0.w_0": "mlp_AR.linear_1.weight", "linear_0.b_0": "mlp_AR.linear_1.bias", "linear_1.w_0": "mlp_AR.linear_2.weight", "linear_1.b_0": "mlp_AR.linear_2.bias", "conv2d_0.w_0": "visual.vision_model.embeddings.patch_embedding.weight", "conv2d_0.b_0": "visual.vision_model.embeddings.patch_embedding.bias", "embedding_0.w_0": "visual.vision_model.embeddings.position_embedding.weight", "embedding_1.w_0": "visual.vision_model.embeddings.packing_position_embedding.weight", "layer_norm_1.w_0": "visual.vision_model.encoder.layers.0.layer_norm1.weight", "layer_norm_1.b_0": "visual.vision_model.encoder.layers.0.layer_norm1.bias", "linear_2.w_0": "visual.vision_model.encoder.layers.0.self_attn.k_proj.weight", "linear_2.b_0": "visual.vision_model.encoder.layers.0.self_attn.k_proj.bias", "linear_3.w_0": "visual.vision_model.encoder.layers.0.self_attn.v_proj.weight", "linear_3.b_0": "visual.vision_model.encoder.layers.0.self_attn.v_proj.bias", "linear_4.w_0": "visual.vision_model.encoder.layers.0.self_attn.q_proj.weight", "linear_4.b_0": "visual.vision_model.encoder.layers.0.self_attn.q_proj.bias", "linear_5.w_0": "visual.vision_model.encoder.layers.0.self_attn.out_proj.weight", "linear_5.b_0": "visual.vision_model.encoder.layers.0.self_attn.out_proj.bias", "layer_norm_2.w_0": "visual.vision_model.encoder.layers.0.layer_norm2.weight", "layer_norm_2.b_0": "visual.vision_model.encoder.layers.0.layer_norm2.bias", "linear_6.w_0": "visual.vision_model.encoder.layers.0.mlp.fc1.weight", "linear_6.b_0": "visual.vision_model.encoder.layers.0.mlp.fc1.bias", "linear_7.w_0": "visual.vision_model.encoder.layers.0.mlp.fc2.weight", "linear_7.b_0": "visual.vision_model.encoder.layers.0.mlp.fc2.bias", "layer_norm_3.w_0": "visual.vision_model.encoder.layers.1.layer_norm1.weight", "layer_norm_3.b_0": "visual.vision_model.encoder.layers.1.layer_norm1.bias", "linear_8.w_0": "visual.vision_model.encoder.layers.1.self_attn.k_proj.weight", "linear_8.b_0": "visual.vision_model.encoder.layers.1.self_attn.k_proj.bias", "linear_9.w_0": "visual.vision_model.encoder.layers.1.self_attn.v_proj.weight", "linear_9.b_0": "visual.vision_model.encoder.layers.1.self_attn.v_proj.bias", "linear_10.w_0": "visual.vision_model.encoder.layers.1.self_attn.q_proj.weight", "linear_10.b_0": "visual.vision_model.encoder.layers.1.self_attn.q_proj.bias", "linear_11.w_0": "visual.vision_model.encoder.layers.1.self_attn.out_proj.weight", "linear_11.b_0": "visual.vision_model.encoder.layers.1.self_attn.out_proj.bias", "layer_norm_4.w_0": "visual.vision_model.encoder.layers.1.layer_norm2.weight", "layer_norm_4.b_0": "visual.vision_model.encoder.layers.1.layer_norm2.bias", "linear_12.w_0": "visual.vision_model.encoder.layers.1.mlp.fc1.weight", "linear_12.b_0": "visual.vision_model.encoder.layers.1.mlp.fc1.bias", "linear_13.w_0": "visual.vision_model.encoder.layers.1.mlp.fc2.weight", "linear_13.b_0": "visual.vision_model.encoder.layers.1.mlp.fc2.bias", "layer_norm_5.w_0": "visual.vision_model.encoder.layers.2.layer_norm1.weight", "layer_norm_5.b_0": "visual.vision_model.encoder.layers.2.layer_norm1.bias", "linear_14.w_0": "visual.vision_model.encoder.layers.2.self_attn.k_proj.weight", "linear_14.b_0": "visual.vision_model.encoder.layers.2.self_attn.k_proj.bias", "linear_15.w_0": "visual.vision_model.encoder.layers.2.self_attn.v_proj.weight", "linear_15.b_0": "visual.vision_model.encoder.layers.2.self_attn.v_proj.bias", "linear_16.w_0": "visual.vision_model.encoder.layers.2.self_attn.q_proj.weight", "linear_16.b_0": "visual.vision_model.encoder.layers.2.self_attn.q_proj.bias", "linear_17.w_0": "visual.vision_model.encoder.layers.2.self_attn.out_proj.weight", "linear_17.b_0": "visual.vision_model.encoder.layers.2.self_attn.out_proj.bias", "layer_norm_6.w_0": "visual.vision_model.encoder.layers.2.layer_norm2.weight", "layer_norm_6.b_0": "visual.vision_model.encoder.layers.2.layer_norm2.bias", "linear_18.w_0": "visual.vision_model.encoder.layers.2.mlp.fc1.weight", "linear_18.b_0": "visual.vision_model.encoder.layers.2.mlp.fc1.bias", "linear_19.w_0": "visual.vision_model.encoder.layers.2.mlp.fc2.weight", "linear_19.b_0": "visual.vision_model.encoder.layers.2.mlp.fc2.bias", "layer_norm_7.w_0": "visual.vision_model.encoder.layers.3.layer_norm1.weight", "layer_norm_7.b_0": "visual.vision_model.encoder.layers.3.layer_norm1.bias", "linear_20.w_0": "visual.vision_model.encoder.layers.3.self_attn.k_proj.weight", "linear_20.b_0": "visual.vision_model.encoder.layers.3.self_attn.k_proj.bias", "linear_21.w_0": "visual.vision_model.encoder.layers.3.self_attn.v_proj.weight", "linear_21.b_0": "visual.vision_model.encoder.layers.3.self_attn.v_proj.bias", "linear_22.w_0": "visual.vision_model.encoder.layers.3.self_attn.q_proj.weight", "linear_22.b_0": "visual.vision_model.encoder.layers.3.self_attn.q_proj.bias", "linear_23.w_0": "visual.vision_model.encoder.layers.3.self_attn.out_proj.weight", "linear_23.b_0": "visual.vision_model.encoder.layers.3.self_attn.out_proj.bias", "layer_norm_8.w_0": "visual.vision_model.encoder.layers.3.layer_norm2.weight", "layer_norm_8.b_0": "visual.vision_model.encoder.layers.3.layer_norm2.bias", "linear_24.w_0": "visual.vision_model.encoder.layers.3.mlp.fc1.weight", "linear_24.b_0": "visual.vision_model.encoder.layers.3.mlp.fc1.bias", "linear_25.w_0": "visual.vision_model.encoder.layers.3.mlp.fc2.weight", "linear_25.b_0": "visual.vision_model.encoder.layers.3.mlp.fc2.bias", "layer_norm_9.w_0": "visual.vision_model.encoder.layers.4.layer_norm1.weight", "layer_norm_9.b_0": "visual.vision_model.encoder.layers.4.layer_norm1.bias", "linear_26.w_0": "visual.vision_model.encoder.layers.4.self_attn.k_proj.weight", "linear_26.b_0": "visual.vision_model.encoder.layers.4.self_attn.k_proj.bias", "linear_27.w_0": "visual.vision_model.encoder.layers.4.self_attn.v_proj.weight", "linear_27.b_0": "visual.vision_model.encoder.layers.4.self_attn.v_proj.bias", "linear_28.w_0": "visual.vision_model.encoder.layers.4.self_attn.q_proj.weight", "linear_28.b_0": "visual.vision_model.encoder.layers.4.self_attn.q_proj.bias", "linear_29.w_0": "visual.vision_model.encoder.layers.4.self_attn.out_proj.weight", "linear_29.b_0": "visual.vision_model.encoder.layers.4.self_attn.out_proj.bias", "layer_norm_10.w_0": "visual.vision_model.encoder.layers.4.layer_norm2.weight", "layer_norm_10.b_0": "visual.vision_model.encoder.layers.4.layer_norm2.bias", "linear_30.w_0": "visual.vision_model.encoder.layers.4.mlp.fc1.weight", "linear_30.b_0": "visual.vision_model.encoder.layers.4.mlp.fc1.bias", "linear_31.w_0": "visual.vision_model.encoder.layers.4.mlp.fc2.weight", "linear_31.b_0": "visual.vision_model.encoder.layers.4.mlp.fc2.bias", "layer_norm_11.w_0": "visual.vision_model.encoder.layers.5.layer_norm1.weight", "layer_norm_11.b_0": "visual.vision_model.encoder.layers.5.layer_norm1.bias", "linear_32.w_0": "visual.vision_model.encoder.layers.5.self_attn.k_proj.weight", "linear_32.b_0": "visual.vision_model.encoder.layers.5.self_attn.k_proj.bias", "linear_33.w_0": "visual.vision_model.encoder.layers.5.self_attn.v_proj.weight", "linear_33.b_0": "visual.vision_model.encoder.layers.5.self_attn.v_proj.bias", "linear_34.w_0": "visual.vision_model.encoder.layers.5.self_attn.q_proj.weight", "linear_34.b_0": "visual.vision_model.encoder.layers.5.self_attn.q_proj.bias", "linear_35.w_0": "visual.vision_model.encoder.layers.5.self_attn.out_proj.weight", "linear_35.b_0": "visual.vision_model.encoder.layers.5.self_attn.out_proj.bias", "layer_norm_12.w_0": "visual.vision_model.encoder.layers.5.layer_norm2.weight", "layer_norm_12.b_0": "visual.vision_model.encoder.layers.5.layer_norm2.bias", "linear_36.w_0": "visual.vision_model.encoder.layers.5.mlp.fc1.weight", "linear_36.b_0": "visual.vision_model.encoder.layers.5.mlp.fc1.bias", "linear_37.w_0": "visual.vision_model.encoder.layers.5.mlp.fc2.weight", "linear_37.b_0": "visual.vision_model.encoder.layers.5.mlp.fc2.bias", "layer_norm_13.w_0": "visual.vision_model.encoder.layers.6.layer_norm1.weight", "layer_norm_13.b_0": "visual.vision_model.encoder.layers.6.layer_norm1.bias", "linear_38.w_0": "visual.vision_model.encoder.layers.6.self_attn.k_proj.weight", "linear_38.b_0": "visual.vision_model.encoder.layers.6.self_attn.k_proj.bias", "linear_39.w_0": "visual.vision_model.encoder.layers.6.self_attn.v_proj.weight", "linear_39.b_0": "visual.vision_model.encoder.layers.6.self_attn.v_proj.bias", "linear_40.w_0": "visual.vision_model.encoder.layers.6.self_attn.q_proj.weight", "linear_40.b_0": "visual.vision_model.encoder.layers.6.self_attn.q_proj.bias", "linear_41.w_0": "visual.vision_model.encoder.layers.6.self_attn.out_proj.weight", "linear_41.b_0": "visual.vision_model.encoder.layers.6.self_attn.out_proj.bias", "layer_norm_14.w_0": "visual.vision_model.encoder.layers.6.layer_norm2.weight", "layer_norm_14.b_0": "visual.vision_model.encoder.layers.6.layer_norm2.bias", "linear_42.w_0": "visual.vision_model.encoder.layers.6.mlp.fc1.weight", "linear_42.b_0": "visual.vision_model.encoder.layers.6.mlp.fc1.bias", "linear_43.w_0": "visual.vision_model.encoder.layers.6.mlp.fc2.weight", "linear_43.b_0": "visual.vision_model.encoder.layers.6.mlp.fc2.bias", "layer_norm_15.w_0": "visual.vision_model.encoder.layers.7.layer_norm1.weight", "layer_norm_15.b_0": "visual.vision_model.encoder.layers.7.layer_norm1.bias", "linear_44.w_0": "visual.vision_model.encoder.layers.7.self_attn.k_proj.weight", "linear_44.b_0": "visual.vision_model.encoder.layers.7.self_attn.k_proj.bias", "linear_45.w_0": "visual.vision_model.encoder.layers.7.self_attn.v_proj.weight", "linear_45.b_0": "visual.vision_model.encoder.layers.7.self_attn.v_proj.bias", "linear_46.w_0": "visual.vision_model.encoder.layers.7.self_attn.q_proj.weight", "linear_46.b_0": "visual.vision_model.encoder.layers.7.self_attn.q_proj.bias", "linear_47.w_0": "visual.vision_model.encoder.layers.7.self_attn.out_proj.weight", "linear_47.b_0": "visual.vision_model.encoder.layers.7.self_attn.out_proj.bias", "layer_norm_16.w_0": "visual.vision_model.encoder.layers.7.layer_norm2.weight", "layer_norm_16.b_0": "visual.vision_model.encoder.layers.7.layer_norm2.bias", "linear_48.w_0": "visual.vision_model.encoder.layers.7.mlp.fc1.weight", "linear_48.b_0": "visual.vision_model.encoder.layers.7.mlp.fc1.bias", "linear_49.w_0": "visual.vision_model.encoder.layers.7.mlp.fc2.weight", "linear_49.b_0": "visual.vision_model.encoder.layers.7.mlp.fc2.bias", "layer_norm_17.w_0": "visual.vision_model.encoder.layers.8.layer_norm1.weight", "layer_norm_17.b_0": "visual.vision_model.encoder.layers.8.layer_norm1.bias", "linear_50.w_0": "visual.vision_model.encoder.layers.8.self_attn.k_proj.weight", "linear_50.b_0": "visual.vision_model.encoder.layers.8.self_attn.k_proj.bias", "linear_51.w_0": "visual.vision_model.encoder.layers.8.self_attn.v_proj.weight", "linear_51.b_0": "visual.vision_model.encoder.layers.8.self_attn.v_proj.bias", "linear_52.w_0": "visual.vision_model.encoder.layers.8.self_attn.q_proj.weight", "linear_52.b_0": "visual.vision_model.encoder.layers.8.self_attn.q_proj.bias", "linear_53.w_0": "visual.vision_model.encoder.layers.8.self_attn.out_proj.weight", "linear_53.b_0": "visual.vision_model.encoder.layers.8.self_attn.out_proj.bias", "layer_norm_18.w_0": "visual.vision_model.encoder.layers.8.layer_norm2.weight", "layer_norm_18.b_0": "visual.vision_model.encoder.layers.8.layer_norm2.bias", "linear_54.w_0": "visual.vision_model.encoder.layers.8.mlp.fc1.weight", "linear_54.b_0": "visual.vision_model.encoder.layers.8.mlp.fc1.bias", "linear_55.w_0": "visual.vision_model.encoder.layers.8.mlp.fc2.weight", "linear_55.b_0": "visual.vision_model.encoder.layers.8.mlp.fc2.bias", "layer_norm_19.w_0": "visual.vision_model.encoder.layers.9.layer_norm1.weight", "layer_norm_19.b_0": "visual.vision_model.encoder.layers.9.layer_norm1.bias", "linear_56.w_0": "visual.vision_model.encoder.layers.9.self_attn.k_proj.weight", "linear_56.b_0": "visual.vision_model.encoder.layers.9.self_attn.k_proj.bias", "linear_57.w_0": "visual.vision_model.encoder.layers.9.self_attn.v_proj.weight", "linear_57.b_0": "visual.vision_model.encoder.layers.9.self_attn.v_proj.bias", "linear_58.w_0": "visual.vision_model.encoder.layers.9.self_attn.q_proj.weight", "linear_58.b_0": "visual.vision_model.encoder.layers.9.self_attn.q_proj.bias", "linear_59.w_0": "visual.vision_model.encoder.layers.9.self_attn.out_proj.weight", "linear_59.b_0": "visual.vision_model.encoder.layers.9.self_attn.out_proj.bias", "layer_norm_20.w_0": "visual.vision_model.encoder.layers.9.layer_norm2.weight", "layer_norm_20.b_0": "visual.vision_model.encoder.layers.9.layer_norm2.bias", "linear_60.w_0": "visual.vision_model.encoder.layers.9.mlp.fc1.weight", "linear_60.b_0": "visual.vision_model.encoder.layers.9.mlp.fc1.bias", "linear_61.w_0": "visual.vision_model.encoder.layers.9.mlp.fc2.weight", "linear_61.b_0": "visual.vision_model.encoder.layers.9.mlp.fc2.bias", "layer_norm_21.w_0": "visual.vision_model.encoder.layers.10.layer_norm1.weight", "layer_norm_21.b_0": "visual.vision_model.encoder.layers.10.layer_norm1.bias", "linear_62.w_0": "visual.vision_model.encoder.layers.10.self_attn.k_proj.weight", "linear_62.b_0": "visual.vision_model.encoder.layers.10.self_attn.k_proj.bias", "linear_63.w_0": "visual.vision_model.encoder.layers.10.self_attn.v_proj.weight", "linear_63.b_0": "visual.vision_model.encoder.layers.10.self_attn.v_proj.bias", "linear_64.w_0": "visual.vision_model.encoder.layers.10.self_attn.q_proj.weight", "linear_64.b_0": "visual.vision_model.encoder.layers.10.self_attn.q_proj.bias", "linear_65.w_0": "visual.vision_model.encoder.layers.10.self_attn.out_proj.weight", "linear_65.b_0": "visual.vision_model.encoder.layers.10.self_attn.out_proj.bias", "layer_norm_22.w_0": "visual.vision_model.encoder.layers.10.layer_norm2.weight", "layer_norm_22.b_0": "visual.vision_model.encoder.layers.10.layer_norm2.bias", "linear_66.w_0": "visual.vision_model.encoder.layers.10.mlp.fc1.weight", "linear_66.b_0": "visual.vision_model.encoder.layers.10.mlp.fc1.bias", "linear_67.w_0": "visual.vision_model.encoder.layers.10.mlp.fc2.weight", "linear_67.b_0": "visual.vision_model.encoder.layers.10.mlp.fc2.bias", "layer_norm_23.w_0": "visual.vision_model.encoder.layers.11.layer_norm1.weight", "layer_norm_23.b_0": "visual.vision_model.encoder.layers.11.layer_norm1.bias", "linear_68.w_0": "visual.vision_model.encoder.layers.11.self_attn.k_proj.weight", "linear_68.b_0": "visual.vision_model.encoder.layers.11.self_attn.k_proj.bias", "linear_69.w_0": "visual.vision_model.encoder.layers.11.self_attn.v_proj.weight", "linear_69.b_0": "visual.vision_model.encoder.layers.11.self_attn.v_proj.bias", "linear_70.w_0": "visual.vision_model.encoder.layers.11.self_attn.q_proj.weight", "linear_70.b_0": "visual.vision_model.encoder.layers.11.self_attn.q_proj.bias", "linear_71.w_0": "visual.vision_model.encoder.layers.11.self_attn.out_proj.weight", "linear_71.b_0": "visual.vision_model.encoder.layers.11.self_attn.out_proj.bias", "layer_norm_24.w_0": "visual.vision_model.encoder.layers.11.layer_norm2.weight", "layer_norm_24.b_0": "visual.vision_model.encoder.layers.11.layer_norm2.bias", "linear_72.w_0": "visual.vision_model.encoder.layers.11.mlp.fc1.weight", "linear_72.b_0": "visual.vision_model.encoder.layers.11.mlp.fc1.bias", "linear_73.w_0": "visual.vision_model.encoder.layers.11.mlp.fc2.weight", "linear_73.b_0": "visual.vision_model.encoder.layers.11.mlp.fc2.bias", "layer_norm_25.w_0": "visual.vision_model.encoder.layers.12.layer_norm1.weight", "layer_norm_25.b_0": "visual.vision_model.encoder.layers.12.layer_norm1.bias", "linear_74.w_0": "visual.vision_model.encoder.layers.12.self_attn.k_proj.weight", "linear_74.b_0": "visual.vision_model.encoder.layers.12.self_attn.k_proj.bias", "linear_75.w_0": "visual.vision_model.encoder.layers.12.self_attn.v_proj.weight", "linear_75.b_0": "visual.vision_model.encoder.layers.12.self_attn.v_proj.bias", "linear_76.w_0": "visual.vision_model.encoder.layers.12.self_attn.q_proj.weight", "linear_76.b_0": "visual.vision_model.encoder.layers.12.self_attn.q_proj.bias", "linear_77.w_0": "visual.vision_model.encoder.layers.12.self_attn.out_proj.weight", "linear_77.b_0": "visual.vision_model.encoder.layers.12.self_attn.out_proj.bias", "layer_norm_26.w_0": "visual.vision_model.encoder.layers.12.layer_norm2.weight", "layer_norm_26.b_0": "visual.vision_model.encoder.layers.12.layer_norm2.bias", "linear_78.w_0": "visual.vision_model.encoder.layers.12.mlp.fc1.weight", "linear_78.b_0": "visual.vision_model.encoder.layers.12.mlp.fc1.bias", "linear_79.w_0": "visual.vision_model.encoder.layers.12.mlp.fc2.weight", "linear_79.b_0": "visual.vision_model.encoder.layers.12.mlp.fc2.bias", "layer_norm_27.w_0": "visual.vision_model.encoder.layers.13.layer_norm1.weight", "layer_norm_27.b_0": "visual.vision_model.encoder.layers.13.layer_norm1.bias", "linear_80.w_0": "visual.vision_model.encoder.layers.13.self_attn.k_proj.weight", "linear_80.b_0": "visual.vision_model.encoder.layers.13.self_attn.k_proj.bias", "linear_81.w_0": "visual.vision_model.encoder.layers.13.self_attn.v_proj.weight", "linear_81.b_0": "visual.vision_model.encoder.layers.13.self_attn.v_proj.bias", "linear_82.w_0": "visual.vision_model.encoder.layers.13.self_attn.q_proj.weight", "linear_82.b_0": "visual.vision_model.encoder.layers.13.self_attn.q_proj.bias", "linear_83.w_0": "visual.vision_model.encoder.layers.13.self_attn.out_proj.weight", "linear_83.b_0": "visual.vision_model.encoder.layers.13.self_attn.out_proj.bias", "layer_norm_28.w_0": "visual.vision_model.encoder.layers.13.layer_norm2.weight", "layer_norm_28.b_0": "visual.vision_model.encoder.layers.13.layer_norm2.bias", "linear_84.w_0": "visual.vision_model.encoder.layers.13.mlp.fc1.weight", "linear_84.b_0": "visual.vision_model.encoder.layers.13.mlp.fc1.bias", "linear_85.w_0": "visual.vision_model.encoder.layers.13.mlp.fc2.weight", "linear_85.b_0": "visual.vision_model.encoder.layers.13.mlp.fc2.bias", "layer_norm_29.w_0": "visual.vision_model.encoder.layers.14.layer_norm1.weight", "layer_norm_29.b_0": "visual.vision_model.encoder.layers.14.layer_norm1.bias", "linear_86.w_0": "visual.vision_model.encoder.layers.14.self_attn.k_proj.weight", "linear_86.b_0": "visual.vision_model.encoder.layers.14.self_attn.k_proj.bias", "linear_87.w_0": "visual.vision_model.encoder.layers.14.self_attn.v_proj.weight", "linear_87.b_0": "visual.vision_model.encoder.layers.14.self_attn.v_proj.bias", "linear_88.w_0": "visual.vision_model.encoder.layers.14.self_attn.q_proj.weight", "linear_88.b_0": "visual.vision_model.encoder.layers.14.self_attn.q_proj.bias", "linear_89.w_0": "visual.vision_model.encoder.layers.14.self_attn.out_proj.weight", "linear_89.b_0": "visual.vision_model.encoder.layers.14.self_attn.out_proj.bias", "layer_norm_30.w_0": "visual.vision_model.encoder.layers.14.layer_norm2.weight", "layer_norm_30.b_0": "visual.vision_model.encoder.layers.14.layer_norm2.bias", "linear_90.w_0": "visual.vision_model.encoder.layers.14.mlp.fc1.weight", "linear_90.b_0": "visual.vision_model.encoder.layers.14.mlp.fc1.bias", "linear_91.w_0": "visual.vision_model.encoder.layers.14.mlp.fc2.weight", "linear_91.b_0": "visual.vision_model.encoder.layers.14.mlp.fc2.bias", "layer_norm_31.w_0": "visual.vision_model.encoder.layers.15.layer_norm1.weight", "layer_norm_31.b_0": "visual.vision_model.encoder.layers.15.layer_norm1.bias", "linear_92.w_0": "visual.vision_model.encoder.layers.15.self_attn.k_proj.weight", "linear_92.b_0": "visual.vision_model.encoder.layers.15.self_attn.k_proj.bias", "linear_93.w_0": "visual.vision_model.encoder.layers.15.self_attn.v_proj.weight", "linear_93.b_0": "visual.vision_model.encoder.layers.15.self_attn.v_proj.bias", "linear_94.w_0": "visual.vision_model.encoder.layers.15.self_attn.q_proj.weight", "linear_94.b_0": "visual.vision_model.encoder.layers.15.self_attn.q_proj.bias", "linear_95.w_0": "visual.vision_model.encoder.layers.15.self_attn.out_proj.weight", "linear_95.b_0": "visual.vision_model.encoder.layers.15.self_attn.out_proj.bias", "layer_norm_32.w_0": "visual.vision_model.encoder.layers.15.layer_norm2.weight", "layer_norm_32.b_0": "visual.vision_model.encoder.layers.15.layer_norm2.bias", "linear_96.w_0": "visual.vision_model.encoder.layers.15.mlp.fc1.weight", "linear_96.b_0": "visual.vision_model.encoder.layers.15.mlp.fc1.bias", "linear_97.w_0": "visual.vision_model.encoder.layers.15.mlp.fc2.weight", "linear_97.b_0": "visual.vision_model.encoder.layers.15.mlp.fc2.bias", "layer_norm_33.w_0": "visual.vision_model.encoder.layers.16.layer_norm1.weight", "layer_norm_33.b_0": "visual.vision_model.encoder.layers.16.layer_norm1.bias", "linear_98.w_0": "visual.vision_model.encoder.layers.16.self_attn.k_proj.weight", "linear_98.b_0": "visual.vision_model.encoder.layers.16.self_attn.k_proj.bias", "linear_99.w_0": "visual.vision_model.encoder.layers.16.self_attn.v_proj.weight", "linear_99.b_0": "visual.vision_model.encoder.layers.16.self_attn.v_proj.bias", "linear_100.w_0": "visual.vision_model.encoder.layers.16.self_attn.q_proj.weight", "linear_100.b_0": "visual.vision_model.encoder.layers.16.self_attn.q_proj.bias", "linear_101.w_0": "visual.vision_model.encoder.layers.16.self_attn.out_proj.weight", "linear_101.b_0": "visual.vision_model.encoder.layers.16.self_attn.out_proj.bias", "layer_norm_34.w_0": "visual.vision_model.encoder.layers.16.layer_norm2.weight", "layer_norm_34.b_0": "visual.vision_model.encoder.layers.16.layer_norm2.bias", "linear_102.w_0": "visual.vision_model.encoder.layers.16.mlp.fc1.weight", "linear_102.b_0": "visual.vision_model.encoder.layers.16.mlp.fc1.bias", "linear_103.w_0": "visual.vision_model.encoder.layers.16.mlp.fc2.weight", "linear_103.b_0": "visual.vision_model.encoder.layers.16.mlp.fc2.bias", "layer_norm_35.w_0": "visual.vision_model.encoder.layers.17.layer_norm1.weight", "layer_norm_35.b_0": "visual.vision_model.encoder.layers.17.layer_norm1.bias", "linear_104.w_0": "visual.vision_model.encoder.layers.17.self_attn.k_proj.weight", "linear_104.b_0": "visual.vision_model.encoder.layers.17.self_attn.k_proj.bias", "linear_105.w_0": "visual.vision_model.encoder.layers.17.self_attn.v_proj.weight", "linear_105.b_0": "visual.vision_model.encoder.layers.17.self_attn.v_proj.bias", "linear_106.w_0": "visual.vision_model.encoder.layers.17.self_attn.q_proj.weight", "linear_106.b_0": "visual.vision_model.encoder.layers.17.self_attn.q_proj.bias", "linear_107.w_0": "visual.vision_model.encoder.layers.17.self_attn.out_proj.weight", "linear_107.b_0": "visual.vision_model.encoder.layers.17.self_attn.out_proj.bias", "layer_norm_36.w_0": "visual.vision_model.encoder.layers.17.layer_norm2.weight", "layer_norm_36.b_0": "visual.vision_model.encoder.layers.17.layer_norm2.bias", "linear_108.w_0": "visual.vision_model.encoder.layers.17.mlp.fc1.weight", "linear_108.b_0": "visual.vision_model.encoder.layers.17.mlp.fc1.bias", "linear_109.w_0": "visual.vision_model.encoder.layers.17.mlp.fc2.weight", "linear_109.b_0": "visual.vision_model.encoder.layers.17.mlp.fc2.bias", "layer_norm_37.w_0": "visual.vision_model.encoder.layers.18.layer_norm1.weight", "layer_norm_37.b_0": "visual.vision_model.encoder.layers.18.layer_norm1.bias", "linear_110.w_0": "visual.vision_model.encoder.layers.18.self_attn.k_proj.weight", "linear_110.b_0": "visual.vision_model.encoder.layers.18.self_attn.k_proj.bias", "linear_111.w_0": "visual.vision_model.encoder.layers.18.self_attn.v_proj.weight", "linear_111.b_0": "visual.vision_model.encoder.layers.18.self_attn.v_proj.bias", "linear_112.w_0": "visual.vision_model.encoder.layers.18.self_attn.q_proj.weight", "linear_112.b_0": "visual.vision_model.encoder.layers.18.self_attn.q_proj.bias", "linear_113.w_0": "visual.vision_model.encoder.layers.18.self_attn.out_proj.weight", "linear_113.b_0": "visual.vision_model.encoder.layers.18.self_attn.out_proj.bias", "layer_norm_38.w_0": "visual.vision_model.encoder.layers.18.layer_norm2.weight", "layer_norm_38.b_0": "visual.vision_model.encoder.layers.18.layer_norm2.bias", "linear_114.w_0": "visual.vision_model.encoder.layers.18.mlp.fc1.weight", "linear_114.b_0": "visual.vision_model.encoder.layers.18.mlp.fc1.bias", "linear_115.w_0": "visual.vision_model.encoder.layers.18.mlp.fc2.weight", "linear_115.b_0": "visual.vision_model.encoder.layers.18.mlp.fc2.bias", "layer_norm_39.w_0": "visual.vision_model.encoder.layers.19.layer_norm1.weight", "layer_norm_39.b_0": "visual.vision_model.encoder.layers.19.layer_norm1.bias", "linear_116.w_0": "visual.vision_model.encoder.layers.19.self_attn.k_proj.weight", "linear_116.b_0": "visual.vision_model.encoder.layers.19.self_attn.k_proj.bias", "linear_117.w_0": "visual.vision_model.encoder.layers.19.self_attn.v_proj.weight", "linear_117.b_0": "visual.vision_model.encoder.layers.19.self_attn.v_proj.bias", "linear_118.w_0": "visual.vision_model.encoder.layers.19.self_attn.q_proj.weight", "linear_118.b_0": "visual.vision_model.encoder.layers.19.self_attn.q_proj.bias", "linear_119.w_0": "visual.vision_model.encoder.layers.19.self_attn.out_proj.weight", "linear_119.b_0": "visual.vision_model.encoder.layers.19.self_attn.out_proj.bias", "layer_norm_40.w_0": "visual.vision_model.encoder.layers.19.layer_norm2.weight", "layer_norm_40.b_0": "visual.vision_model.encoder.layers.19.layer_norm2.bias", "linear_120.w_0": "visual.vision_model.encoder.layers.19.mlp.fc1.weight", "linear_120.b_0": "visual.vision_model.encoder.layers.19.mlp.fc1.bias", "linear_121.w_0": "visual.vision_model.encoder.layers.19.mlp.fc2.weight", "linear_121.b_0": "visual.vision_model.encoder.layers.19.mlp.fc2.bias", "layer_norm_41.w_0": "visual.vision_model.encoder.layers.20.layer_norm1.weight", "layer_norm_41.b_0": "visual.vision_model.encoder.layers.20.layer_norm1.bias", "linear_122.w_0": "visual.vision_model.encoder.layers.20.self_attn.k_proj.weight", "linear_122.b_0": "visual.vision_model.encoder.layers.20.self_attn.k_proj.bias", "linear_123.w_0": "visual.vision_model.encoder.layers.20.self_attn.v_proj.weight", "linear_123.b_0": "visual.vision_model.encoder.layers.20.self_attn.v_proj.bias", "linear_124.w_0": "visual.vision_model.encoder.layers.20.self_attn.q_proj.weight", "linear_124.b_0": "visual.vision_model.encoder.layers.20.self_attn.q_proj.bias", "linear_125.w_0": "visual.vision_model.encoder.layers.20.self_attn.out_proj.weight", "linear_125.b_0": "visual.vision_model.encoder.layers.20.self_attn.out_proj.bias", "layer_norm_42.w_0": "visual.vision_model.encoder.layers.20.layer_norm2.weight", "layer_norm_42.b_0": "visual.vision_model.encoder.layers.20.layer_norm2.bias", "linear_126.w_0": "visual.vision_model.encoder.layers.20.mlp.fc1.weight", "linear_126.b_0": "visual.vision_model.encoder.layers.20.mlp.fc1.bias", "linear_127.w_0": "visual.vision_model.encoder.layers.20.mlp.fc2.weight", "linear_127.b_0": "visual.vision_model.encoder.layers.20.mlp.fc2.bias", "layer_norm_43.w_0": "visual.vision_model.encoder.layers.21.layer_norm1.weight", "layer_norm_43.b_0": "visual.vision_model.encoder.layers.21.layer_norm1.bias", "linear_128.w_0": "visual.vision_model.encoder.layers.21.self_attn.k_proj.weight", "linear_128.b_0": "visual.vision_model.encoder.layers.21.self_attn.k_proj.bias", "linear_129.w_0": "visual.vision_model.encoder.layers.21.self_attn.v_proj.weight", "linear_129.b_0": "visual.vision_model.encoder.layers.21.self_attn.v_proj.bias", "linear_130.w_0": "visual.vision_model.encoder.layers.21.self_attn.q_proj.weight", "linear_130.b_0": "visual.vision_model.encoder.layers.21.self_attn.q_proj.bias", "linear_131.w_0": "visual.vision_model.encoder.layers.21.self_attn.out_proj.weight", "linear_131.b_0": "visual.vision_model.encoder.layers.21.self_attn.out_proj.bias", "layer_norm_44.w_0": "visual.vision_model.encoder.layers.21.layer_norm2.weight", "layer_norm_44.b_0": "visual.vision_model.encoder.layers.21.layer_norm2.bias", "linear_132.w_0": "visual.vision_model.encoder.layers.21.mlp.fc1.weight", "linear_132.b_0": "visual.vision_model.encoder.layers.21.mlp.fc1.bias", "linear_133.w_0": "visual.vision_model.encoder.layers.21.mlp.fc2.weight", "linear_133.b_0": "visual.vision_model.encoder.layers.21.mlp.fc2.bias", "layer_norm_45.w_0": "visual.vision_model.encoder.layers.22.layer_norm1.weight", "layer_norm_45.b_0": "visual.vision_model.encoder.layers.22.layer_norm1.bias", "linear_134.w_0": "visual.vision_model.encoder.layers.22.self_attn.k_proj.weight", "linear_134.b_0": "visual.vision_model.encoder.layers.22.self_attn.k_proj.bias", "linear_135.w_0": "visual.vision_model.encoder.layers.22.self_attn.v_proj.weight", "linear_135.b_0": "visual.vision_model.encoder.layers.22.self_attn.v_proj.bias", "linear_136.w_0": "visual.vision_model.encoder.layers.22.self_attn.q_proj.weight", "linear_136.b_0": "visual.vision_model.encoder.layers.22.self_attn.q_proj.bias", "linear_137.w_0": "visual.vision_model.encoder.layers.22.self_attn.out_proj.weight", "linear_137.b_0": "visual.vision_model.encoder.layers.22.self_attn.out_proj.bias", "layer_norm_46.w_0": "visual.vision_model.encoder.layers.22.layer_norm2.weight", "layer_norm_46.b_0": "visual.vision_model.encoder.layers.22.layer_norm2.bias", "linear_138.w_0": "visual.vision_model.encoder.layers.22.mlp.fc1.weight", "linear_138.b_0": "visual.vision_model.encoder.layers.22.mlp.fc1.bias", "linear_139.w_0": "visual.vision_model.encoder.layers.22.mlp.fc2.weight", "linear_139.b_0": "visual.vision_model.encoder.layers.22.mlp.fc2.bias", "layer_norm_47.w_0": "visual.vision_model.encoder.layers.23.layer_norm1.weight", "layer_norm_47.b_0": "visual.vision_model.encoder.layers.23.layer_norm1.bias", "linear_140.w_0": "visual.vision_model.encoder.layers.23.self_attn.k_proj.weight", "linear_140.b_0": "visual.vision_model.encoder.layers.23.self_attn.k_proj.bias", "linear_141.w_0": "visual.vision_model.encoder.layers.23.self_attn.v_proj.weight", "linear_141.b_0": "visual.vision_model.encoder.layers.23.self_attn.v_proj.bias", "linear_142.w_0": "visual.vision_model.encoder.layers.23.self_attn.q_proj.weight", "linear_142.b_0": "visual.vision_model.encoder.layers.23.self_attn.q_proj.bias", "linear_143.w_0": "visual.vision_model.encoder.layers.23.self_attn.out_proj.weight", "linear_143.b_0": "visual.vision_model.encoder.layers.23.self_attn.out_proj.bias", "layer_norm_48.w_0": "visual.vision_model.encoder.layers.23.layer_norm2.weight", "layer_norm_48.b_0": "visual.vision_model.encoder.layers.23.layer_norm2.bias", "linear_144.w_0": "visual.vision_model.encoder.layers.23.mlp.fc1.weight", "linear_144.b_0": "visual.vision_model.encoder.layers.23.mlp.fc1.bias", "linear_145.w_0": "visual.vision_model.encoder.layers.23.mlp.fc2.weight", "linear_145.b_0": "visual.vision_model.encoder.layers.23.mlp.fc2.bias", "layer_norm_49.w_0": "visual.vision_model.encoder.layers.24.layer_norm1.weight", "layer_norm_49.b_0": "visual.vision_model.encoder.layers.24.layer_norm1.bias", "linear_146.w_0": "visual.vision_model.encoder.layers.24.self_attn.k_proj.weight", "linear_146.b_0": "visual.vision_model.encoder.layers.24.self_attn.k_proj.bias", "linear_147.w_0": "visual.vision_model.encoder.layers.24.self_attn.v_proj.weight", "linear_147.b_0": "visual.vision_model.encoder.layers.24.self_attn.v_proj.bias", "linear_148.w_0": "visual.vision_model.encoder.layers.24.self_attn.q_proj.weight", "linear_148.b_0": "visual.vision_model.encoder.layers.24.self_attn.q_proj.bias", "linear_149.w_0": "visual.vision_model.encoder.layers.24.self_attn.out_proj.weight", "linear_149.b_0": "visual.vision_model.encoder.layers.24.self_attn.out_proj.bias", "layer_norm_50.w_0": "visual.vision_model.encoder.layers.24.layer_norm2.weight", "layer_norm_50.b_0": "visual.vision_model.encoder.layers.24.layer_norm2.bias", "linear_150.w_0": "visual.vision_model.encoder.layers.24.mlp.fc1.weight", "linear_150.b_0": "visual.vision_model.encoder.layers.24.mlp.fc1.bias", "linear_151.w_0": "visual.vision_model.encoder.layers.24.mlp.fc2.weight", "linear_151.b_0": "visual.vision_model.encoder.layers.24.mlp.fc2.bias", "layer_norm_51.w_0": "visual.vision_model.encoder.layers.25.layer_norm1.weight", "layer_norm_51.b_0": "visual.vision_model.encoder.layers.25.layer_norm1.bias", "linear_152.w_0": "visual.vision_model.encoder.layers.25.self_attn.k_proj.weight", "linear_152.b_0": "visual.vision_model.encoder.layers.25.self_attn.k_proj.bias", "linear_153.w_0": "visual.vision_model.encoder.layers.25.self_attn.v_proj.weight", "linear_153.b_0": "visual.vision_model.encoder.layers.25.self_attn.v_proj.bias", "linear_154.w_0": "visual.vision_model.encoder.layers.25.self_attn.q_proj.weight", "linear_154.b_0": "visual.vision_model.encoder.layers.25.self_attn.q_proj.bias", "linear_155.w_0": "visual.vision_model.encoder.layers.25.self_attn.out_proj.weight", "linear_155.b_0": "visual.vision_model.encoder.layers.25.self_attn.out_proj.bias", "layer_norm_52.w_0": "visual.vision_model.encoder.layers.25.layer_norm2.weight", "layer_norm_52.b_0": "visual.vision_model.encoder.layers.25.layer_norm2.bias", "linear_156.w_0": "visual.vision_model.encoder.layers.25.mlp.fc1.weight", "linear_156.b_0": "visual.vision_model.encoder.layers.25.mlp.fc1.bias", "linear_157.w_0": "visual.vision_model.encoder.layers.25.mlp.fc2.weight", "linear_157.b_0": "visual.vision_model.encoder.layers.25.mlp.fc2.bias", "layer_norm_53.w_0": "visual.vision_model.encoder.layers.26.layer_norm1.weight", "layer_norm_53.b_0": "visual.vision_model.encoder.layers.26.layer_norm1.bias", "linear_158.w_0": "visual.vision_model.encoder.layers.26.self_attn.k_proj.weight", "linear_158.b_0": "visual.vision_model.encoder.layers.26.self_attn.k_proj.bias", "linear_159.w_0": "visual.vision_model.encoder.layers.26.self_attn.v_proj.weight", "linear_159.b_0": "visual.vision_model.encoder.layers.26.self_attn.v_proj.bias", "linear_160.w_0": "visual.vision_model.encoder.layers.26.self_attn.q_proj.weight", "linear_160.b_0": "visual.vision_model.encoder.layers.26.self_attn.q_proj.bias", "linear_161.w_0": "visual.vision_model.encoder.layers.26.self_attn.out_proj.weight", "linear_161.b_0": "visual.vision_model.encoder.layers.26.self_attn.out_proj.bias", "layer_norm_54.w_0": "visual.vision_model.encoder.layers.26.layer_norm2.weight", "layer_norm_54.b_0": "visual.vision_model.encoder.layers.26.layer_norm2.bias", "linear_162.w_0": "visual.vision_model.encoder.layers.26.mlp.fc1.weight", "linear_162.b_0": "visual.vision_model.encoder.layers.26.mlp.fc1.bias", "linear_163.w_0": "visual.vision_model.encoder.layers.26.mlp.fc2.weight", "linear_163.b_0": "visual.vision_model.encoder.layers.26.mlp.fc2.bias", "layer_norm_55.w_0": "visual.vision_model.post_layernorm.weight", "layer_norm_55.b_0": "visual.vision_model.post_layernorm.bias", "siglip_multihead_attention_pooling_head_0.w_0": "visual.vision_model.head.probe", "multi_head_attention_0.w_0": "visual.vision_model.head.attention.in_proj_weight", "multi_head_attention_0.w_1": "visual.vision_model.head.attention.in_proj_bias", "linear_164.w_0": "visual.vision_model.head.attention.out_proj.weight", "linear_164.b_0": "visual.vision_model.head.attention.out_proj.bias", "layer_norm_56.w_0": "visual.vision_model.head.layernorm.weight", "layer_norm_56.b_0": "visual.vision_model.head.layernorm.bias", "linear_165.w_0": "visual.vision_model.head.mlp.fc1.weight", "linear_165.b_0": "visual.vision_model.head.mlp.fc1.bias", "linear_166.w_0": "visual.vision_model.head.mlp.fc2.weight", "linear_166.b_0": "visual.vision_model.head.mlp.fc2.bias", "embedding_2.w_0": "model.embed_tokens.weight", "linear_167.w_0": "model.layers.0.self_attn.q_proj.weight", "linear_168.w_0": "model.layers.0.self_attn.k_proj.weight", "linear_169.w_0": "model.layers.0.self_attn.v_proj.weight", "linear_170.w_0": "model.layers.0.self_attn.o_proj.weight", "linear_171.w_0": "model.layers.0.mlp.gate_proj.weight", "linear_172.w_0": "model.layers.0.mlp.up_proj.weight", "linear_173.w_0": "model.layers.0.mlp.down_proj.weight", "create_parameter_0.w_0": "model.layers.0.input_layernorm.weight", "create_parameter_1.w_0": "model.layers.0.post_attention_layernorm.weight", "linear_174.w_0": "model.layers.1.self_attn.q_proj.weight", "linear_175.w_0": "model.layers.1.self_attn.k_proj.weight", "linear_176.w_0": "model.layers.1.self_attn.v_proj.weight", "linear_177.w_0": "model.layers.1.self_attn.o_proj.weight", "linear_178.w_0": "model.layers.1.mlp.gate_proj.weight", "linear_179.w_0": "model.layers.1.mlp.up_proj.weight", "linear_180.w_0": "model.layers.1.mlp.down_proj.weight", "create_parameter_2.w_0": "model.layers.1.input_layernorm.weight", "create_parameter_3.w_0": "model.layers.1.post_attention_layernorm.weight", "linear_181.w_0": "model.layers.2.self_attn.q_proj.weight", "linear_182.w_0": "model.layers.2.self_attn.k_proj.weight", "linear_183.w_0": "model.layers.2.self_attn.v_proj.weight", "linear_184.w_0": "model.layers.2.self_attn.o_proj.weight", "linear_185.w_0": "model.layers.2.mlp.gate_proj.weight", "linear_186.w_0": "model.layers.2.mlp.up_proj.weight", "linear_187.w_0": "model.layers.2.mlp.down_proj.weight", "create_parameter_4.w_0": "model.layers.2.input_layernorm.weight", "create_parameter_5.w_0": "model.layers.2.post_attention_layernorm.weight", "linear_188.w_0": "model.layers.3.self_attn.q_proj.weight", "linear_189.w_0": "model.layers.3.self_attn.k_proj.weight", "linear_190.w_0": "model.layers.3.self_attn.v_proj.weight", "linear_191.w_0": "model.layers.3.self_attn.o_proj.weight", "linear_192.w_0": "model.layers.3.mlp.gate_proj.weight", "linear_193.w_0": "model.layers.3.mlp.up_proj.weight", "linear_194.w_0": "model.layers.3.mlp.down_proj.weight", "create_parameter_6.w_0": "model.layers.3.input_layernorm.weight", "create_parameter_7.w_0": "model.layers.3.post_attention_layernorm.weight", "linear_195.w_0": "model.layers.4.self_attn.q_proj.weight", "linear_196.w_0": "model.layers.4.self_attn.k_proj.weight", "linear_197.w_0": "model.layers.4.self_attn.v_proj.weight", "linear_198.w_0": "model.layers.4.self_attn.o_proj.weight", "linear_199.w_0": "model.layers.4.mlp.gate_proj.weight", "linear_200.w_0": "model.layers.4.mlp.up_proj.weight", "linear_201.w_0": "model.layers.4.mlp.down_proj.weight", "create_parameter_8.w_0": "model.layers.4.input_layernorm.weight", "create_parameter_9.w_0": "model.layers.4.post_attention_layernorm.weight", "linear_202.w_0": "model.layers.5.self_attn.q_proj.weight", "linear_203.w_0": "model.layers.5.self_attn.k_proj.weight", "linear_204.w_0": "model.layers.5.self_attn.v_proj.weight", "linear_205.w_0": "model.layers.5.self_attn.o_proj.weight", "linear_206.w_0": "model.layers.5.mlp.gate_proj.weight", "linear_207.w_0": "model.layers.5.mlp.up_proj.weight", "linear_208.w_0": "model.layers.5.mlp.down_proj.weight", "create_parameter_10.w_0": "model.layers.5.input_layernorm.weight", "create_parameter_11.w_0": "model.layers.5.post_attention_layernorm.weight", "linear_209.w_0": "model.layers.6.self_attn.q_proj.weight", "linear_210.w_0": "model.layers.6.self_attn.k_proj.weight", "linear_211.w_0": "model.layers.6.self_attn.v_proj.weight", "linear_212.w_0": "model.layers.6.self_attn.o_proj.weight", "linear_213.w_0": "model.layers.6.mlp.gate_proj.weight", "linear_214.w_0": "model.layers.6.mlp.up_proj.weight", "linear_215.w_0": "model.layers.6.mlp.down_proj.weight", "create_parameter_12.w_0": "model.layers.6.input_layernorm.weight", "create_parameter_13.w_0": "model.layers.6.post_attention_layernorm.weight", "linear_216.w_0": "model.layers.7.self_attn.q_proj.weight", "linear_217.w_0": "model.layers.7.self_attn.k_proj.weight", "linear_218.w_0": "model.layers.7.self_attn.v_proj.weight", "linear_219.w_0": "model.layers.7.self_attn.o_proj.weight", "linear_220.w_0": "model.layers.7.mlp.gate_proj.weight", "linear_221.w_0": "model.layers.7.mlp.up_proj.weight", "linear_222.w_0": "model.layers.7.mlp.down_proj.weight", "create_parameter_14.w_0": "model.layers.7.input_layernorm.weight", "create_parameter_15.w_0": "model.layers.7.post_attention_layernorm.weight", "linear_223.w_0": "model.layers.8.self_attn.q_proj.weight", "linear_224.w_0": "model.layers.8.self_attn.k_proj.weight", "linear_225.w_0": "model.layers.8.self_attn.v_proj.weight", "linear_226.w_0": "model.layers.8.self_attn.o_proj.weight", "linear_227.w_0": "model.layers.8.mlp.gate_proj.weight", "linear_228.w_0": "model.layers.8.mlp.up_proj.weight", "linear_229.w_0": "model.layers.8.mlp.down_proj.weight", "create_parameter_16.w_0": "model.layers.8.input_layernorm.weight", "create_parameter_17.w_0": "model.layers.8.post_attention_layernorm.weight", "linear_230.w_0": "model.layers.9.self_attn.q_proj.weight", "linear_231.w_0": "model.layers.9.self_attn.k_proj.weight", "linear_232.w_0": "model.layers.9.self_attn.v_proj.weight", "linear_233.w_0": "model.layers.9.self_attn.o_proj.weight", "linear_234.w_0": "model.layers.9.mlp.gate_proj.weight", "linear_235.w_0": "model.layers.9.mlp.up_proj.weight", "linear_236.w_0": "model.layers.9.mlp.down_proj.weight", "create_parameter_18.w_0": "model.layers.9.input_layernorm.weight", "create_parameter_19.w_0": "model.layers.9.post_attention_layernorm.weight", "linear_237.w_0": "model.layers.10.self_attn.q_proj.weight", "linear_238.w_0": "model.layers.10.self_attn.k_proj.weight", "linear_239.w_0": "model.layers.10.self_attn.v_proj.weight", "linear_240.w_0": "model.layers.10.self_attn.o_proj.weight", "linear_241.w_0": "model.layers.10.mlp.gate_proj.weight", "linear_242.w_0": "model.layers.10.mlp.up_proj.weight", "linear_243.w_0": "model.layers.10.mlp.down_proj.weight", "create_parameter_20.w_0": "model.layers.10.input_layernorm.weight", "create_parameter_21.w_0": "model.layers.10.post_attention_layernorm.weight", "linear_244.w_0": "model.layers.11.self_attn.q_proj.weight", "linear_245.w_0": "model.layers.11.self_attn.k_proj.weight", "linear_246.w_0": "model.layers.11.self_attn.v_proj.weight", "linear_247.w_0": "model.layers.11.self_attn.o_proj.weight", "linear_248.w_0": "model.layers.11.mlp.gate_proj.weight", "linear_249.w_0": "model.layers.11.mlp.up_proj.weight", "linear_250.w_0": "model.layers.11.mlp.down_proj.weight", "create_parameter_22.w_0": "model.layers.11.input_layernorm.weight", "create_parameter_23.w_0": "model.layers.11.post_attention_layernorm.weight", "linear_251.w_0": "model.layers.12.self_attn.q_proj.weight", "linear_252.w_0": "model.layers.12.self_attn.k_proj.weight", "linear_253.w_0": "model.layers.12.self_attn.v_proj.weight", "linear_254.w_0": "model.layers.12.self_attn.o_proj.weight", "linear_255.w_0": "model.layers.12.mlp.gate_proj.weight", "linear_256.w_0": "model.layers.12.mlp.up_proj.weight", "linear_257.w_0": "model.layers.12.mlp.down_proj.weight", "create_parameter_24.w_0": "model.layers.12.input_layernorm.weight", "create_parameter_25.w_0": "model.layers.12.post_attention_layernorm.weight", "linear_258.w_0": "model.layers.13.self_attn.q_proj.weight", "linear_259.w_0": "model.layers.13.self_attn.k_proj.weight", "linear_260.w_0": "model.layers.13.self_attn.v_proj.weight", "linear_261.w_0": "model.layers.13.self_attn.o_proj.weight", "linear_262.w_0": "model.layers.13.mlp.gate_proj.weight", "linear_263.w_0": "model.layers.13.mlp.up_proj.weight", "linear_264.w_0": "model.layers.13.mlp.down_proj.weight", "create_parameter_26.w_0": "model.layers.13.input_layernorm.weight", "create_parameter_27.w_0": "model.layers.13.post_attention_layernorm.weight", "linear_265.w_0": "model.layers.14.self_attn.q_proj.weight", "linear_266.w_0": "model.layers.14.self_attn.k_proj.weight", "linear_267.w_0": "model.layers.14.self_attn.v_proj.weight", "linear_268.w_0": "model.layers.14.self_attn.o_proj.weight", "linear_269.w_0": "model.layers.14.mlp.gate_proj.weight", "linear_270.w_0": "model.layers.14.mlp.up_proj.weight", "linear_271.w_0": "model.layers.14.mlp.down_proj.weight", "create_parameter_28.w_0": "model.layers.14.input_layernorm.weight", "create_parameter_29.w_0": "model.layers.14.post_attention_layernorm.weight", "linear_272.w_0": "model.layers.15.self_attn.q_proj.weight", "linear_273.w_0": "model.layers.15.self_attn.k_proj.weight", "linear_274.w_0": "model.layers.15.self_attn.v_proj.weight", "linear_275.w_0": "model.layers.15.self_attn.o_proj.weight", "linear_276.w_0": "model.layers.15.mlp.gate_proj.weight", "linear_277.w_0": "model.layers.15.mlp.up_proj.weight", "linear_278.w_0": "model.layers.15.mlp.down_proj.weight", "create_parameter_30.w_0": "model.layers.15.input_layernorm.weight", "create_parameter_31.w_0": "model.layers.15.post_attention_layernorm.weight", "linear_279.w_0": "model.layers.16.self_attn.q_proj.weight", "linear_280.w_0": "model.layers.16.self_attn.k_proj.weight", "linear_281.w_0": "model.layers.16.self_attn.v_proj.weight", "linear_282.w_0": "model.layers.16.self_attn.o_proj.weight", "linear_283.w_0": "model.layers.16.mlp.gate_proj.weight", "linear_284.w_0": "model.layers.16.mlp.up_proj.weight", "linear_285.w_0": "model.layers.16.mlp.down_proj.weight", "create_parameter_32.w_0": "model.layers.16.input_layernorm.weight", "create_parameter_33.w_0": "model.layers.16.post_attention_layernorm.weight", "linear_286.w_0": "model.layers.17.self_attn.q_proj.weight", "linear_287.w_0": "model.layers.17.self_attn.k_proj.weight", "linear_288.w_0": "model.layers.17.self_attn.v_proj.weight", "linear_289.w_0": "model.layers.17.self_attn.o_proj.weight", "linear_290.w_0": "model.layers.17.mlp.gate_proj.weight", "linear_291.w_0": "model.layers.17.mlp.up_proj.weight", "linear_292.w_0": "model.layers.17.mlp.down_proj.weight", "create_parameter_34.w_0": "model.layers.17.input_layernorm.weight", "create_parameter_35.w_0": "model.layers.17.post_attention_layernorm.weight", "create_parameter_36.w_0": "model.norm.weight", "linear_293.w_0": "lm_head.weight"}
|
checkpoint-378/tokenizer.model
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:34ef7db83df785924fb83d7b887b6e822a031c56e15cff40aaf9b982988180df
|
| 3 |
+
size 1614363
|
checkpoint-378/tokenizer_config.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
checkpoint-378/trainer_state.json
ADDED
|
@@ -0,0 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"best_metric": null,
|
| 3 |
+
"best_model_checkpoint": null,
|
| 4 |
+
"consumed_samples": 0,
|
| 5 |
+
"epoch": 0.8978622327790974,
|
| 6 |
+
"global_step": 378,
|
| 7 |
+
"is_local_process_zero": true,
|
| 8 |
+
"is_world_process_zero": true,
|
| 9 |
+
"log_history": [],
|
| 10 |
+
"max_steps": 421,
|
| 11 |
+
"num_train_epochs": 9223372036854775807,
|
| 12 |
+
"total_flos": 0,
|
| 13 |
+
"trial_name": null,
|
| 14 |
+
"trial_params": null
|
| 15 |
+
}
|
checkpoint-378/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0888f0d3f65facdf04c57f4bd4b2036e3fec9976c17b7eb281a443129536c8e7
|
| 3 |
+
size 6425
|
checkpoint-420/added_tokens.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"<unk>": 0, "<s>": 1, "</s>": 2, "0": 3, "1": 4, "2": 5, "3": 6, "4": 7, "5": 8, "6": 9, "7": 10, "8": 11, "9": 12, "<|end_of_sentence|>": 100272, "<|begin_of_sentence|>": 100273, "<mask:1>": 100274, "<mask:4>": 100277, "<mask:5>": 100278, "<mask:6>": 100279, "<mask:7>": 100280, "<|IMAGE_PLACEHOLDER|>": 100295, "<|AUDIO_PLACEHOLDER|>": 100296, "<|LOC_0|>": 100297, "<|LOC_1|>": 100298, "<|LOC_2|>": 100299, "<|LOC_3|>": 100300, "<|LOC_4|>": 100301, "<|LOC_5|>": 100302, "<|LOC_6|>": 100303, "<|LOC_7|>": 100304, "<|LOC_8|>": 100305, "<|LOC_9|>": 100306, "<|LOC_10|>": 100307, "<|LOC_11|>": 100308, "<|LOC_12|>": 100309, "<|LOC_13|>": 100310, "<|LOC_14|>": 100311, "<|LOC_15|>": 100312, "<|LOC_16|>": 100313, "<|LOC_17|>": 100314, "<|LOC_18|>": 100315, "<|LOC_19|>": 100316, "<|LOC_20|>": 100317, "<|LOC_21|>": 100318, "<|LOC_22|>": 100319, "<|LOC_23|>": 100320, "<|LOC_24|>": 100321, "<|LOC_25|>": 100322, "<|LOC_26|>": 100323, "<|LOC_27|>": 100324, "<|LOC_28|>": 100325, "<|LOC_29|>": 100326, "<|LOC_30|>": 100327, "<|LOC_31|>": 100328, "<|LOC_32|>": 100329, "<|LOC_33|>": 100330, "<|LOC_34|>": 100331, "<|LOC_35|>": 100332, "<|LOC_36|>": 100333, "<|LOC_37|>": 100334, "<|LOC_38|>": 100335, "<|LOC_39|>": 100336, "<|LOC_40|>": 100337, "<|LOC_41|>": 100338, "<|LOC_42|>": 100339, "<|LOC_43|>": 100340, "<|LOC_44|>": 100341, "<|LOC_45|>": 100342, "<|LOC_46|>": 100343, "<|LOC_47|>": 100344, "<|LOC_48|>": 100345, "<|LOC_49|>": 100346, "<|LOC_50|>": 100347, "<|LOC_51|>": 100348, "<|LOC_52|>": 100349, "<|LOC_53|>": 100350, "<|LOC_54|>": 100351, "<|LOC_55|>": 100352, "<|LOC_56|>": 100353, "<|LOC_57|>": 100354, "<|LOC_58|>": 100355, "<|LOC_59|>": 100356, "<|LOC_60|>": 100357, "<|LOC_61|>": 100358, "<|LOC_62|>": 100359, "<|LOC_63|>": 100360, "<|LOC_64|>": 100361, "<|LOC_65|>": 100362, "<|LOC_66|>": 100363, "<|LOC_67|>": 100364, "<|LOC_68|>": 100365, "<|LOC_69|>": 100366, "<|LOC_70|>": 100367, "<|LOC_71|>": 100368, "<|LOC_72|>": 100369, "<|LOC_73|>": 100370, "<|LOC_74|>": 100371, "<|LOC_75|>": 100372, "<|LOC_76|>": 100373, "<|LOC_77|>": 100374, "<|LOC_78|>": 100375, "<|LOC_79|>": 100376, "<|LOC_80|>": 100377, "<|LOC_81|>": 100378, "<|LOC_82|>": 100379, "<|LOC_83|>": 100380, "<|LOC_84|>": 100381, "<|LOC_85|>": 100382, "<|LOC_86|>": 100383, "<|LOC_87|>": 100384, "<|LOC_88|>": 100385, "<|LOC_89|>": 100386, "<|LOC_90|>": 100387, "<|LOC_91|>": 100388, "<|LOC_92|>": 100389, "<|LOC_93|>": 100390, "<|LOC_94|>": 100391, "<|LOC_95|>": 100392, "<|LOC_96|>": 100393, "<|LOC_97|>": 100394, "<|LOC_98|>": 100395, "<|LOC_99|>": 100396, "<|LOC_100|>": 100397, "<|LOC_101|>": 100398, "<|LOC_102|>": 100399, "<|LOC_103|>": 100400, "<|LOC_104|>": 100401, "<|LOC_105|>": 100402, "<|LOC_106|>": 100403, "<|LOC_107|>": 100404, "<|LOC_108|>": 100405, "<|LOC_109|>": 100406, "<|LOC_110|>": 100407, "<|LOC_111|>": 100408, "<|LOC_112|>": 100409, "<|LOC_113|>": 100410, "<|LOC_114|>": 100411, "<|LOC_115|>": 100412, "<|LOC_116|>": 100413, "<|LOC_117|>": 100414, "<|LOC_118|>": 100415, "<|LOC_119|>": 100416, "<|LOC_120|>": 100417, "<|LOC_121|>": 100418, "<|LOC_122|>": 100419, "<|LOC_123|>": 100420, "<|LOC_124|>": 100421, "<|LOC_125|>": 100422, "<|LOC_126|>": 100423, "<|LOC_127|>": 100424, "<|LOC_128|>": 100425, "<|LOC_129|>": 100426, "<|LOC_130|>": 100427, "<|LOC_131|>": 100428, "<|LOC_132|>": 100429, "<|LOC_133|>": 100430, "<|LOC_134|>": 100431, "<|LOC_135|>": 100432, "<|LOC_136|>": 100433, "<|LOC_137|>": 100434, "<|LOC_138|>": 100435, "<|LOC_139|>": 100436, "<|LOC_140|>": 100437, "<|LOC_141|>": 100438, "<|LOC_142|>": 100439, "<|LOC_143|>": 100440, "<|LOC_144|>": 100441, "<|LOC_145|>": 100442, "<|LOC_146|>": 100443, "<|LOC_147|>": 100444, "<|LOC_148|>": 100445, "<|LOC_149|>": 100446, "<|LOC_150|>": 100447, "<|LOC_151|>": 100448, "<|LOC_152|>": 100449, "<|LOC_153|>": 100450, "<|LOC_154|>": 100451, "<|LOC_155|>": 100452, "<|LOC_156|>": 100453, "<|LOC_157|>": 100454, "<|LOC_158|>": 100455, "<|LOC_159|>": 100456, "<|LOC_160|>": 100457, "<|LOC_161|>": 100458, "<|LOC_162|>": 100459, "<|LOC_163|>": 100460, "<|LOC_164|>": 100461, "<|LOC_165|>": 100462, "<|LOC_166|>": 100463, "<|LOC_167|>": 100464, "<|LOC_168|>": 100465, "<|LOC_169|>": 100466, "<|LOC_170|>": 100467, "<|LOC_171|>": 100468, "<|LOC_172|>": 100469, "<|LOC_173|>": 100470, "<|LOC_174|>": 100471, "<|LOC_175|>": 100472, "<|LOC_176|>": 100473, "<|LOC_177|>": 100474, "<|LOC_178|>": 100475, "<|LOC_179|>": 100476, "<|LOC_180|>": 100477, "<|LOC_181|>": 100478, "<|LOC_182|>": 100479, "<|LOC_183|>": 100480, "<|LOC_184|>": 100481, "<|LOC_185|>": 100482, "<|LOC_186|>": 100483, "<|LOC_187|>": 100484, "<|LOC_188|>": 100485, "<|LOC_189|>": 100486, "<|LOC_190|>": 100487, "<|LOC_191|>": 100488, "<|LOC_192|>": 100489, "<|LOC_193|>": 100490, "<|LOC_194|>": 100491, "<|LOC_195|>": 100492, "<|LOC_196|>": 100493, "<|LOC_197|>": 100494, "<|LOC_198|>": 100495, "<|LOC_199|>": 100496, "<|LOC_200|>": 100497, "<|LOC_201|>": 100498, "<|LOC_202|>": 100499, "<|LOC_203|>": 100500, "<|LOC_204|>": 100501, "<|LOC_205|>": 100502, "<|LOC_206|>": 100503, "<|LOC_207|>": 100504, "<|LOC_208|>": 100505, "<|LOC_209|>": 100506, "<|LOC_210|>": 100507, "<|LOC_211|>": 100508, "<|LOC_212|>": 100509, "<|LOC_213|>": 100510, "<|LOC_214|>": 100511, "<|LOC_215|>": 100512, "<|LOC_216|>": 100513, "<|LOC_217|>": 100514, "<|LOC_218|>": 100515, "<|LOC_219|>": 100516, "<|LOC_220|>": 100517, "<|LOC_221|>": 100518, "<|LOC_222|>": 100519, "<|LOC_223|>": 100520, "<|LOC_224|>": 100521, "<|LOC_225|>": 100522, "<|LOC_226|>": 100523, "<|LOC_227|>": 100524, "<|LOC_228|>": 100525, "<|LOC_229|>": 100526, "<|LOC_230|>": 100527, "<|LOC_231|>": 100528, "<|LOC_232|>": 100529, "<|LOC_233|>": 100530, "<|LOC_234|>": 100531, "<|LOC_235|>": 100532, "<|LOC_236|>": 100533, "<|LOC_237|>": 100534, "<|LOC_238|>": 100535, "<|LOC_239|>": 100536, "<|LOC_240|>": 100537, "<|LOC_241|>": 100538, "<|LOC_242|>": 100539, "<|LOC_243|>": 100540, "<|LOC_244|>": 100541, "<|LOC_245|>": 100542, "<|LOC_246|>": 100543, "<|LOC_247|>": 100544, "<|LOC_248|>": 100545, "<|LOC_249|>": 100546, "<|LOC_250|>": 100547, "<|LOC_251|>": 100548, "<|LOC_252|>": 100549, "<|LOC_253|>": 100550, "<|LOC_254|>": 100551, "<|LOC_255|>": 100552, "<|LOC_256|>": 100553, "<|LOC_257|>": 100554, "<|LOC_258|>": 100555, "<|LOC_259|>": 100556, "<|LOC_260|>": 100557, "<|LOC_261|>": 100558, "<|LOC_262|>": 100559, "<|LOC_263|>": 100560, "<|LOC_264|>": 100561, "<|LOC_265|>": 100562, "<|LOC_266|>": 100563, "<|LOC_267|>": 100564, "<|LOC_268|>": 100565, "<|LOC_269|>": 100566, "<|LOC_270|>": 100567, "<|LOC_271|>": 100568, "<|LOC_272|>": 100569, "<|LOC_273|>": 100570, "<|LOC_274|>": 100571, "<|LOC_275|>": 100572, "<|LOC_276|>": 100573, "<|LOC_277|>": 100574, "<|LOC_278|>": 100575, "<|LOC_279|>": 100576, "<|LOC_280|>": 100577, "<|LOC_281|>": 100578, "<|LOC_282|>": 100579, "<|LOC_283|>": 100580, "<|LOC_284|>": 100581, "<|LOC_285|>": 100582, "<|LOC_286|>": 100583, "<|LOC_287|>": 100584, "<|LOC_288|>": 100585, "<|LOC_289|>": 100586, "<|LOC_290|>": 100587, "<|LOC_291|>": 100588, "<|LOC_292|>": 100589, "<|LOC_293|>": 100590, "<|LOC_294|>": 100591, "<|LOC_295|>": 100592, "<|LOC_296|>": 100593, "<|LOC_297|>": 100594, "<|LOC_298|>": 100595, "<|LOC_299|>": 100596, "<|LOC_300|>": 100597, "<|LOC_301|>": 100598, "<|LOC_302|>": 100599, "<|LOC_303|>": 100600, "<|LOC_304|>": 100601, "<|LOC_305|>": 100602, "<|LOC_306|>": 100603, "<|LOC_307|>": 100604, "<|LOC_308|>": 100605, "<|LOC_309|>": 100606, "<|LOC_310|>": 100607, "<|LOC_311|>": 100608, "<|LOC_312|>": 100609, "<|LOC_313|>": 100610, "<|LOC_314|>": 100611, "<|LOC_315|>": 100612, "<|LOC_316|>": 100613, "<|LOC_317|>": 100614, "<|LOC_318|>": 100615, "<|LOC_319|>": 100616, "<|LOC_320|>": 100617, "<|LOC_321|>": 100618, "<|LOC_322|>": 100619, "<|LOC_323|>": 100620, "<|LOC_324|>": 100621, "<|LOC_325|>": 100622, "<|LOC_326|>": 100623, "<|LOC_327|>": 100624, "<|LOC_328|>": 100625, "<|LOC_329|>": 100626, "<|LOC_330|>": 100627, "<|LOC_331|>": 100628, "<|LOC_332|>": 100629, "<|LOC_333|>": 100630, "<|LOC_334|>": 100631, "<|LOC_335|>": 100632, "<|LOC_336|>": 100633, "<|LOC_337|>": 100634, "<|LOC_338|>": 100635, "<|LOC_339|>": 100636, "<|LOC_340|>": 100637, "<|LOC_341|>": 100638, "<|LOC_342|>": 100639, "<|LOC_343|>": 100640, "<|LOC_344|>": 100641, "<|LOC_345|>": 100642, "<|LOC_346|>": 100643, "<|LOC_347|>": 100644, "<|LOC_348|>": 100645, "<|LOC_349|>": 100646, "<|LOC_350|>": 100647, "<|LOC_351|>": 100648, "<|LOC_352|>": 100649, "<|LOC_353|>": 100650, "<|LOC_354|>": 100651, "<|LOC_355|>": 100652, "<|LOC_356|>": 100653, "<|LOC_357|>": 100654, "<|LOC_358|>": 100655, "<|LOC_359|>": 100656, "<|LOC_360|>": 100657, "<|LOC_361|>": 100658, "<|LOC_362|>": 100659, "<|LOC_363|>": 100660, "<|LOC_364|>": 100661, "<|LOC_365|>": 100662, "<|LOC_366|>": 100663, "<|LOC_367|>": 100664, "<|LOC_368|>": 100665, "<|LOC_369|>": 100666, "<|LOC_370|>": 100667, "<|LOC_371|>": 100668, "<|LOC_372|>": 100669, "<|LOC_373|>": 100670, "<|LOC_374|>": 100671, "<|LOC_375|>": 100672, "<|LOC_376|>": 100673, "<|LOC_377|>": 100674, "<|LOC_378|>": 100675, "<|LOC_379|>": 100676, "<|LOC_380|>": 100677, "<|LOC_381|>": 100678, "<|LOC_382|>": 100679, "<|LOC_383|>": 100680, "<|LOC_384|>": 100681, "<|LOC_385|>": 100682, "<|LOC_386|>": 100683, "<|LOC_387|>": 100684, "<|LOC_388|>": 100685, "<|LOC_389|>": 100686, "<|LOC_390|>": 100687, "<|LOC_391|>": 100688, "<|LOC_392|>": 100689, "<|LOC_393|>": 100690, "<|LOC_394|>": 100691, "<|LOC_395|>": 100692, "<|LOC_396|>": 100693, "<|LOC_397|>": 100694, "<|LOC_398|>": 100695, "<|LOC_399|>": 100696, "<|LOC_400|>": 100697, "<|LOC_401|>": 100698, "<|LOC_402|>": 100699, "<|LOC_403|>": 100700, "<|LOC_404|>": 100701, "<|LOC_405|>": 100702, "<|LOC_406|>": 100703, "<|LOC_407|>": 100704, "<|LOC_408|>": 100705, "<|LOC_409|>": 100706, "<|LOC_410|>": 100707, "<|LOC_411|>": 100708, "<|LOC_412|>": 100709, "<|LOC_413|>": 100710, "<|LOC_414|>": 100711, "<|LOC_415|>": 100712, "<|LOC_416|>": 100713, "<|LOC_417|>": 100714, "<|LOC_418|>": 100715, "<|LOC_419|>": 100716, "<|LOC_420|>": 100717, "<|LOC_421|>": 100718, "<|LOC_422|>": 100719, "<|LOC_423|>": 100720, "<|LOC_424|>": 100721, "<|LOC_425|>": 100722, "<|LOC_426|>": 100723, "<|LOC_427|>": 100724, "<|LOC_428|>": 100725, "<|LOC_429|>": 100726, "<|LOC_430|>": 100727, "<|LOC_431|>": 100728, "<|LOC_432|>": 100729, "<|LOC_433|>": 100730, "<|LOC_434|>": 100731, "<|LOC_435|>": 100732, "<|LOC_436|>": 100733, "<|LOC_437|>": 100734, "<|LOC_438|>": 100735, "<|LOC_439|>": 100736, "<|LOC_440|>": 100737, "<|LOC_441|>": 100738, "<|LOC_442|>": 100739, "<|LOC_443|>": 100740, "<|LOC_444|>": 100741, "<|LOC_445|>": 100742, "<|LOC_446|>": 100743, "<|LOC_447|>": 100744, "<|LOC_448|>": 100745, "<|LOC_449|>": 100746, "<|LOC_450|>": 100747, "<|LOC_451|>": 100748, "<|LOC_452|>": 100749, "<|LOC_453|>": 100750, "<|LOC_454|>": 100751, "<|LOC_455|>": 100752, "<|LOC_456|>": 100753, "<|LOC_457|>": 100754, "<|LOC_458|>": 100755, "<|LOC_459|>": 100756, "<|LOC_460|>": 100757, "<|LOC_461|>": 100758, "<|LOC_462|>": 100759, "<|LOC_463|>": 100760, "<|LOC_464|>": 100761, "<|LOC_465|>": 100762, "<|LOC_466|>": 100763, "<|LOC_467|>": 100764, "<|LOC_468|>": 100765, "<|LOC_469|>": 100766, "<|LOC_470|>": 100767, "<|LOC_471|>": 100768, "<|LOC_472|>": 100769, "<|LOC_473|>": 100770, "<|LOC_474|>": 100771, "<|LOC_475|>": 100772, "<|LOC_476|>": 100773, "<|LOC_477|>": 100774, "<|LOC_478|>": 100775, "<|LOC_479|>": 100776, "<|LOC_480|>": 100777, "<|LOC_481|>": 100778, "<|LOC_482|>": 100779, "<|LOC_483|>": 100780, "<|LOC_484|>": 100781, "<|LOC_485|>": 100782, "<|LOC_486|>": 100783, "<|LOC_487|>": 100784, "<|LOC_488|>": 100785, "<|LOC_489|>": 100786, "<|LOC_490|>": 100787, "<|LOC_491|>": 100788, "<|LOC_492|>": 100789, "<|LOC_493|>": 100790, "<|LOC_494|>": 100791, "<|LOC_495|>": 100792, "<|LOC_496|>": 100793, "<|LOC_497|>": 100794, "<|LOC_498|>": 100795, "<|LOC_499|>": 100796, "<|LOC_500|>": 100797, "<|LOC_501|>": 100798, "<|LOC_502|>": 100799, "<|LOC_503|>": 100800, "<|LOC_504|>": 100801, "<|LOC_505|>": 100802, "<|LOC_506|>": 100803, "<|LOC_507|>": 100804, "<|LOC_508|>": 100805, "<|LOC_509|>": 100806, "<|LOC_510|>": 100807, "<|LOC_511|>": 100808, "<|LOC_512|>": 100809, "<|LOC_513|>": 100810, "<|LOC_514|>": 100811, "<|LOC_515|>": 100812, "<|LOC_516|>": 100813, "<|LOC_517|>": 100814, "<|LOC_518|>": 100815, "<|LOC_519|>": 100816, "<|LOC_520|>": 100817, "<|LOC_521|>": 100818, "<|LOC_522|>": 100819, "<|LOC_523|>": 100820, "<|LOC_524|>": 100821, "<|LOC_525|>": 100822, "<|LOC_526|>": 100823, "<|LOC_527|>": 100824, "<|LOC_528|>": 100825, "<|LOC_529|>": 100826, "<|LOC_530|>": 100827, "<|LOC_531|>": 100828, "<|LOC_532|>": 100829, "<|LOC_533|>": 100830, "<|LOC_534|>": 100831, "<|LOC_535|>": 100832, "<|LOC_536|>": 100833, "<|LOC_537|>": 100834, "<|LOC_538|>": 100835, "<|LOC_539|>": 100836, "<|LOC_540|>": 100837, "<|LOC_541|>": 100838, "<|LOC_542|>": 100839, "<|LOC_543|>": 100840, "<|LOC_544|>": 100841, "<|LOC_545|>": 100842, "<|LOC_546|>": 100843, "<|LOC_547|>": 100844, "<|LOC_548|>": 100845, "<|LOC_549|>": 100846, "<|LOC_550|>": 100847, "<|LOC_551|>": 100848, "<|LOC_552|>": 100849, "<|LOC_553|>": 100850, "<|LOC_554|>": 100851, "<|LOC_555|>": 100852, "<|LOC_556|>": 100853, "<|LOC_557|>": 100854, "<|LOC_558|>": 100855, "<|LOC_559|>": 100856, "<|LOC_560|>": 100857, "<|LOC_561|>": 100858, "<|LOC_562|>": 100859, "<|LOC_563|>": 100860, "<|LOC_564|>": 100861, "<|LOC_565|>": 100862, "<|LOC_566|>": 100863, "<|LOC_567|>": 100864, "<|LOC_568|>": 100865, "<|LOC_569|>": 100866, "<|LOC_570|>": 100867, "<|LOC_571|>": 100868, "<|LOC_572|>": 100869, "<|LOC_573|>": 100870, "<|LOC_574|>": 100871, "<|LOC_575|>": 100872, "<|LOC_576|>": 100873, "<|LOC_577|>": 100874, "<|LOC_578|>": 100875, "<|LOC_579|>": 100876, "<|LOC_580|>": 100877, "<|LOC_581|>": 100878, "<|LOC_582|>": 100879, "<|LOC_583|>": 100880, "<|LOC_584|>": 100881, "<|LOC_585|>": 100882, "<|LOC_586|>": 100883, "<|LOC_587|>": 100884, "<|LOC_588|>": 100885, "<|LOC_589|>": 100886, "<|LOC_590|>": 100887, "<|LOC_591|>": 100888, "<|LOC_592|>": 100889, "<|LOC_593|>": 100890, "<|LOC_594|>": 100891, "<|LOC_595|>": 100892, "<|LOC_596|>": 100893, "<|LOC_597|>": 100894, "<|LOC_598|>": 100895, "<|LOC_599|>": 100896, "<|LOC_600|>": 100897, "<|LOC_601|>": 100898, "<|LOC_602|>": 100899, "<|LOC_603|>": 100900, "<|LOC_604|>": 100901, "<|LOC_605|>": 100902, "<|LOC_606|>": 100903, "<|LOC_607|>": 100904, "<|LOC_608|>": 100905, "<|LOC_609|>": 100906, "<|LOC_610|>": 100907, "<|LOC_611|>": 100908, "<|LOC_612|>": 100909, "<|LOC_613|>": 100910, "<|LOC_614|>": 100911, "<|LOC_615|>": 100912, "<|LOC_616|>": 100913, "<|LOC_617|>": 100914, "<|LOC_618|>": 100915, "<|LOC_619|>": 100916, "<|LOC_620|>": 100917, "<|LOC_621|>": 100918, "<|LOC_622|>": 100919, "<|LOC_623|>": 100920, "<|LOC_624|>": 100921, "<|LOC_625|>": 100922, "<|LOC_626|>": 100923, "<|LOC_627|>": 100924, "<|LOC_628|>": 100925, "<|LOC_629|>": 100926, "<|LOC_630|>": 100927, "<|LOC_631|>": 100928, "<|LOC_632|>": 100929, "<|LOC_633|>": 100930, "<|LOC_634|>": 100931, "<|LOC_635|>": 100932, "<|LOC_636|>": 100933, "<|LOC_637|>": 100934, "<|LOC_638|>": 100935, "<|LOC_639|>": 100936, "<|LOC_640|>": 100937, "<|LOC_641|>": 100938, "<|LOC_642|>": 100939, "<|LOC_643|>": 100940, "<|LOC_644|>": 100941, "<|LOC_645|>": 100942, "<|LOC_646|>": 100943, "<|LOC_647|>": 100944, "<|LOC_648|>": 100945, "<|LOC_649|>": 100946, "<|LOC_650|>": 100947, "<|LOC_651|>": 100948, "<|LOC_652|>": 100949, "<|LOC_653|>": 100950, "<|LOC_654|>": 100951, "<|LOC_655|>": 100952, "<|LOC_656|>": 100953, "<|LOC_657|>": 100954, "<|LOC_658|>": 100955, "<|LOC_659|>": 100956, "<|LOC_660|>": 100957, "<|LOC_661|>": 100958, "<|LOC_662|>": 100959, "<|LOC_663|>": 100960, "<|LOC_664|>": 100961, "<|LOC_665|>": 100962, "<|LOC_666|>": 100963, "<|LOC_667|>": 100964, "<|LOC_668|>": 100965, "<|LOC_669|>": 100966, "<|LOC_670|>": 100967, "<|LOC_671|>": 100968, "<|LOC_672|>": 100969, "<|LOC_673|>": 100970, "<|LOC_674|>": 100971, "<|LOC_675|>": 100972, "<|LOC_676|>": 100973, "<|LOC_677|>": 100974, "<|LOC_678|>": 100975, "<|LOC_679|>": 100976, "<|LOC_680|>": 100977, "<|LOC_681|>": 100978, "<|LOC_682|>": 100979, "<|LOC_683|>": 100980, "<|LOC_684|>": 100981, "<|LOC_685|>": 100982, "<|LOC_686|>": 100983, "<|LOC_687|>": 100984, "<|LOC_688|>": 100985, "<|LOC_689|>": 100986, "<|LOC_690|>": 100987, "<|LOC_691|>": 100988, "<|LOC_692|>": 100989, "<|LOC_693|>": 100990, "<|LOC_694|>": 100991, "<|LOC_695|>": 100992, "<|LOC_696|>": 100993, "<|LOC_697|>": 100994, "<|LOC_698|>": 100995, "<|LOC_699|>": 100996, "<|LOC_700|>": 100997, "<|LOC_701|>": 100998, "<|LOC_702|>": 100999, "<|LOC_703|>": 101000, "<|LOC_704|>": 101001, "<|LOC_705|>": 101002, "<|LOC_706|>": 101003, "<|LOC_707|>": 101004, "<|LOC_708|>": 101005, "<|LOC_709|>": 101006, "<|LOC_710|>": 101007, "<|LOC_711|>": 101008, "<|LOC_712|>": 101009, "<|LOC_713|>": 101010, "<|LOC_714|>": 101011, "<|LOC_715|>": 101012, "<|LOC_716|>": 101013, "<|LOC_717|>": 101014, "<|LOC_718|>": 101015, "<|LOC_719|>": 101016, "<|LOC_720|>": 101017, "<|LOC_721|>": 101018, "<|LOC_722|>": 101019, "<|LOC_723|>": 101020, "<|LOC_724|>": 101021, "<|LOC_725|>": 101022, "<|LOC_726|>": 101023, "<|LOC_727|>": 101024, "<|LOC_728|>": 101025, "<|LOC_729|>": 101026, "<|LOC_730|>": 101027, "<|LOC_731|>": 101028, "<|LOC_732|>": 101029, "<|LOC_733|>": 101030, "<|LOC_734|>": 101031, "<|LOC_735|>": 101032, "<|LOC_736|>": 101033, "<|LOC_737|>": 101034, "<|LOC_738|>": 101035, "<|LOC_739|>": 101036, "<|LOC_740|>": 101037, "<|LOC_741|>": 101038, "<|LOC_742|>": 101039, "<|LOC_743|>": 101040, "<|LOC_744|>": 101041, "<|LOC_745|>": 101042, "<|LOC_746|>": 101043, "<|LOC_747|>": 101044, "<|LOC_748|>": 101045, "<|LOC_749|>": 101046, "<|LOC_750|>": 101047, "<|LOC_751|>": 101048, "<|LOC_752|>": 101049, "<|LOC_753|>": 101050, "<|LOC_754|>": 101051, "<|LOC_755|>": 101052, "<|LOC_756|>": 101053, "<|LOC_757|>": 101054, "<|LOC_758|>": 101055, "<|LOC_759|>": 101056, "<|LOC_760|>": 101057, "<|LOC_761|>": 101058, "<|LOC_762|>": 101059, "<|LOC_763|>": 101060, "<|LOC_764|>": 101061, "<|LOC_765|>": 101062, "<|LOC_766|>": 101063, "<|LOC_767|>": 101064, "<|LOC_768|>": 101065, "<|LOC_769|>": 101066, "<|LOC_770|>": 101067, "<|LOC_771|>": 101068, "<|LOC_772|>": 101069, "<|LOC_773|>": 101070, "<|LOC_774|>": 101071, "<|LOC_775|>": 101072, "<|LOC_776|>": 101073, "<|LOC_777|>": 101074, "<|LOC_778|>": 101075, "<|LOC_779|>": 101076, "<|LOC_780|>": 101077, "<|LOC_781|>": 101078, "<|LOC_782|>": 101079, "<|LOC_783|>": 101080, "<|LOC_784|>": 101081, "<|LOC_785|>": 101082, "<|LOC_786|>": 101083, "<|LOC_787|>": 101084, "<|LOC_788|>": 101085, "<|LOC_789|>": 101086, "<|LOC_790|>": 101087, "<|LOC_791|>": 101088, "<|LOC_792|>": 101089, "<|LOC_793|>": 101090, "<|LOC_794|>": 101091, "<|LOC_795|>": 101092, "<|LOC_796|>": 101093, "<|LOC_797|>": 101094, "<|LOC_798|>": 101095, "<|LOC_799|>": 101096, "<|LOC_800|>": 101097, "<|LOC_801|>": 101098, "<|LOC_802|>": 101099, "<|LOC_803|>": 101100, "<|LOC_804|>": 101101, "<|LOC_805|>": 101102, "<|LOC_806|>": 101103, "<|LOC_807|>": 101104, "<|LOC_808|>": 101105, "<|LOC_809|>": 101106, "<|LOC_810|>": 101107, "<|LOC_811|>": 101108, "<|LOC_812|>": 101109, "<|LOC_813|>": 101110, "<|LOC_814|>": 101111, "<|LOC_815|>": 101112, "<|LOC_816|>": 101113, "<|LOC_817|>": 101114, "<|LOC_818|>": 101115, "<|LOC_819|>": 101116, "<|LOC_820|>": 101117, "<|LOC_821|>": 101118, "<|LOC_822|>": 101119, "<|LOC_823|>": 101120, "<|LOC_824|>": 101121, "<|LOC_825|>": 101122, "<|LOC_826|>": 101123, "<|LOC_827|>": 101124, "<|LOC_828|>": 101125, "<|LOC_829|>": 101126, "<|LOC_830|>": 101127, "<|LOC_831|>": 101128, "<|LOC_832|>": 101129, "<|LOC_833|>": 101130, "<|LOC_834|>": 101131, "<|LOC_835|>": 101132, "<|LOC_836|>": 101133, "<|LOC_837|>": 101134, "<|LOC_838|>": 101135, "<|LOC_839|>": 101136, "<|LOC_840|>": 101137, "<|LOC_841|>": 101138, "<|LOC_842|>": 101139, "<|LOC_843|>": 101140, "<|LOC_844|>": 101141, "<|LOC_845|>": 101142, "<|LOC_846|>": 101143, "<|LOC_847|>": 101144, "<|LOC_848|>": 101145, "<|LOC_849|>": 101146, "<|LOC_850|>": 101147, "<|LOC_851|>": 101148, "<|LOC_852|>": 101149, "<|LOC_853|>": 101150, "<|LOC_854|>": 101151, "<|LOC_855|>": 101152, "<|LOC_856|>": 101153, "<|LOC_857|>": 101154, "<|LOC_858|>": 101155, "<|LOC_859|>": 101156, "<|LOC_860|>": 101157, "<|LOC_861|>": 101158, "<|LOC_862|>": 101159, "<|LOC_863|>": 101160, "<|LOC_864|>": 101161, "<|LOC_865|>": 101162, "<|LOC_866|>": 101163, "<|LOC_867|>": 101164, "<|LOC_868|>": 101165, "<|LOC_869|>": 101166, "<|LOC_870|>": 101167, "<|LOC_871|>": 101168, "<|LOC_872|>": 101169, "<|LOC_873|>": 101170, "<|LOC_874|>": 101171, "<|LOC_875|>": 101172, "<|LOC_876|>": 101173, "<|LOC_877|>": 101174, "<|LOC_878|>": 101175, "<|LOC_879|>": 101176, "<|LOC_880|>": 101177, "<|LOC_881|>": 101178, "<|LOC_882|>": 101179, "<|LOC_883|>": 101180, "<|LOC_884|>": 101181, "<|LOC_885|>": 101182, "<|LOC_886|>": 101183, "<|LOC_887|>": 101184, "<|LOC_888|>": 101185, "<|LOC_889|>": 101186, "<|LOC_890|>": 101187, "<|LOC_891|>": 101188, "<|LOC_892|>": 101189, "<|LOC_893|>": 101190, "<|LOC_894|>": 101191, "<|LOC_895|>": 101192, "<|LOC_896|>": 101193, "<|LOC_897|>": 101194, "<|LOC_898|>": 101195, "<|LOC_899|>": 101196, "<|LOC_900|>": 101197, "<|LOC_901|>": 101198, "<|LOC_902|>": 101199, "<|LOC_903|>": 101200, "<|LOC_904|>": 101201, "<|LOC_905|>": 101202, "<|LOC_906|>": 101203, "<|LOC_907|>": 101204, "<|LOC_908|>": 101205, "<|LOC_909|>": 101206, "<|LOC_910|>": 101207, "<|LOC_911|>": 101208, "<|LOC_912|>": 101209, "<|LOC_913|>": 101210, "<|LOC_914|>": 101211, "<|LOC_915|>": 101212, "<|LOC_916|>": 101213, "<|LOC_917|>": 101214, "<|LOC_918|>": 101215, "<|LOC_919|>": 101216, "<|LOC_920|>": 101217, "<|LOC_921|>": 101218, "<|LOC_922|>": 101219, "<|LOC_923|>": 101220, "<|LOC_924|>": 101221, "<|LOC_925|>": 101222, "<|LOC_926|>": 101223, "<|LOC_927|>": 101224, "<|LOC_928|>": 101225, "<|LOC_929|>": 101226, "<|LOC_930|>": 101227, "<|LOC_931|>": 101228, "<|LOC_932|>": 101229, "<|LOC_933|>": 101230, "<|LOC_934|>": 101231, "<|LOC_935|>": 101232, "<|LOC_936|>": 101233, "<|LOC_937|>": 101234, "<|LOC_938|>": 101235, "<|LOC_939|>": 101236, "<|LOC_940|>": 101237, "<|LOC_941|>": 101238, "<|LOC_942|>": 101239, "<|LOC_943|>": 101240, "<|LOC_944|>": 101241, "<|LOC_945|>": 101242, "<|LOC_946|>": 101243, "<|LOC_947|>": 101244, "<|LOC_948|>": 101245, "<|LOC_949|>": 101246, "<|LOC_950|>": 101247, "<|LOC_951|>": 101248, "<|LOC_952|>": 101249, "<|LOC_953|>": 101250, "<|LOC_954|>": 101251, "<|LOC_955|>": 101252, "<|LOC_956|>": 101253, "<|LOC_957|>": 101254, "<|LOC_958|>": 101255, "<|LOC_959|>": 101256, "<|LOC_960|>": 101257, "<|LOC_961|>": 101258, "<|LOC_962|>": 101259, "<|LOC_963|>": 101260, "<|LOC_964|>": 101261, "<|LOC_965|>": 101262, "<|LOC_966|>": 101263, "<|LOC_967|>": 101264, "<|LOC_968|>": 101265, "<|LOC_969|>": 101266, "<|LOC_970|>": 101267, "<|LOC_971|>": 101268, "<|LOC_972|>": 101269, "<|LOC_973|>": 101270, "<|LOC_974|>": 101271, "<|LOC_975|>": 101272, "<|LOC_976|>": 101273, "<|LOC_977|>": 101274, "<|LOC_978|>": 101275, "<|LOC_979|>": 101276, "<|LOC_980|>": 101277, "<|LOC_981|>": 101278, "<|LOC_982|>": 101279, "<|LOC_983|>": 101280, "<|LOC_984|>": 101281, "<|LOC_985|>": 101282, "<|LOC_986|>": 101283, "<|LOC_987|>": 101284, "<|LOC_988|>": 101285, "<|LOC_989|>": 101286, "<|LOC_990|>": 101287, "<|LOC_991|>": 101288, "<|LOC_992|>": 101289, "<|LOC_993|>": 101290, "<|LOC_994|>": 101291, "<|LOC_995|>": 101292, "<|LOC_996|>": 101293, "<|LOC_997|>": 101294, "<|LOC_998|>": 101295, "<|LOC_999|>": 101296, "<|LOC_1000|>": 101297, "<|LOC_BEGIN|>": 101298, "<|LOC_END|>": 101299, "<|LOC_SEP|>": 101300, "<|CROP_COL_SEP|>": 101301, "<|CROP_ROW_SEP|>": 101302, "<|IMAGE_SEP|>": 101303, "<|image_pad|>": 101304, "<|IMAGE_START|>": 101305, "<|IMAGE_END|>": 101306, "<|video_pad|>": 101307, "<ecel>": 101308, "<fcel>": 101309, "<xcel>": 101310, "<lcel>": 101311, "<ucel>": 101312, "<nl>": 101313}
|
checkpoint-420/config.json
ADDED
|
@@ -0,0 +1,223 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"architectures": [
|
| 3 |
+
"PaddleOCRVLForConditionalGeneration"
|
| 4 |
+
],
|
| 5 |
+
"attention_probs_dropout_prob": 0.0,
|
| 6 |
+
"auto_map": {
|
| 7 |
+
"AutoConfig": "configuration_paddleocr_vl.PaddleOCRVLConfig",
|
| 8 |
+
"AutoModel": "modeling_paddleocr_vl.PaddleOCRVLForConditionalGeneration",
|
| 9 |
+
"AutoModelForCausalLM": "modeling_paddleocr_vl.PaddleOCRVLForConditionalGeneration"
|
| 10 |
+
},
|
| 11 |
+
"compression_ratio": 1.0,
|
| 12 |
+
"disable_pipeline_warmup": false,
|
| 13 |
+
"enable_mtp_magic_send": false,
|
| 14 |
+
"fp16_opt_level": "O2",
|
| 15 |
+
"freq_allocation": 20,
|
| 16 |
+
"fuse_ln": false,
|
| 17 |
+
"fuse_rms_norm": true,
|
| 18 |
+
"head_dim": 128,
|
| 19 |
+
"hidden_act": "silu",
|
| 20 |
+
"hidden_dropout_prob": 0.0,
|
| 21 |
+
"hidden_size": 1024,
|
| 22 |
+
"ignored_index": -100,
|
| 23 |
+
"im_patch_id": 100295,
|
| 24 |
+
"image_token_id": 100295,
|
| 25 |
+
"intermediate_size": 3072,
|
| 26 |
+
"max_position_embeddings": 131072,
|
| 27 |
+
"max_text_id": 100295,
|
| 28 |
+
"model_type": "paddleocr_vl",
|
| 29 |
+
"moe_dropout_prob": 0.0,
|
| 30 |
+
"moe_multimodal_dispatch_use_allgather": "v2-alltoall-unpad",
|
| 31 |
+
"num_attention_heads": 16,
|
| 32 |
+
"num_hidden_layers": 18,
|
| 33 |
+
"num_key_value_heads": 2,
|
| 34 |
+
"paddleformers_version": "0.4.0",
|
| 35 |
+
"pixel_hidden_size": 1152,
|
| 36 |
+
"rms_norm_eps": 1e-05,
|
| 37 |
+
"rope_is_neox_style": true,
|
| 38 |
+
"rope_scaling": {
|
| 39 |
+
"mrope_section": [
|
| 40 |
+
16,
|
| 41 |
+
24,
|
| 42 |
+
24
|
| 43 |
+
],
|
| 44 |
+
"rope_type": "default",
|
| 45 |
+
"type": "default"
|
| 46 |
+
},
|
| 47 |
+
"rope_theta": 500000,
|
| 48 |
+
"scale_qk_coeff": 1.0,
|
| 49 |
+
"seqlen": 16384,
|
| 50 |
+
"sliding_window": null,
|
| 51 |
+
"tie_word_embeddings": false,
|
| 52 |
+
"token_balance_loss": false,
|
| 53 |
+
"token_balance_seqlen": 16384,
|
| 54 |
+
"torch_dtype": "bfloat16",
|
| 55 |
+
"use_3d_rope": true,
|
| 56 |
+
"use_bias": false,
|
| 57 |
+
"use_flash_attn_with_mask": true,
|
| 58 |
+
"use_fp8": false,
|
| 59 |
+
"use_mem_eff_attn": true,
|
| 60 |
+
"use_recompute_moe": false,
|
| 61 |
+
"use_rmsnorm": true,
|
| 62 |
+
"video_token_id": 101307,
|
| 63 |
+
"vision_config": {
|
| 64 |
+
"_attn_implementation": "eager",
|
| 65 |
+
"_name_or_path": "",
|
| 66 |
+
"_save_to_hf": false,
|
| 67 |
+
"add_cross_attention": false,
|
| 68 |
+
"add_tail_layers": 0,
|
| 69 |
+
"architectures": [
|
| 70 |
+
"PaddleOCRVisionModel"
|
| 71 |
+
],
|
| 72 |
+
"attention_dropout": 0.0,
|
| 73 |
+
"auto_map": {
|
| 74 |
+
"AutoConfig": "configuration_paddleocr_vl.PaddleOCRVLConfig",
|
| 75 |
+
"AutoModel": "modeling_paddleocr_vl.PaddleOCRVisionModel"
|
| 76 |
+
},
|
| 77 |
+
"bad_words_ids": null,
|
| 78 |
+
"begin_suppress_tokens": null,
|
| 79 |
+
"bos_token_id": null,
|
| 80 |
+
"chunk_size_feed_forward": 0,
|
| 81 |
+
"classifier_dropout": null,
|
| 82 |
+
"context_parallel_degree": 1,
|
| 83 |
+
"cross_attention_hidden_size": null,
|
| 84 |
+
"decoder_start_token_id": null,
|
| 85 |
+
"diversity_penalty": 0.0,
|
| 86 |
+
"do_sample": false,
|
| 87 |
+
"dpo_config": null,
|
| 88 |
+
"dtype": "bfloat16",
|
| 89 |
+
"early_stopping": false,
|
| 90 |
+
"encoder_no_repeat_ngram_size": 0,
|
| 91 |
+
"eos_token_id": null,
|
| 92 |
+
"ep_communication_type": "deepep",
|
| 93 |
+
"exponential_decay_length_penalty": null,
|
| 94 |
+
"finetuning_task": null,
|
| 95 |
+
"forced_bos_token_id": null,
|
| 96 |
+
"forced_eos_token_id": null,
|
| 97 |
+
"fuse_attention_ffn": false,
|
| 98 |
+
"fuse_attention_qkv": false,
|
| 99 |
+
"fuse_linear": false,
|
| 100 |
+
"fuse_rope": false,
|
| 101 |
+
"fuse_sequence_parallel_allreduce": false,
|
| 102 |
+
"fuse_swiglu": false,
|
| 103 |
+
"hidden_act": "gelu_new",
|
| 104 |
+
"hidden_size": 1152,
|
| 105 |
+
"id2label": {
|
| 106 |
+
"0": "LABEL_0",
|
| 107 |
+
"1": "LABEL_1"
|
| 108 |
+
},
|
| 109 |
+
"image_size": 384,
|
| 110 |
+
"intermediate_size": 4304,
|
| 111 |
+
"is_decoder": false,
|
| 112 |
+
"is_encoder_decoder": false,
|
| 113 |
+
"kto_config": null,
|
| 114 |
+
"label2id": {
|
| 115 |
+
"LABEL_0": 0,
|
| 116 |
+
"LABEL_1": 1
|
| 117 |
+
},
|
| 118 |
+
"layer_norm_eps": 1e-06,
|
| 119 |
+
"length_penalty": 1.0,
|
| 120 |
+
"loss_subbatch_sequence_length": -1,
|
| 121 |
+
"max_length": 20,
|
| 122 |
+
"min_length": 0,
|
| 123 |
+
"model_type": "paddleocr_vl",
|
| 124 |
+
"moe_subbatch_token_num": 0,
|
| 125 |
+
"no_recompute_layers": null,
|
| 126 |
+
"no_repeat_ngram_size": 0,
|
| 127 |
+
"num_attention_heads": 16,
|
| 128 |
+
"num_beam_groups": 1,
|
| 129 |
+
"num_beams": 1,
|
| 130 |
+
"num_channels": 3,
|
| 131 |
+
"num_choices": null,
|
| 132 |
+
"num_hidden_layers": 27,
|
| 133 |
+
"num_nextn_predict_layers": 0,
|
| 134 |
+
"num_return_sequences": 1,
|
| 135 |
+
"offload_recompute_inputs": false,
|
| 136 |
+
"output_attentions": false,
|
| 137 |
+
"output_hidden_states": false,
|
| 138 |
+
"output_scores": false,
|
| 139 |
+
"pad_token_id": 0,
|
| 140 |
+
"patch_size": 14,
|
| 141 |
+
"pipeline_parallel_degree": 1,
|
| 142 |
+
"pp_recompute_interval": 1,
|
| 143 |
+
"prefix": null,
|
| 144 |
+
"problem_type": null,
|
| 145 |
+
"pruned_heads": {},
|
| 146 |
+
"quantization_config": {
|
| 147 |
+
"act_quant_method": "abs_max",
|
| 148 |
+
"activation_scheme": null,
|
| 149 |
+
"actscale_moving_rate": 0.01,
|
| 150 |
+
"apply_hadamard": false,
|
| 151 |
+
"apply_online_actscale_step": 200,
|
| 152 |
+
"dense_quant_type": "",
|
| 153 |
+
"dtype": null,
|
| 154 |
+
"fmt": null,
|
| 155 |
+
"fp8_format_type": "hybrid",
|
| 156 |
+
"group_size": -1,
|
| 157 |
+
"hadamard_block_size": 32,
|
| 158 |
+
"ignore_modules": null,
|
| 159 |
+
"llm_int8_threshold": 6.0,
|
| 160 |
+
"moe_quant_type": "",
|
| 161 |
+
"qlora_weight_blocksize": 64,
|
| 162 |
+
"qlora_weight_double_quant": false,
|
| 163 |
+
"qlora_weight_double_quant_block_size": 256,
|
| 164 |
+
"quant_input_grad": false,
|
| 165 |
+
"quant_method": null,
|
| 166 |
+
"quant_round_type": 0,
|
| 167 |
+
"quant_type": null,
|
| 168 |
+
"quant_weight_grad": false,
|
| 169 |
+
"quantization": "",
|
| 170 |
+
"scale_epsilon": 1e-08,
|
| 171 |
+
"shift": false,
|
| 172 |
+
"shift_smooth_all_linears": false,
|
| 173 |
+
"smooth": false,
|
| 174 |
+
"weight_block_size": null,
|
| 175 |
+
"weight_quant_method": "abs_max_channel_wise",
|
| 176 |
+
"weight_quantize_algo": null
|
| 177 |
+
},
|
| 178 |
+
"recompute": true,
|
| 179 |
+
"recompute_granularity": "full",
|
| 180 |
+
"recompute_use_reentrant": false,
|
| 181 |
+
"refined_recompute": "",
|
| 182 |
+
"remove_invalid_values": false,
|
| 183 |
+
"repetition_penalty": 1.0,
|
| 184 |
+
"return_dict": false,
|
| 185 |
+
"return_dict_in_generate": false,
|
| 186 |
+
"sep_parallel_degree": 1,
|
| 187 |
+
"sep_token_id": null,
|
| 188 |
+
"sequence_parallel": false,
|
| 189 |
+
"spatial_merge_size": 2,
|
| 190 |
+
"suppress_tokens": null,
|
| 191 |
+
"task_specific_params": null,
|
| 192 |
+
"temperature": 1.0,
|
| 193 |
+
"temporal_patch_size": 2,
|
| 194 |
+
"tensor_parallel_degree": 1,
|
| 195 |
+
"tensor_parallel_output": true,
|
| 196 |
+
"tensor_parallel_rank": 0,
|
| 197 |
+
"tie_encoder_decoder": false,
|
| 198 |
+
"tie_word_embeddings": true,
|
| 199 |
+
"tokenizer_class": null,
|
| 200 |
+
"tokens_per_second": 2,
|
| 201 |
+
"top_k": 50,
|
| 202 |
+
"top_p": 1.0,
|
| 203 |
+
"typical_p": 1.0,
|
| 204 |
+
"use_cache": false,
|
| 205 |
+
"use_filtered_label_loss": false,
|
| 206 |
+
"use_flash_attention": true,
|
| 207 |
+
"use_fused_dropout_add": false,
|
| 208 |
+
"use_fused_head_and_loss_fn": false,
|
| 209 |
+
"use_fused_linear": false,
|
| 210 |
+
"use_fused_linear_cross_entropy": false,
|
| 211 |
+
"use_fused_rms_norm": false,
|
| 212 |
+
"use_fused_rope": false,
|
| 213 |
+
"use_sparse_flash_attn": true,
|
| 214 |
+
"use_sparse_head_and_loss_fn": false,
|
| 215 |
+
"use_unified_moe": false,
|
| 216 |
+
"using_fake_gate": false,
|
| 217 |
+
"virtual_pp_degree": 1
|
| 218 |
+
},
|
| 219 |
+
"vision_end_token_id": 101306,
|
| 220 |
+
"vision_start_token_id": 101305,
|
| 221 |
+
"vocab_size": 103424,
|
| 222 |
+
"weight_share_add_bias": true
|
| 223 |
+
}
|
checkpoint-420/generation_config.json
ADDED
|
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_from_model_config": true,
|
| 3 |
+
"eos_token_id": 2,
|
| 4 |
+
"pad_token_id": 0,
|
| 5 |
+
"use_cache": false
|
| 6 |
+
}
|
checkpoint-420/master_weights-00001-of-00001.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8c939d1ffb8d46bfb1b37af1fa66b6781f66ce0dbeda32eafb408fb1a114ae62
|
| 3 |
+
size 3622483352
|
checkpoint-420/master_weights.safetensors.index.json
ADDED
|
@@ -0,0 +1,615 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"metadata": {
|
| 3 |
+
"total_size": 3622406592
|
| 4 |
+
},
|
| 5 |
+
"weight_map": {
|
| 6 |
+
"mlp_AR.pre_norm.weight": "master_weights-00001-of-00001.safetensors",
|
| 7 |
+
"mlp_AR.pre_norm.bias": "master_weights-00001-of-00001.safetensors",
|
| 8 |
+
"mlp_AR.linear_1.weight": "master_weights-00001-of-00001.safetensors",
|
| 9 |
+
"mlp_AR.linear_1.bias": "master_weights-00001-of-00001.safetensors",
|
| 10 |
+
"mlp_AR.linear_2.weight": "master_weights-00001-of-00001.safetensors",
|
| 11 |
+
"mlp_AR.linear_2.bias": "master_weights-00001-of-00001.safetensors",
|
| 12 |
+
"visual.vision_model.embeddings.patch_embedding.weight": "master_weights-00001-of-00001.safetensors",
|
| 13 |
+
"visual.vision_model.embeddings.patch_embedding.bias": "master_weights-00001-of-00001.safetensors",
|
| 14 |
+
"visual.vision_model.embeddings.position_embedding.weight": "master_weights-00001-of-00001.safetensors",
|
| 15 |
+
"visual.vision_model.encoder.layers.0.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 16 |
+
"visual.vision_model.encoder.layers.0.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 17 |
+
"visual.vision_model.encoder.layers.0.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 18 |
+
"visual.vision_model.encoder.layers.0.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 19 |
+
"visual.vision_model.encoder.layers.0.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 20 |
+
"visual.vision_model.encoder.layers.0.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 21 |
+
"visual.vision_model.encoder.layers.0.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 22 |
+
"visual.vision_model.encoder.layers.0.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 23 |
+
"visual.vision_model.encoder.layers.0.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 24 |
+
"visual.vision_model.encoder.layers.0.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 25 |
+
"visual.vision_model.encoder.layers.0.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 26 |
+
"visual.vision_model.encoder.layers.0.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 27 |
+
"visual.vision_model.encoder.layers.0.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 28 |
+
"visual.vision_model.encoder.layers.0.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 29 |
+
"visual.vision_model.encoder.layers.0.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 30 |
+
"visual.vision_model.encoder.layers.0.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 31 |
+
"visual.vision_model.encoder.layers.1.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 32 |
+
"visual.vision_model.encoder.layers.1.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 33 |
+
"visual.vision_model.encoder.layers.1.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 34 |
+
"visual.vision_model.encoder.layers.1.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 35 |
+
"visual.vision_model.encoder.layers.1.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 36 |
+
"visual.vision_model.encoder.layers.1.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 37 |
+
"visual.vision_model.encoder.layers.1.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 38 |
+
"visual.vision_model.encoder.layers.1.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 39 |
+
"visual.vision_model.encoder.layers.1.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 40 |
+
"visual.vision_model.encoder.layers.1.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 41 |
+
"visual.vision_model.encoder.layers.1.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 42 |
+
"visual.vision_model.encoder.layers.1.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 43 |
+
"visual.vision_model.encoder.layers.1.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 44 |
+
"visual.vision_model.encoder.layers.1.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 45 |
+
"visual.vision_model.encoder.layers.1.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 46 |
+
"visual.vision_model.encoder.layers.1.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 47 |
+
"visual.vision_model.encoder.layers.2.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 48 |
+
"visual.vision_model.encoder.layers.2.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 49 |
+
"visual.vision_model.encoder.layers.2.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 50 |
+
"visual.vision_model.encoder.layers.2.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 51 |
+
"visual.vision_model.encoder.layers.2.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 52 |
+
"visual.vision_model.encoder.layers.2.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 53 |
+
"visual.vision_model.encoder.layers.2.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 54 |
+
"visual.vision_model.encoder.layers.2.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 55 |
+
"visual.vision_model.encoder.layers.2.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 56 |
+
"visual.vision_model.encoder.layers.2.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 57 |
+
"visual.vision_model.encoder.layers.2.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 58 |
+
"visual.vision_model.encoder.layers.2.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 59 |
+
"visual.vision_model.encoder.layers.2.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 60 |
+
"visual.vision_model.encoder.layers.2.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 61 |
+
"visual.vision_model.encoder.layers.2.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 62 |
+
"visual.vision_model.encoder.layers.2.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 63 |
+
"visual.vision_model.encoder.layers.3.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 64 |
+
"visual.vision_model.encoder.layers.3.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 65 |
+
"visual.vision_model.encoder.layers.3.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 66 |
+
"visual.vision_model.encoder.layers.3.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 67 |
+
"visual.vision_model.encoder.layers.3.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 68 |
+
"visual.vision_model.encoder.layers.3.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 69 |
+
"visual.vision_model.encoder.layers.3.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 70 |
+
"visual.vision_model.encoder.layers.3.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 71 |
+
"visual.vision_model.encoder.layers.3.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 72 |
+
"visual.vision_model.encoder.layers.3.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 73 |
+
"visual.vision_model.encoder.layers.3.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 74 |
+
"visual.vision_model.encoder.layers.3.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 75 |
+
"visual.vision_model.encoder.layers.3.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 76 |
+
"visual.vision_model.encoder.layers.3.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 77 |
+
"visual.vision_model.encoder.layers.3.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 78 |
+
"visual.vision_model.encoder.layers.3.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 79 |
+
"visual.vision_model.encoder.layers.4.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 80 |
+
"visual.vision_model.encoder.layers.4.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 81 |
+
"visual.vision_model.encoder.layers.4.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 82 |
+
"visual.vision_model.encoder.layers.4.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 83 |
+
"visual.vision_model.encoder.layers.4.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 84 |
+
"visual.vision_model.encoder.layers.4.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 85 |
+
"visual.vision_model.encoder.layers.4.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 86 |
+
"visual.vision_model.encoder.layers.4.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 87 |
+
"visual.vision_model.encoder.layers.4.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 88 |
+
"visual.vision_model.encoder.layers.4.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 89 |
+
"visual.vision_model.encoder.layers.4.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 90 |
+
"visual.vision_model.encoder.layers.4.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 91 |
+
"visual.vision_model.encoder.layers.4.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 92 |
+
"visual.vision_model.encoder.layers.4.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 93 |
+
"visual.vision_model.encoder.layers.4.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 94 |
+
"visual.vision_model.encoder.layers.4.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 95 |
+
"visual.vision_model.encoder.layers.5.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 96 |
+
"visual.vision_model.encoder.layers.5.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 97 |
+
"visual.vision_model.encoder.layers.5.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 98 |
+
"visual.vision_model.encoder.layers.5.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 99 |
+
"visual.vision_model.encoder.layers.5.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 100 |
+
"visual.vision_model.encoder.layers.5.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 101 |
+
"visual.vision_model.encoder.layers.5.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 102 |
+
"visual.vision_model.encoder.layers.5.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 103 |
+
"visual.vision_model.encoder.layers.5.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 104 |
+
"visual.vision_model.encoder.layers.5.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 105 |
+
"visual.vision_model.encoder.layers.5.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 106 |
+
"visual.vision_model.encoder.layers.5.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 107 |
+
"visual.vision_model.encoder.layers.5.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 108 |
+
"visual.vision_model.encoder.layers.5.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 109 |
+
"visual.vision_model.encoder.layers.5.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 110 |
+
"visual.vision_model.encoder.layers.5.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 111 |
+
"visual.vision_model.encoder.layers.6.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 112 |
+
"visual.vision_model.encoder.layers.6.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 113 |
+
"visual.vision_model.encoder.layers.6.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 114 |
+
"visual.vision_model.encoder.layers.6.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 115 |
+
"visual.vision_model.encoder.layers.6.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 116 |
+
"visual.vision_model.encoder.layers.6.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 117 |
+
"visual.vision_model.encoder.layers.6.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 118 |
+
"visual.vision_model.encoder.layers.6.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 119 |
+
"visual.vision_model.encoder.layers.6.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 120 |
+
"visual.vision_model.encoder.layers.6.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 121 |
+
"visual.vision_model.encoder.layers.6.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 122 |
+
"visual.vision_model.encoder.layers.6.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 123 |
+
"visual.vision_model.encoder.layers.6.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 124 |
+
"visual.vision_model.encoder.layers.6.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 125 |
+
"visual.vision_model.encoder.layers.6.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 126 |
+
"visual.vision_model.encoder.layers.6.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 127 |
+
"visual.vision_model.encoder.layers.7.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 128 |
+
"visual.vision_model.encoder.layers.7.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 129 |
+
"visual.vision_model.encoder.layers.7.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 130 |
+
"visual.vision_model.encoder.layers.7.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 131 |
+
"visual.vision_model.encoder.layers.7.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 132 |
+
"visual.vision_model.encoder.layers.7.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 133 |
+
"visual.vision_model.encoder.layers.7.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 134 |
+
"visual.vision_model.encoder.layers.7.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 135 |
+
"visual.vision_model.encoder.layers.7.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 136 |
+
"visual.vision_model.encoder.layers.7.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 137 |
+
"visual.vision_model.encoder.layers.7.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 138 |
+
"visual.vision_model.encoder.layers.7.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 139 |
+
"visual.vision_model.encoder.layers.7.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 140 |
+
"visual.vision_model.encoder.layers.7.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 141 |
+
"visual.vision_model.encoder.layers.7.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 142 |
+
"visual.vision_model.encoder.layers.7.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 143 |
+
"visual.vision_model.encoder.layers.8.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 144 |
+
"visual.vision_model.encoder.layers.8.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 145 |
+
"visual.vision_model.encoder.layers.8.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 146 |
+
"visual.vision_model.encoder.layers.8.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 147 |
+
"visual.vision_model.encoder.layers.8.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 148 |
+
"visual.vision_model.encoder.layers.8.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 149 |
+
"visual.vision_model.encoder.layers.8.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 150 |
+
"visual.vision_model.encoder.layers.8.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 151 |
+
"visual.vision_model.encoder.layers.8.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 152 |
+
"visual.vision_model.encoder.layers.8.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 153 |
+
"visual.vision_model.encoder.layers.8.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 154 |
+
"visual.vision_model.encoder.layers.8.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 155 |
+
"visual.vision_model.encoder.layers.8.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 156 |
+
"visual.vision_model.encoder.layers.8.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 157 |
+
"visual.vision_model.encoder.layers.8.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 158 |
+
"visual.vision_model.encoder.layers.8.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 159 |
+
"visual.vision_model.encoder.layers.9.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 160 |
+
"visual.vision_model.encoder.layers.9.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 161 |
+
"visual.vision_model.encoder.layers.9.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 162 |
+
"visual.vision_model.encoder.layers.9.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 163 |
+
"visual.vision_model.encoder.layers.9.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 164 |
+
"visual.vision_model.encoder.layers.9.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 165 |
+
"visual.vision_model.encoder.layers.9.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 166 |
+
"visual.vision_model.encoder.layers.9.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 167 |
+
"visual.vision_model.encoder.layers.9.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 168 |
+
"visual.vision_model.encoder.layers.9.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 169 |
+
"visual.vision_model.encoder.layers.9.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 170 |
+
"visual.vision_model.encoder.layers.9.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 171 |
+
"visual.vision_model.encoder.layers.9.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 172 |
+
"visual.vision_model.encoder.layers.9.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 173 |
+
"visual.vision_model.encoder.layers.9.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 174 |
+
"visual.vision_model.encoder.layers.9.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 175 |
+
"visual.vision_model.encoder.layers.10.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 176 |
+
"visual.vision_model.encoder.layers.10.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 177 |
+
"visual.vision_model.encoder.layers.10.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 178 |
+
"visual.vision_model.encoder.layers.10.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 179 |
+
"visual.vision_model.encoder.layers.10.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 180 |
+
"visual.vision_model.encoder.layers.10.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 181 |
+
"visual.vision_model.encoder.layers.10.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 182 |
+
"visual.vision_model.encoder.layers.10.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 183 |
+
"visual.vision_model.encoder.layers.10.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 184 |
+
"visual.vision_model.encoder.layers.10.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 185 |
+
"visual.vision_model.encoder.layers.10.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 186 |
+
"visual.vision_model.encoder.layers.10.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 187 |
+
"visual.vision_model.encoder.layers.10.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 188 |
+
"visual.vision_model.encoder.layers.10.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 189 |
+
"visual.vision_model.encoder.layers.10.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 190 |
+
"visual.vision_model.encoder.layers.10.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 191 |
+
"visual.vision_model.encoder.layers.11.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 192 |
+
"visual.vision_model.encoder.layers.11.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 193 |
+
"visual.vision_model.encoder.layers.11.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 194 |
+
"visual.vision_model.encoder.layers.11.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 195 |
+
"visual.vision_model.encoder.layers.11.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 196 |
+
"visual.vision_model.encoder.layers.11.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 197 |
+
"visual.vision_model.encoder.layers.11.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 198 |
+
"visual.vision_model.encoder.layers.11.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 199 |
+
"visual.vision_model.encoder.layers.11.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 200 |
+
"visual.vision_model.encoder.layers.11.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 201 |
+
"visual.vision_model.encoder.layers.11.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 202 |
+
"visual.vision_model.encoder.layers.11.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 203 |
+
"visual.vision_model.encoder.layers.11.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 204 |
+
"visual.vision_model.encoder.layers.11.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 205 |
+
"visual.vision_model.encoder.layers.11.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 206 |
+
"visual.vision_model.encoder.layers.11.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 207 |
+
"visual.vision_model.encoder.layers.12.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 208 |
+
"visual.vision_model.encoder.layers.12.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 209 |
+
"visual.vision_model.encoder.layers.12.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 210 |
+
"visual.vision_model.encoder.layers.12.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 211 |
+
"visual.vision_model.encoder.layers.12.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 212 |
+
"visual.vision_model.encoder.layers.12.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 213 |
+
"visual.vision_model.encoder.layers.12.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 214 |
+
"visual.vision_model.encoder.layers.12.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 215 |
+
"visual.vision_model.encoder.layers.12.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 216 |
+
"visual.vision_model.encoder.layers.12.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 217 |
+
"visual.vision_model.encoder.layers.12.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 218 |
+
"visual.vision_model.encoder.layers.12.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 219 |
+
"visual.vision_model.encoder.layers.12.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 220 |
+
"visual.vision_model.encoder.layers.12.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 221 |
+
"visual.vision_model.encoder.layers.12.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 222 |
+
"visual.vision_model.encoder.layers.12.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 223 |
+
"visual.vision_model.encoder.layers.13.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 224 |
+
"visual.vision_model.encoder.layers.13.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 225 |
+
"visual.vision_model.encoder.layers.13.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 226 |
+
"visual.vision_model.encoder.layers.13.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 227 |
+
"visual.vision_model.encoder.layers.13.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 228 |
+
"visual.vision_model.encoder.layers.13.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 229 |
+
"visual.vision_model.encoder.layers.13.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 230 |
+
"visual.vision_model.encoder.layers.13.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 231 |
+
"visual.vision_model.encoder.layers.13.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 232 |
+
"visual.vision_model.encoder.layers.13.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 233 |
+
"visual.vision_model.encoder.layers.13.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 234 |
+
"visual.vision_model.encoder.layers.13.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 235 |
+
"visual.vision_model.encoder.layers.13.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 236 |
+
"visual.vision_model.encoder.layers.13.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 237 |
+
"visual.vision_model.encoder.layers.13.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 238 |
+
"visual.vision_model.encoder.layers.13.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 239 |
+
"visual.vision_model.encoder.layers.14.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 240 |
+
"visual.vision_model.encoder.layers.14.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 241 |
+
"visual.vision_model.encoder.layers.14.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 242 |
+
"visual.vision_model.encoder.layers.14.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 243 |
+
"visual.vision_model.encoder.layers.14.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 244 |
+
"visual.vision_model.encoder.layers.14.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 245 |
+
"visual.vision_model.encoder.layers.14.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 246 |
+
"visual.vision_model.encoder.layers.14.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 247 |
+
"visual.vision_model.encoder.layers.14.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 248 |
+
"visual.vision_model.encoder.layers.14.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 249 |
+
"visual.vision_model.encoder.layers.14.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 250 |
+
"visual.vision_model.encoder.layers.14.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 251 |
+
"visual.vision_model.encoder.layers.14.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 252 |
+
"visual.vision_model.encoder.layers.14.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 253 |
+
"visual.vision_model.encoder.layers.14.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 254 |
+
"visual.vision_model.encoder.layers.14.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 255 |
+
"visual.vision_model.encoder.layers.15.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 256 |
+
"visual.vision_model.encoder.layers.15.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 257 |
+
"visual.vision_model.encoder.layers.15.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 258 |
+
"visual.vision_model.encoder.layers.15.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 259 |
+
"visual.vision_model.encoder.layers.15.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 260 |
+
"visual.vision_model.encoder.layers.15.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 261 |
+
"visual.vision_model.encoder.layers.15.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 262 |
+
"visual.vision_model.encoder.layers.15.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 263 |
+
"visual.vision_model.encoder.layers.15.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 264 |
+
"visual.vision_model.encoder.layers.15.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 265 |
+
"visual.vision_model.encoder.layers.15.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 266 |
+
"visual.vision_model.encoder.layers.15.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 267 |
+
"visual.vision_model.encoder.layers.15.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 268 |
+
"visual.vision_model.encoder.layers.15.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 269 |
+
"visual.vision_model.encoder.layers.15.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 270 |
+
"visual.vision_model.encoder.layers.15.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 271 |
+
"visual.vision_model.encoder.layers.16.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 272 |
+
"visual.vision_model.encoder.layers.16.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 273 |
+
"visual.vision_model.encoder.layers.16.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 274 |
+
"visual.vision_model.encoder.layers.16.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 275 |
+
"visual.vision_model.encoder.layers.16.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 276 |
+
"visual.vision_model.encoder.layers.16.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 277 |
+
"visual.vision_model.encoder.layers.16.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 278 |
+
"visual.vision_model.encoder.layers.16.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 279 |
+
"visual.vision_model.encoder.layers.16.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 280 |
+
"visual.vision_model.encoder.layers.16.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 281 |
+
"visual.vision_model.encoder.layers.16.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 282 |
+
"visual.vision_model.encoder.layers.16.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 283 |
+
"visual.vision_model.encoder.layers.16.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 284 |
+
"visual.vision_model.encoder.layers.16.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 285 |
+
"visual.vision_model.encoder.layers.16.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 286 |
+
"visual.vision_model.encoder.layers.16.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 287 |
+
"visual.vision_model.encoder.layers.17.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 288 |
+
"visual.vision_model.encoder.layers.17.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 289 |
+
"visual.vision_model.encoder.layers.17.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 290 |
+
"visual.vision_model.encoder.layers.17.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 291 |
+
"visual.vision_model.encoder.layers.17.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 292 |
+
"visual.vision_model.encoder.layers.17.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 293 |
+
"visual.vision_model.encoder.layers.17.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 294 |
+
"visual.vision_model.encoder.layers.17.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 295 |
+
"visual.vision_model.encoder.layers.17.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 296 |
+
"visual.vision_model.encoder.layers.17.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 297 |
+
"visual.vision_model.encoder.layers.17.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 298 |
+
"visual.vision_model.encoder.layers.17.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 299 |
+
"visual.vision_model.encoder.layers.17.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 300 |
+
"visual.vision_model.encoder.layers.17.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 301 |
+
"visual.vision_model.encoder.layers.17.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 302 |
+
"visual.vision_model.encoder.layers.17.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 303 |
+
"visual.vision_model.encoder.layers.18.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 304 |
+
"visual.vision_model.encoder.layers.18.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 305 |
+
"visual.vision_model.encoder.layers.18.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 306 |
+
"visual.vision_model.encoder.layers.18.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 307 |
+
"visual.vision_model.encoder.layers.18.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 308 |
+
"visual.vision_model.encoder.layers.18.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 309 |
+
"visual.vision_model.encoder.layers.18.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 310 |
+
"visual.vision_model.encoder.layers.18.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 311 |
+
"visual.vision_model.encoder.layers.18.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 312 |
+
"visual.vision_model.encoder.layers.18.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 313 |
+
"visual.vision_model.encoder.layers.18.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 314 |
+
"visual.vision_model.encoder.layers.18.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 315 |
+
"visual.vision_model.encoder.layers.18.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 316 |
+
"visual.vision_model.encoder.layers.18.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 317 |
+
"visual.vision_model.encoder.layers.18.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 318 |
+
"visual.vision_model.encoder.layers.18.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 319 |
+
"visual.vision_model.encoder.layers.19.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 320 |
+
"visual.vision_model.encoder.layers.19.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 321 |
+
"visual.vision_model.encoder.layers.19.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 322 |
+
"visual.vision_model.encoder.layers.19.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 323 |
+
"visual.vision_model.encoder.layers.19.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 324 |
+
"visual.vision_model.encoder.layers.19.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 325 |
+
"visual.vision_model.encoder.layers.19.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 326 |
+
"visual.vision_model.encoder.layers.19.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 327 |
+
"visual.vision_model.encoder.layers.19.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 328 |
+
"visual.vision_model.encoder.layers.19.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 329 |
+
"visual.vision_model.encoder.layers.19.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 330 |
+
"visual.vision_model.encoder.layers.19.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 331 |
+
"visual.vision_model.encoder.layers.19.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 332 |
+
"visual.vision_model.encoder.layers.19.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 333 |
+
"visual.vision_model.encoder.layers.19.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 334 |
+
"visual.vision_model.encoder.layers.19.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 335 |
+
"visual.vision_model.encoder.layers.20.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 336 |
+
"visual.vision_model.encoder.layers.20.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 337 |
+
"visual.vision_model.encoder.layers.20.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 338 |
+
"visual.vision_model.encoder.layers.20.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 339 |
+
"visual.vision_model.encoder.layers.20.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 340 |
+
"visual.vision_model.encoder.layers.20.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 341 |
+
"visual.vision_model.encoder.layers.20.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 342 |
+
"visual.vision_model.encoder.layers.20.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 343 |
+
"visual.vision_model.encoder.layers.20.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 344 |
+
"visual.vision_model.encoder.layers.20.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 345 |
+
"visual.vision_model.encoder.layers.20.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 346 |
+
"visual.vision_model.encoder.layers.20.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 347 |
+
"visual.vision_model.encoder.layers.20.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 348 |
+
"visual.vision_model.encoder.layers.20.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 349 |
+
"visual.vision_model.encoder.layers.20.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 350 |
+
"visual.vision_model.encoder.layers.20.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 351 |
+
"visual.vision_model.encoder.layers.21.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 352 |
+
"visual.vision_model.encoder.layers.21.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 353 |
+
"visual.vision_model.encoder.layers.21.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 354 |
+
"visual.vision_model.encoder.layers.21.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 355 |
+
"visual.vision_model.encoder.layers.21.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 356 |
+
"visual.vision_model.encoder.layers.21.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 357 |
+
"visual.vision_model.encoder.layers.21.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 358 |
+
"visual.vision_model.encoder.layers.21.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 359 |
+
"visual.vision_model.encoder.layers.21.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 360 |
+
"visual.vision_model.encoder.layers.21.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 361 |
+
"visual.vision_model.encoder.layers.21.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 362 |
+
"visual.vision_model.encoder.layers.21.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 363 |
+
"visual.vision_model.encoder.layers.21.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 364 |
+
"visual.vision_model.encoder.layers.21.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 365 |
+
"visual.vision_model.encoder.layers.21.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 366 |
+
"visual.vision_model.encoder.layers.21.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 367 |
+
"visual.vision_model.encoder.layers.22.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 368 |
+
"visual.vision_model.encoder.layers.22.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 369 |
+
"visual.vision_model.encoder.layers.22.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 370 |
+
"visual.vision_model.encoder.layers.22.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 371 |
+
"visual.vision_model.encoder.layers.22.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 372 |
+
"visual.vision_model.encoder.layers.22.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 373 |
+
"visual.vision_model.encoder.layers.22.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 374 |
+
"visual.vision_model.encoder.layers.22.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 375 |
+
"visual.vision_model.encoder.layers.22.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 376 |
+
"visual.vision_model.encoder.layers.22.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 377 |
+
"visual.vision_model.encoder.layers.22.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 378 |
+
"visual.vision_model.encoder.layers.22.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 379 |
+
"visual.vision_model.encoder.layers.22.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 380 |
+
"visual.vision_model.encoder.layers.22.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 381 |
+
"visual.vision_model.encoder.layers.22.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 382 |
+
"visual.vision_model.encoder.layers.22.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 383 |
+
"visual.vision_model.encoder.layers.23.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 384 |
+
"visual.vision_model.encoder.layers.23.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 385 |
+
"visual.vision_model.encoder.layers.23.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 386 |
+
"visual.vision_model.encoder.layers.23.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 387 |
+
"visual.vision_model.encoder.layers.23.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 388 |
+
"visual.vision_model.encoder.layers.23.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 389 |
+
"visual.vision_model.encoder.layers.23.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 390 |
+
"visual.vision_model.encoder.layers.23.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 391 |
+
"visual.vision_model.encoder.layers.23.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 392 |
+
"visual.vision_model.encoder.layers.23.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 393 |
+
"visual.vision_model.encoder.layers.23.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 394 |
+
"visual.vision_model.encoder.layers.23.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 395 |
+
"visual.vision_model.encoder.layers.23.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 396 |
+
"visual.vision_model.encoder.layers.23.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 397 |
+
"visual.vision_model.encoder.layers.23.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 398 |
+
"visual.vision_model.encoder.layers.23.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 399 |
+
"visual.vision_model.encoder.layers.24.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 400 |
+
"visual.vision_model.encoder.layers.24.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 401 |
+
"visual.vision_model.encoder.layers.24.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 402 |
+
"visual.vision_model.encoder.layers.24.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 403 |
+
"visual.vision_model.encoder.layers.24.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 404 |
+
"visual.vision_model.encoder.layers.24.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 405 |
+
"visual.vision_model.encoder.layers.24.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 406 |
+
"visual.vision_model.encoder.layers.24.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 407 |
+
"visual.vision_model.encoder.layers.24.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 408 |
+
"visual.vision_model.encoder.layers.24.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 409 |
+
"visual.vision_model.encoder.layers.24.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 410 |
+
"visual.vision_model.encoder.layers.24.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 411 |
+
"visual.vision_model.encoder.layers.24.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 412 |
+
"visual.vision_model.encoder.layers.24.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 413 |
+
"visual.vision_model.encoder.layers.24.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 414 |
+
"visual.vision_model.encoder.layers.24.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 415 |
+
"visual.vision_model.encoder.layers.25.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 416 |
+
"visual.vision_model.encoder.layers.25.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 417 |
+
"visual.vision_model.encoder.layers.25.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 418 |
+
"visual.vision_model.encoder.layers.25.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 419 |
+
"visual.vision_model.encoder.layers.25.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 420 |
+
"visual.vision_model.encoder.layers.25.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 421 |
+
"visual.vision_model.encoder.layers.25.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 422 |
+
"visual.vision_model.encoder.layers.25.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 423 |
+
"visual.vision_model.encoder.layers.25.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 424 |
+
"visual.vision_model.encoder.layers.25.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 425 |
+
"visual.vision_model.encoder.layers.25.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 426 |
+
"visual.vision_model.encoder.layers.25.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 427 |
+
"visual.vision_model.encoder.layers.25.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 428 |
+
"visual.vision_model.encoder.layers.25.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 429 |
+
"visual.vision_model.encoder.layers.25.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 430 |
+
"visual.vision_model.encoder.layers.25.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 431 |
+
"visual.vision_model.encoder.layers.26.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
|
| 432 |
+
"visual.vision_model.encoder.layers.26.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
|
| 433 |
+
"visual.vision_model.encoder.layers.26.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 434 |
+
"visual.vision_model.encoder.layers.26.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 435 |
+
"visual.vision_model.encoder.layers.26.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 436 |
+
"visual.vision_model.encoder.layers.26.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 437 |
+
"visual.vision_model.encoder.layers.26.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 438 |
+
"visual.vision_model.encoder.layers.26.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 439 |
+
"visual.vision_model.encoder.layers.26.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 440 |
+
"visual.vision_model.encoder.layers.26.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
|
| 441 |
+
"visual.vision_model.encoder.layers.26.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
|
| 442 |
+
"visual.vision_model.encoder.layers.26.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
|
| 443 |
+
"visual.vision_model.encoder.layers.26.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
|
| 444 |
+
"visual.vision_model.encoder.layers.26.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
|
| 445 |
+
"visual.vision_model.encoder.layers.26.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
|
| 446 |
+
"visual.vision_model.encoder.layers.26.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
|
| 447 |
+
"visual.vision_model.post_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 448 |
+
"visual.vision_model.post_layernorm.bias": "master_weights-00001-of-00001.safetensors",
|
| 449 |
+
"model.embed_tokens.weight": "master_weights-00001-of-00001.safetensors",
|
| 450 |
+
"model.layers.0.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 451 |
+
"model.layers.0.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 452 |
+
"model.layers.0.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 453 |
+
"model.layers.0.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 454 |
+
"model.layers.0.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 455 |
+
"model.layers.0.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 456 |
+
"model.layers.0.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 457 |
+
"model.layers.0.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 458 |
+
"model.layers.0.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 459 |
+
"model.layers.1.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 460 |
+
"model.layers.1.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 461 |
+
"model.layers.1.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 462 |
+
"model.layers.1.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 463 |
+
"model.layers.1.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 464 |
+
"model.layers.1.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 465 |
+
"model.layers.1.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 466 |
+
"model.layers.1.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 467 |
+
"model.layers.1.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 468 |
+
"model.layers.2.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 469 |
+
"model.layers.2.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 470 |
+
"model.layers.2.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 471 |
+
"model.layers.2.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 472 |
+
"model.layers.2.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 473 |
+
"model.layers.2.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 474 |
+
"model.layers.2.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 475 |
+
"model.layers.2.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 476 |
+
"model.layers.2.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 477 |
+
"model.layers.3.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 478 |
+
"model.layers.3.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 479 |
+
"model.layers.3.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 480 |
+
"model.layers.3.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 481 |
+
"model.layers.3.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 482 |
+
"model.layers.3.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 483 |
+
"model.layers.3.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 484 |
+
"model.layers.3.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 485 |
+
"model.layers.3.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 486 |
+
"model.layers.4.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 487 |
+
"model.layers.4.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 488 |
+
"model.layers.4.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 489 |
+
"model.layers.4.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 490 |
+
"model.layers.4.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 491 |
+
"model.layers.4.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 492 |
+
"model.layers.4.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 493 |
+
"model.layers.4.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 494 |
+
"model.layers.4.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 495 |
+
"model.layers.5.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 496 |
+
"model.layers.5.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 497 |
+
"model.layers.5.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 498 |
+
"model.layers.5.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 499 |
+
"model.layers.5.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 500 |
+
"model.layers.5.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 501 |
+
"model.layers.5.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 502 |
+
"model.layers.5.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 503 |
+
"model.layers.5.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 504 |
+
"model.layers.6.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 505 |
+
"model.layers.6.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 506 |
+
"model.layers.6.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 507 |
+
"model.layers.6.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 508 |
+
"model.layers.6.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 509 |
+
"model.layers.6.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 510 |
+
"model.layers.6.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 511 |
+
"model.layers.6.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 512 |
+
"model.layers.6.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 513 |
+
"model.layers.7.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 514 |
+
"model.layers.7.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 515 |
+
"model.layers.7.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 516 |
+
"model.layers.7.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 517 |
+
"model.layers.7.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 518 |
+
"model.layers.7.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 519 |
+
"model.layers.7.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 520 |
+
"model.layers.7.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 521 |
+
"model.layers.7.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 522 |
+
"model.layers.8.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 523 |
+
"model.layers.8.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 524 |
+
"model.layers.8.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 525 |
+
"model.layers.8.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 526 |
+
"model.layers.8.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 527 |
+
"model.layers.8.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 528 |
+
"model.layers.8.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 529 |
+
"model.layers.8.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 530 |
+
"model.layers.8.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 531 |
+
"model.layers.9.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 532 |
+
"model.layers.9.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 533 |
+
"model.layers.9.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 534 |
+
"model.layers.9.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 535 |
+
"model.layers.9.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 536 |
+
"model.layers.9.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 537 |
+
"model.layers.9.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 538 |
+
"model.layers.9.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 539 |
+
"model.layers.9.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 540 |
+
"model.layers.10.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 541 |
+
"model.layers.10.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 542 |
+
"model.layers.10.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 543 |
+
"model.layers.10.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 544 |
+
"model.layers.10.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 545 |
+
"model.layers.10.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 546 |
+
"model.layers.10.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 547 |
+
"model.layers.10.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 548 |
+
"model.layers.10.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 549 |
+
"model.layers.11.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 550 |
+
"model.layers.11.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 551 |
+
"model.layers.11.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 552 |
+
"model.layers.11.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 553 |
+
"model.layers.11.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 554 |
+
"model.layers.11.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 555 |
+
"model.layers.11.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 556 |
+
"model.layers.11.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 557 |
+
"model.layers.11.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 558 |
+
"model.layers.12.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 559 |
+
"model.layers.12.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 560 |
+
"model.layers.12.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 561 |
+
"model.layers.12.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 562 |
+
"model.layers.12.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 563 |
+
"model.layers.12.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 564 |
+
"model.layers.12.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 565 |
+
"model.layers.12.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 566 |
+
"model.layers.12.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 567 |
+
"model.layers.13.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 568 |
+
"model.layers.13.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 569 |
+
"model.layers.13.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 570 |
+
"model.layers.13.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 571 |
+
"model.layers.13.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 572 |
+
"model.layers.13.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 573 |
+
"model.layers.13.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 574 |
+
"model.layers.13.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 575 |
+
"model.layers.13.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 576 |
+
"model.layers.14.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 577 |
+
"model.layers.14.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 578 |
+
"model.layers.14.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 579 |
+
"model.layers.14.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 580 |
+
"model.layers.14.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 581 |
+
"model.layers.14.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 582 |
+
"model.layers.14.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 583 |
+
"model.layers.14.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 584 |
+
"model.layers.14.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 585 |
+
"model.layers.15.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 586 |
+
"model.layers.15.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 587 |
+
"model.layers.15.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 588 |
+
"model.layers.15.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 589 |
+
"model.layers.15.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 590 |
+
"model.layers.15.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 591 |
+
"model.layers.15.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 592 |
+
"model.layers.15.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 593 |
+
"model.layers.15.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 594 |
+
"model.layers.16.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 595 |
+
"model.layers.16.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 596 |
+
"model.layers.16.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 597 |
+
"model.layers.16.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 598 |
+
"model.layers.16.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 599 |
+
"model.layers.16.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 600 |
+
"model.layers.16.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 601 |
+
"model.layers.16.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 602 |
+
"model.layers.16.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 603 |
+
"model.layers.17.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 604 |
+
"model.layers.17.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 605 |
+
"model.layers.17.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 606 |
+
"model.layers.17.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 607 |
+
"model.layers.17.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 608 |
+
"model.layers.17.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 609 |
+
"model.layers.17.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
|
| 610 |
+
"model.layers.17.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 611 |
+
"model.layers.17.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
|
| 612 |
+
"model.norm.weight": "master_weights-00001-of-00001.safetensors",
|
| 613 |
+
"lm_head.weight": "master_weights-00001-of-00001.safetensors"
|
| 614 |
+
}
|
| 615 |
+
}
|
checkpoint-420/model-00001-of-00001.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2c94d703a0352affef9bd2203aee6cc51e4dfb33b7b294542eb6ae19def2237d
|
| 3 |
+
size 1917255968
|
checkpoint-420/model.safetensors.index.json
ADDED
|
@@ -0,0 +1,627 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"metadata": {
|
| 3 |
+
"total_size": 1917177472
|
| 4 |
+
},
|
| 5 |
+
"weight_map": {
|
| 6 |
+
"mlp_AR.pre_norm.weight": "model-00001-of-00001.safetensors",
|
| 7 |
+
"mlp_AR.pre_norm.bias": "model-00001-of-00001.safetensors",
|
| 8 |
+
"mlp_AR.linear_1.weight": "model-00001-of-00001.safetensors",
|
| 9 |
+
"mlp_AR.linear_1.bias": "model-00001-of-00001.safetensors",
|
| 10 |
+
"mlp_AR.linear_2.weight": "model-00001-of-00001.safetensors",
|
| 11 |
+
"mlp_AR.linear_2.bias": "model-00001-of-00001.safetensors",
|
| 12 |
+
"visual.vision_model.embeddings.patch_embedding.weight": "model-00001-of-00001.safetensors",
|
| 13 |
+
"visual.vision_model.embeddings.patch_embedding.bias": "model-00001-of-00001.safetensors",
|
| 14 |
+
"visual.vision_model.embeddings.position_embedding.weight": "model-00001-of-00001.safetensors",
|
| 15 |
+
"visual.vision_model.embeddings.packing_position_embedding.weight": "model-00001-of-00001.safetensors",
|
| 16 |
+
"visual.vision_model.encoder.layers.0.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 17 |
+
"visual.vision_model.encoder.layers.0.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 18 |
+
"visual.vision_model.encoder.layers.0.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 19 |
+
"visual.vision_model.encoder.layers.0.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 20 |
+
"visual.vision_model.encoder.layers.0.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 21 |
+
"visual.vision_model.encoder.layers.0.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 22 |
+
"visual.vision_model.encoder.layers.0.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 23 |
+
"visual.vision_model.encoder.layers.0.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 24 |
+
"visual.vision_model.encoder.layers.0.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 25 |
+
"visual.vision_model.encoder.layers.0.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 26 |
+
"visual.vision_model.encoder.layers.0.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 27 |
+
"visual.vision_model.encoder.layers.0.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 28 |
+
"visual.vision_model.encoder.layers.0.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 29 |
+
"visual.vision_model.encoder.layers.0.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 30 |
+
"visual.vision_model.encoder.layers.0.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 31 |
+
"visual.vision_model.encoder.layers.0.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 32 |
+
"visual.vision_model.encoder.layers.1.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 33 |
+
"visual.vision_model.encoder.layers.1.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 34 |
+
"visual.vision_model.encoder.layers.1.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 35 |
+
"visual.vision_model.encoder.layers.1.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 36 |
+
"visual.vision_model.encoder.layers.1.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 37 |
+
"visual.vision_model.encoder.layers.1.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 38 |
+
"visual.vision_model.encoder.layers.1.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 39 |
+
"visual.vision_model.encoder.layers.1.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 40 |
+
"visual.vision_model.encoder.layers.1.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 41 |
+
"visual.vision_model.encoder.layers.1.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 42 |
+
"visual.vision_model.encoder.layers.1.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 43 |
+
"visual.vision_model.encoder.layers.1.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 44 |
+
"visual.vision_model.encoder.layers.1.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 45 |
+
"visual.vision_model.encoder.layers.1.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 46 |
+
"visual.vision_model.encoder.layers.1.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 47 |
+
"visual.vision_model.encoder.layers.1.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 48 |
+
"visual.vision_model.encoder.layers.2.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 49 |
+
"visual.vision_model.encoder.layers.2.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 50 |
+
"visual.vision_model.encoder.layers.2.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 51 |
+
"visual.vision_model.encoder.layers.2.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 52 |
+
"visual.vision_model.encoder.layers.2.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 53 |
+
"visual.vision_model.encoder.layers.2.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 54 |
+
"visual.vision_model.encoder.layers.2.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 55 |
+
"visual.vision_model.encoder.layers.2.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 56 |
+
"visual.vision_model.encoder.layers.2.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 57 |
+
"visual.vision_model.encoder.layers.2.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 58 |
+
"visual.vision_model.encoder.layers.2.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 59 |
+
"visual.vision_model.encoder.layers.2.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 60 |
+
"visual.vision_model.encoder.layers.2.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 61 |
+
"visual.vision_model.encoder.layers.2.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 62 |
+
"visual.vision_model.encoder.layers.2.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 63 |
+
"visual.vision_model.encoder.layers.2.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 64 |
+
"visual.vision_model.encoder.layers.3.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 65 |
+
"visual.vision_model.encoder.layers.3.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 66 |
+
"visual.vision_model.encoder.layers.3.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 67 |
+
"visual.vision_model.encoder.layers.3.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 68 |
+
"visual.vision_model.encoder.layers.3.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 69 |
+
"visual.vision_model.encoder.layers.3.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 70 |
+
"visual.vision_model.encoder.layers.3.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 71 |
+
"visual.vision_model.encoder.layers.3.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 72 |
+
"visual.vision_model.encoder.layers.3.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 73 |
+
"visual.vision_model.encoder.layers.3.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 74 |
+
"visual.vision_model.encoder.layers.3.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 75 |
+
"visual.vision_model.encoder.layers.3.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 76 |
+
"visual.vision_model.encoder.layers.3.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 77 |
+
"visual.vision_model.encoder.layers.3.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 78 |
+
"visual.vision_model.encoder.layers.3.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 79 |
+
"visual.vision_model.encoder.layers.3.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 80 |
+
"visual.vision_model.encoder.layers.4.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 81 |
+
"visual.vision_model.encoder.layers.4.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 82 |
+
"visual.vision_model.encoder.layers.4.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 83 |
+
"visual.vision_model.encoder.layers.4.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 84 |
+
"visual.vision_model.encoder.layers.4.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 85 |
+
"visual.vision_model.encoder.layers.4.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 86 |
+
"visual.vision_model.encoder.layers.4.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 87 |
+
"visual.vision_model.encoder.layers.4.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 88 |
+
"visual.vision_model.encoder.layers.4.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 89 |
+
"visual.vision_model.encoder.layers.4.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 90 |
+
"visual.vision_model.encoder.layers.4.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 91 |
+
"visual.vision_model.encoder.layers.4.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 92 |
+
"visual.vision_model.encoder.layers.4.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 93 |
+
"visual.vision_model.encoder.layers.4.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 94 |
+
"visual.vision_model.encoder.layers.4.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 95 |
+
"visual.vision_model.encoder.layers.4.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 96 |
+
"visual.vision_model.encoder.layers.5.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 97 |
+
"visual.vision_model.encoder.layers.5.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 98 |
+
"visual.vision_model.encoder.layers.5.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 99 |
+
"visual.vision_model.encoder.layers.5.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 100 |
+
"visual.vision_model.encoder.layers.5.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 101 |
+
"visual.vision_model.encoder.layers.5.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 102 |
+
"visual.vision_model.encoder.layers.5.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 103 |
+
"visual.vision_model.encoder.layers.5.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 104 |
+
"visual.vision_model.encoder.layers.5.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 105 |
+
"visual.vision_model.encoder.layers.5.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 106 |
+
"visual.vision_model.encoder.layers.5.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 107 |
+
"visual.vision_model.encoder.layers.5.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 108 |
+
"visual.vision_model.encoder.layers.5.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 109 |
+
"visual.vision_model.encoder.layers.5.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 110 |
+
"visual.vision_model.encoder.layers.5.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 111 |
+
"visual.vision_model.encoder.layers.5.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 112 |
+
"visual.vision_model.encoder.layers.6.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 113 |
+
"visual.vision_model.encoder.layers.6.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 114 |
+
"visual.vision_model.encoder.layers.6.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 115 |
+
"visual.vision_model.encoder.layers.6.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 116 |
+
"visual.vision_model.encoder.layers.6.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 117 |
+
"visual.vision_model.encoder.layers.6.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 118 |
+
"visual.vision_model.encoder.layers.6.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 119 |
+
"visual.vision_model.encoder.layers.6.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 120 |
+
"visual.vision_model.encoder.layers.6.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 121 |
+
"visual.vision_model.encoder.layers.6.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 122 |
+
"visual.vision_model.encoder.layers.6.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 123 |
+
"visual.vision_model.encoder.layers.6.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 124 |
+
"visual.vision_model.encoder.layers.6.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 125 |
+
"visual.vision_model.encoder.layers.6.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 126 |
+
"visual.vision_model.encoder.layers.6.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 127 |
+
"visual.vision_model.encoder.layers.6.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 128 |
+
"visual.vision_model.encoder.layers.7.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 129 |
+
"visual.vision_model.encoder.layers.7.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 130 |
+
"visual.vision_model.encoder.layers.7.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 131 |
+
"visual.vision_model.encoder.layers.7.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 132 |
+
"visual.vision_model.encoder.layers.7.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 133 |
+
"visual.vision_model.encoder.layers.7.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 134 |
+
"visual.vision_model.encoder.layers.7.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 135 |
+
"visual.vision_model.encoder.layers.7.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 136 |
+
"visual.vision_model.encoder.layers.7.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 137 |
+
"visual.vision_model.encoder.layers.7.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 138 |
+
"visual.vision_model.encoder.layers.7.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 139 |
+
"visual.vision_model.encoder.layers.7.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 140 |
+
"visual.vision_model.encoder.layers.7.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 141 |
+
"visual.vision_model.encoder.layers.7.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 142 |
+
"visual.vision_model.encoder.layers.7.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 143 |
+
"visual.vision_model.encoder.layers.7.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 144 |
+
"visual.vision_model.encoder.layers.8.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 145 |
+
"visual.vision_model.encoder.layers.8.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 146 |
+
"visual.vision_model.encoder.layers.8.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 147 |
+
"visual.vision_model.encoder.layers.8.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 148 |
+
"visual.vision_model.encoder.layers.8.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 149 |
+
"visual.vision_model.encoder.layers.8.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 150 |
+
"visual.vision_model.encoder.layers.8.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 151 |
+
"visual.vision_model.encoder.layers.8.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 152 |
+
"visual.vision_model.encoder.layers.8.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 153 |
+
"visual.vision_model.encoder.layers.8.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 154 |
+
"visual.vision_model.encoder.layers.8.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 155 |
+
"visual.vision_model.encoder.layers.8.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 156 |
+
"visual.vision_model.encoder.layers.8.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 157 |
+
"visual.vision_model.encoder.layers.8.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 158 |
+
"visual.vision_model.encoder.layers.8.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 159 |
+
"visual.vision_model.encoder.layers.8.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 160 |
+
"visual.vision_model.encoder.layers.9.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 161 |
+
"visual.vision_model.encoder.layers.9.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 162 |
+
"visual.vision_model.encoder.layers.9.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 163 |
+
"visual.vision_model.encoder.layers.9.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 164 |
+
"visual.vision_model.encoder.layers.9.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 165 |
+
"visual.vision_model.encoder.layers.9.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 166 |
+
"visual.vision_model.encoder.layers.9.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 167 |
+
"visual.vision_model.encoder.layers.9.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 168 |
+
"visual.vision_model.encoder.layers.9.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 169 |
+
"visual.vision_model.encoder.layers.9.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 170 |
+
"visual.vision_model.encoder.layers.9.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 171 |
+
"visual.vision_model.encoder.layers.9.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 172 |
+
"visual.vision_model.encoder.layers.9.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 173 |
+
"visual.vision_model.encoder.layers.9.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 174 |
+
"visual.vision_model.encoder.layers.9.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 175 |
+
"visual.vision_model.encoder.layers.9.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 176 |
+
"visual.vision_model.encoder.layers.10.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 177 |
+
"visual.vision_model.encoder.layers.10.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 178 |
+
"visual.vision_model.encoder.layers.10.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 179 |
+
"visual.vision_model.encoder.layers.10.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 180 |
+
"visual.vision_model.encoder.layers.10.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 181 |
+
"visual.vision_model.encoder.layers.10.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 182 |
+
"visual.vision_model.encoder.layers.10.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 183 |
+
"visual.vision_model.encoder.layers.10.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 184 |
+
"visual.vision_model.encoder.layers.10.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 185 |
+
"visual.vision_model.encoder.layers.10.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 186 |
+
"visual.vision_model.encoder.layers.10.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 187 |
+
"visual.vision_model.encoder.layers.10.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 188 |
+
"visual.vision_model.encoder.layers.10.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 189 |
+
"visual.vision_model.encoder.layers.10.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 190 |
+
"visual.vision_model.encoder.layers.10.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 191 |
+
"visual.vision_model.encoder.layers.10.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 192 |
+
"visual.vision_model.encoder.layers.11.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 193 |
+
"visual.vision_model.encoder.layers.11.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 194 |
+
"visual.vision_model.encoder.layers.11.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 195 |
+
"visual.vision_model.encoder.layers.11.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 196 |
+
"visual.vision_model.encoder.layers.11.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 197 |
+
"visual.vision_model.encoder.layers.11.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 198 |
+
"visual.vision_model.encoder.layers.11.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 199 |
+
"visual.vision_model.encoder.layers.11.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 200 |
+
"visual.vision_model.encoder.layers.11.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 201 |
+
"visual.vision_model.encoder.layers.11.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 202 |
+
"visual.vision_model.encoder.layers.11.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 203 |
+
"visual.vision_model.encoder.layers.11.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 204 |
+
"visual.vision_model.encoder.layers.11.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 205 |
+
"visual.vision_model.encoder.layers.11.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 206 |
+
"visual.vision_model.encoder.layers.11.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 207 |
+
"visual.vision_model.encoder.layers.11.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 208 |
+
"visual.vision_model.encoder.layers.12.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 209 |
+
"visual.vision_model.encoder.layers.12.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 210 |
+
"visual.vision_model.encoder.layers.12.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 211 |
+
"visual.vision_model.encoder.layers.12.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 212 |
+
"visual.vision_model.encoder.layers.12.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 213 |
+
"visual.vision_model.encoder.layers.12.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 214 |
+
"visual.vision_model.encoder.layers.12.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 215 |
+
"visual.vision_model.encoder.layers.12.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 216 |
+
"visual.vision_model.encoder.layers.12.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 217 |
+
"visual.vision_model.encoder.layers.12.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 218 |
+
"visual.vision_model.encoder.layers.12.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 219 |
+
"visual.vision_model.encoder.layers.12.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 220 |
+
"visual.vision_model.encoder.layers.12.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 221 |
+
"visual.vision_model.encoder.layers.12.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 222 |
+
"visual.vision_model.encoder.layers.12.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 223 |
+
"visual.vision_model.encoder.layers.12.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 224 |
+
"visual.vision_model.encoder.layers.13.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 225 |
+
"visual.vision_model.encoder.layers.13.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 226 |
+
"visual.vision_model.encoder.layers.13.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 227 |
+
"visual.vision_model.encoder.layers.13.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 228 |
+
"visual.vision_model.encoder.layers.13.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 229 |
+
"visual.vision_model.encoder.layers.13.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 230 |
+
"visual.vision_model.encoder.layers.13.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 231 |
+
"visual.vision_model.encoder.layers.13.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 232 |
+
"visual.vision_model.encoder.layers.13.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 233 |
+
"visual.vision_model.encoder.layers.13.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 234 |
+
"visual.vision_model.encoder.layers.13.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 235 |
+
"visual.vision_model.encoder.layers.13.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 236 |
+
"visual.vision_model.encoder.layers.13.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 237 |
+
"visual.vision_model.encoder.layers.13.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 238 |
+
"visual.vision_model.encoder.layers.13.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 239 |
+
"visual.vision_model.encoder.layers.13.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 240 |
+
"visual.vision_model.encoder.layers.14.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 241 |
+
"visual.vision_model.encoder.layers.14.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 242 |
+
"visual.vision_model.encoder.layers.14.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 243 |
+
"visual.vision_model.encoder.layers.14.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 244 |
+
"visual.vision_model.encoder.layers.14.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 245 |
+
"visual.vision_model.encoder.layers.14.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 246 |
+
"visual.vision_model.encoder.layers.14.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 247 |
+
"visual.vision_model.encoder.layers.14.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 248 |
+
"visual.vision_model.encoder.layers.14.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 249 |
+
"visual.vision_model.encoder.layers.14.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 250 |
+
"visual.vision_model.encoder.layers.14.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 251 |
+
"visual.vision_model.encoder.layers.14.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 252 |
+
"visual.vision_model.encoder.layers.14.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 253 |
+
"visual.vision_model.encoder.layers.14.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 254 |
+
"visual.vision_model.encoder.layers.14.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 255 |
+
"visual.vision_model.encoder.layers.14.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 256 |
+
"visual.vision_model.encoder.layers.15.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 257 |
+
"visual.vision_model.encoder.layers.15.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 258 |
+
"visual.vision_model.encoder.layers.15.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 259 |
+
"visual.vision_model.encoder.layers.15.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 260 |
+
"visual.vision_model.encoder.layers.15.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 261 |
+
"visual.vision_model.encoder.layers.15.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 262 |
+
"visual.vision_model.encoder.layers.15.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 263 |
+
"visual.vision_model.encoder.layers.15.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 264 |
+
"visual.vision_model.encoder.layers.15.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 265 |
+
"visual.vision_model.encoder.layers.15.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 266 |
+
"visual.vision_model.encoder.layers.15.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 267 |
+
"visual.vision_model.encoder.layers.15.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 268 |
+
"visual.vision_model.encoder.layers.15.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 269 |
+
"visual.vision_model.encoder.layers.15.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 270 |
+
"visual.vision_model.encoder.layers.15.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 271 |
+
"visual.vision_model.encoder.layers.15.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 272 |
+
"visual.vision_model.encoder.layers.16.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 273 |
+
"visual.vision_model.encoder.layers.16.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 274 |
+
"visual.vision_model.encoder.layers.16.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 275 |
+
"visual.vision_model.encoder.layers.16.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 276 |
+
"visual.vision_model.encoder.layers.16.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 277 |
+
"visual.vision_model.encoder.layers.16.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 278 |
+
"visual.vision_model.encoder.layers.16.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 279 |
+
"visual.vision_model.encoder.layers.16.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 280 |
+
"visual.vision_model.encoder.layers.16.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 281 |
+
"visual.vision_model.encoder.layers.16.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 282 |
+
"visual.vision_model.encoder.layers.16.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 283 |
+
"visual.vision_model.encoder.layers.16.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 284 |
+
"visual.vision_model.encoder.layers.16.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 285 |
+
"visual.vision_model.encoder.layers.16.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 286 |
+
"visual.vision_model.encoder.layers.16.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 287 |
+
"visual.vision_model.encoder.layers.16.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 288 |
+
"visual.vision_model.encoder.layers.17.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 289 |
+
"visual.vision_model.encoder.layers.17.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 290 |
+
"visual.vision_model.encoder.layers.17.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 291 |
+
"visual.vision_model.encoder.layers.17.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 292 |
+
"visual.vision_model.encoder.layers.17.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 293 |
+
"visual.vision_model.encoder.layers.17.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 294 |
+
"visual.vision_model.encoder.layers.17.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 295 |
+
"visual.vision_model.encoder.layers.17.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 296 |
+
"visual.vision_model.encoder.layers.17.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 297 |
+
"visual.vision_model.encoder.layers.17.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 298 |
+
"visual.vision_model.encoder.layers.17.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 299 |
+
"visual.vision_model.encoder.layers.17.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 300 |
+
"visual.vision_model.encoder.layers.17.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 301 |
+
"visual.vision_model.encoder.layers.17.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 302 |
+
"visual.vision_model.encoder.layers.17.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 303 |
+
"visual.vision_model.encoder.layers.17.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 304 |
+
"visual.vision_model.encoder.layers.18.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 305 |
+
"visual.vision_model.encoder.layers.18.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 306 |
+
"visual.vision_model.encoder.layers.18.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 307 |
+
"visual.vision_model.encoder.layers.18.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 308 |
+
"visual.vision_model.encoder.layers.18.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 309 |
+
"visual.vision_model.encoder.layers.18.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 310 |
+
"visual.vision_model.encoder.layers.18.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 311 |
+
"visual.vision_model.encoder.layers.18.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 312 |
+
"visual.vision_model.encoder.layers.18.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 313 |
+
"visual.vision_model.encoder.layers.18.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 314 |
+
"visual.vision_model.encoder.layers.18.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 315 |
+
"visual.vision_model.encoder.layers.18.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 316 |
+
"visual.vision_model.encoder.layers.18.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 317 |
+
"visual.vision_model.encoder.layers.18.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 318 |
+
"visual.vision_model.encoder.layers.18.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 319 |
+
"visual.vision_model.encoder.layers.18.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 320 |
+
"visual.vision_model.encoder.layers.19.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 321 |
+
"visual.vision_model.encoder.layers.19.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 322 |
+
"visual.vision_model.encoder.layers.19.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 323 |
+
"visual.vision_model.encoder.layers.19.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 324 |
+
"visual.vision_model.encoder.layers.19.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 325 |
+
"visual.vision_model.encoder.layers.19.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 326 |
+
"visual.vision_model.encoder.layers.19.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 327 |
+
"visual.vision_model.encoder.layers.19.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 328 |
+
"visual.vision_model.encoder.layers.19.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 329 |
+
"visual.vision_model.encoder.layers.19.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 330 |
+
"visual.vision_model.encoder.layers.19.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 331 |
+
"visual.vision_model.encoder.layers.19.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 332 |
+
"visual.vision_model.encoder.layers.19.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 333 |
+
"visual.vision_model.encoder.layers.19.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 334 |
+
"visual.vision_model.encoder.layers.19.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 335 |
+
"visual.vision_model.encoder.layers.19.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 336 |
+
"visual.vision_model.encoder.layers.20.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 337 |
+
"visual.vision_model.encoder.layers.20.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 338 |
+
"visual.vision_model.encoder.layers.20.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 339 |
+
"visual.vision_model.encoder.layers.20.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 340 |
+
"visual.vision_model.encoder.layers.20.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 341 |
+
"visual.vision_model.encoder.layers.20.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 342 |
+
"visual.vision_model.encoder.layers.20.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 343 |
+
"visual.vision_model.encoder.layers.20.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 344 |
+
"visual.vision_model.encoder.layers.20.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 345 |
+
"visual.vision_model.encoder.layers.20.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 346 |
+
"visual.vision_model.encoder.layers.20.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 347 |
+
"visual.vision_model.encoder.layers.20.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 348 |
+
"visual.vision_model.encoder.layers.20.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 349 |
+
"visual.vision_model.encoder.layers.20.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 350 |
+
"visual.vision_model.encoder.layers.20.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 351 |
+
"visual.vision_model.encoder.layers.20.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 352 |
+
"visual.vision_model.encoder.layers.21.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 353 |
+
"visual.vision_model.encoder.layers.21.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 354 |
+
"visual.vision_model.encoder.layers.21.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 355 |
+
"visual.vision_model.encoder.layers.21.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 356 |
+
"visual.vision_model.encoder.layers.21.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 357 |
+
"visual.vision_model.encoder.layers.21.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 358 |
+
"visual.vision_model.encoder.layers.21.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 359 |
+
"visual.vision_model.encoder.layers.21.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 360 |
+
"visual.vision_model.encoder.layers.21.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 361 |
+
"visual.vision_model.encoder.layers.21.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 362 |
+
"visual.vision_model.encoder.layers.21.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 363 |
+
"visual.vision_model.encoder.layers.21.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 364 |
+
"visual.vision_model.encoder.layers.21.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 365 |
+
"visual.vision_model.encoder.layers.21.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 366 |
+
"visual.vision_model.encoder.layers.21.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 367 |
+
"visual.vision_model.encoder.layers.21.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 368 |
+
"visual.vision_model.encoder.layers.22.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 369 |
+
"visual.vision_model.encoder.layers.22.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 370 |
+
"visual.vision_model.encoder.layers.22.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 371 |
+
"visual.vision_model.encoder.layers.22.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 372 |
+
"visual.vision_model.encoder.layers.22.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 373 |
+
"visual.vision_model.encoder.layers.22.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 374 |
+
"visual.vision_model.encoder.layers.22.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 375 |
+
"visual.vision_model.encoder.layers.22.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 376 |
+
"visual.vision_model.encoder.layers.22.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 377 |
+
"visual.vision_model.encoder.layers.22.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 378 |
+
"visual.vision_model.encoder.layers.22.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 379 |
+
"visual.vision_model.encoder.layers.22.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 380 |
+
"visual.vision_model.encoder.layers.22.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 381 |
+
"visual.vision_model.encoder.layers.22.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 382 |
+
"visual.vision_model.encoder.layers.22.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 383 |
+
"visual.vision_model.encoder.layers.22.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 384 |
+
"visual.vision_model.encoder.layers.23.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 385 |
+
"visual.vision_model.encoder.layers.23.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 386 |
+
"visual.vision_model.encoder.layers.23.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 387 |
+
"visual.vision_model.encoder.layers.23.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 388 |
+
"visual.vision_model.encoder.layers.23.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 389 |
+
"visual.vision_model.encoder.layers.23.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 390 |
+
"visual.vision_model.encoder.layers.23.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 391 |
+
"visual.vision_model.encoder.layers.23.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 392 |
+
"visual.vision_model.encoder.layers.23.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 393 |
+
"visual.vision_model.encoder.layers.23.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 394 |
+
"visual.vision_model.encoder.layers.23.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 395 |
+
"visual.vision_model.encoder.layers.23.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 396 |
+
"visual.vision_model.encoder.layers.23.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 397 |
+
"visual.vision_model.encoder.layers.23.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 398 |
+
"visual.vision_model.encoder.layers.23.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 399 |
+
"visual.vision_model.encoder.layers.23.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 400 |
+
"visual.vision_model.encoder.layers.24.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 401 |
+
"visual.vision_model.encoder.layers.24.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 402 |
+
"visual.vision_model.encoder.layers.24.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 403 |
+
"visual.vision_model.encoder.layers.24.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 404 |
+
"visual.vision_model.encoder.layers.24.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 405 |
+
"visual.vision_model.encoder.layers.24.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 406 |
+
"visual.vision_model.encoder.layers.24.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 407 |
+
"visual.vision_model.encoder.layers.24.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 408 |
+
"visual.vision_model.encoder.layers.24.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 409 |
+
"visual.vision_model.encoder.layers.24.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 410 |
+
"visual.vision_model.encoder.layers.24.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 411 |
+
"visual.vision_model.encoder.layers.24.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 412 |
+
"visual.vision_model.encoder.layers.24.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 413 |
+
"visual.vision_model.encoder.layers.24.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 414 |
+
"visual.vision_model.encoder.layers.24.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 415 |
+
"visual.vision_model.encoder.layers.24.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 416 |
+
"visual.vision_model.encoder.layers.25.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 417 |
+
"visual.vision_model.encoder.layers.25.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 418 |
+
"visual.vision_model.encoder.layers.25.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 419 |
+
"visual.vision_model.encoder.layers.25.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 420 |
+
"visual.vision_model.encoder.layers.25.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 421 |
+
"visual.vision_model.encoder.layers.25.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 422 |
+
"visual.vision_model.encoder.layers.25.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 423 |
+
"visual.vision_model.encoder.layers.25.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 424 |
+
"visual.vision_model.encoder.layers.25.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 425 |
+
"visual.vision_model.encoder.layers.25.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 426 |
+
"visual.vision_model.encoder.layers.25.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 427 |
+
"visual.vision_model.encoder.layers.25.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 428 |
+
"visual.vision_model.encoder.layers.25.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 429 |
+
"visual.vision_model.encoder.layers.25.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 430 |
+
"visual.vision_model.encoder.layers.25.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 431 |
+
"visual.vision_model.encoder.layers.25.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 432 |
+
"visual.vision_model.encoder.layers.26.layer_norm1.weight": "model-00001-of-00001.safetensors",
|
| 433 |
+
"visual.vision_model.encoder.layers.26.layer_norm1.bias": "model-00001-of-00001.safetensors",
|
| 434 |
+
"visual.vision_model.encoder.layers.26.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 435 |
+
"visual.vision_model.encoder.layers.26.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
|
| 436 |
+
"visual.vision_model.encoder.layers.26.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 437 |
+
"visual.vision_model.encoder.layers.26.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
|
| 438 |
+
"visual.vision_model.encoder.layers.26.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 439 |
+
"visual.vision_model.encoder.layers.26.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
|
| 440 |
+
"visual.vision_model.encoder.layers.26.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 441 |
+
"visual.vision_model.encoder.layers.26.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 442 |
+
"visual.vision_model.encoder.layers.26.layer_norm2.weight": "model-00001-of-00001.safetensors",
|
| 443 |
+
"visual.vision_model.encoder.layers.26.layer_norm2.bias": "model-00001-of-00001.safetensors",
|
| 444 |
+
"visual.vision_model.encoder.layers.26.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 445 |
+
"visual.vision_model.encoder.layers.26.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 446 |
+
"visual.vision_model.encoder.layers.26.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 447 |
+
"visual.vision_model.encoder.layers.26.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 448 |
+
"visual.vision_model.post_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 449 |
+
"visual.vision_model.post_layernorm.bias": "model-00001-of-00001.safetensors",
|
| 450 |
+
"visual.vision_model.head.probe": "model-00001-of-00001.safetensors",
|
| 451 |
+
"visual.vision_model.head.attention.in_proj_weight": "model-00001-of-00001.safetensors",
|
| 452 |
+
"visual.vision_model.head.attention.in_proj_bias": "model-00001-of-00001.safetensors",
|
| 453 |
+
"visual.vision_model.head.attention.out_proj.weight": "model-00001-of-00001.safetensors",
|
| 454 |
+
"visual.vision_model.head.attention.out_proj.bias": "model-00001-of-00001.safetensors",
|
| 455 |
+
"visual.vision_model.head.layernorm.weight": "model-00001-of-00001.safetensors",
|
| 456 |
+
"visual.vision_model.head.layernorm.bias": "model-00001-of-00001.safetensors",
|
| 457 |
+
"visual.vision_model.head.mlp.fc1.weight": "model-00001-of-00001.safetensors",
|
| 458 |
+
"visual.vision_model.head.mlp.fc1.bias": "model-00001-of-00001.safetensors",
|
| 459 |
+
"visual.vision_model.head.mlp.fc2.weight": "model-00001-of-00001.safetensors",
|
| 460 |
+
"visual.vision_model.head.mlp.fc2.bias": "model-00001-of-00001.safetensors",
|
| 461 |
+
"model.embed_tokens.weight": "model-00001-of-00001.safetensors",
|
| 462 |
+
"model.layers.0.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 463 |
+
"model.layers.0.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 464 |
+
"model.layers.0.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 465 |
+
"model.layers.0.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 466 |
+
"model.layers.0.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 467 |
+
"model.layers.0.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 468 |
+
"model.layers.0.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 469 |
+
"model.layers.0.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 470 |
+
"model.layers.0.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 471 |
+
"model.layers.1.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 472 |
+
"model.layers.1.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 473 |
+
"model.layers.1.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 474 |
+
"model.layers.1.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 475 |
+
"model.layers.1.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 476 |
+
"model.layers.1.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 477 |
+
"model.layers.1.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 478 |
+
"model.layers.1.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 479 |
+
"model.layers.1.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 480 |
+
"model.layers.2.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 481 |
+
"model.layers.2.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 482 |
+
"model.layers.2.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 483 |
+
"model.layers.2.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 484 |
+
"model.layers.2.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 485 |
+
"model.layers.2.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 486 |
+
"model.layers.2.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 487 |
+
"model.layers.2.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 488 |
+
"model.layers.2.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 489 |
+
"model.layers.3.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 490 |
+
"model.layers.3.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 491 |
+
"model.layers.3.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 492 |
+
"model.layers.3.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 493 |
+
"model.layers.3.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 494 |
+
"model.layers.3.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 495 |
+
"model.layers.3.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 496 |
+
"model.layers.3.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 497 |
+
"model.layers.3.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 498 |
+
"model.layers.4.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 499 |
+
"model.layers.4.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 500 |
+
"model.layers.4.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 501 |
+
"model.layers.4.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 502 |
+
"model.layers.4.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 503 |
+
"model.layers.4.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 504 |
+
"model.layers.4.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 505 |
+
"model.layers.4.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 506 |
+
"model.layers.4.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 507 |
+
"model.layers.5.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 508 |
+
"model.layers.5.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 509 |
+
"model.layers.5.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 510 |
+
"model.layers.5.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 511 |
+
"model.layers.5.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 512 |
+
"model.layers.5.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 513 |
+
"model.layers.5.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 514 |
+
"model.layers.5.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 515 |
+
"model.layers.5.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 516 |
+
"model.layers.6.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 517 |
+
"model.layers.6.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 518 |
+
"model.layers.6.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 519 |
+
"model.layers.6.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 520 |
+
"model.layers.6.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 521 |
+
"model.layers.6.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 522 |
+
"model.layers.6.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 523 |
+
"model.layers.6.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 524 |
+
"model.layers.6.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 525 |
+
"model.layers.7.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 526 |
+
"model.layers.7.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 527 |
+
"model.layers.7.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 528 |
+
"model.layers.7.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 529 |
+
"model.layers.7.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 530 |
+
"model.layers.7.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 531 |
+
"model.layers.7.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 532 |
+
"model.layers.7.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 533 |
+
"model.layers.7.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 534 |
+
"model.layers.8.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 535 |
+
"model.layers.8.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 536 |
+
"model.layers.8.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 537 |
+
"model.layers.8.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 538 |
+
"model.layers.8.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 539 |
+
"model.layers.8.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 540 |
+
"model.layers.8.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 541 |
+
"model.layers.8.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 542 |
+
"model.layers.8.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 543 |
+
"model.layers.9.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 544 |
+
"model.layers.9.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 545 |
+
"model.layers.9.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 546 |
+
"model.layers.9.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 547 |
+
"model.layers.9.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 548 |
+
"model.layers.9.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 549 |
+
"model.layers.9.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 550 |
+
"model.layers.9.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 551 |
+
"model.layers.9.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 552 |
+
"model.layers.10.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 553 |
+
"model.layers.10.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 554 |
+
"model.layers.10.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 555 |
+
"model.layers.10.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 556 |
+
"model.layers.10.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 557 |
+
"model.layers.10.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 558 |
+
"model.layers.10.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 559 |
+
"model.layers.10.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 560 |
+
"model.layers.10.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 561 |
+
"model.layers.11.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 562 |
+
"model.layers.11.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 563 |
+
"model.layers.11.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 564 |
+
"model.layers.11.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 565 |
+
"model.layers.11.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 566 |
+
"model.layers.11.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 567 |
+
"model.layers.11.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 568 |
+
"model.layers.11.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 569 |
+
"model.layers.11.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 570 |
+
"model.layers.12.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 571 |
+
"model.layers.12.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 572 |
+
"model.layers.12.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 573 |
+
"model.layers.12.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 574 |
+
"model.layers.12.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 575 |
+
"model.layers.12.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 576 |
+
"model.layers.12.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 577 |
+
"model.layers.12.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 578 |
+
"model.layers.12.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 579 |
+
"model.layers.13.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 580 |
+
"model.layers.13.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 581 |
+
"model.layers.13.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 582 |
+
"model.layers.13.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 583 |
+
"model.layers.13.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 584 |
+
"model.layers.13.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 585 |
+
"model.layers.13.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 586 |
+
"model.layers.13.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 587 |
+
"model.layers.13.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 588 |
+
"model.layers.14.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 589 |
+
"model.layers.14.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 590 |
+
"model.layers.14.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 591 |
+
"model.layers.14.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 592 |
+
"model.layers.14.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 593 |
+
"model.layers.14.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 594 |
+
"model.layers.14.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 595 |
+
"model.layers.14.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 596 |
+
"model.layers.14.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 597 |
+
"model.layers.15.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 598 |
+
"model.layers.15.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 599 |
+
"model.layers.15.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 600 |
+
"model.layers.15.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 601 |
+
"model.layers.15.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 602 |
+
"model.layers.15.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 603 |
+
"model.layers.15.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 604 |
+
"model.layers.15.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 605 |
+
"model.layers.15.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 606 |
+
"model.layers.16.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 607 |
+
"model.layers.16.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 608 |
+
"model.layers.16.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 609 |
+
"model.layers.16.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 610 |
+
"model.layers.16.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 611 |
+
"model.layers.16.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 612 |
+
"model.layers.16.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 613 |
+
"model.layers.16.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 614 |
+
"model.layers.16.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 615 |
+
"model.layers.17.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
|
| 616 |
+
"model.layers.17.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
|
| 617 |
+
"model.layers.17.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
|
| 618 |
+
"model.layers.17.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
|
| 619 |
+
"model.layers.17.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
|
| 620 |
+
"model.layers.17.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
|
| 621 |
+
"model.layers.17.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
|
| 622 |
+
"model.layers.17.input_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 623 |
+
"model.layers.17.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
|
| 624 |
+
"model.norm.weight": "model-00001-of-00001.safetensors",
|
| 625 |
+
"lm_head.weight": "model-00001-of-00001.safetensors"
|
| 626 |
+
}
|
| 627 |
+
}
|
checkpoint-420/optimizer-00001-of-00001.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1c47f3e32e7bf9d312ec8cee5d22be648a75a5ab0172892f752f0761f24d2d93
|
| 3 |
+
size 7245150152
|
checkpoint-420/optimizer.safetensors.index.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
checkpoint-420/preprocessor_config.json
ADDED
|
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"do_convert_rgb": true,
|
| 3 |
+
"do_normalize": true,
|
| 4 |
+
"do_rescale": true,
|
| 5 |
+
"do_resize": true,
|
| 6 |
+
"image_mean": [
|
| 7 |
+
0.5,
|
| 8 |
+
0.5,
|
| 9 |
+
0.5
|
| 10 |
+
],
|
| 11 |
+
"image_processor_type": "SiglipImageProcessor",
|
| 12 |
+
"image_std": [
|
| 13 |
+
0.5,
|
| 14 |
+
0.5,
|
| 15 |
+
0.5
|
| 16 |
+
],
|
| 17 |
+
"max_pixels": 2822400,
|
| 18 |
+
"merge_size": 2,
|
| 19 |
+
"min_pixels": 147384,
|
| 20 |
+
"patch_size": 14,
|
| 21 |
+
"resample": 3,
|
| 22 |
+
"rescale_factor": 0.00392156862745098,
|
| 23 |
+
"size": {
|
| 24 |
+
"max_pixels": 2822400,
|
| 25 |
+
"min_pixels": 147384
|
| 26 |
+
},
|
| 27 |
+
"temporal_conv_size": 1,
|
| 28 |
+
"temporal_patch_size": 1
|
| 29 |
+
}
|
checkpoint-420/rng_state_0.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:29cde6478b49d2682dfa8c033ed7da4f33ab76ccd1941cfe950453d3923dfec3
|
| 3 |
+
size 32008
|
checkpoint-420/scheduler.pdparams
ADDED
|
Binary file (51 Bytes). View file
|
|
|