Deepesh-001 commited on
Commit
0a1683d
·
verified ·
1 Parent(s): e6190ec

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. added_tokens.json +1 -0
  2. all_results.json +7 -0
  3. checkpoint-336/added_tokens.json +1 -0
  4. checkpoint-336/config.json +223 -0
  5. checkpoint-336/generation_config.json +6 -0
  6. checkpoint-336/master_weights-00001-of-00001.safetensors +3 -0
  7. checkpoint-336/master_weights.safetensors.index.json +615 -0
  8. checkpoint-336/model-00001-of-00001.safetensors +3 -0
  9. checkpoint-336/model.safetensors.index.json +627 -0
  10. checkpoint-336/optimizer-00001-of-00001.safetensors +3 -0
  11. checkpoint-336/optimizer.safetensors.index.json +0 -0
  12. checkpoint-336/preprocessor_config.json +29 -0
  13. checkpoint-336/rng_state_0.pth +3 -0
  14. checkpoint-336/scheduler.pdparams +0 -0
  15. checkpoint-336/special_tokens_map.json +1 -0
  16. checkpoint-336/static_name_to_dyg_name.json +1 -0
  17. checkpoint-336/tokenizer.model +3 -0
  18. checkpoint-336/tokenizer_config.json +0 -0
  19. checkpoint-336/trainer_state.json +15 -0
  20. checkpoint-336/training_args.bin +3 -0
  21. checkpoint-378/added_tokens.json +1 -0
  22. checkpoint-378/config.json +223 -0
  23. checkpoint-378/generation_config.json +6 -0
  24. checkpoint-378/master_weights-00001-of-00001.safetensors +3 -0
  25. checkpoint-378/master_weights.safetensors.index.json +615 -0
  26. checkpoint-378/model-00001-of-00001.safetensors +3 -0
  27. checkpoint-378/model.safetensors.index.json +627 -0
  28. checkpoint-378/optimizer-00001-of-00001.safetensors +3 -0
  29. checkpoint-378/optimizer.safetensors.index.json +0 -0
  30. checkpoint-378/preprocessor_config.json +29 -0
  31. checkpoint-378/rng_state_0.pth +3 -0
  32. checkpoint-378/scheduler.pdparams +0 -0
  33. checkpoint-378/special_tokens_map.json +1 -0
  34. checkpoint-378/static_name_to_dyg_name.json +1 -0
  35. checkpoint-378/tokenizer.model +3 -0
  36. checkpoint-378/tokenizer_config.json +0 -0
  37. checkpoint-378/trainer_state.json +15 -0
  38. checkpoint-378/training_args.bin +3 -0
  39. checkpoint-420/added_tokens.json +1 -0
  40. checkpoint-420/config.json +223 -0
  41. checkpoint-420/generation_config.json +6 -0
  42. checkpoint-420/master_weights-00001-of-00001.safetensors +3 -0
  43. checkpoint-420/master_weights.safetensors.index.json +615 -0
  44. checkpoint-420/model-00001-of-00001.safetensors +3 -0
  45. checkpoint-420/model.safetensors.index.json +627 -0
  46. checkpoint-420/optimizer-00001-of-00001.safetensors +3 -0
  47. checkpoint-420/optimizer.safetensors.index.json +0 -0
  48. checkpoint-420/preprocessor_config.json +29 -0
  49. checkpoint-420/rng_state_0.pth +3 -0
  50. checkpoint-420/scheduler.pdparams +0 -0
added_tokens.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"<unk>": 0, "<s>": 1, "</s>": 2, "0": 3, "1": 4, "2": 5, "3": 6, "4": 7, "5": 8, "6": 9, "7": 10, "8": 11, "9": 12, "<|end_of_sentence|>": 100272, "<|begin_of_sentence|>": 100273, "<mask:1>": 100274, "<mask:4>": 100277, "<mask:5>": 100278, "<mask:6>": 100279, "<mask:7>": 100280, "<|IMAGE_PLACEHOLDER|>": 100295, "<|AUDIO_PLACEHOLDER|>": 100296, "<|LOC_0|>": 100297, "<|LOC_1|>": 100298, "<|LOC_2|>": 100299, "<|LOC_3|>": 100300, "<|LOC_4|>": 100301, "<|LOC_5|>": 100302, "<|LOC_6|>": 100303, "<|LOC_7|>": 100304, "<|LOC_8|>": 100305, "<|LOC_9|>": 100306, "<|LOC_10|>": 100307, "<|LOC_11|>": 100308, "<|LOC_12|>": 100309, "<|LOC_13|>": 100310, "<|LOC_14|>": 100311, "<|LOC_15|>": 100312, "<|LOC_16|>": 100313, "<|LOC_17|>": 100314, "<|LOC_18|>": 100315, "<|LOC_19|>": 100316, "<|LOC_20|>": 100317, "<|LOC_21|>": 100318, "<|LOC_22|>": 100319, "<|LOC_23|>": 100320, "<|LOC_24|>": 100321, "<|LOC_25|>": 100322, "<|LOC_26|>": 100323, "<|LOC_27|>": 100324, "<|LOC_28|>": 100325, "<|LOC_29|>": 100326, "<|LOC_30|>": 100327, "<|LOC_31|>": 100328, "<|LOC_32|>": 100329, "<|LOC_33|>": 100330, "<|LOC_34|>": 100331, "<|LOC_35|>": 100332, "<|LOC_36|>": 100333, "<|LOC_37|>": 100334, "<|LOC_38|>": 100335, "<|LOC_39|>": 100336, "<|LOC_40|>": 100337, "<|LOC_41|>": 100338, "<|LOC_42|>": 100339, "<|LOC_43|>": 100340, "<|LOC_44|>": 100341, "<|LOC_45|>": 100342, "<|LOC_46|>": 100343, "<|LOC_47|>": 100344, "<|LOC_48|>": 100345, "<|LOC_49|>": 100346, "<|LOC_50|>": 100347, "<|LOC_51|>": 100348, "<|LOC_52|>": 100349, "<|LOC_53|>": 100350, "<|LOC_54|>": 100351, "<|LOC_55|>": 100352, "<|LOC_56|>": 100353, "<|LOC_57|>": 100354, "<|LOC_58|>": 100355, "<|LOC_59|>": 100356, "<|LOC_60|>": 100357, "<|LOC_61|>": 100358, "<|LOC_62|>": 100359, "<|LOC_63|>": 100360, "<|LOC_64|>": 100361, "<|LOC_65|>": 100362, "<|LOC_66|>": 100363, "<|LOC_67|>": 100364, "<|LOC_68|>": 100365, "<|LOC_69|>": 100366, "<|LOC_70|>": 100367, "<|LOC_71|>": 100368, "<|LOC_72|>": 100369, "<|LOC_73|>": 100370, "<|LOC_74|>": 100371, "<|LOC_75|>": 100372, "<|LOC_76|>": 100373, "<|LOC_77|>": 100374, "<|LOC_78|>": 100375, "<|LOC_79|>": 100376, "<|LOC_80|>": 100377, "<|LOC_81|>": 100378, "<|LOC_82|>": 100379, "<|LOC_83|>": 100380, "<|LOC_84|>": 100381, "<|LOC_85|>": 100382, "<|LOC_86|>": 100383, "<|LOC_87|>": 100384, "<|LOC_88|>": 100385, "<|LOC_89|>": 100386, "<|LOC_90|>": 100387, "<|LOC_91|>": 100388, "<|LOC_92|>": 100389, "<|LOC_93|>": 100390, "<|LOC_94|>": 100391, "<|LOC_95|>": 100392, "<|LOC_96|>": 100393, "<|LOC_97|>": 100394, "<|LOC_98|>": 100395, "<|LOC_99|>": 100396, "<|LOC_100|>": 100397, "<|LOC_101|>": 100398, "<|LOC_102|>": 100399, "<|LOC_103|>": 100400, "<|LOC_104|>": 100401, "<|LOC_105|>": 100402, "<|LOC_106|>": 100403, "<|LOC_107|>": 100404, "<|LOC_108|>": 100405, "<|LOC_109|>": 100406, "<|LOC_110|>": 100407, "<|LOC_111|>": 100408, "<|LOC_112|>": 100409, "<|LOC_113|>": 100410, "<|LOC_114|>": 100411, "<|LOC_115|>": 100412, "<|LOC_116|>": 100413, "<|LOC_117|>": 100414, "<|LOC_118|>": 100415, "<|LOC_119|>": 100416, "<|LOC_120|>": 100417, "<|LOC_121|>": 100418, "<|LOC_122|>": 100419, "<|LOC_123|>": 100420, "<|LOC_124|>": 100421, "<|LOC_125|>": 100422, "<|LOC_126|>": 100423, "<|LOC_127|>": 100424, "<|LOC_128|>": 100425, "<|LOC_129|>": 100426, "<|LOC_130|>": 100427, "<|LOC_131|>": 100428, "<|LOC_132|>": 100429, "<|LOC_133|>": 100430, "<|LOC_134|>": 100431, "<|LOC_135|>": 100432, "<|LOC_136|>": 100433, "<|LOC_137|>": 100434, "<|LOC_138|>": 100435, "<|LOC_139|>": 100436, "<|LOC_140|>": 100437, "<|LOC_141|>": 100438, "<|LOC_142|>": 100439, "<|LOC_143|>": 100440, "<|LOC_144|>": 100441, "<|LOC_145|>": 100442, "<|LOC_146|>": 100443, "<|LOC_147|>": 100444, "<|LOC_148|>": 100445, "<|LOC_149|>": 100446, "<|LOC_150|>": 100447, "<|LOC_151|>": 100448, "<|LOC_152|>": 100449, "<|LOC_153|>": 100450, "<|LOC_154|>": 100451, "<|LOC_155|>": 100452, "<|LOC_156|>": 100453, "<|LOC_157|>": 100454, "<|LOC_158|>": 100455, "<|LOC_159|>": 100456, "<|LOC_160|>": 100457, "<|LOC_161|>": 100458, "<|LOC_162|>": 100459, "<|LOC_163|>": 100460, "<|LOC_164|>": 100461, "<|LOC_165|>": 100462, "<|LOC_166|>": 100463, "<|LOC_167|>": 100464, "<|LOC_168|>": 100465, "<|LOC_169|>": 100466, "<|LOC_170|>": 100467, "<|LOC_171|>": 100468, "<|LOC_172|>": 100469, "<|LOC_173|>": 100470, "<|LOC_174|>": 100471, "<|LOC_175|>": 100472, "<|LOC_176|>": 100473, "<|LOC_177|>": 100474, "<|LOC_178|>": 100475, "<|LOC_179|>": 100476, "<|LOC_180|>": 100477, "<|LOC_181|>": 100478, "<|LOC_182|>": 100479, "<|LOC_183|>": 100480, "<|LOC_184|>": 100481, "<|LOC_185|>": 100482, "<|LOC_186|>": 100483, "<|LOC_187|>": 100484, "<|LOC_188|>": 100485, "<|LOC_189|>": 100486, "<|LOC_190|>": 100487, "<|LOC_191|>": 100488, "<|LOC_192|>": 100489, "<|LOC_193|>": 100490, "<|LOC_194|>": 100491, "<|LOC_195|>": 100492, "<|LOC_196|>": 100493, "<|LOC_197|>": 100494, "<|LOC_198|>": 100495, "<|LOC_199|>": 100496, "<|LOC_200|>": 100497, "<|LOC_201|>": 100498, "<|LOC_202|>": 100499, "<|LOC_203|>": 100500, "<|LOC_204|>": 100501, "<|LOC_205|>": 100502, "<|LOC_206|>": 100503, "<|LOC_207|>": 100504, "<|LOC_208|>": 100505, "<|LOC_209|>": 100506, "<|LOC_210|>": 100507, "<|LOC_211|>": 100508, "<|LOC_212|>": 100509, "<|LOC_213|>": 100510, "<|LOC_214|>": 100511, "<|LOC_215|>": 100512, "<|LOC_216|>": 100513, "<|LOC_217|>": 100514, "<|LOC_218|>": 100515, "<|LOC_219|>": 100516, "<|LOC_220|>": 100517, "<|LOC_221|>": 100518, "<|LOC_222|>": 100519, "<|LOC_223|>": 100520, "<|LOC_224|>": 100521, "<|LOC_225|>": 100522, "<|LOC_226|>": 100523, "<|LOC_227|>": 100524, "<|LOC_228|>": 100525, "<|LOC_229|>": 100526, "<|LOC_230|>": 100527, "<|LOC_231|>": 100528, "<|LOC_232|>": 100529, "<|LOC_233|>": 100530, "<|LOC_234|>": 100531, "<|LOC_235|>": 100532, "<|LOC_236|>": 100533, "<|LOC_237|>": 100534, "<|LOC_238|>": 100535, "<|LOC_239|>": 100536, "<|LOC_240|>": 100537, "<|LOC_241|>": 100538, "<|LOC_242|>": 100539, "<|LOC_243|>": 100540, "<|LOC_244|>": 100541, "<|LOC_245|>": 100542, "<|LOC_246|>": 100543, "<|LOC_247|>": 100544, "<|LOC_248|>": 100545, "<|LOC_249|>": 100546, "<|LOC_250|>": 100547, "<|LOC_251|>": 100548, "<|LOC_252|>": 100549, "<|LOC_253|>": 100550, "<|LOC_254|>": 100551, "<|LOC_255|>": 100552, "<|LOC_256|>": 100553, "<|LOC_257|>": 100554, "<|LOC_258|>": 100555, "<|LOC_259|>": 100556, "<|LOC_260|>": 100557, "<|LOC_261|>": 100558, "<|LOC_262|>": 100559, "<|LOC_263|>": 100560, "<|LOC_264|>": 100561, "<|LOC_265|>": 100562, "<|LOC_266|>": 100563, "<|LOC_267|>": 100564, "<|LOC_268|>": 100565, "<|LOC_269|>": 100566, "<|LOC_270|>": 100567, "<|LOC_271|>": 100568, "<|LOC_272|>": 100569, "<|LOC_273|>": 100570, "<|LOC_274|>": 100571, "<|LOC_275|>": 100572, "<|LOC_276|>": 100573, "<|LOC_277|>": 100574, "<|LOC_278|>": 100575, "<|LOC_279|>": 100576, "<|LOC_280|>": 100577, "<|LOC_281|>": 100578, "<|LOC_282|>": 100579, "<|LOC_283|>": 100580, "<|LOC_284|>": 100581, "<|LOC_285|>": 100582, "<|LOC_286|>": 100583, "<|LOC_287|>": 100584, "<|LOC_288|>": 100585, "<|LOC_289|>": 100586, "<|LOC_290|>": 100587, "<|LOC_291|>": 100588, "<|LOC_292|>": 100589, "<|LOC_293|>": 100590, "<|LOC_294|>": 100591, "<|LOC_295|>": 100592, "<|LOC_296|>": 100593, "<|LOC_297|>": 100594, "<|LOC_298|>": 100595, "<|LOC_299|>": 100596, "<|LOC_300|>": 100597, "<|LOC_301|>": 100598, "<|LOC_302|>": 100599, "<|LOC_303|>": 100600, "<|LOC_304|>": 100601, "<|LOC_305|>": 100602, "<|LOC_306|>": 100603, "<|LOC_307|>": 100604, "<|LOC_308|>": 100605, "<|LOC_309|>": 100606, "<|LOC_310|>": 100607, "<|LOC_311|>": 100608, "<|LOC_312|>": 100609, "<|LOC_313|>": 100610, "<|LOC_314|>": 100611, "<|LOC_315|>": 100612, "<|LOC_316|>": 100613, "<|LOC_317|>": 100614, "<|LOC_318|>": 100615, "<|LOC_319|>": 100616, "<|LOC_320|>": 100617, "<|LOC_321|>": 100618, "<|LOC_322|>": 100619, "<|LOC_323|>": 100620, "<|LOC_324|>": 100621, "<|LOC_325|>": 100622, "<|LOC_326|>": 100623, "<|LOC_327|>": 100624, "<|LOC_328|>": 100625, "<|LOC_329|>": 100626, "<|LOC_330|>": 100627, "<|LOC_331|>": 100628, "<|LOC_332|>": 100629, "<|LOC_333|>": 100630, "<|LOC_334|>": 100631, "<|LOC_335|>": 100632, "<|LOC_336|>": 100633, "<|LOC_337|>": 100634, "<|LOC_338|>": 100635, "<|LOC_339|>": 100636, "<|LOC_340|>": 100637, "<|LOC_341|>": 100638, "<|LOC_342|>": 100639, "<|LOC_343|>": 100640, "<|LOC_344|>": 100641, "<|LOC_345|>": 100642, "<|LOC_346|>": 100643, "<|LOC_347|>": 100644, "<|LOC_348|>": 100645, "<|LOC_349|>": 100646, "<|LOC_350|>": 100647, "<|LOC_351|>": 100648, "<|LOC_352|>": 100649, "<|LOC_353|>": 100650, "<|LOC_354|>": 100651, "<|LOC_355|>": 100652, "<|LOC_356|>": 100653, "<|LOC_357|>": 100654, "<|LOC_358|>": 100655, "<|LOC_359|>": 100656, "<|LOC_360|>": 100657, "<|LOC_361|>": 100658, "<|LOC_362|>": 100659, "<|LOC_363|>": 100660, "<|LOC_364|>": 100661, "<|LOC_365|>": 100662, "<|LOC_366|>": 100663, "<|LOC_367|>": 100664, "<|LOC_368|>": 100665, "<|LOC_369|>": 100666, "<|LOC_370|>": 100667, "<|LOC_371|>": 100668, "<|LOC_372|>": 100669, "<|LOC_373|>": 100670, "<|LOC_374|>": 100671, "<|LOC_375|>": 100672, "<|LOC_376|>": 100673, "<|LOC_377|>": 100674, "<|LOC_378|>": 100675, "<|LOC_379|>": 100676, "<|LOC_380|>": 100677, "<|LOC_381|>": 100678, "<|LOC_382|>": 100679, "<|LOC_383|>": 100680, "<|LOC_384|>": 100681, "<|LOC_385|>": 100682, "<|LOC_386|>": 100683, "<|LOC_387|>": 100684, "<|LOC_388|>": 100685, "<|LOC_389|>": 100686, "<|LOC_390|>": 100687, "<|LOC_391|>": 100688, "<|LOC_392|>": 100689, "<|LOC_393|>": 100690, "<|LOC_394|>": 100691, "<|LOC_395|>": 100692, "<|LOC_396|>": 100693, "<|LOC_397|>": 100694, "<|LOC_398|>": 100695, "<|LOC_399|>": 100696, "<|LOC_400|>": 100697, "<|LOC_401|>": 100698, "<|LOC_402|>": 100699, "<|LOC_403|>": 100700, "<|LOC_404|>": 100701, "<|LOC_405|>": 100702, "<|LOC_406|>": 100703, "<|LOC_407|>": 100704, "<|LOC_408|>": 100705, "<|LOC_409|>": 100706, "<|LOC_410|>": 100707, "<|LOC_411|>": 100708, "<|LOC_412|>": 100709, "<|LOC_413|>": 100710, "<|LOC_414|>": 100711, "<|LOC_415|>": 100712, "<|LOC_416|>": 100713, "<|LOC_417|>": 100714, "<|LOC_418|>": 100715, "<|LOC_419|>": 100716, "<|LOC_420|>": 100717, "<|LOC_421|>": 100718, "<|LOC_422|>": 100719, "<|LOC_423|>": 100720, "<|LOC_424|>": 100721, "<|LOC_425|>": 100722, "<|LOC_426|>": 100723, "<|LOC_427|>": 100724, "<|LOC_428|>": 100725, "<|LOC_429|>": 100726, "<|LOC_430|>": 100727, "<|LOC_431|>": 100728, "<|LOC_432|>": 100729, "<|LOC_433|>": 100730, "<|LOC_434|>": 100731, "<|LOC_435|>": 100732, "<|LOC_436|>": 100733, "<|LOC_437|>": 100734, "<|LOC_438|>": 100735, "<|LOC_439|>": 100736, "<|LOC_440|>": 100737, "<|LOC_441|>": 100738, "<|LOC_442|>": 100739, "<|LOC_443|>": 100740, "<|LOC_444|>": 100741, "<|LOC_445|>": 100742, "<|LOC_446|>": 100743, "<|LOC_447|>": 100744, "<|LOC_448|>": 100745, "<|LOC_449|>": 100746, "<|LOC_450|>": 100747, "<|LOC_451|>": 100748, "<|LOC_452|>": 100749, "<|LOC_453|>": 100750, "<|LOC_454|>": 100751, "<|LOC_455|>": 100752, "<|LOC_456|>": 100753, "<|LOC_457|>": 100754, "<|LOC_458|>": 100755, "<|LOC_459|>": 100756, "<|LOC_460|>": 100757, "<|LOC_461|>": 100758, "<|LOC_462|>": 100759, "<|LOC_463|>": 100760, "<|LOC_464|>": 100761, "<|LOC_465|>": 100762, "<|LOC_466|>": 100763, "<|LOC_467|>": 100764, "<|LOC_468|>": 100765, "<|LOC_469|>": 100766, "<|LOC_470|>": 100767, "<|LOC_471|>": 100768, "<|LOC_472|>": 100769, "<|LOC_473|>": 100770, "<|LOC_474|>": 100771, "<|LOC_475|>": 100772, "<|LOC_476|>": 100773, "<|LOC_477|>": 100774, "<|LOC_478|>": 100775, "<|LOC_479|>": 100776, "<|LOC_480|>": 100777, "<|LOC_481|>": 100778, "<|LOC_482|>": 100779, "<|LOC_483|>": 100780, "<|LOC_484|>": 100781, "<|LOC_485|>": 100782, "<|LOC_486|>": 100783, "<|LOC_487|>": 100784, "<|LOC_488|>": 100785, "<|LOC_489|>": 100786, "<|LOC_490|>": 100787, "<|LOC_491|>": 100788, "<|LOC_492|>": 100789, "<|LOC_493|>": 100790, "<|LOC_494|>": 100791, "<|LOC_495|>": 100792, "<|LOC_496|>": 100793, "<|LOC_497|>": 100794, "<|LOC_498|>": 100795, "<|LOC_499|>": 100796, "<|LOC_500|>": 100797, "<|LOC_501|>": 100798, "<|LOC_502|>": 100799, "<|LOC_503|>": 100800, "<|LOC_504|>": 100801, "<|LOC_505|>": 100802, "<|LOC_506|>": 100803, "<|LOC_507|>": 100804, "<|LOC_508|>": 100805, "<|LOC_509|>": 100806, "<|LOC_510|>": 100807, "<|LOC_511|>": 100808, "<|LOC_512|>": 100809, "<|LOC_513|>": 100810, "<|LOC_514|>": 100811, "<|LOC_515|>": 100812, "<|LOC_516|>": 100813, "<|LOC_517|>": 100814, "<|LOC_518|>": 100815, "<|LOC_519|>": 100816, "<|LOC_520|>": 100817, "<|LOC_521|>": 100818, "<|LOC_522|>": 100819, "<|LOC_523|>": 100820, "<|LOC_524|>": 100821, "<|LOC_525|>": 100822, "<|LOC_526|>": 100823, "<|LOC_527|>": 100824, "<|LOC_528|>": 100825, "<|LOC_529|>": 100826, "<|LOC_530|>": 100827, "<|LOC_531|>": 100828, "<|LOC_532|>": 100829, "<|LOC_533|>": 100830, "<|LOC_534|>": 100831, "<|LOC_535|>": 100832, "<|LOC_536|>": 100833, "<|LOC_537|>": 100834, "<|LOC_538|>": 100835, "<|LOC_539|>": 100836, "<|LOC_540|>": 100837, "<|LOC_541|>": 100838, "<|LOC_542|>": 100839, "<|LOC_543|>": 100840, "<|LOC_544|>": 100841, "<|LOC_545|>": 100842, "<|LOC_546|>": 100843, "<|LOC_547|>": 100844, "<|LOC_548|>": 100845, "<|LOC_549|>": 100846, "<|LOC_550|>": 100847, "<|LOC_551|>": 100848, "<|LOC_552|>": 100849, "<|LOC_553|>": 100850, "<|LOC_554|>": 100851, "<|LOC_555|>": 100852, "<|LOC_556|>": 100853, "<|LOC_557|>": 100854, "<|LOC_558|>": 100855, "<|LOC_559|>": 100856, "<|LOC_560|>": 100857, "<|LOC_561|>": 100858, "<|LOC_562|>": 100859, "<|LOC_563|>": 100860, "<|LOC_564|>": 100861, "<|LOC_565|>": 100862, "<|LOC_566|>": 100863, "<|LOC_567|>": 100864, "<|LOC_568|>": 100865, "<|LOC_569|>": 100866, "<|LOC_570|>": 100867, "<|LOC_571|>": 100868, "<|LOC_572|>": 100869, "<|LOC_573|>": 100870, "<|LOC_574|>": 100871, "<|LOC_575|>": 100872, "<|LOC_576|>": 100873, "<|LOC_577|>": 100874, "<|LOC_578|>": 100875, "<|LOC_579|>": 100876, "<|LOC_580|>": 100877, "<|LOC_581|>": 100878, "<|LOC_582|>": 100879, "<|LOC_583|>": 100880, "<|LOC_584|>": 100881, "<|LOC_585|>": 100882, "<|LOC_586|>": 100883, "<|LOC_587|>": 100884, "<|LOC_588|>": 100885, "<|LOC_589|>": 100886, "<|LOC_590|>": 100887, "<|LOC_591|>": 100888, "<|LOC_592|>": 100889, "<|LOC_593|>": 100890, "<|LOC_594|>": 100891, "<|LOC_595|>": 100892, "<|LOC_596|>": 100893, "<|LOC_597|>": 100894, "<|LOC_598|>": 100895, "<|LOC_599|>": 100896, "<|LOC_600|>": 100897, "<|LOC_601|>": 100898, "<|LOC_602|>": 100899, "<|LOC_603|>": 100900, "<|LOC_604|>": 100901, "<|LOC_605|>": 100902, "<|LOC_606|>": 100903, "<|LOC_607|>": 100904, "<|LOC_608|>": 100905, "<|LOC_609|>": 100906, "<|LOC_610|>": 100907, "<|LOC_611|>": 100908, "<|LOC_612|>": 100909, "<|LOC_613|>": 100910, "<|LOC_614|>": 100911, "<|LOC_615|>": 100912, "<|LOC_616|>": 100913, "<|LOC_617|>": 100914, "<|LOC_618|>": 100915, "<|LOC_619|>": 100916, "<|LOC_620|>": 100917, "<|LOC_621|>": 100918, "<|LOC_622|>": 100919, "<|LOC_623|>": 100920, "<|LOC_624|>": 100921, "<|LOC_625|>": 100922, "<|LOC_626|>": 100923, "<|LOC_627|>": 100924, "<|LOC_628|>": 100925, "<|LOC_629|>": 100926, "<|LOC_630|>": 100927, "<|LOC_631|>": 100928, "<|LOC_632|>": 100929, "<|LOC_633|>": 100930, "<|LOC_634|>": 100931, "<|LOC_635|>": 100932, "<|LOC_636|>": 100933, "<|LOC_637|>": 100934, "<|LOC_638|>": 100935, "<|LOC_639|>": 100936, "<|LOC_640|>": 100937, "<|LOC_641|>": 100938, "<|LOC_642|>": 100939, "<|LOC_643|>": 100940, "<|LOC_644|>": 100941, "<|LOC_645|>": 100942, "<|LOC_646|>": 100943, "<|LOC_647|>": 100944, "<|LOC_648|>": 100945, "<|LOC_649|>": 100946, "<|LOC_650|>": 100947, "<|LOC_651|>": 100948, "<|LOC_652|>": 100949, "<|LOC_653|>": 100950, "<|LOC_654|>": 100951, "<|LOC_655|>": 100952, "<|LOC_656|>": 100953, "<|LOC_657|>": 100954, "<|LOC_658|>": 100955, "<|LOC_659|>": 100956, "<|LOC_660|>": 100957, "<|LOC_661|>": 100958, "<|LOC_662|>": 100959, "<|LOC_663|>": 100960, "<|LOC_664|>": 100961, "<|LOC_665|>": 100962, "<|LOC_666|>": 100963, "<|LOC_667|>": 100964, "<|LOC_668|>": 100965, "<|LOC_669|>": 100966, "<|LOC_670|>": 100967, "<|LOC_671|>": 100968, "<|LOC_672|>": 100969, "<|LOC_673|>": 100970, "<|LOC_674|>": 100971, "<|LOC_675|>": 100972, "<|LOC_676|>": 100973, "<|LOC_677|>": 100974, "<|LOC_678|>": 100975, "<|LOC_679|>": 100976, "<|LOC_680|>": 100977, "<|LOC_681|>": 100978, "<|LOC_682|>": 100979, "<|LOC_683|>": 100980, "<|LOC_684|>": 100981, "<|LOC_685|>": 100982, "<|LOC_686|>": 100983, "<|LOC_687|>": 100984, "<|LOC_688|>": 100985, "<|LOC_689|>": 100986, "<|LOC_690|>": 100987, "<|LOC_691|>": 100988, "<|LOC_692|>": 100989, "<|LOC_693|>": 100990, "<|LOC_694|>": 100991, "<|LOC_695|>": 100992, "<|LOC_696|>": 100993, "<|LOC_697|>": 100994, "<|LOC_698|>": 100995, "<|LOC_699|>": 100996, "<|LOC_700|>": 100997, "<|LOC_701|>": 100998, "<|LOC_702|>": 100999, "<|LOC_703|>": 101000, "<|LOC_704|>": 101001, "<|LOC_705|>": 101002, "<|LOC_706|>": 101003, "<|LOC_707|>": 101004, "<|LOC_708|>": 101005, "<|LOC_709|>": 101006, "<|LOC_710|>": 101007, "<|LOC_711|>": 101008, "<|LOC_712|>": 101009, "<|LOC_713|>": 101010, "<|LOC_714|>": 101011, "<|LOC_715|>": 101012, "<|LOC_716|>": 101013, "<|LOC_717|>": 101014, "<|LOC_718|>": 101015, "<|LOC_719|>": 101016, "<|LOC_720|>": 101017, "<|LOC_721|>": 101018, "<|LOC_722|>": 101019, "<|LOC_723|>": 101020, "<|LOC_724|>": 101021, "<|LOC_725|>": 101022, "<|LOC_726|>": 101023, "<|LOC_727|>": 101024, "<|LOC_728|>": 101025, "<|LOC_729|>": 101026, "<|LOC_730|>": 101027, "<|LOC_731|>": 101028, "<|LOC_732|>": 101029, "<|LOC_733|>": 101030, "<|LOC_734|>": 101031, "<|LOC_735|>": 101032, "<|LOC_736|>": 101033, "<|LOC_737|>": 101034, "<|LOC_738|>": 101035, "<|LOC_739|>": 101036, "<|LOC_740|>": 101037, "<|LOC_741|>": 101038, "<|LOC_742|>": 101039, "<|LOC_743|>": 101040, "<|LOC_744|>": 101041, "<|LOC_745|>": 101042, "<|LOC_746|>": 101043, "<|LOC_747|>": 101044, "<|LOC_748|>": 101045, "<|LOC_749|>": 101046, "<|LOC_750|>": 101047, "<|LOC_751|>": 101048, "<|LOC_752|>": 101049, "<|LOC_753|>": 101050, "<|LOC_754|>": 101051, "<|LOC_755|>": 101052, "<|LOC_756|>": 101053, "<|LOC_757|>": 101054, "<|LOC_758|>": 101055, "<|LOC_759|>": 101056, "<|LOC_760|>": 101057, "<|LOC_761|>": 101058, "<|LOC_762|>": 101059, "<|LOC_763|>": 101060, "<|LOC_764|>": 101061, "<|LOC_765|>": 101062, "<|LOC_766|>": 101063, "<|LOC_767|>": 101064, "<|LOC_768|>": 101065, "<|LOC_769|>": 101066, "<|LOC_770|>": 101067, "<|LOC_771|>": 101068, "<|LOC_772|>": 101069, "<|LOC_773|>": 101070, "<|LOC_774|>": 101071, "<|LOC_775|>": 101072, "<|LOC_776|>": 101073, "<|LOC_777|>": 101074, "<|LOC_778|>": 101075, "<|LOC_779|>": 101076, "<|LOC_780|>": 101077, "<|LOC_781|>": 101078, "<|LOC_782|>": 101079, "<|LOC_783|>": 101080, "<|LOC_784|>": 101081, "<|LOC_785|>": 101082, "<|LOC_786|>": 101083, "<|LOC_787|>": 101084, "<|LOC_788|>": 101085, "<|LOC_789|>": 101086, "<|LOC_790|>": 101087, "<|LOC_791|>": 101088, "<|LOC_792|>": 101089, "<|LOC_793|>": 101090, "<|LOC_794|>": 101091, "<|LOC_795|>": 101092, "<|LOC_796|>": 101093, "<|LOC_797|>": 101094, "<|LOC_798|>": 101095, "<|LOC_799|>": 101096, "<|LOC_800|>": 101097, "<|LOC_801|>": 101098, "<|LOC_802|>": 101099, "<|LOC_803|>": 101100, "<|LOC_804|>": 101101, "<|LOC_805|>": 101102, "<|LOC_806|>": 101103, "<|LOC_807|>": 101104, "<|LOC_808|>": 101105, "<|LOC_809|>": 101106, "<|LOC_810|>": 101107, "<|LOC_811|>": 101108, "<|LOC_812|>": 101109, "<|LOC_813|>": 101110, "<|LOC_814|>": 101111, "<|LOC_815|>": 101112, "<|LOC_816|>": 101113, "<|LOC_817|>": 101114, "<|LOC_818|>": 101115, "<|LOC_819|>": 101116, "<|LOC_820|>": 101117, "<|LOC_821|>": 101118, "<|LOC_822|>": 101119, "<|LOC_823|>": 101120, "<|LOC_824|>": 101121, "<|LOC_825|>": 101122, "<|LOC_826|>": 101123, "<|LOC_827|>": 101124, "<|LOC_828|>": 101125, "<|LOC_829|>": 101126, "<|LOC_830|>": 101127, "<|LOC_831|>": 101128, "<|LOC_832|>": 101129, "<|LOC_833|>": 101130, "<|LOC_834|>": 101131, "<|LOC_835|>": 101132, "<|LOC_836|>": 101133, "<|LOC_837|>": 101134, "<|LOC_838|>": 101135, "<|LOC_839|>": 101136, "<|LOC_840|>": 101137, "<|LOC_841|>": 101138, "<|LOC_842|>": 101139, "<|LOC_843|>": 101140, "<|LOC_844|>": 101141, "<|LOC_845|>": 101142, "<|LOC_846|>": 101143, "<|LOC_847|>": 101144, "<|LOC_848|>": 101145, "<|LOC_849|>": 101146, "<|LOC_850|>": 101147, "<|LOC_851|>": 101148, "<|LOC_852|>": 101149, "<|LOC_853|>": 101150, "<|LOC_854|>": 101151, "<|LOC_855|>": 101152, "<|LOC_856|>": 101153, "<|LOC_857|>": 101154, "<|LOC_858|>": 101155, "<|LOC_859|>": 101156, "<|LOC_860|>": 101157, "<|LOC_861|>": 101158, "<|LOC_862|>": 101159, "<|LOC_863|>": 101160, "<|LOC_864|>": 101161, "<|LOC_865|>": 101162, "<|LOC_866|>": 101163, "<|LOC_867|>": 101164, "<|LOC_868|>": 101165, "<|LOC_869|>": 101166, "<|LOC_870|>": 101167, "<|LOC_871|>": 101168, "<|LOC_872|>": 101169, "<|LOC_873|>": 101170, "<|LOC_874|>": 101171, "<|LOC_875|>": 101172, "<|LOC_876|>": 101173, "<|LOC_877|>": 101174, "<|LOC_878|>": 101175, "<|LOC_879|>": 101176, "<|LOC_880|>": 101177, "<|LOC_881|>": 101178, "<|LOC_882|>": 101179, "<|LOC_883|>": 101180, "<|LOC_884|>": 101181, "<|LOC_885|>": 101182, "<|LOC_886|>": 101183, "<|LOC_887|>": 101184, "<|LOC_888|>": 101185, "<|LOC_889|>": 101186, "<|LOC_890|>": 101187, "<|LOC_891|>": 101188, "<|LOC_892|>": 101189, "<|LOC_893|>": 101190, "<|LOC_894|>": 101191, "<|LOC_895|>": 101192, "<|LOC_896|>": 101193, "<|LOC_897|>": 101194, "<|LOC_898|>": 101195, "<|LOC_899|>": 101196, "<|LOC_900|>": 101197, "<|LOC_901|>": 101198, "<|LOC_902|>": 101199, "<|LOC_903|>": 101200, "<|LOC_904|>": 101201, "<|LOC_905|>": 101202, "<|LOC_906|>": 101203, "<|LOC_907|>": 101204, "<|LOC_908|>": 101205, "<|LOC_909|>": 101206, "<|LOC_910|>": 101207, "<|LOC_911|>": 101208, "<|LOC_912|>": 101209, "<|LOC_913|>": 101210, "<|LOC_914|>": 101211, "<|LOC_915|>": 101212, "<|LOC_916|>": 101213, "<|LOC_917|>": 101214, "<|LOC_918|>": 101215, "<|LOC_919|>": 101216, "<|LOC_920|>": 101217, "<|LOC_921|>": 101218, "<|LOC_922|>": 101219, "<|LOC_923|>": 101220, "<|LOC_924|>": 101221, "<|LOC_925|>": 101222, "<|LOC_926|>": 101223, "<|LOC_927|>": 101224, "<|LOC_928|>": 101225, "<|LOC_929|>": 101226, "<|LOC_930|>": 101227, "<|LOC_931|>": 101228, "<|LOC_932|>": 101229, "<|LOC_933|>": 101230, "<|LOC_934|>": 101231, "<|LOC_935|>": 101232, "<|LOC_936|>": 101233, "<|LOC_937|>": 101234, "<|LOC_938|>": 101235, "<|LOC_939|>": 101236, "<|LOC_940|>": 101237, "<|LOC_941|>": 101238, "<|LOC_942|>": 101239, "<|LOC_943|>": 101240, "<|LOC_944|>": 101241, "<|LOC_945|>": 101242, "<|LOC_946|>": 101243, "<|LOC_947|>": 101244, "<|LOC_948|>": 101245, "<|LOC_949|>": 101246, "<|LOC_950|>": 101247, "<|LOC_951|>": 101248, "<|LOC_952|>": 101249, "<|LOC_953|>": 101250, "<|LOC_954|>": 101251, "<|LOC_955|>": 101252, "<|LOC_956|>": 101253, "<|LOC_957|>": 101254, "<|LOC_958|>": 101255, "<|LOC_959|>": 101256, "<|LOC_960|>": 101257, "<|LOC_961|>": 101258, "<|LOC_962|>": 101259, "<|LOC_963|>": 101260, "<|LOC_964|>": 101261, "<|LOC_965|>": 101262, "<|LOC_966|>": 101263, "<|LOC_967|>": 101264, "<|LOC_968|>": 101265, "<|LOC_969|>": 101266, "<|LOC_970|>": 101267, "<|LOC_971|>": 101268, "<|LOC_972|>": 101269, "<|LOC_973|>": 101270, "<|LOC_974|>": 101271, "<|LOC_975|>": 101272, "<|LOC_976|>": 101273, "<|LOC_977|>": 101274, "<|LOC_978|>": 101275, "<|LOC_979|>": 101276, "<|LOC_980|>": 101277, "<|LOC_981|>": 101278, "<|LOC_982|>": 101279, "<|LOC_983|>": 101280, "<|LOC_984|>": 101281, "<|LOC_985|>": 101282, "<|LOC_986|>": 101283, "<|LOC_987|>": 101284, "<|LOC_988|>": 101285, "<|LOC_989|>": 101286, "<|LOC_990|>": 101287, "<|LOC_991|>": 101288, "<|LOC_992|>": 101289, "<|LOC_993|>": 101290, "<|LOC_994|>": 101291, "<|LOC_995|>": 101292, "<|LOC_996|>": 101293, "<|LOC_997|>": 101294, "<|LOC_998|>": 101295, "<|LOC_999|>": 101296, "<|LOC_1000|>": 101297, "<|LOC_BEGIN|>": 101298, "<|LOC_END|>": 101299, "<|LOC_SEP|>": 101300, "<|CROP_COL_SEP|>": 101301, "<|CROP_ROW_SEP|>": 101302, "<|IMAGE_SEP|>": 101303, "<|image_pad|>": 101304, "<|IMAGE_START|>": 101305, "<|IMAGE_END|>": 101306, "<|video_pad|>": 101307, "<ecel>": 101308, "<fcel>": 101309, "<xcel>": 101310, "<lcel>": 101311, "<ucel>": 101312, "<nl>": 101313}
all_results.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "progress_or_epoch": 1.0,
3
+ "train_loss": 0.6455322598193434,
4
+ "train_runtime": 2255.4152,
5
+ "train_samples_per_second": 1.4933,
6
+ "train_steps_per_second": 0.1867
7
+ }
checkpoint-336/added_tokens.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"<unk>": 0, "<s>": 1, "</s>": 2, "0": 3, "1": 4, "2": 5, "3": 6, "4": 7, "5": 8, "6": 9, "7": 10, "8": 11, "9": 12, "<|end_of_sentence|>": 100272, "<|begin_of_sentence|>": 100273, "<mask:1>": 100274, "<mask:4>": 100277, "<mask:5>": 100278, "<mask:6>": 100279, "<mask:7>": 100280, "<|IMAGE_PLACEHOLDER|>": 100295, "<|AUDIO_PLACEHOLDER|>": 100296, "<|LOC_0|>": 100297, "<|LOC_1|>": 100298, "<|LOC_2|>": 100299, "<|LOC_3|>": 100300, "<|LOC_4|>": 100301, "<|LOC_5|>": 100302, "<|LOC_6|>": 100303, "<|LOC_7|>": 100304, "<|LOC_8|>": 100305, "<|LOC_9|>": 100306, "<|LOC_10|>": 100307, "<|LOC_11|>": 100308, "<|LOC_12|>": 100309, "<|LOC_13|>": 100310, "<|LOC_14|>": 100311, "<|LOC_15|>": 100312, "<|LOC_16|>": 100313, "<|LOC_17|>": 100314, "<|LOC_18|>": 100315, "<|LOC_19|>": 100316, "<|LOC_20|>": 100317, "<|LOC_21|>": 100318, "<|LOC_22|>": 100319, "<|LOC_23|>": 100320, "<|LOC_24|>": 100321, "<|LOC_25|>": 100322, "<|LOC_26|>": 100323, "<|LOC_27|>": 100324, "<|LOC_28|>": 100325, "<|LOC_29|>": 100326, "<|LOC_30|>": 100327, "<|LOC_31|>": 100328, "<|LOC_32|>": 100329, "<|LOC_33|>": 100330, "<|LOC_34|>": 100331, "<|LOC_35|>": 100332, "<|LOC_36|>": 100333, "<|LOC_37|>": 100334, "<|LOC_38|>": 100335, "<|LOC_39|>": 100336, "<|LOC_40|>": 100337, "<|LOC_41|>": 100338, "<|LOC_42|>": 100339, "<|LOC_43|>": 100340, "<|LOC_44|>": 100341, "<|LOC_45|>": 100342, "<|LOC_46|>": 100343, "<|LOC_47|>": 100344, "<|LOC_48|>": 100345, "<|LOC_49|>": 100346, "<|LOC_50|>": 100347, "<|LOC_51|>": 100348, "<|LOC_52|>": 100349, "<|LOC_53|>": 100350, "<|LOC_54|>": 100351, "<|LOC_55|>": 100352, "<|LOC_56|>": 100353, "<|LOC_57|>": 100354, "<|LOC_58|>": 100355, "<|LOC_59|>": 100356, "<|LOC_60|>": 100357, "<|LOC_61|>": 100358, "<|LOC_62|>": 100359, "<|LOC_63|>": 100360, "<|LOC_64|>": 100361, "<|LOC_65|>": 100362, "<|LOC_66|>": 100363, "<|LOC_67|>": 100364, "<|LOC_68|>": 100365, "<|LOC_69|>": 100366, "<|LOC_70|>": 100367, "<|LOC_71|>": 100368, "<|LOC_72|>": 100369, "<|LOC_73|>": 100370, "<|LOC_74|>": 100371, "<|LOC_75|>": 100372, "<|LOC_76|>": 100373, "<|LOC_77|>": 100374, "<|LOC_78|>": 100375, "<|LOC_79|>": 100376, "<|LOC_80|>": 100377, "<|LOC_81|>": 100378, "<|LOC_82|>": 100379, "<|LOC_83|>": 100380, "<|LOC_84|>": 100381, "<|LOC_85|>": 100382, "<|LOC_86|>": 100383, "<|LOC_87|>": 100384, "<|LOC_88|>": 100385, "<|LOC_89|>": 100386, "<|LOC_90|>": 100387, "<|LOC_91|>": 100388, "<|LOC_92|>": 100389, "<|LOC_93|>": 100390, "<|LOC_94|>": 100391, "<|LOC_95|>": 100392, "<|LOC_96|>": 100393, "<|LOC_97|>": 100394, "<|LOC_98|>": 100395, "<|LOC_99|>": 100396, "<|LOC_100|>": 100397, "<|LOC_101|>": 100398, "<|LOC_102|>": 100399, "<|LOC_103|>": 100400, "<|LOC_104|>": 100401, "<|LOC_105|>": 100402, "<|LOC_106|>": 100403, "<|LOC_107|>": 100404, "<|LOC_108|>": 100405, "<|LOC_109|>": 100406, "<|LOC_110|>": 100407, "<|LOC_111|>": 100408, "<|LOC_112|>": 100409, "<|LOC_113|>": 100410, "<|LOC_114|>": 100411, "<|LOC_115|>": 100412, "<|LOC_116|>": 100413, "<|LOC_117|>": 100414, "<|LOC_118|>": 100415, "<|LOC_119|>": 100416, "<|LOC_120|>": 100417, "<|LOC_121|>": 100418, "<|LOC_122|>": 100419, "<|LOC_123|>": 100420, "<|LOC_124|>": 100421, "<|LOC_125|>": 100422, "<|LOC_126|>": 100423, "<|LOC_127|>": 100424, "<|LOC_128|>": 100425, "<|LOC_129|>": 100426, "<|LOC_130|>": 100427, "<|LOC_131|>": 100428, "<|LOC_132|>": 100429, "<|LOC_133|>": 100430, "<|LOC_134|>": 100431, "<|LOC_135|>": 100432, "<|LOC_136|>": 100433, "<|LOC_137|>": 100434, "<|LOC_138|>": 100435, "<|LOC_139|>": 100436, "<|LOC_140|>": 100437, "<|LOC_141|>": 100438, "<|LOC_142|>": 100439, "<|LOC_143|>": 100440, "<|LOC_144|>": 100441, "<|LOC_145|>": 100442, "<|LOC_146|>": 100443, "<|LOC_147|>": 100444, "<|LOC_148|>": 100445, "<|LOC_149|>": 100446, "<|LOC_150|>": 100447, "<|LOC_151|>": 100448, "<|LOC_152|>": 100449, "<|LOC_153|>": 100450, "<|LOC_154|>": 100451, "<|LOC_155|>": 100452, "<|LOC_156|>": 100453, "<|LOC_157|>": 100454, "<|LOC_158|>": 100455, "<|LOC_159|>": 100456, "<|LOC_160|>": 100457, "<|LOC_161|>": 100458, "<|LOC_162|>": 100459, "<|LOC_163|>": 100460, "<|LOC_164|>": 100461, "<|LOC_165|>": 100462, "<|LOC_166|>": 100463, "<|LOC_167|>": 100464, "<|LOC_168|>": 100465, "<|LOC_169|>": 100466, "<|LOC_170|>": 100467, "<|LOC_171|>": 100468, "<|LOC_172|>": 100469, "<|LOC_173|>": 100470, "<|LOC_174|>": 100471, "<|LOC_175|>": 100472, "<|LOC_176|>": 100473, "<|LOC_177|>": 100474, "<|LOC_178|>": 100475, "<|LOC_179|>": 100476, "<|LOC_180|>": 100477, "<|LOC_181|>": 100478, "<|LOC_182|>": 100479, "<|LOC_183|>": 100480, "<|LOC_184|>": 100481, "<|LOC_185|>": 100482, "<|LOC_186|>": 100483, "<|LOC_187|>": 100484, "<|LOC_188|>": 100485, "<|LOC_189|>": 100486, "<|LOC_190|>": 100487, "<|LOC_191|>": 100488, "<|LOC_192|>": 100489, "<|LOC_193|>": 100490, "<|LOC_194|>": 100491, "<|LOC_195|>": 100492, "<|LOC_196|>": 100493, "<|LOC_197|>": 100494, "<|LOC_198|>": 100495, "<|LOC_199|>": 100496, "<|LOC_200|>": 100497, "<|LOC_201|>": 100498, "<|LOC_202|>": 100499, "<|LOC_203|>": 100500, "<|LOC_204|>": 100501, "<|LOC_205|>": 100502, "<|LOC_206|>": 100503, "<|LOC_207|>": 100504, "<|LOC_208|>": 100505, "<|LOC_209|>": 100506, "<|LOC_210|>": 100507, "<|LOC_211|>": 100508, "<|LOC_212|>": 100509, "<|LOC_213|>": 100510, "<|LOC_214|>": 100511, "<|LOC_215|>": 100512, "<|LOC_216|>": 100513, "<|LOC_217|>": 100514, "<|LOC_218|>": 100515, "<|LOC_219|>": 100516, "<|LOC_220|>": 100517, "<|LOC_221|>": 100518, "<|LOC_222|>": 100519, "<|LOC_223|>": 100520, "<|LOC_224|>": 100521, "<|LOC_225|>": 100522, "<|LOC_226|>": 100523, "<|LOC_227|>": 100524, "<|LOC_228|>": 100525, "<|LOC_229|>": 100526, "<|LOC_230|>": 100527, "<|LOC_231|>": 100528, "<|LOC_232|>": 100529, "<|LOC_233|>": 100530, "<|LOC_234|>": 100531, "<|LOC_235|>": 100532, "<|LOC_236|>": 100533, "<|LOC_237|>": 100534, "<|LOC_238|>": 100535, "<|LOC_239|>": 100536, "<|LOC_240|>": 100537, "<|LOC_241|>": 100538, "<|LOC_242|>": 100539, "<|LOC_243|>": 100540, "<|LOC_244|>": 100541, "<|LOC_245|>": 100542, "<|LOC_246|>": 100543, "<|LOC_247|>": 100544, "<|LOC_248|>": 100545, "<|LOC_249|>": 100546, "<|LOC_250|>": 100547, "<|LOC_251|>": 100548, "<|LOC_252|>": 100549, "<|LOC_253|>": 100550, "<|LOC_254|>": 100551, "<|LOC_255|>": 100552, "<|LOC_256|>": 100553, "<|LOC_257|>": 100554, "<|LOC_258|>": 100555, "<|LOC_259|>": 100556, "<|LOC_260|>": 100557, "<|LOC_261|>": 100558, "<|LOC_262|>": 100559, "<|LOC_263|>": 100560, "<|LOC_264|>": 100561, "<|LOC_265|>": 100562, "<|LOC_266|>": 100563, "<|LOC_267|>": 100564, "<|LOC_268|>": 100565, "<|LOC_269|>": 100566, "<|LOC_270|>": 100567, "<|LOC_271|>": 100568, "<|LOC_272|>": 100569, "<|LOC_273|>": 100570, "<|LOC_274|>": 100571, "<|LOC_275|>": 100572, "<|LOC_276|>": 100573, "<|LOC_277|>": 100574, "<|LOC_278|>": 100575, "<|LOC_279|>": 100576, "<|LOC_280|>": 100577, "<|LOC_281|>": 100578, "<|LOC_282|>": 100579, "<|LOC_283|>": 100580, "<|LOC_284|>": 100581, "<|LOC_285|>": 100582, "<|LOC_286|>": 100583, "<|LOC_287|>": 100584, "<|LOC_288|>": 100585, "<|LOC_289|>": 100586, "<|LOC_290|>": 100587, "<|LOC_291|>": 100588, "<|LOC_292|>": 100589, "<|LOC_293|>": 100590, "<|LOC_294|>": 100591, "<|LOC_295|>": 100592, "<|LOC_296|>": 100593, "<|LOC_297|>": 100594, "<|LOC_298|>": 100595, "<|LOC_299|>": 100596, "<|LOC_300|>": 100597, "<|LOC_301|>": 100598, "<|LOC_302|>": 100599, "<|LOC_303|>": 100600, "<|LOC_304|>": 100601, "<|LOC_305|>": 100602, "<|LOC_306|>": 100603, "<|LOC_307|>": 100604, "<|LOC_308|>": 100605, "<|LOC_309|>": 100606, "<|LOC_310|>": 100607, "<|LOC_311|>": 100608, "<|LOC_312|>": 100609, "<|LOC_313|>": 100610, "<|LOC_314|>": 100611, "<|LOC_315|>": 100612, "<|LOC_316|>": 100613, "<|LOC_317|>": 100614, "<|LOC_318|>": 100615, "<|LOC_319|>": 100616, "<|LOC_320|>": 100617, "<|LOC_321|>": 100618, "<|LOC_322|>": 100619, "<|LOC_323|>": 100620, "<|LOC_324|>": 100621, "<|LOC_325|>": 100622, "<|LOC_326|>": 100623, "<|LOC_327|>": 100624, "<|LOC_328|>": 100625, "<|LOC_329|>": 100626, "<|LOC_330|>": 100627, "<|LOC_331|>": 100628, "<|LOC_332|>": 100629, "<|LOC_333|>": 100630, "<|LOC_334|>": 100631, "<|LOC_335|>": 100632, "<|LOC_336|>": 100633, "<|LOC_337|>": 100634, "<|LOC_338|>": 100635, "<|LOC_339|>": 100636, "<|LOC_340|>": 100637, "<|LOC_341|>": 100638, "<|LOC_342|>": 100639, "<|LOC_343|>": 100640, "<|LOC_344|>": 100641, "<|LOC_345|>": 100642, "<|LOC_346|>": 100643, "<|LOC_347|>": 100644, "<|LOC_348|>": 100645, "<|LOC_349|>": 100646, "<|LOC_350|>": 100647, "<|LOC_351|>": 100648, "<|LOC_352|>": 100649, "<|LOC_353|>": 100650, "<|LOC_354|>": 100651, "<|LOC_355|>": 100652, "<|LOC_356|>": 100653, "<|LOC_357|>": 100654, "<|LOC_358|>": 100655, "<|LOC_359|>": 100656, "<|LOC_360|>": 100657, "<|LOC_361|>": 100658, "<|LOC_362|>": 100659, "<|LOC_363|>": 100660, "<|LOC_364|>": 100661, "<|LOC_365|>": 100662, "<|LOC_366|>": 100663, "<|LOC_367|>": 100664, "<|LOC_368|>": 100665, "<|LOC_369|>": 100666, "<|LOC_370|>": 100667, "<|LOC_371|>": 100668, "<|LOC_372|>": 100669, "<|LOC_373|>": 100670, "<|LOC_374|>": 100671, "<|LOC_375|>": 100672, "<|LOC_376|>": 100673, "<|LOC_377|>": 100674, "<|LOC_378|>": 100675, "<|LOC_379|>": 100676, "<|LOC_380|>": 100677, "<|LOC_381|>": 100678, "<|LOC_382|>": 100679, "<|LOC_383|>": 100680, "<|LOC_384|>": 100681, "<|LOC_385|>": 100682, "<|LOC_386|>": 100683, "<|LOC_387|>": 100684, "<|LOC_388|>": 100685, "<|LOC_389|>": 100686, "<|LOC_390|>": 100687, "<|LOC_391|>": 100688, "<|LOC_392|>": 100689, "<|LOC_393|>": 100690, "<|LOC_394|>": 100691, "<|LOC_395|>": 100692, "<|LOC_396|>": 100693, "<|LOC_397|>": 100694, "<|LOC_398|>": 100695, "<|LOC_399|>": 100696, "<|LOC_400|>": 100697, "<|LOC_401|>": 100698, "<|LOC_402|>": 100699, "<|LOC_403|>": 100700, "<|LOC_404|>": 100701, "<|LOC_405|>": 100702, "<|LOC_406|>": 100703, "<|LOC_407|>": 100704, "<|LOC_408|>": 100705, "<|LOC_409|>": 100706, "<|LOC_410|>": 100707, "<|LOC_411|>": 100708, "<|LOC_412|>": 100709, "<|LOC_413|>": 100710, "<|LOC_414|>": 100711, "<|LOC_415|>": 100712, "<|LOC_416|>": 100713, "<|LOC_417|>": 100714, "<|LOC_418|>": 100715, "<|LOC_419|>": 100716, "<|LOC_420|>": 100717, "<|LOC_421|>": 100718, "<|LOC_422|>": 100719, "<|LOC_423|>": 100720, "<|LOC_424|>": 100721, "<|LOC_425|>": 100722, "<|LOC_426|>": 100723, "<|LOC_427|>": 100724, "<|LOC_428|>": 100725, "<|LOC_429|>": 100726, "<|LOC_430|>": 100727, "<|LOC_431|>": 100728, "<|LOC_432|>": 100729, "<|LOC_433|>": 100730, "<|LOC_434|>": 100731, "<|LOC_435|>": 100732, "<|LOC_436|>": 100733, "<|LOC_437|>": 100734, "<|LOC_438|>": 100735, "<|LOC_439|>": 100736, "<|LOC_440|>": 100737, "<|LOC_441|>": 100738, "<|LOC_442|>": 100739, "<|LOC_443|>": 100740, "<|LOC_444|>": 100741, "<|LOC_445|>": 100742, "<|LOC_446|>": 100743, "<|LOC_447|>": 100744, "<|LOC_448|>": 100745, "<|LOC_449|>": 100746, "<|LOC_450|>": 100747, "<|LOC_451|>": 100748, "<|LOC_452|>": 100749, "<|LOC_453|>": 100750, "<|LOC_454|>": 100751, "<|LOC_455|>": 100752, "<|LOC_456|>": 100753, "<|LOC_457|>": 100754, "<|LOC_458|>": 100755, "<|LOC_459|>": 100756, "<|LOC_460|>": 100757, "<|LOC_461|>": 100758, "<|LOC_462|>": 100759, "<|LOC_463|>": 100760, "<|LOC_464|>": 100761, "<|LOC_465|>": 100762, "<|LOC_466|>": 100763, "<|LOC_467|>": 100764, "<|LOC_468|>": 100765, "<|LOC_469|>": 100766, "<|LOC_470|>": 100767, "<|LOC_471|>": 100768, "<|LOC_472|>": 100769, "<|LOC_473|>": 100770, "<|LOC_474|>": 100771, "<|LOC_475|>": 100772, "<|LOC_476|>": 100773, "<|LOC_477|>": 100774, "<|LOC_478|>": 100775, "<|LOC_479|>": 100776, "<|LOC_480|>": 100777, "<|LOC_481|>": 100778, "<|LOC_482|>": 100779, "<|LOC_483|>": 100780, "<|LOC_484|>": 100781, "<|LOC_485|>": 100782, "<|LOC_486|>": 100783, "<|LOC_487|>": 100784, "<|LOC_488|>": 100785, "<|LOC_489|>": 100786, "<|LOC_490|>": 100787, "<|LOC_491|>": 100788, "<|LOC_492|>": 100789, "<|LOC_493|>": 100790, "<|LOC_494|>": 100791, "<|LOC_495|>": 100792, "<|LOC_496|>": 100793, "<|LOC_497|>": 100794, "<|LOC_498|>": 100795, "<|LOC_499|>": 100796, "<|LOC_500|>": 100797, "<|LOC_501|>": 100798, "<|LOC_502|>": 100799, "<|LOC_503|>": 100800, "<|LOC_504|>": 100801, "<|LOC_505|>": 100802, "<|LOC_506|>": 100803, "<|LOC_507|>": 100804, "<|LOC_508|>": 100805, "<|LOC_509|>": 100806, "<|LOC_510|>": 100807, "<|LOC_511|>": 100808, "<|LOC_512|>": 100809, "<|LOC_513|>": 100810, "<|LOC_514|>": 100811, "<|LOC_515|>": 100812, "<|LOC_516|>": 100813, "<|LOC_517|>": 100814, "<|LOC_518|>": 100815, "<|LOC_519|>": 100816, "<|LOC_520|>": 100817, "<|LOC_521|>": 100818, "<|LOC_522|>": 100819, "<|LOC_523|>": 100820, "<|LOC_524|>": 100821, "<|LOC_525|>": 100822, "<|LOC_526|>": 100823, "<|LOC_527|>": 100824, "<|LOC_528|>": 100825, "<|LOC_529|>": 100826, "<|LOC_530|>": 100827, "<|LOC_531|>": 100828, "<|LOC_532|>": 100829, "<|LOC_533|>": 100830, "<|LOC_534|>": 100831, "<|LOC_535|>": 100832, "<|LOC_536|>": 100833, "<|LOC_537|>": 100834, "<|LOC_538|>": 100835, "<|LOC_539|>": 100836, "<|LOC_540|>": 100837, "<|LOC_541|>": 100838, "<|LOC_542|>": 100839, "<|LOC_543|>": 100840, "<|LOC_544|>": 100841, "<|LOC_545|>": 100842, "<|LOC_546|>": 100843, "<|LOC_547|>": 100844, "<|LOC_548|>": 100845, "<|LOC_549|>": 100846, "<|LOC_550|>": 100847, "<|LOC_551|>": 100848, "<|LOC_552|>": 100849, "<|LOC_553|>": 100850, "<|LOC_554|>": 100851, "<|LOC_555|>": 100852, "<|LOC_556|>": 100853, "<|LOC_557|>": 100854, "<|LOC_558|>": 100855, "<|LOC_559|>": 100856, "<|LOC_560|>": 100857, "<|LOC_561|>": 100858, "<|LOC_562|>": 100859, "<|LOC_563|>": 100860, "<|LOC_564|>": 100861, "<|LOC_565|>": 100862, "<|LOC_566|>": 100863, "<|LOC_567|>": 100864, "<|LOC_568|>": 100865, "<|LOC_569|>": 100866, "<|LOC_570|>": 100867, "<|LOC_571|>": 100868, "<|LOC_572|>": 100869, "<|LOC_573|>": 100870, "<|LOC_574|>": 100871, "<|LOC_575|>": 100872, "<|LOC_576|>": 100873, "<|LOC_577|>": 100874, "<|LOC_578|>": 100875, "<|LOC_579|>": 100876, "<|LOC_580|>": 100877, "<|LOC_581|>": 100878, "<|LOC_582|>": 100879, "<|LOC_583|>": 100880, "<|LOC_584|>": 100881, "<|LOC_585|>": 100882, "<|LOC_586|>": 100883, "<|LOC_587|>": 100884, "<|LOC_588|>": 100885, "<|LOC_589|>": 100886, "<|LOC_590|>": 100887, "<|LOC_591|>": 100888, "<|LOC_592|>": 100889, "<|LOC_593|>": 100890, "<|LOC_594|>": 100891, "<|LOC_595|>": 100892, "<|LOC_596|>": 100893, "<|LOC_597|>": 100894, "<|LOC_598|>": 100895, "<|LOC_599|>": 100896, "<|LOC_600|>": 100897, "<|LOC_601|>": 100898, "<|LOC_602|>": 100899, "<|LOC_603|>": 100900, "<|LOC_604|>": 100901, "<|LOC_605|>": 100902, "<|LOC_606|>": 100903, "<|LOC_607|>": 100904, "<|LOC_608|>": 100905, "<|LOC_609|>": 100906, "<|LOC_610|>": 100907, "<|LOC_611|>": 100908, "<|LOC_612|>": 100909, "<|LOC_613|>": 100910, "<|LOC_614|>": 100911, "<|LOC_615|>": 100912, "<|LOC_616|>": 100913, "<|LOC_617|>": 100914, "<|LOC_618|>": 100915, "<|LOC_619|>": 100916, "<|LOC_620|>": 100917, "<|LOC_621|>": 100918, "<|LOC_622|>": 100919, "<|LOC_623|>": 100920, "<|LOC_624|>": 100921, "<|LOC_625|>": 100922, "<|LOC_626|>": 100923, "<|LOC_627|>": 100924, "<|LOC_628|>": 100925, "<|LOC_629|>": 100926, "<|LOC_630|>": 100927, "<|LOC_631|>": 100928, "<|LOC_632|>": 100929, "<|LOC_633|>": 100930, "<|LOC_634|>": 100931, "<|LOC_635|>": 100932, "<|LOC_636|>": 100933, "<|LOC_637|>": 100934, "<|LOC_638|>": 100935, "<|LOC_639|>": 100936, "<|LOC_640|>": 100937, "<|LOC_641|>": 100938, "<|LOC_642|>": 100939, "<|LOC_643|>": 100940, "<|LOC_644|>": 100941, "<|LOC_645|>": 100942, "<|LOC_646|>": 100943, "<|LOC_647|>": 100944, "<|LOC_648|>": 100945, "<|LOC_649|>": 100946, "<|LOC_650|>": 100947, "<|LOC_651|>": 100948, "<|LOC_652|>": 100949, "<|LOC_653|>": 100950, "<|LOC_654|>": 100951, "<|LOC_655|>": 100952, "<|LOC_656|>": 100953, "<|LOC_657|>": 100954, "<|LOC_658|>": 100955, "<|LOC_659|>": 100956, "<|LOC_660|>": 100957, "<|LOC_661|>": 100958, "<|LOC_662|>": 100959, "<|LOC_663|>": 100960, "<|LOC_664|>": 100961, "<|LOC_665|>": 100962, "<|LOC_666|>": 100963, "<|LOC_667|>": 100964, "<|LOC_668|>": 100965, "<|LOC_669|>": 100966, "<|LOC_670|>": 100967, "<|LOC_671|>": 100968, "<|LOC_672|>": 100969, "<|LOC_673|>": 100970, "<|LOC_674|>": 100971, "<|LOC_675|>": 100972, "<|LOC_676|>": 100973, "<|LOC_677|>": 100974, "<|LOC_678|>": 100975, "<|LOC_679|>": 100976, "<|LOC_680|>": 100977, "<|LOC_681|>": 100978, "<|LOC_682|>": 100979, "<|LOC_683|>": 100980, "<|LOC_684|>": 100981, "<|LOC_685|>": 100982, "<|LOC_686|>": 100983, "<|LOC_687|>": 100984, "<|LOC_688|>": 100985, "<|LOC_689|>": 100986, "<|LOC_690|>": 100987, "<|LOC_691|>": 100988, "<|LOC_692|>": 100989, "<|LOC_693|>": 100990, "<|LOC_694|>": 100991, "<|LOC_695|>": 100992, "<|LOC_696|>": 100993, "<|LOC_697|>": 100994, "<|LOC_698|>": 100995, "<|LOC_699|>": 100996, "<|LOC_700|>": 100997, "<|LOC_701|>": 100998, "<|LOC_702|>": 100999, "<|LOC_703|>": 101000, "<|LOC_704|>": 101001, "<|LOC_705|>": 101002, "<|LOC_706|>": 101003, "<|LOC_707|>": 101004, "<|LOC_708|>": 101005, "<|LOC_709|>": 101006, "<|LOC_710|>": 101007, "<|LOC_711|>": 101008, "<|LOC_712|>": 101009, "<|LOC_713|>": 101010, "<|LOC_714|>": 101011, "<|LOC_715|>": 101012, "<|LOC_716|>": 101013, "<|LOC_717|>": 101014, "<|LOC_718|>": 101015, "<|LOC_719|>": 101016, "<|LOC_720|>": 101017, "<|LOC_721|>": 101018, "<|LOC_722|>": 101019, "<|LOC_723|>": 101020, "<|LOC_724|>": 101021, "<|LOC_725|>": 101022, "<|LOC_726|>": 101023, "<|LOC_727|>": 101024, "<|LOC_728|>": 101025, "<|LOC_729|>": 101026, "<|LOC_730|>": 101027, "<|LOC_731|>": 101028, "<|LOC_732|>": 101029, "<|LOC_733|>": 101030, "<|LOC_734|>": 101031, "<|LOC_735|>": 101032, "<|LOC_736|>": 101033, "<|LOC_737|>": 101034, "<|LOC_738|>": 101035, "<|LOC_739|>": 101036, "<|LOC_740|>": 101037, "<|LOC_741|>": 101038, "<|LOC_742|>": 101039, "<|LOC_743|>": 101040, "<|LOC_744|>": 101041, "<|LOC_745|>": 101042, "<|LOC_746|>": 101043, "<|LOC_747|>": 101044, "<|LOC_748|>": 101045, "<|LOC_749|>": 101046, "<|LOC_750|>": 101047, "<|LOC_751|>": 101048, "<|LOC_752|>": 101049, "<|LOC_753|>": 101050, "<|LOC_754|>": 101051, "<|LOC_755|>": 101052, "<|LOC_756|>": 101053, "<|LOC_757|>": 101054, "<|LOC_758|>": 101055, "<|LOC_759|>": 101056, "<|LOC_760|>": 101057, "<|LOC_761|>": 101058, "<|LOC_762|>": 101059, "<|LOC_763|>": 101060, "<|LOC_764|>": 101061, "<|LOC_765|>": 101062, "<|LOC_766|>": 101063, "<|LOC_767|>": 101064, "<|LOC_768|>": 101065, "<|LOC_769|>": 101066, "<|LOC_770|>": 101067, "<|LOC_771|>": 101068, "<|LOC_772|>": 101069, "<|LOC_773|>": 101070, "<|LOC_774|>": 101071, "<|LOC_775|>": 101072, "<|LOC_776|>": 101073, "<|LOC_777|>": 101074, "<|LOC_778|>": 101075, "<|LOC_779|>": 101076, "<|LOC_780|>": 101077, "<|LOC_781|>": 101078, "<|LOC_782|>": 101079, "<|LOC_783|>": 101080, "<|LOC_784|>": 101081, "<|LOC_785|>": 101082, "<|LOC_786|>": 101083, "<|LOC_787|>": 101084, "<|LOC_788|>": 101085, "<|LOC_789|>": 101086, "<|LOC_790|>": 101087, "<|LOC_791|>": 101088, "<|LOC_792|>": 101089, "<|LOC_793|>": 101090, "<|LOC_794|>": 101091, "<|LOC_795|>": 101092, "<|LOC_796|>": 101093, "<|LOC_797|>": 101094, "<|LOC_798|>": 101095, "<|LOC_799|>": 101096, "<|LOC_800|>": 101097, "<|LOC_801|>": 101098, "<|LOC_802|>": 101099, "<|LOC_803|>": 101100, "<|LOC_804|>": 101101, "<|LOC_805|>": 101102, "<|LOC_806|>": 101103, "<|LOC_807|>": 101104, "<|LOC_808|>": 101105, "<|LOC_809|>": 101106, "<|LOC_810|>": 101107, "<|LOC_811|>": 101108, "<|LOC_812|>": 101109, "<|LOC_813|>": 101110, "<|LOC_814|>": 101111, "<|LOC_815|>": 101112, "<|LOC_816|>": 101113, "<|LOC_817|>": 101114, "<|LOC_818|>": 101115, "<|LOC_819|>": 101116, "<|LOC_820|>": 101117, "<|LOC_821|>": 101118, "<|LOC_822|>": 101119, "<|LOC_823|>": 101120, "<|LOC_824|>": 101121, "<|LOC_825|>": 101122, "<|LOC_826|>": 101123, "<|LOC_827|>": 101124, "<|LOC_828|>": 101125, "<|LOC_829|>": 101126, "<|LOC_830|>": 101127, "<|LOC_831|>": 101128, "<|LOC_832|>": 101129, "<|LOC_833|>": 101130, "<|LOC_834|>": 101131, "<|LOC_835|>": 101132, "<|LOC_836|>": 101133, "<|LOC_837|>": 101134, "<|LOC_838|>": 101135, "<|LOC_839|>": 101136, "<|LOC_840|>": 101137, "<|LOC_841|>": 101138, "<|LOC_842|>": 101139, "<|LOC_843|>": 101140, "<|LOC_844|>": 101141, "<|LOC_845|>": 101142, "<|LOC_846|>": 101143, "<|LOC_847|>": 101144, "<|LOC_848|>": 101145, "<|LOC_849|>": 101146, "<|LOC_850|>": 101147, "<|LOC_851|>": 101148, "<|LOC_852|>": 101149, "<|LOC_853|>": 101150, "<|LOC_854|>": 101151, "<|LOC_855|>": 101152, "<|LOC_856|>": 101153, "<|LOC_857|>": 101154, "<|LOC_858|>": 101155, "<|LOC_859|>": 101156, "<|LOC_860|>": 101157, "<|LOC_861|>": 101158, "<|LOC_862|>": 101159, "<|LOC_863|>": 101160, "<|LOC_864|>": 101161, "<|LOC_865|>": 101162, "<|LOC_866|>": 101163, "<|LOC_867|>": 101164, "<|LOC_868|>": 101165, "<|LOC_869|>": 101166, "<|LOC_870|>": 101167, "<|LOC_871|>": 101168, "<|LOC_872|>": 101169, "<|LOC_873|>": 101170, "<|LOC_874|>": 101171, "<|LOC_875|>": 101172, "<|LOC_876|>": 101173, "<|LOC_877|>": 101174, "<|LOC_878|>": 101175, "<|LOC_879|>": 101176, "<|LOC_880|>": 101177, "<|LOC_881|>": 101178, "<|LOC_882|>": 101179, "<|LOC_883|>": 101180, "<|LOC_884|>": 101181, "<|LOC_885|>": 101182, "<|LOC_886|>": 101183, "<|LOC_887|>": 101184, "<|LOC_888|>": 101185, "<|LOC_889|>": 101186, "<|LOC_890|>": 101187, "<|LOC_891|>": 101188, "<|LOC_892|>": 101189, "<|LOC_893|>": 101190, "<|LOC_894|>": 101191, "<|LOC_895|>": 101192, "<|LOC_896|>": 101193, "<|LOC_897|>": 101194, "<|LOC_898|>": 101195, "<|LOC_899|>": 101196, "<|LOC_900|>": 101197, "<|LOC_901|>": 101198, "<|LOC_902|>": 101199, "<|LOC_903|>": 101200, "<|LOC_904|>": 101201, "<|LOC_905|>": 101202, "<|LOC_906|>": 101203, "<|LOC_907|>": 101204, "<|LOC_908|>": 101205, "<|LOC_909|>": 101206, "<|LOC_910|>": 101207, "<|LOC_911|>": 101208, "<|LOC_912|>": 101209, "<|LOC_913|>": 101210, "<|LOC_914|>": 101211, "<|LOC_915|>": 101212, "<|LOC_916|>": 101213, "<|LOC_917|>": 101214, "<|LOC_918|>": 101215, "<|LOC_919|>": 101216, "<|LOC_920|>": 101217, "<|LOC_921|>": 101218, "<|LOC_922|>": 101219, "<|LOC_923|>": 101220, "<|LOC_924|>": 101221, "<|LOC_925|>": 101222, "<|LOC_926|>": 101223, "<|LOC_927|>": 101224, "<|LOC_928|>": 101225, "<|LOC_929|>": 101226, "<|LOC_930|>": 101227, "<|LOC_931|>": 101228, "<|LOC_932|>": 101229, "<|LOC_933|>": 101230, "<|LOC_934|>": 101231, "<|LOC_935|>": 101232, "<|LOC_936|>": 101233, "<|LOC_937|>": 101234, "<|LOC_938|>": 101235, "<|LOC_939|>": 101236, "<|LOC_940|>": 101237, "<|LOC_941|>": 101238, "<|LOC_942|>": 101239, "<|LOC_943|>": 101240, "<|LOC_944|>": 101241, "<|LOC_945|>": 101242, "<|LOC_946|>": 101243, "<|LOC_947|>": 101244, "<|LOC_948|>": 101245, "<|LOC_949|>": 101246, "<|LOC_950|>": 101247, "<|LOC_951|>": 101248, "<|LOC_952|>": 101249, "<|LOC_953|>": 101250, "<|LOC_954|>": 101251, "<|LOC_955|>": 101252, "<|LOC_956|>": 101253, "<|LOC_957|>": 101254, "<|LOC_958|>": 101255, "<|LOC_959|>": 101256, "<|LOC_960|>": 101257, "<|LOC_961|>": 101258, "<|LOC_962|>": 101259, "<|LOC_963|>": 101260, "<|LOC_964|>": 101261, "<|LOC_965|>": 101262, "<|LOC_966|>": 101263, "<|LOC_967|>": 101264, "<|LOC_968|>": 101265, "<|LOC_969|>": 101266, "<|LOC_970|>": 101267, "<|LOC_971|>": 101268, "<|LOC_972|>": 101269, "<|LOC_973|>": 101270, "<|LOC_974|>": 101271, "<|LOC_975|>": 101272, "<|LOC_976|>": 101273, "<|LOC_977|>": 101274, "<|LOC_978|>": 101275, "<|LOC_979|>": 101276, "<|LOC_980|>": 101277, "<|LOC_981|>": 101278, "<|LOC_982|>": 101279, "<|LOC_983|>": 101280, "<|LOC_984|>": 101281, "<|LOC_985|>": 101282, "<|LOC_986|>": 101283, "<|LOC_987|>": 101284, "<|LOC_988|>": 101285, "<|LOC_989|>": 101286, "<|LOC_990|>": 101287, "<|LOC_991|>": 101288, "<|LOC_992|>": 101289, "<|LOC_993|>": 101290, "<|LOC_994|>": 101291, "<|LOC_995|>": 101292, "<|LOC_996|>": 101293, "<|LOC_997|>": 101294, "<|LOC_998|>": 101295, "<|LOC_999|>": 101296, "<|LOC_1000|>": 101297, "<|LOC_BEGIN|>": 101298, "<|LOC_END|>": 101299, "<|LOC_SEP|>": 101300, "<|CROP_COL_SEP|>": 101301, "<|CROP_ROW_SEP|>": 101302, "<|IMAGE_SEP|>": 101303, "<|image_pad|>": 101304, "<|IMAGE_START|>": 101305, "<|IMAGE_END|>": 101306, "<|video_pad|>": 101307, "<ecel>": 101308, "<fcel>": 101309, "<xcel>": 101310, "<lcel>": 101311, "<ucel>": 101312, "<nl>": 101313}
checkpoint-336/config.json ADDED
@@ -0,0 +1,223 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "PaddleOCRVLForConditionalGeneration"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.0,
6
+ "auto_map": {
7
+ "AutoConfig": "configuration_paddleocr_vl.PaddleOCRVLConfig",
8
+ "AutoModel": "modeling_paddleocr_vl.PaddleOCRVLForConditionalGeneration",
9
+ "AutoModelForCausalLM": "modeling_paddleocr_vl.PaddleOCRVLForConditionalGeneration"
10
+ },
11
+ "compression_ratio": 1.0,
12
+ "disable_pipeline_warmup": false,
13
+ "enable_mtp_magic_send": false,
14
+ "fp16_opt_level": "O2",
15
+ "freq_allocation": 20,
16
+ "fuse_ln": false,
17
+ "fuse_rms_norm": true,
18
+ "head_dim": 128,
19
+ "hidden_act": "silu",
20
+ "hidden_dropout_prob": 0.0,
21
+ "hidden_size": 1024,
22
+ "ignored_index": -100,
23
+ "im_patch_id": 100295,
24
+ "image_token_id": 100295,
25
+ "intermediate_size": 3072,
26
+ "max_position_embeddings": 131072,
27
+ "max_text_id": 100295,
28
+ "model_type": "paddleocr_vl",
29
+ "moe_dropout_prob": 0.0,
30
+ "moe_multimodal_dispatch_use_allgather": "v2-alltoall-unpad",
31
+ "num_attention_heads": 16,
32
+ "num_hidden_layers": 18,
33
+ "num_key_value_heads": 2,
34
+ "paddleformers_version": "0.4.0",
35
+ "pixel_hidden_size": 1152,
36
+ "rms_norm_eps": 1e-05,
37
+ "rope_is_neox_style": true,
38
+ "rope_scaling": {
39
+ "mrope_section": [
40
+ 16,
41
+ 24,
42
+ 24
43
+ ],
44
+ "rope_type": "default",
45
+ "type": "default"
46
+ },
47
+ "rope_theta": 500000,
48
+ "scale_qk_coeff": 1.0,
49
+ "seqlen": 16384,
50
+ "sliding_window": null,
51
+ "tie_word_embeddings": false,
52
+ "token_balance_loss": false,
53
+ "token_balance_seqlen": 16384,
54
+ "torch_dtype": "bfloat16",
55
+ "use_3d_rope": true,
56
+ "use_bias": false,
57
+ "use_flash_attn_with_mask": true,
58
+ "use_fp8": false,
59
+ "use_mem_eff_attn": true,
60
+ "use_recompute_moe": false,
61
+ "use_rmsnorm": true,
62
+ "video_token_id": 101307,
63
+ "vision_config": {
64
+ "_attn_implementation": "eager",
65
+ "_name_or_path": "",
66
+ "_save_to_hf": false,
67
+ "add_cross_attention": false,
68
+ "add_tail_layers": 0,
69
+ "architectures": [
70
+ "PaddleOCRVisionModel"
71
+ ],
72
+ "attention_dropout": 0.0,
73
+ "auto_map": {
74
+ "AutoConfig": "configuration_paddleocr_vl.PaddleOCRVLConfig",
75
+ "AutoModel": "modeling_paddleocr_vl.PaddleOCRVisionModel"
76
+ },
77
+ "bad_words_ids": null,
78
+ "begin_suppress_tokens": null,
79
+ "bos_token_id": null,
80
+ "chunk_size_feed_forward": 0,
81
+ "classifier_dropout": null,
82
+ "context_parallel_degree": 1,
83
+ "cross_attention_hidden_size": null,
84
+ "decoder_start_token_id": null,
85
+ "diversity_penalty": 0.0,
86
+ "do_sample": false,
87
+ "dpo_config": null,
88
+ "dtype": "bfloat16",
89
+ "early_stopping": false,
90
+ "encoder_no_repeat_ngram_size": 0,
91
+ "eos_token_id": null,
92
+ "ep_communication_type": "deepep",
93
+ "exponential_decay_length_penalty": null,
94
+ "finetuning_task": null,
95
+ "forced_bos_token_id": null,
96
+ "forced_eos_token_id": null,
97
+ "fuse_attention_ffn": false,
98
+ "fuse_attention_qkv": false,
99
+ "fuse_linear": false,
100
+ "fuse_rope": false,
101
+ "fuse_sequence_parallel_allreduce": false,
102
+ "fuse_swiglu": false,
103
+ "hidden_act": "gelu_new",
104
+ "hidden_size": 1152,
105
+ "id2label": {
106
+ "0": "LABEL_0",
107
+ "1": "LABEL_1"
108
+ },
109
+ "image_size": 384,
110
+ "intermediate_size": 4304,
111
+ "is_decoder": false,
112
+ "is_encoder_decoder": false,
113
+ "kto_config": null,
114
+ "label2id": {
115
+ "LABEL_0": 0,
116
+ "LABEL_1": 1
117
+ },
118
+ "layer_norm_eps": 1e-06,
119
+ "length_penalty": 1.0,
120
+ "loss_subbatch_sequence_length": -1,
121
+ "max_length": 20,
122
+ "min_length": 0,
123
+ "model_type": "paddleocr_vl",
124
+ "moe_subbatch_token_num": 0,
125
+ "no_recompute_layers": null,
126
+ "no_repeat_ngram_size": 0,
127
+ "num_attention_heads": 16,
128
+ "num_beam_groups": 1,
129
+ "num_beams": 1,
130
+ "num_channels": 3,
131
+ "num_choices": null,
132
+ "num_hidden_layers": 27,
133
+ "num_nextn_predict_layers": 0,
134
+ "num_return_sequences": 1,
135
+ "offload_recompute_inputs": false,
136
+ "output_attentions": false,
137
+ "output_hidden_states": false,
138
+ "output_scores": false,
139
+ "pad_token_id": 0,
140
+ "patch_size": 14,
141
+ "pipeline_parallel_degree": 1,
142
+ "pp_recompute_interval": 1,
143
+ "prefix": null,
144
+ "problem_type": null,
145
+ "pruned_heads": {},
146
+ "quantization_config": {
147
+ "act_quant_method": "abs_max",
148
+ "activation_scheme": null,
149
+ "actscale_moving_rate": 0.01,
150
+ "apply_hadamard": false,
151
+ "apply_online_actscale_step": 200,
152
+ "dense_quant_type": "",
153
+ "dtype": null,
154
+ "fmt": null,
155
+ "fp8_format_type": "hybrid",
156
+ "group_size": -1,
157
+ "hadamard_block_size": 32,
158
+ "ignore_modules": null,
159
+ "llm_int8_threshold": 6.0,
160
+ "moe_quant_type": "",
161
+ "qlora_weight_blocksize": 64,
162
+ "qlora_weight_double_quant": false,
163
+ "qlora_weight_double_quant_block_size": 256,
164
+ "quant_input_grad": false,
165
+ "quant_method": null,
166
+ "quant_round_type": 0,
167
+ "quant_type": null,
168
+ "quant_weight_grad": false,
169
+ "quantization": "",
170
+ "scale_epsilon": 1e-08,
171
+ "shift": false,
172
+ "shift_smooth_all_linears": false,
173
+ "smooth": false,
174
+ "weight_block_size": null,
175
+ "weight_quant_method": "abs_max_channel_wise",
176
+ "weight_quantize_algo": null
177
+ },
178
+ "recompute": true,
179
+ "recompute_granularity": "full",
180
+ "recompute_use_reentrant": false,
181
+ "refined_recompute": "",
182
+ "remove_invalid_values": false,
183
+ "repetition_penalty": 1.0,
184
+ "return_dict": false,
185
+ "return_dict_in_generate": false,
186
+ "sep_parallel_degree": 1,
187
+ "sep_token_id": null,
188
+ "sequence_parallel": false,
189
+ "spatial_merge_size": 2,
190
+ "suppress_tokens": null,
191
+ "task_specific_params": null,
192
+ "temperature": 1.0,
193
+ "temporal_patch_size": 2,
194
+ "tensor_parallel_degree": 1,
195
+ "tensor_parallel_output": true,
196
+ "tensor_parallel_rank": 0,
197
+ "tie_encoder_decoder": false,
198
+ "tie_word_embeddings": true,
199
+ "tokenizer_class": null,
200
+ "tokens_per_second": 2,
201
+ "top_k": 50,
202
+ "top_p": 1.0,
203
+ "typical_p": 1.0,
204
+ "use_cache": false,
205
+ "use_filtered_label_loss": false,
206
+ "use_flash_attention": true,
207
+ "use_fused_dropout_add": false,
208
+ "use_fused_head_and_loss_fn": false,
209
+ "use_fused_linear": false,
210
+ "use_fused_linear_cross_entropy": false,
211
+ "use_fused_rms_norm": false,
212
+ "use_fused_rope": false,
213
+ "use_sparse_flash_attn": true,
214
+ "use_sparse_head_and_loss_fn": false,
215
+ "use_unified_moe": false,
216
+ "using_fake_gate": false,
217
+ "virtual_pp_degree": 1
218
+ },
219
+ "vision_end_token_id": 101306,
220
+ "vision_start_token_id": 101305,
221
+ "vocab_size": 103424,
222
+ "weight_share_add_bias": true
223
+ }
checkpoint-336/generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "eos_token_id": 2,
4
+ "pad_token_id": 0,
5
+ "use_cache": false
6
+ }
checkpoint-336/master_weights-00001-of-00001.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a0c4828027d40f6474016cae12335498d8c612eb2531e2fdbedb34699f503ce
3
+ size 3622483352
checkpoint-336/master_weights.safetensors.index.json ADDED
@@ -0,0 +1,615 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 3622406592
4
+ },
5
+ "weight_map": {
6
+ "mlp_AR.pre_norm.weight": "master_weights-00001-of-00001.safetensors",
7
+ "mlp_AR.pre_norm.bias": "master_weights-00001-of-00001.safetensors",
8
+ "mlp_AR.linear_1.weight": "master_weights-00001-of-00001.safetensors",
9
+ "mlp_AR.linear_1.bias": "master_weights-00001-of-00001.safetensors",
10
+ "mlp_AR.linear_2.weight": "master_weights-00001-of-00001.safetensors",
11
+ "mlp_AR.linear_2.bias": "master_weights-00001-of-00001.safetensors",
12
+ "visual.vision_model.embeddings.patch_embedding.weight": "master_weights-00001-of-00001.safetensors",
13
+ "visual.vision_model.embeddings.patch_embedding.bias": "master_weights-00001-of-00001.safetensors",
14
+ "visual.vision_model.embeddings.position_embedding.weight": "master_weights-00001-of-00001.safetensors",
15
+ "visual.vision_model.encoder.layers.0.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
16
+ "visual.vision_model.encoder.layers.0.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
17
+ "visual.vision_model.encoder.layers.0.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
18
+ "visual.vision_model.encoder.layers.0.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
19
+ "visual.vision_model.encoder.layers.0.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
20
+ "visual.vision_model.encoder.layers.0.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
21
+ "visual.vision_model.encoder.layers.0.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
22
+ "visual.vision_model.encoder.layers.0.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
23
+ "visual.vision_model.encoder.layers.0.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
24
+ "visual.vision_model.encoder.layers.0.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
25
+ "visual.vision_model.encoder.layers.0.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
26
+ "visual.vision_model.encoder.layers.0.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
27
+ "visual.vision_model.encoder.layers.0.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
28
+ "visual.vision_model.encoder.layers.0.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
29
+ "visual.vision_model.encoder.layers.0.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
30
+ "visual.vision_model.encoder.layers.0.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
31
+ "visual.vision_model.encoder.layers.1.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
32
+ "visual.vision_model.encoder.layers.1.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
33
+ "visual.vision_model.encoder.layers.1.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
34
+ "visual.vision_model.encoder.layers.1.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
35
+ "visual.vision_model.encoder.layers.1.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
36
+ "visual.vision_model.encoder.layers.1.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
37
+ "visual.vision_model.encoder.layers.1.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
38
+ "visual.vision_model.encoder.layers.1.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
39
+ "visual.vision_model.encoder.layers.1.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
40
+ "visual.vision_model.encoder.layers.1.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
41
+ "visual.vision_model.encoder.layers.1.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
42
+ "visual.vision_model.encoder.layers.1.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
43
+ "visual.vision_model.encoder.layers.1.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
44
+ "visual.vision_model.encoder.layers.1.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
45
+ "visual.vision_model.encoder.layers.1.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
46
+ "visual.vision_model.encoder.layers.1.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
47
+ "visual.vision_model.encoder.layers.2.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
48
+ "visual.vision_model.encoder.layers.2.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
49
+ "visual.vision_model.encoder.layers.2.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
50
+ "visual.vision_model.encoder.layers.2.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
51
+ "visual.vision_model.encoder.layers.2.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
52
+ "visual.vision_model.encoder.layers.2.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
53
+ "visual.vision_model.encoder.layers.2.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
54
+ "visual.vision_model.encoder.layers.2.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
55
+ "visual.vision_model.encoder.layers.2.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
56
+ "visual.vision_model.encoder.layers.2.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
57
+ "visual.vision_model.encoder.layers.2.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
58
+ "visual.vision_model.encoder.layers.2.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
59
+ "visual.vision_model.encoder.layers.2.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
60
+ "visual.vision_model.encoder.layers.2.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
61
+ "visual.vision_model.encoder.layers.2.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
62
+ "visual.vision_model.encoder.layers.2.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
63
+ "visual.vision_model.encoder.layers.3.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
64
+ "visual.vision_model.encoder.layers.3.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
65
+ "visual.vision_model.encoder.layers.3.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
66
+ "visual.vision_model.encoder.layers.3.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
67
+ "visual.vision_model.encoder.layers.3.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
68
+ "visual.vision_model.encoder.layers.3.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
69
+ "visual.vision_model.encoder.layers.3.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
70
+ "visual.vision_model.encoder.layers.3.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
71
+ "visual.vision_model.encoder.layers.3.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
72
+ "visual.vision_model.encoder.layers.3.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
73
+ "visual.vision_model.encoder.layers.3.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
74
+ "visual.vision_model.encoder.layers.3.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
75
+ "visual.vision_model.encoder.layers.3.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
76
+ "visual.vision_model.encoder.layers.3.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
77
+ "visual.vision_model.encoder.layers.3.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
78
+ "visual.vision_model.encoder.layers.3.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
79
+ "visual.vision_model.encoder.layers.4.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
80
+ "visual.vision_model.encoder.layers.4.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
81
+ "visual.vision_model.encoder.layers.4.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
82
+ "visual.vision_model.encoder.layers.4.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
83
+ "visual.vision_model.encoder.layers.4.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
84
+ "visual.vision_model.encoder.layers.4.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
85
+ "visual.vision_model.encoder.layers.4.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
86
+ "visual.vision_model.encoder.layers.4.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
87
+ "visual.vision_model.encoder.layers.4.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
88
+ "visual.vision_model.encoder.layers.4.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
89
+ "visual.vision_model.encoder.layers.4.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
90
+ "visual.vision_model.encoder.layers.4.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
91
+ "visual.vision_model.encoder.layers.4.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
92
+ "visual.vision_model.encoder.layers.4.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
93
+ "visual.vision_model.encoder.layers.4.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
94
+ "visual.vision_model.encoder.layers.4.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
95
+ "visual.vision_model.encoder.layers.5.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
96
+ "visual.vision_model.encoder.layers.5.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
97
+ "visual.vision_model.encoder.layers.5.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
98
+ "visual.vision_model.encoder.layers.5.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
99
+ "visual.vision_model.encoder.layers.5.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
100
+ "visual.vision_model.encoder.layers.5.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
101
+ "visual.vision_model.encoder.layers.5.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
102
+ "visual.vision_model.encoder.layers.5.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
103
+ "visual.vision_model.encoder.layers.5.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
104
+ "visual.vision_model.encoder.layers.5.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
105
+ "visual.vision_model.encoder.layers.5.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
106
+ "visual.vision_model.encoder.layers.5.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
107
+ "visual.vision_model.encoder.layers.5.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
108
+ "visual.vision_model.encoder.layers.5.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
109
+ "visual.vision_model.encoder.layers.5.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
110
+ "visual.vision_model.encoder.layers.5.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
111
+ "visual.vision_model.encoder.layers.6.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
112
+ "visual.vision_model.encoder.layers.6.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
113
+ "visual.vision_model.encoder.layers.6.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
114
+ "visual.vision_model.encoder.layers.6.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
115
+ "visual.vision_model.encoder.layers.6.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
116
+ "visual.vision_model.encoder.layers.6.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
117
+ "visual.vision_model.encoder.layers.6.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
118
+ "visual.vision_model.encoder.layers.6.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
119
+ "visual.vision_model.encoder.layers.6.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
120
+ "visual.vision_model.encoder.layers.6.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
121
+ "visual.vision_model.encoder.layers.6.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
122
+ "visual.vision_model.encoder.layers.6.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
123
+ "visual.vision_model.encoder.layers.6.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
124
+ "visual.vision_model.encoder.layers.6.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
125
+ "visual.vision_model.encoder.layers.6.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
126
+ "visual.vision_model.encoder.layers.6.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
127
+ "visual.vision_model.encoder.layers.7.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
128
+ "visual.vision_model.encoder.layers.7.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
129
+ "visual.vision_model.encoder.layers.7.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
130
+ "visual.vision_model.encoder.layers.7.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
131
+ "visual.vision_model.encoder.layers.7.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
132
+ "visual.vision_model.encoder.layers.7.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
133
+ "visual.vision_model.encoder.layers.7.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
134
+ "visual.vision_model.encoder.layers.7.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
135
+ "visual.vision_model.encoder.layers.7.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
136
+ "visual.vision_model.encoder.layers.7.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
137
+ "visual.vision_model.encoder.layers.7.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
138
+ "visual.vision_model.encoder.layers.7.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
139
+ "visual.vision_model.encoder.layers.7.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
140
+ "visual.vision_model.encoder.layers.7.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
141
+ "visual.vision_model.encoder.layers.7.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
142
+ "visual.vision_model.encoder.layers.7.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
143
+ "visual.vision_model.encoder.layers.8.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
144
+ "visual.vision_model.encoder.layers.8.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
145
+ "visual.vision_model.encoder.layers.8.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
146
+ "visual.vision_model.encoder.layers.8.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
147
+ "visual.vision_model.encoder.layers.8.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
148
+ "visual.vision_model.encoder.layers.8.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
149
+ "visual.vision_model.encoder.layers.8.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
150
+ "visual.vision_model.encoder.layers.8.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
151
+ "visual.vision_model.encoder.layers.8.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
152
+ "visual.vision_model.encoder.layers.8.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
153
+ "visual.vision_model.encoder.layers.8.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
154
+ "visual.vision_model.encoder.layers.8.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
155
+ "visual.vision_model.encoder.layers.8.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
156
+ "visual.vision_model.encoder.layers.8.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
157
+ "visual.vision_model.encoder.layers.8.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
158
+ "visual.vision_model.encoder.layers.8.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
159
+ "visual.vision_model.encoder.layers.9.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
160
+ "visual.vision_model.encoder.layers.9.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
161
+ "visual.vision_model.encoder.layers.9.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
162
+ "visual.vision_model.encoder.layers.9.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
163
+ "visual.vision_model.encoder.layers.9.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
164
+ "visual.vision_model.encoder.layers.9.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
165
+ "visual.vision_model.encoder.layers.9.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
166
+ "visual.vision_model.encoder.layers.9.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
167
+ "visual.vision_model.encoder.layers.9.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
168
+ "visual.vision_model.encoder.layers.9.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
169
+ "visual.vision_model.encoder.layers.9.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
170
+ "visual.vision_model.encoder.layers.9.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
171
+ "visual.vision_model.encoder.layers.9.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
172
+ "visual.vision_model.encoder.layers.9.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
173
+ "visual.vision_model.encoder.layers.9.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
174
+ "visual.vision_model.encoder.layers.9.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
175
+ "visual.vision_model.encoder.layers.10.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
176
+ "visual.vision_model.encoder.layers.10.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
177
+ "visual.vision_model.encoder.layers.10.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
178
+ "visual.vision_model.encoder.layers.10.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
179
+ "visual.vision_model.encoder.layers.10.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
180
+ "visual.vision_model.encoder.layers.10.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
181
+ "visual.vision_model.encoder.layers.10.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
182
+ "visual.vision_model.encoder.layers.10.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
183
+ "visual.vision_model.encoder.layers.10.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
184
+ "visual.vision_model.encoder.layers.10.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
185
+ "visual.vision_model.encoder.layers.10.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
186
+ "visual.vision_model.encoder.layers.10.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
187
+ "visual.vision_model.encoder.layers.10.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
188
+ "visual.vision_model.encoder.layers.10.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
189
+ "visual.vision_model.encoder.layers.10.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
190
+ "visual.vision_model.encoder.layers.10.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
191
+ "visual.vision_model.encoder.layers.11.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
192
+ "visual.vision_model.encoder.layers.11.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
193
+ "visual.vision_model.encoder.layers.11.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
194
+ "visual.vision_model.encoder.layers.11.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
195
+ "visual.vision_model.encoder.layers.11.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
196
+ "visual.vision_model.encoder.layers.11.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
197
+ "visual.vision_model.encoder.layers.11.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
198
+ "visual.vision_model.encoder.layers.11.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
199
+ "visual.vision_model.encoder.layers.11.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
200
+ "visual.vision_model.encoder.layers.11.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
201
+ "visual.vision_model.encoder.layers.11.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
202
+ "visual.vision_model.encoder.layers.11.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
203
+ "visual.vision_model.encoder.layers.11.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
204
+ "visual.vision_model.encoder.layers.11.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
205
+ "visual.vision_model.encoder.layers.11.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
206
+ "visual.vision_model.encoder.layers.11.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
207
+ "visual.vision_model.encoder.layers.12.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
208
+ "visual.vision_model.encoder.layers.12.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
209
+ "visual.vision_model.encoder.layers.12.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
210
+ "visual.vision_model.encoder.layers.12.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
211
+ "visual.vision_model.encoder.layers.12.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
212
+ "visual.vision_model.encoder.layers.12.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
213
+ "visual.vision_model.encoder.layers.12.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
214
+ "visual.vision_model.encoder.layers.12.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
215
+ "visual.vision_model.encoder.layers.12.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
216
+ "visual.vision_model.encoder.layers.12.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
217
+ "visual.vision_model.encoder.layers.12.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
218
+ "visual.vision_model.encoder.layers.12.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
219
+ "visual.vision_model.encoder.layers.12.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
220
+ "visual.vision_model.encoder.layers.12.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
221
+ "visual.vision_model.encoder.layers.12.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
222
+ "visual.vision_model.encoder.layers.12.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
223
+ "visual.vision_model.encoder.layers.13.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
224
+ "visual.vision_model.encoder.layers.13.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
225
+ "visual.vision_model.encoder.layers.13.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
226
+ "visual.vision_model.encoder.layers.13.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
227
+ "visual.vision_model.encoder.layers.13.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
228
+ "visual.vision_model.encoder.layers.13.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
229
+ "visual.vision_model.encoder.layers.13.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
230
+ "visual.vision_model.encoder.layers.13.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
231
+ "visual.vision_model.encoder.layers.13.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
232
+ "visual.vision_model.encoder.layers.13.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
233
+ "visual.vision_model.encoder.layers.13.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
234
+ "visual.vision_model.encoder.layers.13.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
235
+ "visual.vision_model.encoder.layers.13.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
236
+ "visual.vision_model.encoder.layers.13.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
237
+ "visual.vision_model.encoder.layers.13.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
238
+ "visual.vision_model.encoder.layers.13.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
239
+ "visual.vision_model.encoder.layers.14.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
240
+ "visual.vision_model.encoder.layers.14.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
241
+ "visual.vision_model.encoder.layers.14.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
242
+ "visual.vision_model.encoder.layers.14.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
243
+ "visual.vision_model.encoder.layers.14.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
244
+ "visual.vision_model.encoder.layers.14.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
245
+ "visual.vision_model.encoder.layers.14.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
246
+ "visual.vision_model.encoder.layers.14.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
247
+ "visual.vision_model.encoder.layers.14.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
248
+ "visual.vision_model.encoder.layers.14.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
249
+ "visual.vision_model.encoder.layers.14.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
250
+ "visual.vision_model.encoder.layers.14.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
251
+ "visual.vision_model.encoder.layers.14.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
252
+ "visual.vision_model.encoder.layers.14.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
253
+ "visual.vision_model.encoder.layers.14.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
254
+ "visual.vision_model.encoder.layers.14.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
255
+ "visual.vision_model.encoder.layers.15.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
256
+ "visual.vision_model.encoder.layers.15.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
257
+ "visual.vision_model.encoder.layers.15.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
258
+ "visual.vision_model.encoder.layers.15.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
259
+ "visual.vision_model.encoder.layers.15.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
260
+ "visual.vision_model.encoder.layers.15.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
261
+ "visual.vision_model.encoder.layers.15.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
262
+ "visual.vision_model.encoder.layers.15.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
263
+ "visual.vision_model.encoder.layers.15.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
264
+ "visual.vision_model.encoder.layers.15.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
265
+ "visual.vision_model.encoder.layers.15.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
266
+ "visual.vision_model.encoder.layers.15.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
267
+ "visual.vision_model.encoder.layers.15.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
268
+ "visual.vision_model.encoder.layers.15.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
269
+ "visual.vision_model.encoder.layers.15.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
270
+ "visual.vision_model.encoder.layers.15.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
271
+ "visual.vision_model.encoder.layers.16.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
272
+ "visual.vision_model.encoder.layers.16.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
273
+ "visual.vision_model.encoder.layers.16.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
274
+ "visual.vision_model.encoder.layers.16.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
275
+ "visual.vision_model.encoder.layers.16.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
276
+ "visual.vision_model.encoder.layers.16.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
277
+ "visual.vision_model.encoder.layers.16.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
278
+ "visual.vision_model.encoder.layers.16.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
279
+ "visual.vision_model.encoder.layers.16.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
280
+ "visual.vision_model.encoder.layers.16.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
281
+ "visual.vision_model.encoder.layers.16.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
282
+ "visual.vision_model.encoder.layers.16.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
283
+ "visual.vision_model.encoder.layers.16.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
284
+ "visual.vision_model.encoder.layers.16.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
285
+ "visual.vision_model.encoder.layers.16.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
286
+ "visual.vision_model.encoder.layers.16.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
287
+ "visual.vision_model.encoder.layers.17.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
288
+ "visual.vision_model.encoder.layers.17.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
289
+ "visual.vision_model.encoder.layers.17.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
290
+ "visual.vision_model.encoder.layers.17.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
291
+ "visual.vision_model.encoder.layers.17.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
292
+ "visual.vision_model.encoder.layers.17.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
293
+ "visual.vision_model.encoder.layers.17.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
294
+ "visual.vision_model.encoder.layers.17.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
295
+ "visual.vision_model.encoder.layers.17.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
296
+ "visual.vision_model.encoder.layers.17.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
297
+ "visual.vision_model.encoder.layers.17.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
298
+ "visual.vision_model.encoder.layers.17.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
299
+ "visual.vision_model.encoder.layers.17.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
300
+ "visual.vision_model.encoder.layers.17.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
301
+ "visual.vision_model.encoder.layers.17.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
302
+ "visual.vision_model.encoder.layers.17.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
303
+ "visual.vision_model.encoder.layers.18.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
304
+ "visual.vision_model.encoder.layers.18.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
305
+ "visual.vision_model.encoder.layers.18.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
306
+ "visual.vision_model.encoder.layers.18.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
307
+ "visual.vision_model.encoder.layers.18.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
308
+ "visual.vision_model.encoder.layers.18.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
309
+ "visual.vision_model.encoder.layers.18.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
310
+ "visual.vision_model.encoder.layers.18.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
311
+ "visual.vision_model.encoder.layers.18.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
312
+ "visual.vision_model.encoder.layers.18.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
313
+ "visual.vision_model.encoder.layers.18.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
314
+ "visual.vision_model.encoder.layers.18.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
315
+ "visual.vision_model.encoder.layers.18.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
316
+ "visual.vision_model.encoder.layers.18.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
317
+ "visual.vision_model.encoder.layers.18.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
318
+ "visual.vision_model.encoder.layers.18.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
319
+ "visual.vision_model.encoder.layers.19.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
320
+ "visual.vision_model.encoder.layers.19.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
321
+ "visual.vision_model.encoder.layers.19.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
322
+ "visual.vision_model.encoder.layers.19.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
323
+ "visual.vision_model.encoder.layers.19.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
324
+ "visual.vision_model.encoder.layers.19.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
325
+ "visual.vision_model.encoder.layers.19.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
326
+ "visual.vision_model.encoder.layers.19.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
327
+ "visual.vision_model.encoder.layers.19.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
328
+ "visual.vision_model.encoder.layers.19.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
329
+ "visual.vision_model.encoder.layers.19.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
330
+ "visual.vision_model.encoder.layers.19.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
331
+ "visual.vision_model.encoder.layers.19.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
332
+ "visual.vision_model.encoder.layers.19.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
333
+ "visual.vision_model.encoder.layers.19.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
334
+ "visual.vision_model.encoder.layers.19.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
335
+ "visual.vision_model.encoder.layers.20.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
336
+ "visual.vision_model.encoder.layers.20.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
337
+ "visual.vision_model.encoder.layers.20.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
338
+ "visual.vision_model.encoder.layers.20.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
339
+ "visual.vision_model.encoder.layers.20.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
340
+ "visual.vision_model.encoder.layers.20.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
341
+ "visual.vision_model.encoder.layers.20.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
342
+ "visual.vision_model.encoder.layers.20.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
343
+ "visual.vision_model.encoder.layers.20.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
344
+ "visual.vision_model.encoder.layers.20.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
345
+ "visual.vision_model.encoder.layers.20.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
346
+ "visual.vision_model.encoder.layers.20.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
347
+ "visual.vision_model.encoder.layers.20.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
348
+ "visual.vision_model.encoder.layers.20.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
349
+ "visual.vision_model.encoder.layers.20.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
350
+ "visual.vision_model.encoder.layers.20.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
351
+ "visual.vision_model.encoder.layers.21.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
352
+ "visual.vision_model.encoder.layers.21.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
353
+ "visual.vision_model.encoder.layers.21.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
354
+ "visual.vision_model.encoder.layers.21.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
355
+ "visual.vision_model.encoder.layers.21.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
356
+ "visual.vision_model.encoder.layers.21.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
357
+ "visual.vision_model.encoder.layers.21.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
358
+ "visual.vision_model.encoder.layers.21.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
359
+ "visual.vision_model.encoder.layers.21.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
360
+ "visual.vision_model.encoder.layers.21.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
361
+ "visual.vision_model.encoder.layers.21.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
362
+ "visual.vision_model.encoder.layers.21.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
363
+ "visual.vision_model.encoder.layers.21.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
364
+ "visual.vision_model.encoder.layers.21.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
365
+ "visual.vision_model.encoder.layers.21.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
366
+ "visual.vision_model.encoder.layers.21.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
367
+ "visual.vision_model.encoder.layers.22.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
368
+ "visual.vision_model.encoder.layers.22.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
369
+ "visual.vision_model.encoder.layers.22.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
370
+ "visual.vision_model.encoder.layers.22.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
371
+ "visual.vision_model.encoder.layers.22.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
372
+ "visual.vision_model.encoder.layers.22.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
373
+ "visual.vision_model.encoder.layers.22.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
374
+ "visual.vision_model.encoder.layers.22.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
375
+ "visual.vision_model.encoder.layers.22.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
376
+ "visual.vision_model.encoder.layers.22.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
377
+ "visual.vision_model.encoder.layers.22.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
378
+ "visual.vision_model.encoder.layers.22.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
379
+ "visual.vision_model.encoder.layers.22.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
380
+ "visual.vision_model.encoder.layers.22.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
381
+ "visual.vision_model.encoder.layers.22.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
382
+ "visual.vision_model.encoder.layers.22.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
383
+ "visual.vision_model.encoder.layers.23.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
384
+ "visual.vision_model.encoder.layers.23.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
385
+ "visual.vision_model.encoder.layers.23.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
386
+ "visual.vision_model.encoder.layers.23.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
387
+ "visual.vision_model.encoder.layers.23.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
388
+ "visual.vision_model.encoder.layers.23.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
389
+ "visual.vision_model.encoder.layers.23.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
390
+ "visual.vision_model.encoder.layers.23.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
391
+ "visual.vision_model.encoder.layers.23.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
392
+ "visual.vision_model.encoder.layers.23.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
393
+ "visual.vision_model.encoder.layers.23.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
394
+ "visual.vision_model.encoder.layers.23.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
395
+ "visual.vision_model.encoder.layers.23.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
396
+ "visual.vision_model.encoder.layers.23.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
397
+ "visual.vision_model.encoder.layers.23.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
398
+ "visual.vision_model.encoder.layers.23.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
399
+ "visual.vision_model.encoder.layers.24.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
400
+ "visual.vision_model.encoder.layers.24.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
401
+ "visual.vision_model.encoder.layers.24.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
402
+ "visual.vision_model.encoder.layers.24.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
403
+ "visual.vision_model.encoder.layers.24.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
404
+ "visual.vision_model.encoder.layers.24.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
405
+ "visual.vision_model.encoder.layers.24.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
406
+ "visual.vision_model.encoder.layers.24.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
407
+ "visual.vision_model.encoder.layers.24.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
408
+ "visual.vision_model.encoder.layers.24.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
409
+ "visual.vision_model.encoder.layers.24.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
410
+ "visual.vision_model.encoder.layers.24.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
411
+ "visual.vision_model.encoder.layers.24.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
412
+ "visual.vision_model.encoder.layers.24.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
413
+ "visual.vision_model.encoder.layers.24.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
414
+ "visual.vision_model.encoder.layers.24.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
415
+ "visual.vision_model.encoder.layers.25.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
416
+ "visual.vision_model.encoder.layers.25.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
417
+ "visual.vision_model.encoder.layers.25.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
418
+ "visual.vision_model.encoder.layers.25.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
419
+ "visual.vision_model.encoder.layers.25.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
420
+ "visual.vision_model.encoder.layers.25.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
421
+ "visual.vision_model.encoder.layers.25.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
422
+ "visual.vision_model.encoder.layers.25.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
423
+ "visual.vision_model.encoder.layers.25.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
424
+ "visual.vision_model.encoder.layers.25.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
425
+ "visual.vision_model.encoder.layers.25.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
426
+ "visual.vision_model.encoder.layers.25.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
427
+ "visual.vision_model.encoder.layers.25.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
428
+ "visual.vision_model.encoder.layers.25.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
429
+ "visual.vision_model.encoder.layers.25.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
430
+ "visual.vision_model.encoder.layers.25.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
431
+ "visual.vision_model.encoder.layers.26.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
432
+ "visual.vision_model.encoder.layers.26.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
433
+ "visual.vision_model.encoder.layers.26.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
434
+ "visual.vision_model.encoder.layers.26.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
435
+ "visual.vision_model.encoder.layers.26.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
436
+ "visual.vision_model.encoder.layers.26.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
437
+ "visual.vision_model.encoder.layers.26.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
438
+ "visual.vision_model.encoder.layers.26.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
439
+ "visual.vision_model.encoder.layers.26.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
440
+ "visual.vision_model.encoder.layers.26.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
441
+ "visual.vision_model.encoder.layers.26.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
442
+ "visual.vision_model.encoder.layers.26.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
443
+ "visual.vision_model.encoder.layers.26.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
444
+ "visual.vision_model.encoder.layers.26.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
445
+ "visual.vision_model.encoder.layers.26.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
446
+ "visual.vision_model.encoder.layers.26.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
447
+ "visual.vision_model.post_layernorm.weight": "master_weights-00001-of-00001.safetensors",
448
+ "visual.vision_model.post_layernorm.bias": "master_weights-00001-of-00001.safetensors",
449
+ "model.embed_tokens.weight": "master_weights-00001-of-00001.safetensors",
450
+ "model.layers.0.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
451
+ "model.layers.0.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
452
+ "model.layers.0.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
453
+ "model.layers.0.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
454
+ "model.layers.0.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
455
+ "model.layers.0.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
456
+ "model.layers.0.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
457
+ "model.layers.0.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
458
+ "model.layers.0.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
459
+ "model.layers.1.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
460
+ "model.layers.1.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
461
+ "model.layers.1.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
462
+ "model.layers.1.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
463
+ "model.layers.1.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
464
+ "model.layers.1.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
465
+ "model.layers.1.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
466
+ "model.layers.1.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
467
+ "model.layers.1.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
468
+ "model.layers.2.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
469
+ "model.layers.2.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
470
+ "model.layers.2.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
471
+ "model.layers.2.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
472
+ "model.layers.2.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
473
+ "model.layers.2.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
474
+ "model.layers.2.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
475
+ "model.layers.2.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
476
+ "model.layers.2.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
477
+ "model.layers.3.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
478
+ "model.layers.3.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
479
+ "model.layers.3.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
480
+ "model.layers.3.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
481
+ "model.layers.3.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
482
+ "model.layers.3.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
483
+ "model.layers.3.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
484
+ "model.layers.3.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
485
+ "model.layers.3.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
486
+ "model.layers.4.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
487
+ "model.layers.4.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
488
+ "model.layers.4.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
489
+ "model.layers.4.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
490
+ "model.layers.4.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
491
+ "model.layers.4.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
492
+ "model.layers.4.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
493
+ "model.layers.4.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
494
+ "model.layers.4.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
495
+ "model.layers.5.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
496
+ "model.layers.5.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
497
+ "model.layers.5.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
498
+ "model.layers.5.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
499
+ "model.layers.5.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
500
+ "model.layers.5.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
501
+ "model.layers.5.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
502
+ "model.layers.5.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
503
+ "model.layers.5.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
504
+ "model.layers.6.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
505
+ "model.layers.6.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
506
+ "model.layers.6.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
507
+ "model.layers.6.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
508
+ "model.layers.6.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
509
+ "model.layers.6.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
510
+ "model.layers.6.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
511
+ "model.layers.6.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
512
+ "model.layers.6.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
513
+ "model.layers.7.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
514
+ "model.layers.7.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
515
+ "model.layers.7.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
516
+ "model.layers.7.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
517
+ "model.layers.7.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
518
+ "model.layers.7.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
519
+ "model.layers.7.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
520
+ "model.layers.7.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
521
+ "model.layers.7.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
522
+ "model.layers.8.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
523
+ "model.layers.8.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
524
+ "model.layers.8.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
525
+ "model.layers.8.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
526
+ "model.layers.8.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
527
+ "model.layers.8.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
528
+ "model.layers.8.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
529
+ "model.layers.8.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
530
+ "model.layers.8.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
531
+ "model.layers.9.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
532
+ "model.layers.9.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
533
+ "model.layers.9.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
534
+ "model.layers.9.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
535
+ "model.layers.9.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
536
+ "model.layers.9.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
537
+ "model.layers.9.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
538
+ "model.layers.9.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
539
+ "model.layers.9.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
540
+ "model.layers.10.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
541
+ "model.layers.10.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
542
+ "model.layers.10.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
543
+ "model.layers.10.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
544
+ "model.layers.10.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
545
+ "model.layers.10.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
546
+ "model.layers.10.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
547
+ "model.layers.10.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
548
+ "model.layers.10.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
549
+ "model.layers.11.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
550
+ "model.layers.11.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
551
+ "model.layers.11.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
552
+ "model.layers.11.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
553
+ "model.layers.11.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
554
+ "model.layers.11.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
555
+ "model.layers.11.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
556
+ "model.layers.11.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
557
+ "model.layers.11.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
558
+ "model.layers.12.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
559
+ "model.layers.12.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
560
+ "model.layers.12.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
561
+ "model.layers.12.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
562
+ "model.layers.12.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
563
+ "model.layers.12.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
564
+ "model.layers.12.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
565
+ "model.layers.12.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
566
+ "model.layers.12.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
567
+ "model.layers.13.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
568
+ "model.layers.13.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
569
+ "model.layers.13.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
570
+ "model.layers.13.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
571
+ "model.layers.13.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
572
+ "model.layers.13.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
573
+ "model.layers.13.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
574
+ "model.layers.13.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
575
+ "model.layers.13.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
576
+ "model.layers.14.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
577
+ "model.layers.14.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
578
+ "model.layers.14.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
579
+ "model.layers.14.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
580
+ "model.layers.14.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
581
+ "model.layers.14.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
582
+ "model.layers.14.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
583
+ "model.layers.14.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
584
+ "model.layers.14.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
585
+ "model.layers.15.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
586
+ "model.layers.15.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
587
+ "model.layers.15.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
588
+ "model.layers.15.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
589
+ "model.layers.15.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
590
+ "model.layers.15.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
591
+ "model.layers.15.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
592
+ "model.layers.15.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
593
+ "model.layers.15.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
594
+ "model.layers.16.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
595
+ "model.layers.16.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
596
+ "model.layers.16.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
597
+ "model.layers.16.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
598
+ "model.layers.16.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
599
+ "model.layers.16.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
600
+ "model.layers.16.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
601
+ "model.layers.16.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
602
+ "model.layers.16.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
603
+ "model.layers.17.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
604
+ "model.layers.17.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
605
+ "model.layers.17.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
606
+ "model.layers.17.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
607
+ "model.layers.17.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
608
+ "model.layers.17.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
609
+ "model.layers.17.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
610
+ "model.layers.17.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
611
+ "model.layers.17.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
612
+ "model.norm.weight": "master_weights-00001-of-00001.safetensors",
613
+ "lm_head.weight": "master_weights-00001-of-00001.safetensors"
614
+ }
615
+ }
checkpoint-336/model-00001-of-00001.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8765aee223483d0bf21fd28e690f6de5a1e3668de260b7c197b99beac936adfd
3
+ size 1917255968
checkpoint-336/model.safetensors.index.json ADDED
@@ -0,0 +1,627 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 1917177472
4
+ },
5
+ "weight_map": {
6
+ "mlp_AR.pre_norm.weight": "model-00001-of-00001.safetensors",
7
+ "mlp_AR.pre_norm.bias": "model-00001-of-00001.safetensors",
8
+ "mlp_AR.linear_1.weight": "model-00001-of-00001.safetensors",
9
+ "mlp_AR.linear_1.bias": "model-00001-of-00001.safetensors",
10
+ "mlp_AR.linear_2.weight": "model-00001-of-00001.safetensors",
11
+ "mlp_AR.linear_2.bias": "model-00001-of-00001.safetensors",
12
+ "visual.vision_model.embeddings.patch_embedding.weight": "model-00001-of-00001.safetensors",
13
+ "visual.vision_model.embeddings.patch_embedding.bias": "model-00001-of-00001.safetensors",
14
+ "visual.vision_model.embeddings.position_embedding.weight": "model-00001-of-00001.safetensors",
15
+ "visual.vision_model.embeddings.packing_position_embedding.weight": "model-00001-of-00001.safetensors",
16
+ "visual.vision_model.encoder.layers.0.layer_norm1.weight": "model-00001-of-00001.safetensors",
17
+ "visual.vision_model.encoder.layers.0.layer_norm1.bias": "model-00001-of-00001.safetensors",
18
+ "visual.vision_model.encoder.layers.0.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
19
+ "visual.vision_model.encoder.layers.0.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
20
+ "visual.vision_model.encoder.layers.0.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
21
+ "visual.vision_model.encoder.layers.0.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
22
+ "visual.vision_model.encoder.layers.0.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
23
+ "visual.vision_model.encoder.layers.0.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
24
+ "visual.vision_model.encoder.layers.0.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
25
+ "visual.vision_model.encoder.layers.0.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
26
+ "visual.vision_model.encoder.layers.0.layer_norm2.weight": "model-00001-of-00001.safetensors",
27
+ "visual.vision_model.encoder.layers.0.layer_norm2.bias": "model-00001-of-00001.safetensors",
28
+ "visual.vision_model.encoder.layers.0.mlp.fc1.weight": "model-00001-of-00001.safetensors",
29
+ "visual.vision_model.encoder.layers.0.mlp.fc1.bias": "model-00001-of-00001.safetensors",
30
+ "visual.vision_model.encoder.layers.0.mlp.fc2.weight": "model-00001-of-00001.safetensors",
31
+ "visual.vision_model.encoder.layers.0.mlp.fc2.bias": "model-00001-of-00001.safetensors",
32
+ "visual.vision_model.encoder.layers.1.layer_norm1.weight": "model-00001-of-00001.safetensors",
33
+ "visual.vision_model.encoder.layers.1.layer_norm1.bias": "model-00001-of-00001.safetensors",
34
+ "visual.vision_model.encoder.layers.1.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
35
+ "visual.vision_model.encoder.layers.1.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
36
+ "visual.vision_model.encoder.layers.1.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
37
+ "visual.vision_model.encoder.layers.1.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
38
+ "visual.vision_model.encoder.layers.1.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
39
+ "visual.vision_model.encoder.layers.1.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
40
+ "visual.vision_model.encoder.layers.1.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
41
+ "visual.vision_model.encoder.layers.1.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
42
+ "visual.vision_model.encoder.layers.1.layer_norm2.weight": "model-00001-of-00001.safetensors",
43
+ "visual.vision_model.encoder.layers.1.layer_norm2.bias": "model-00001-of-00001.safetensors",
44
+ "visual.vision_model.encoder.layers.1.mlp.fc1.weight": "model-00001-of-00001.safetensors",
45
+ "visual.vision_model.encoder.layers.1.mlp.fc1.bias": "model-00001-of-00001.safetensors",
46
+ "visual.vision_model.encoder.layers.1.mlp.fc2.weight": "model-00001-of-00001.safetensors",
47
+ "visual.vision_model.encoder.layers.1.mlp.fc2.bias": "model-00001-of-00001.safetensors",
48
+ "visual.vision_model.encoder.layers.2.layer_norm1.weight": "model-00001-of-00001.safetensors",
49
+ "visual.vision_model.encoder.layers.2.layer_norm1.bias": "model-00001-of-00001.safetensors",
50
+ "visual.vision_model.encoder.layers.2.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
51
+ "visual.vision_model.encoder.layers.2.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
52
+ "visual.vision_model.encoder.layers.2.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
53
+ "visual.vision_model.encoder.layers.2.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
54
+ "visual.vision_model.encoder.layers.2.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
55
+ "visual.vision_model.encoder.layers.2.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
56
+ "visual.vision_model.encoder.layers.2.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
57
+ "visual.vision_model.encoder.layers.2.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
58
+ "visual.vision_model.encoder.layers.2.layer_norm2.weight": "model-00001-of-00001.safetensors",
59
+ "visual.vision_model.encoder.layers.2.layer_norm2.bias": "model-00001-of-00001.safetensors",
60
+ "visual.vision_model.encoder.layers.2.mlp.fc1.weight": "model-00001-of-00001.safetensors",
61
+ "visual.vision_model.encoder.layers.2.mlp.fc1.bias": "model-00001-of-00001.safetensors",
62
+ "visual.vision_model.encoder.layers.2.mlp.fc2.weight": "model-00001-of-00001.safetensors",
63
+ "visual.vision_model.encoder.layers.2.mlp.fc2.bias": "model-00001-of-00001.safetensors",
64
+ "visual.vision_model.encoder.layers.3.layer_norm1.weight": "model-00001-of-00001.safetensors",
65
+ "visual.vision_model.encoder.layers.3.layer_norm1.bias": "model-00001-of-00001.safetensors",
66
+ "visual.vision_model.encoder.layers.3.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
67
+ "visual.vision_model.encoder.layers.3.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
68
+ "visual.vision_model.encoder.layers.3.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
69
+ "visual.vision_model.encoder.layers.3.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
70
+ "visual.vision_model.encoder.layers.3.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
71
+ "visual.vision_model.encoder.layers.3.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
72
+ "visual.vision_model.encoder.layers.3.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
73
+ "visual.vision_model.encoder.layers.3.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
74
+ "visual.vision_model.encoder.layers.3.layer_norm2.weight": "model-00001-of-00001.safetensors",
75
+ "visual.vision_model.encoder.layers.3.layer_norm2.bias": "model-00001-of-00001.safetensors",
76
+ "visual.vision_model.encoder.layers.3.mlp.fc1.weight": "model-00001-of-00001.safetensors",
77
+ "visual.vision_model.encoder.layers.3.mlp.fc1.bias": "model-00001-of-00001.safetensors",
78
+ "visual.vision_model.encoder.layers.3.mlp.fc2.weight": "model-00001-of-00001.safetensors",
79
+ "visual.vision_model.encoder.layers.3.mlp.fc2.bias": "model-00001-of-00001.safetensors",
80
+ "visual.vision_model.encoder.layers.4.layer_norm1.weight": "model-00001-of-00001.safetensors",
81
+ "visual.vision_model.encoder.layers.4.layer_norm1.bias": "model-00001-of-00001.safetensors",
82
+ "visual.vision_model.encoder.layers.4.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
83
+ "visual.vision_model.encoder.layers.4.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
84
+ "visual.vision_model.encoder.layers.4.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
85
+ "visual.vision_model.encoder.layers.4.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
86
+ "visual.vision_model.encoder.layers.4.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
87
+ "visual.vision_model.encoder.layers.4.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
88
+ "visual.vision_model.encoder.layers.4.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
89
+ "visual.vision_model.encoder.layers.4.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
90
+ "visual.vision_model.encoder.layers.4.layer_norm2.weight": "model-00001-of-00001.safetensors",
91
+ "visual.vision_model.encoder.layers.4.layer_norm2.bias": "model-00001-of-00001.safetensors",
92
+ "visual.vision_model.encoder.layers.4.mlp.fc1.weight": "model-00001-of-00001.safetensors",
93
+ "visual.vision_model.encoder.layers.4.mlp.fc1.bias": "model-00001-of-00001.safetensors",
94
+ "visual.vision_model.encoder.layers.4.mlp.fc2.weight": "model-00001-of-00001.safetensors",
95
+ "visual.vision_model.encoder.layers.4.mlp.fc2.bias": "model-00001-of-00001.safetensors",
96
+ "visual.vision_model.encoder.layers.5.layer_norm1.weight": "model-00001-of-00001.safetensors",
97
+ "visual.vision_model.encoder.layers.5.layer_norm1.bias": "model-00001-of-00001.safetensors",
98
+ "visual.vision_model.encoder.layers.5.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
99
+ "visual.vision_model.encoder.layers.5.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
100
+ "visual.vision_model.encoder.layers.5.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
101
+ "visual.vision_model.encoder.layers.5.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
102
+ "visual.vision_model.encoder.layers.5.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
103
+ "visual.vision_model.encoder.layers.5.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
104
+ "visual.vision_model.encoder.layers.5.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
105
+ "visual.vision_model.encoder.layers.5.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
106
+ "visual.vision_model.encoder.layers.5.layer_norm2.weight": "model-00001-of-00001.safetensors",
107
+ "visual.vision_model.encoder.layers.5.layer_norm2.bias": "model-00001-of-00001.safetensors",
108
+ "visual.vision_model.encoder.layers.5.mlp.fc1.weight": "model-00001-of-00001.safetensors",
109
+ "visual.vision_model.encoder.layers.5.mlp.fc1.bias": "model-00001-of-00001.safetensors",
110
+ "visual.vision_model.encoder.layers.5.mlp.fc2.weight": "model-00001-of-00001.safetensors",
111
+ "visual.vision_model.encoder.layers.5.mlp.fc2.bias": "model-00001-of-00001.safetensors",
112
+ "visual.vision_model.encoder.layers.6.layer_norm1.weight": "model-00001-of-00001.safetensors",
113
+ "visual.vision_model.encoder.layers.6.layer_norm1.bias": "model-00001-of-00001.safetensors",
114
+ "visual.vision_model.encoder.layers.6.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
115
+ "visual.vision_model.encoder.layers.6.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
116
+ "visual.vision_model.encoder.layers.6.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
117
+ "visual.vision_model.encoder.layers.6.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
118
+ "visual.vision_model.encoder.layers.6.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
119
+ "visual.vision_model.encoder.layers.6.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
120
+ "visual.vision_model.encoder.layers.6.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
121
+ "visual.vision_model.encoder.layers.6.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
122
+ "visual.vision_model.encoder.layers.6.layer_norm2.weight": "model-00001-of-00001.safetensors",
123
+ "visual.vision_model.encoder.layers.6.layer_norm2.bias": "model-00001-of-00001.safetensors",
124
+ "visual.vision_model.encoder.layers.6.mlp.fc1.weight": "model-00001-of-00001.safetensors",
125
+ "visual.vision_model.encoder.layers.6.mlp.fc1.bias": "model-00001-of-00001.safetensors",
126
+ "visual.vision_model.encoder.layers.6.mlp.fc2.weight": "model-00001-of-00001.safetensors",
127
+ "visual.vision_model.encoder.layers.6.mlp.fc2.bias": "model-00001-of-00001.safetensors",
128
+ "visual.vision_model.encoder.layers.7.layer_norm1.weight": "model-00001-of-00001.safetensors",
129
+ "visual.vision_model.encoder.layers.7.layer_norm1.bias": "model-00001-of-00001.safetensors",
130
+ "visual.vision_model.encoder.layers.7.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
131
+ "visual.vision_model.encoder.layers.7.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
132
+ "visual.vision_model.encoder.layers.7.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
133
+ "visual.vision_model.encoder.layers.7.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
134
+ "visual.vision_model.encoder.layers.7.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
135
+ "visual.vision_model.encoder.layers.7.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
136
+ "visual.vision_model.encoder.layers.7.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
137
+ "visual.vision_model.encoder.layers.7.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
138
+ "visual.vision_model.encoder.layers.7.layer_norm2.weight": "model-00001-of-00001.safetensors",
139
+ "visual.vision_model.encoder.layers.7.layer_norm2.bias": "model-00001-of-00001.safetensors",
140
+ "visual.vision_model.encoder.layers.7.mlp.fc1.weight": "model-00001-of-00001.safetensors",
141
+ "visual.vision_model.encoder.layers.7.mlp.fc1.bias": "model-00001-of-00001.safetensors",
142
+ "visual.vision_model.encoder.layers.7.mlp.fc2.weight": "model-00001-of-00001.safetensors",
143
+ "visual.vision_model.encoder.layers.7.mlp.fc2.bias": "model-00001-of-00001.safetensors",
144
+ "visual.vision_model.encoder.layers.8.layer_norm1.weight": "model-00001-of-00001.safetensors",
145
+ "visual.vision_model.encoder.layers.8.layer_norm1.bias": "model-00001-of-00001.safetensors",
146
+ "visual.vision_model.encoder.layers.8.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
147
+ "visual.vision_model.encoder.layers.8.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
148
+ "visual.vision_model.encoder.layers.8.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
149
+ "visual.vision_model.encoder.layers.8.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
150
+ "visual.vision_model.encoder.layers.8.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
151
+ "visual.vision_model.encoder.layers.8.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
152
+ "visual.vision_model.encoder.layers.8.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
153
+ "visual.vision_model.encoder.layers.8.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
154
+ "visual.vision_model.encoder.layers.8.layer_norm2.weight": "model-00001-of-00001.safetensors",
155
+ "visual.vision_model.encoder.layers.8.layer_norm2.bias": "model-00001-of-00001.safetensors",
156
+ "visual.vision_model.encoder.layers.8.mlp.fc1.weight": "model-00001-of-00001.safetensors",
157
+ "visual.vision_model.encoder.layers.8.mlp.fc1.bias": "model-00001-of-00001.safetensors",
158
+ "visual.vision_model.encoder.layers.8.mlp.fc2.weight": "model-00001-of-00001.safetensors",
159
+ "visual.vision_model.encoder.layers.8.mlp.fc2.bias": "model-00001-of-00001.safetensors",
160
+ "visual.vision_model.encoder.layers.9.layer_norm1.weight": "model-00001-of-00001.safetensors",
161
+ "visual.vision_model.encoder.layers.9.layer_norm1.bias": "model-00001-of-00001.safetensors",
162
+ "visual.vision_model.encoder.layers.9.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
163
+ "visual.vision_model.encoder.layers.9.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
164
+ "visual.vision_model.encoder.layers.9.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
165
+ "visual.vision_model.encoder.layers.9.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
166
+ "visual.vision_model.encoder.layers.9.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
167
+ "visual.vision_model.encoder.layers.9.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
168
+ "visual.vision_model.encoder.layers.9.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
169
+ "visual.vision_model.encoder.layers.9.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
170
+ "visual.vision_model.encoder.layers.9.layer_norm2.weight": "model-00001-of-00001.safetensors",
171
+ "visual.vision_model.encoder.layers.9.layer_norm2.bias": "model-00001-of-00001.safetensors",
172
+ "visual.vision_model.encoder.layers.9.mlp.fc1.weight": "model-00001-of-00001.safetensors",
173
+ "visual.vision_model.encoder.layers.9.mlp.fc1.bias": "model-00001-of-00001.safetensors",
174
+ "visual.vision_model.encoder.layers.9.mlp.fc2.weight": "model-00001-of-00001.safetensors",
175
+ "visual.vision_model.encoder.layers.9.mlp.fc2.bias": "model-00001-of-00001.safetensors",
176
+ "visual.vision_model.encoder.layers.10.layer_norm1.weight": "model-00001-of-00001.safetensors",
177
+ "visual.vision_model.encoder.layers.10.layer_norm1.bias": "model-00001-of-00001.safetensors",
178
+ "visual.vision_model.encoder.layers.10.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
179
+ "visual.vision_model.encoder.layers.10.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
180
+ "visual.vision_model.encoder.layers.10.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
181
+ "visual.vision_model.encoder.layers.10.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
182
+ "visual.vision_model.encoder.layers.10.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
183
+ "visual.vision_model.encoder.layers.10.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
184
+ "visual.vision_model.encoder.layers.10.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
185
+ "visual.vision_model.encoder.layers.10.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
186
+ "visual.vision_model.encoder.layers.10.layer_norm2.weight": "model-00001-of-00001.safetensors",
187
+ "visual.vision_model.encoder.layers.10.layer_norm2.bias": "model-00001-of-00001.safetensors",
188
+ "visual.vision_model.encoder.layers.10.mlp.fc1.weight": "model-00001-of-00001.safetensors",
189
+ "visual.vision_model.encoder.layers.10.mlp.fc1.bias": "model-00001-of-00001.safetensors",
190
+ "visual.vision_model.encoder.layers.10.mlp.fc2.weight": "model-00001-of-00001.safetensors",
191
+ "visual.vision_model.encoder.layers.10.mlp.fc2.bias": "model-00001-of-00001.safetensors",
192
+ "visual.vision_model.encoder.layers.11.layer_norm1.weight": "model-00001-of-00001.safetensors",
193
+ "visual.vision_model.encoder.layers.11.layer_norm1.bias": "model-00001-of-00001.safetensors",
194
+ "visual.vision_model.encoder.layers.11.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
195
+ "visual.vision_model.encoder.layers.11.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
196
+ "visual.vision_model.encoder.layers.11.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
197
+ "visual.vision_model.encoder.layers.11.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
198
+ "visual.vision_model.encoder.layers.11.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
199
+ "visual.vision_model.encoder.layers.11.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
200
+ "visual.vision_model.encoder.layers.11.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
201
+ "visual.vision_model.encoder.layers.11.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
202
+ "visual.vision_model.encoder.layers.11.layer_norm2.weight": "model-00001-of-00001.safetensors",
203
+ "visual.vision_model.encoder.layers.11.layer_norm2.bias": "model-00001-of-00001.safetensors",
204
+ "visual.vision_model.encoder.layers.11.mlp.fc1.weight": "model-00001-of-00001.safetensors",
205
+ "visual.vision_model.encoder.layers.11.mlp.fc1.bias": "model-00001-of-00001.safetensors",
206
+ "visual.vision_model.encoder.layers.11.mlp.fc2.weight": "model-00001-of-00001.safetensors",
207
+ "visual.vision_model.encoder.layers.11.mlp.fc2.bias": "model-00001-of-00001.safetensors",
208
+ "visual.vision_model.encoder.layers.12.layer_norm1.weight": "model-00001-of-00001.safetensors",
209
+ "visual.vision_model.encoder.layers.12.layer_norm1.bias": "model-00001-of-00001.safetensors",
210
+ "visual.vision_model.encoder.layers.12.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
211
+ "visual.vision_model.encoder.layers.12.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
212
+ "visual.vision_model.encoder.layers.12.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
213
+ "visual.vision_model.encoder.layers.12.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
214
+ "visual.vision_model.encoder.layers.12.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
215
+ "visual.vision_model.encoder.layers.12.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
216
+ "visual.vision_model.encoder.layers.12.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
217
+ "visual.vision_model.encoder.layers.12.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
218
+ "visual.vision_model.encoder.layers.12.layer_norm2.weight": "model-00001-of-00001.safetensors",
219
+ "visual.vision_model.encoder.layers.12.layer_norm2.bias": "model-00001-of-00001.safetensors",
220
+ "visual.vision_model.encoder.layers.12.mlp.fc1.weight": "model-00001-of-00001.safetensors",
221
+ "visual.vision_model.encoder.layers.12.mlp.fc1.bias": "model-00001-of-00001.safetensors",
222
+ "visual.vision_model.encoder.layers.12.mlp.fc2.weight": "model-00001-of-00001.safetensors",
223
+ "visual.vision_model.encoder.layers.12.mlp.fc2.bias": "model-00001-of-00001.safetensors",
224
+ "visual.vision_model.encoder.layers.13.layer_norm1.weight": "model-00001-of-00001.safetensors",
225
+ "visual.vision_model.encoder.layers.13.layer_norm1.bias": "model-00001-of-00001.safetensors",
226
+ "visual.vision_model.encoder.layers.13.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
227
+ "visual.vision_model.encoder.layers.13.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
228
+ "visual.vision_model.encoder.layers.13.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
229
+ "visual.vision_model.encoder.layers.13.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
230
+ "visual.vision_model.encoder.layers.13.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
231
+ "visual.vision_model.encoder.layers.13.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
232
+ "visual.vision_model.encoder.layers.13.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
233
+ "visual.vision_model.encoder.layers.13.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
234
+ "visual.vision_model.encoder.layers.13.layer_norm2.weight": "model-00001-of-00001.safetensors",
235
+ "visual.vision_model.encoder.layers.13.layer_norm2.bias": "model-00001-of-00001.safetensors",
236
+ "visual.vision_model.encoder.layers.13.mlp.fc1.weight": "model-00001-of-00001.safetensors",
237
+ "visual.vision_model.encoder.layers.13.mlp.fc1.bias": "model-00001-of-00001.safetensors",
238
+ "visual.vision_model.encoder.layers.13.mlp.fc2.weight": "model-00001-of-00001.safetensors",
239
+ "visual.vision_model.encoder.layers.13.mlp.fc2.bias": "model-00001-of-00001.safetensors",
240
+ "visual.vision_model.encoder.layers.14.layer_norm1.weight": "model-00001-of-00001.safetensors",
241
+ "visual.vision_model.encoder.layers.14.layer_norm1.bias": "model-00001-of-00001.safetensors",
242
+ "visual.vision_model.encoder.layers.14.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
243
+ "visual.vision_model.encoder.layers.14.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
244
+ "visual.vision_model.encoder.layers.14.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
245
+ "visual.vision_model.encoder.layers.14.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
246
+ "visual.vision_model.encoder.layers.14.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
247
+ "visual.vision_model.encoder.layers.14.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
248
+ "visual.vision_model.encoder.layers.14.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
249
+ "visual.vision_model.encoder.layers.14.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
250
+ "visual.vision_model.encoder.layers.14.layer_norm2.weight": "model-00001-of-00001.safetensors",
251
+ "visual.vision_model.encoder.layers.14.layer_norm2.bias": "model-00001-of-00001.safetensors",
252
+ "visual.vision_model.encoder.layers.14.mlp.fc1.weight": "model-00001-of-00001.safetensors",
253
+ "visual.vision_model.encoder.layers.14.mlp.fc1.bias": "model-00001-of-00001.safetensors",
254
+ "visual.vision_model.encoder.layers.14.mlp.fc2.weight": "model-00001-of-00001.safetensors",
255
+ "visual.vision_model.encoder.layers.14.mlp.fc2.bias": "model-00001-of-00001.safetensors",
256
+ "visual.vision_model.encoder.layers.15.layer_norm1.weight": "model-00001-of-00001.safetensors",
257
+ "visual.vision_model.encoder.layers.15.layer_norm1.bias": "model-00001-of-00001.safetensors",
258
+ "visual.vision_model.encoder.layers.15.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
259
+ "visual.vision_model.encoder.layers.15.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
260
+ "visual.vision_model.encoder.layers.15.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
261
+ "visual.vision_model.encoder.layers.15.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
262
+ "visual.vision_model.encoder.layers.15.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
263
+ "visual.vision_model.encoder.layers.15.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
264
+ "visual.vision_model.encoder.layers.15.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
265
+ "visual.vision_model.encoder.layers.15.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
266
+ "visual.vision_model.encoder.layers.15.layer_norm2.weight": "model-00001-of-00001.safetensors",
267
+ "visual.vision_model.encoder.layers.15.layer_norm2.bias": "model-00001-of-00001.safetensors",
268
+ "visual.vision_model.encoder.layers.15.mlp.fc1.weight": "model-00001-of-00001.safetensors",
269
+ "visual.vision_model.encoder.layers.15.mlp.fc1.bias": "model-00001-of-00001.safetensors",
270
+ "visual.vision_model.encoder.layers.15.mlp.fc2.weight": "model-00001-of-00001.safetensors",
271
+ "visual.vision_model.encoder.layers.15.mlp.fc2.bias": "model-00001-of-00001.safetensors",
272
+ "visual.vision_model.encoder.layers.16.layer_norm1.weight": "model-00001-of-00001.safetensors",
273
+ "visual.vision_model.encoder.layers.16.layer_norm1.bias": "model-00001-of-00001.safetensors",
274
+ "visual.vision_model.encoder.layers.16.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
275
+ "visual.vision_model.encoder.layers.16.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
276
+ "visual.vision_model.encoder.layers.16.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
277
+ "visual.vision_model.encoder.layers.16.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
278
+ "visual.vision_model.encoder.layers.16.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
279
+ "visual.vision_model.encoder.layers.16.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
280
+ "visual.vision_model.encoder.layers.16.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
281
+ "visual.vision_model.encoder.layers.16.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
282
+ "visual.vision_model.encoder.layers.16.layer_norm2.weight": "model-00001-of-00001.safetensors",
283
+ "visual.vision_model.encoder.layers.16.layer_norm2.bias": "model-00001-of-00001.safetensors",
284
+ "visual.vision_model.encoder.layers.16.mlp.fc1.weight": "model-00001-of-00001.safetensors",
285
+ "visual.vision_model.encoder.layers.16.mlp.fc1.bias": "model-00001-of-00001.safetensors",
286
+ "visual.vision_model.encoder.layers.16.mlp.fc2.weight": "model-00001-of-00001.safetensors",
287
+ "visual.vision_model.encoder.layers.16.mlp.fc2.bias": "model-00001-of-00001.safetensors",
288
+ "visual.vision_model.encoder.layers.17.layer_norm1.weight": "model-00001-of-00001.safetensors",
289
+ "visual.vision_model.encoder.layers.17.layer_norm1.bias": "model-00001-of-00001.safetensors",
290
+ "visual.vision_model.encoder.layers.17.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
291
+ "visual.vision_model.encoder.layers.17.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
292
+ "visual.vision_model.encoder.layers.17.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
293
+ "visual.vision_model.encoder.layers.17.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
294
+ "visual.vision_model.encoder.layers.17.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
295
+ "visual.vision_model.encoder.layers.17.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
296
+ "visual.vision_model.encoder.layers.17.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
297
+ "visual.vision_model.encoder.layers.17.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
298
+ "visual.vision_model.encoder.layers.17.layer_norm2.weight": "model-00001-of-00001.safetensors",
299
+ "visual.vision_model.encoder.layers.17.layer_norm2.bias": "model-00001-of-00001.safetensors",
300
+ "visual.vision_model.encoder.layers.17.mlp.fc1.weight": "model-00001-of-00001.safetensors",
301
+ "visual.vision_model.encoder.layers.17.mlp.fc1.bias": "model-00001-of-00001.safetensors",
302
+ "visual.vision_model.encoder.layers.17.mlp.fc2.weight": "model-00001-of-00001.safetensors",
303
+ "visual.vision_model.encoder.layers.17.mlp.fc2.bias": "model-00001-of-00001.safetensors",
304
+ "visual.vision_model.encoder.layers.18.layer_norm1.weight": "model-00001-of-00001.safetensors",
305
+ "visual.vision_model.encoder.layers.18.layer_norm1.bias": "model-00001-of-00001.safetensors",
306
+ "visual.vision_model.encoder.layers.18.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
307
+ "visual.vision_model.encoder.layers.18.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
308
+ "visual.vision_model.encoder.layers.18.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
309
+ "visual.vision_model.encoder.layers.18.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
310
+ "visual.vision_model.encoder.layers.18.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
311
+ "visual.vision_model.encoder.layers.18.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
312
+ "visual.vision_model.encoder.layers.18.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
313
+ "visual.vision_model.encoder.layers.18.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
314
+ "visual.vision_model.encoder.layers.18.layer_norm2.weight": "model-00001-of-00001.safetensors",
315
+ "visual.vision_model.encoder.layers.18.layer_norm2.bias": "model-00001-of-00001.safetensors",
316
+ "visual.vision_model.encoder.layers.18.mlp.fc1.weight": "model-00001-of-00001.safetensors",
317
+ "visual.vision_model.encoder.layers.18.mlp.fc1.bias": "model-00001-of-00001.safetensors",
318
+ "visual.vision_model.encoder.layers.18.mlp.fc2.weight": "model-00001-of-00001.safetensors",
319
+ "visual.vision_model.encoder.layers.18.mlp.fc2.bias": "model-00001-of-00001.safetensors",
320
+ "visual.vision_model.encoder.layers.19.layer_norm1.weight": "model-00001-of-00001.safetensors",
321
+ "visual.vision_model.encoder.layers.19.layer_norm1.bias": "model-00001-of-00001.safetensors",
322
+ "visual.vision_model.encoder.layers.19.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
323
+ "visual.vision_model.encoder.layers.19.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
324
+ "visual.vision_model.encoder.layers.19.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
325
+ "visual.vision_model.encoder.layers.19.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
326
+ "visual.vision_model.encoder.layers.19.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
327
+ "visual.vision_model.encoder.layers.19.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
328
+ "visual.vision_model.encoder.layers.19.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
329
+ "visual.vision_model.encoder.layers.19.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
330
+ "visual.vision_model.encoder.layers.19.layer_norm2.weight": "model-00001-of-00001.safetensors",
331
+ "visual.vision_model.encoder.layers.19.layer_norm2.bias": "model-00001-of-00001.safetensors",
332
+ "visual.vision_model.encoder.layers.19.mlp.fc1.weight": "model-00001-of-00001.safetensors",
333
+ "visual.vision_model.encoder.layers.19.mlp.fc1.bias": "model-00001-of-00001.safetensors",
334
+ "visual.vision_model.encoder.layers.19.mlp.fc2.weight": "model-00001-of-00001.safetensors",
335
+ "visual.vision_model.encoder.layers.19.mlp.fc2.bias": "model-00001-of-00001.safetensors",
336
+ "visual.vision_model.encoder.layers.20.layer_norm1.weight": "model-00001-of-00001.safetensors",
337
+ "visual.vision_model.encoder.layers.20.layer_norm1.bias": "model-00001-of-00001.safetensors",
338
+ "visual.vision_model.encoder.layers.20.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
339
+ "visual.vision_model.encoder.layers.20.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
340
+ "visual.vision_model.encoder.layers.20.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
341
+ "visual.vision_model.encoder.layers.20.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
342
+ "visual.vision_model.encoder.layers.20.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
343
+ "visual.vision_model.encoder.layers.20.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
344
+ "visual.vision_model.encoder.layers.20.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
345
+ "visual.vision_model.encoder.layers.20.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
346
+ "visual.vision_model.encoder.layers.20.layer_norm2.weight": "model-00001-of-00001.safetensors",
347
+ "visual.vision_model.encoder.layers.20.layer_norm2.bias": "model-00001-of-00001.safetensors",
348
+ "visual.vision_model.encoder.layers.20.mlp.fc1.weight": "model-00001-of-00001.safetensors",
349
+ "visual.vision_model.encoder.layers.20.mlp.fc1.bias": "model-00001-of-00001.safetensors",
350
+ "visual.vision_model.encoder.layers.20.mlp.fc2.weight": "model-00001-of-00001.safetensors",
351
+ "visual.vision_model.encoder.layers.20.mlp.fc2.bias": "model-00001-of-00001.safetensors",
352
+ "visual.vision_model.encoder.layers.21.layer_norm1.weight": "model-00001-of-00001.safetensors",
353
+ "visual.vision_model.encoder.layers.21.layer_norm1.bias": "model-00001-of-00001.safetensors",
354
+ "visual.vision_model.encoder.layers.21.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
355
+ "visual.vision_model.encoder.layers.21.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
356
+ "visual.vision_model.encoder.layers.21.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
357
+ "visual.vision_model.encoder.layers.21.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
358
+ "visual.vision_model.encoder.layers.21.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
359
+ "visual.vision_model.encoder.layers.21.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
360
+ "visual.vision_model.encoder.layers.21.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
361
+ "visual.vision_model.encoder.layers.21.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
362
+ "visual.vision_model.encoder.layers.21.layer_norm2.weight": "model-00001-of-00001.safetensors",
363
+ "visual.vision_model.encoder.layers.21.layer_norm2.bias": "model-00001-of-00001.safetensors",
364
+ "visual.vision_model.encoder.layers.21.mlp.fc1.weight": "model-00001-of-00001.safetensors",
365
+ "visual.vision_model.encoder.layers.21.mlp.fc1.bias": "model-00001-of-00001.safetensors",
366
+ "visual.vision_model.encoder.layers.21.mlp.fc2.weight": "model-00001-of-00001.safetensors",
367
+ "visual.vision_model.encoder.layers.21.mlp.fc2.bias": "model-00001-of-00001.safetensors",
368
+ "visual.vision_model.encoder.layers.22.layer_norm1.weight": "model-00001-of-00001.safetensors",
369
+ "visual.vision_model.encoder.layers.22.layer_norm1.bias": "model-00001-of-00001.safetensors",
370
+ "visual.vision_model.encoder.layers.22.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
371
+ "visual.vision_model.encoder.layers.22.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
372
+ "visual.vision_model.encoder.layers.22.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
373
+ "visual.vision_model.encoder.layers.22.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
374
+ "visual.vision_model.encoder.layers.22.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
375
+ "visual.vision_model.encoder.layers.22.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
376
+ "visual.vision_model.encoder.layers.22.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
377
+ "visual.vision_model.encoder.layers.22.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
378
+ "visual.vision_model.encoder.layers.22.layer_norm2.weight": "model-00001-of-00001.safetensors",
379
+ "visual.vision_model.encoder.layers.22.layer_norm2.bias": "model-00001-of-00001.safetensors",
380
+ "visual.vision_model.encoder.layers.22.mlp.fc1.weight": "model-00001-of-00001.safetensors",
381
+ "visual.vision_model.encoder.layers.22.mlp.fc1.bias": "model-00001-of-00001.safetensors",
382
+ "visual.vision_model.encoder.layers.22.mlp.fc2.weight": "model-00001-of-00001.safetensors",
383
+ "visual.vision_model.encoder.layers.22.mlp.fc2.bias": "model-00001-of-00001.safetensors",
384
+ "visual.vision_model.encoder.layers.23.layer_norm1.weight": "model-00001-of-00001.safetensors",
385
+ "visual.vision_model.encoder.layers.23.layer_norm1.bias": "model-00001-of-00001.safetensors",
386
+ "visual.vision_model.encoder.layers.23.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
387
+ "visual.vision_model.encoder.layers.23.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
388
+ "visual.vision_model.encoder.layers.23.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
389
+ "visual.vision_model.encoder.layers.23.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
390
+ "visual.vision_model.encoder.layers.23.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
391
+ "visual.vision_model.encoder.layers.23.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
392
+ "visual.vision_model.encoder.layers.23.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
393
+ "visual.vision_model.encoder.layers.23.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
394
+ "visual.vision_model.encoder.layers.23.layer_norm2.weight": "model-00001-of-00001.safetensors",
395
+ "visual.vision_model.encoder.layers.23.layer_norm2.bias": "model-00001-of-00001.safetensors",
396
+ "visual.vision_model.encoder.layers.23.mlp.fc1.weight": "model-00001-of-00001.safetensors",
397
+ "visual.vision_model.encoder.layers.23.mlp.fc1.bias": "model-00001-of-00001.safetensors",
398
+ "visual.vision_model.encoder.layers.23.mlp.fc2.weight": "model-00001-of-00001.safetensors",
399
+ "visual.vision_model.encoder.layers.23.mlp.fc2.bias": "model-00001-of-00001.safetensors",
400
+ "visual.vision_model.encoder.layers.24.layer_norm1.weight": "model-00001-of-00001.safetensors",
401
+ "visual.vision_model.encoder.layers.24.layer_norm1.bias": "model-00001-of-00001.safetensors",
402
+ "visual.vision_model.encoder.layers.24.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
403
+ "visual.vision_model.encoder.layers.24.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
404
+ "visual.vision_model.encoder.layers.24.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
405
+ "visual.vision_model.encoder.layers.24.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
406
+ "visual.vision_model.encoder.layers.24.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
407
+ "visual.vision_model.encoder.layers.24.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
408
+ "visual.vision_model.encoder.layers.24.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
409
+ "visual.vision_model.encoder.layers.24.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
410
+ "visual.vision_model.encoder.layers.24.layer_norm2.weight": "model-00001-of-00001.safetensors",
411
+ "visual.vision_model.encoder.layers.24.layer_norm2.bias": "model-00001-of-00001.safetensors",
412
+ "visual.vision_model.encoder.layers.24.mlp.fc1.weight": "model-00001-of-00001.safetensors",
413
+ "visual.vision_model.encoder.layers.24.mlp.fc1.bias": "model-00001-of-00001.safetensors",
414
+ "visual.vision_model.encoder.layers.24.mlp.fc2.weight": "model-00001-of-00001.safetensors",
415
+ "visual.vision_model.encoder.layers.24.mlp.fc2.bias": "model-00001-of-00001.safetensors",
416
+ "visual.vision_model.encoder.layers.25.layer_norm1.weight": "model-00001-of-00001.safetensors",
417
+ "visual.vision_model.encoder.layers.25.layer_norm1.bias": "model-00001-of-00001.safetensors",
418
+ "visual.vision_model.encoder.layers.25.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
419
+ "visual.vision_model.encoder.layers.25.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
420
+ "visual.vision_model.encoder.layers.25.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
421
+ "visual.vision_model.encoder.layers.25.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
422
+ "visual.vision_model.encoder.layers.25.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
423
+ "visual.vision_model.encoder.layers.25.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
424
+ "visual.vision_model.encoder.layers.25.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
425
+ "visual.vision_model.encoder.layers.25.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
426
+ "visual.vision_model.encoder.layers.25.layer_norm2.weight": "model-00001-of-00001.safetensors",
427
+ "visual.vision_model.encoder.layers.25.layer_norm2.bias": "model-00001-of-00001.safetensors",
428
+ "visual.vision_model.encoder.layers.25.mlp.fc1.weight": "model-00001-of-00001.safetensors",
429
+ "visual.vision_model.encoder.layers.25.mlp.fc1.bias": "model-00001-of-00001.safetensors",
430
+ "visual.vision_model.encoder.layers.25.mlp.fc2.weight": "model-00001-of-00001.safetensors",
431
+ "visual.vision_model.encoder.layers.25.mlp.fc2.bias": "model-00001-of-00001.safetensors",
432
+ "visual.vision_model.encoder.layers.26.layer_norm1.weight": "model-00001-of-00001.safetensors",
433
+ "visual.vision_model.encoder.layers.26.layer_norm1.bias": "model-00001-of-00001.safetensors",
434
+ "visual.vision_model.encoder.layers.26.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
435
+ "visual.vision_model.encoder.layers.26.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
436
+ "visual.vision_model.encoder.layers.26.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
437
+ "visual.vision_model.encoder.layers.26.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
438
+ "visual.vision_model.encoder.layers.26.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
439
+ "visual.vision_model.encoder.layers.26.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
440
+ "visual.vision_model.encoder.layers.26.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
441
+ "visual.vision_model.encoder.layers.26.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
442
+ "visual.vision_model.encoder.layers.26.layer_norm2.weight": "model-00001-of-00001.safetensors",
443
+ "visual.vision_model.encoder.layers.26.layer_norm2.bias": "model-00001-of-00001.safetensors",
444
+ "visual.vision_model.encoder.layers.26.mlp.fc1.weight": "model-00001-of-00001.safetensors",
445
+ "visual.vision_model.encoder.layers.26.mlp.fc1.bias": "model-00001-of-00001.safetensors",
446
+ "visual.vision_model.encoder.layers.26.mlp.fc2.weight": "model-00001-of-00001.safetensors",
447
+ "visual.vision_model.encoder.layers.26.mlp.fc2.bias": "model-00001-of-00001.safetensors",
448
+ "visual.vision_model.post_layernorm.weight": "model-00001-of-00001.safetensors",
449
+ "visual.vision_model.post_layernorm.bias": "model-00001-of-00001.safetensors",
450
+ "visual.vision_model.head.probe": "model-00001-of-00001.safetensors",
451
+ "visual.vision_model.head.attention.in_proj_weight": "model-00001-of-00001.safetensors",
452
+ "visual.vision_model.head.attention.in_proj_bias": "model-00001-of-00001.safetensors",
453
+ "visual.vision_model.head.attention.out_proj.weight": "model-00001-of-00001.safetensors",
454
+ "visual.vision_model.head.attention.out_proj.bias": "model-00001-of-00001.safetensors",
455
+ "visual.vision_model.head.layernorm.weight": "model-00001-of-00001.safetensors",
456
+ "visual.vision_model.head.layernorm.bias": "model-00001-of-00001.safetensors",
457
+ "visual.vision_model.head.mlp.fc1.weight": "model-00001-of-00001.safetensors",
458
+ "visual.vision_model.head.mlp.fc1.bias": "model-00001-of-00001.safetensors",
459
+ "visual.vision_model.head.mlp.fc2.weight": "model-00001-of-00001.safetensors",
460
+ "visual.vision_model.head.mlp.fc2.bias": "model-00001-of-00001.safetensors",
461
+ "model.embed_tokens.weight": "model-00001-of-00001.safetensors",
462
+ "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
463
+ "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
464
+ "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
465
+ "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
466
+ "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
467
+ "model.layers.0.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
468
+ "model.layers.0.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
469
+ "model.layers.0.input_layernorm.weight": "model-00001-of-00001.safetensors",
470
+ "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
471
+ "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
472
+ "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
473
+ "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
474
+ "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
475
+ "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
476
+ "model.layers.1.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
477
+ "model.layers.1.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
478
+ "model.layers.1.input_layernorm.weight": "model-00001-of-00001.safetensors",
479
+ "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
480
+ "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
481
+ "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
482
+ "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
483
+ "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
484
+ "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
485
+ "model.layers.2.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
486
+ "model.layers.2.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
487
+ "model.layers.2.input_layernorm.weight": "model-00001-of-00001.safetensors",
488
+ "model.layers.2.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
489
+ "model.layers.3.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
490
+ "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
491
+ "model.layers.3.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
492
+ "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
493
+ "model.layers.3.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
494
+ "model.layers.3.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
495
+ "model.layers.3.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
496
+ "model.layers.3.input_layernorm.weight": "model-00001-of-00001.safetensors",
497
+ "model.layers.3.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
498
+ "model.layers.4.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
499
+ "model.layers.4.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
500
+ "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
501
+ "model.layers.4.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
502
+ "model.layers.4.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
503
+ "model.layers.4.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
504
+ "model.layers.4.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
505
+ "model.layers.4.input_layernorm.weight": "model-00001-of-00001.safetensors",
506
+ "model.layers.4.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
507
+ "model.layers.5.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
508
+ "model.layers.5.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
509
+ "model.layers.5.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
510
+ "model.layers.5.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
511
+ "model.layers.5.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
512
+ "model.layers.5.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
513
+ "model.layers.5.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
514
+ "model.layers.5.input_layernorm.weight": "model-00001-of-00001.safetensors",
515
+ "model.layers.5.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
516
+ "model.layers.6.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
517
+ "model.layers.6.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
518
+ "model.layers.6.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
519
+ "model.layers.6.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
520
+ "model.layers.6.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
521
+ "model.layers.6.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
522
+ "model.layers.6.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
523
+ "model.layers.6.input_layernorm.weight": "model-00001-of-00001.safetensors",
524
+ "model.layers.6.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
525
+ "model.layers.7.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
526
+ "model.layers.7.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
527
+ "model.layers.7.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
528
+ "model.layers.7.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
529
+ "model.layers.7.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
530
+ "model.layers.7.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
531
+ "model.layers.7.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
532
+ "model.layers.7.input_layernorm.weight": "model-00001-of-00001.safetensors",
533
+ "model.layers.7.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
534
+ "model.layers.8.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
535
+ "model.layers.8.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
536
+ "model.layers.8.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
537
+ "model.layers.8.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
538
+ "model.layers.8.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
539
+ "model.layers.8.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
540
+ "model.layers.8.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
541
+ "model.layers.8.input_layernorm.weight": "model-00001-of-00001.safetensors",
542
+ "model.layers.8.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
543
+ "model.layers.9.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
544
+ "model.layers.9.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
545
+ "model.layers.9.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
546
+ "model.layers.9.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
547
+ "model.layers.9.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
548
+ "model.layers.9.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
549
+ "model.layers.9.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
550
+ "model.layers.9.input_layernorm.weight": "model-00001-of-00001.safetensors",
551
+ "model.layers.9.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
552
+ "model.layers.10.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
553
+ "model.layers.10.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
554
+ "model.layers.10.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
555
+ "model.layers.10.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
556
+ "model.layers.10.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
557
+ "model.layers.10.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
558
+ "model.layers.10.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
559
+ "model.layers.10.input_layernorm.weight": "model-00001-of-00001.safetensors",
560
+ "model.layers.10.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
561
+ "model.layers.11.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
562
+ "model.layers.11.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
563
+ "model.layers.11.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
564
+ "model.layers.11.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
565
+ "model.layers.11.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
566
+ "model.layers.11.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
567
+ "model.layers.11.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
568
+ "model.layers.11.input_layernorm.weight": "model-00001-of-00001.safetensors",
569
+ "model.layers.11.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
570
+ "model.layers.12.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
571
+ "model.layers.12.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
572
+ "model.layers.12.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
573
+ "model.layers.12.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
574
+ "model.layers.12.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
575
+ "model.layers.12.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
576
+ "model.layers.12.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
577
+ "model.layers.12.input_layernorm.weight": "model-00001-of-00001.safetensors",
578
+ "model.layers.12.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
579
+ "model.layers.13.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
580
+ "model.layers.13.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
581
+ "model.layers.13.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
582
+ "model.layers.13.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
583
+ "model.layers.13.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
584
+ "model.layers.13.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
585
+ "model.layers.13.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
586
+ "model.layers.13.input_layernorm.weight": "model-00001-of-00001.safetensors",
587
+ "model.layers.13.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
588
+ "model.layers.14.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
589
+ "model.layers.14.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
590
+ "model.layers.14.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
591
+ "model.layers.14.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
592
+ "model.layers.14.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
593
+ "model.layers.14.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
594
+ "model.layers.14.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
595
+ "model.layers.14.input_layernorm.weight": "model-00001-of-00001.safetensors",
596
+ "model.layers.14.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
597
+ "model.layers.15.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
598
+ "model.layers.15.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
599
+ "model.layers.15.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
600
+ "model.layers.15.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
601
+ "model.layers.15.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
602
+ "model.layers.15.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
603
+ "model.layers.15.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
604
+ "model.layers.15.input_layernorm.weight": "model-00001-of-00001.safetensors",
605
+ "model.layers.15.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
606
+ "model.layers.16.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
607
+ "model.layers.16.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
608
+ "model.layers.16.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
609
+ "model.layers.16.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
610
+ "model.layers.16.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
611
+ "model.layers.16.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
612
+ "model.layers.16.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
613
+ "model.layers.16.input_layernorm.weight": "model-00001-of-00001.safetensors",
614
+ "model.layers.16.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
615
+ "model.layers.17.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
616
+ "model.layers.17.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
617
+ "model.layers.17.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
618
+ "model.layers.17.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
619
+ "model.layers.17.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
620
+ "model.layers.17.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
621
+ "model.layers.17.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
622
+ "model.layers.17.input_layernorm.weight": "model-00001-of-00001.safetensors",
623
+ "model.layers.17.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
624
+ "model.norm.weight": "model-00001-of-00001.safetensors",
625
+ "lm_head.weight": "model-00001-of-00001.safetensors"
626
+ }
627
+ }
checkpoint-336/optimizer-00001-of-00001.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:09a0535d83eb6a76d79e74bd6c9f2b380e6d5b1d4f8499f0a666f0d82260720a
3
+ size 7245150152
checkpoint-336/optimizer.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-336/preprocessor_config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_convert_rgb": true,
3
+ "do_normalize": true,
4
+ "do_rescale": true,
5
+ "do_resize": true,
6
+ "image_mean": [
7
+ 0.5,
8
+ 0.5,
9
+ 0.5
10
+ ],
11
+ "image_processor_type": "SiglipImageProcessor",
12
+ "image_std": [
13
+ 0.5,
14
+ 0.5,
15
+ 0.5
16
+ ],
17
+ "max_pixels": 2822400,
18
+ "merge_size": 2,
19
+ "min_pixels": 147384,
20
+ "patch_size": 14,
21
+ "resample": 3,
22
+ "rescale_factor": 0.00392156862745098,
23
+ "size": {
24
+ "max_pixels": 2822400,
25
+ "min_pixels": 147384
26
+ },
27
+ "temporal_conv_size": 1,
28
+ "temporal_patch_size": 1
29
+ }
checkpoint-336/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:29cde6478b49d2682dfa8c033ed7da4f33ab76ccd1941cfe950453d3923dfec3
3
+ size 32008
checkpoint-336/scheduler.pdparams ADDED
Binary file (51 Bytes). View file
 
checkpoint-336/special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": false, "special": false}, "eos_token": "</s>", "unk_token": {"content": "<unk>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": false, "special": false}, "sep_token": "<|end_of_sentence|>", "pad_token": {"content": "<unk>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": false, "special": false}, "cls_token": "<|begin_of_sentence|>", "mask_token": {"content": "<mask:1>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": false, "special": false}, "additional_special_tokens": ["<|IMAGE_PLACEHOLDER|>", "<|image_pad|>", "<|IMAGE_START|>", "<|IMAGE_END|>", "<|video_pad|>"]}
checkpoint-336/static_name_to_dyg_name.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"layer_norm_0.w_0": "mlp_AR.pre_norm.weight", "layer_norm_0.b_0": "mlp_AR.pre_norm.bias", "linear_0.w_0": "mlp_AR.linear_1.weight", "linear_0.b_0": "mlp_AR.linear_1.bias", "linear_1.w_0": "mlp_AR.linear_2.weight", "linear_1.b_0": "mlp_AR.linear_2.bias", "conv2d_0.w_0": "visual.vision_model.embeddings.patch_embedding.weight", "conv2d_0.b_0": "visual.vision_model.embeddings.patch_embedding.bias", "embedding_0.w_0": "visual.vision_model.embeddings.position_embedding.weight", "embedding_1.w_0": "visual.vision_model.embeddings.packing_position_embedding.weight", "layer_norm_1.w_0": "visual.vision_model.encoder.layers.0.layer_norm1.weight", "layer_norm_1.b_0": "visual.vision_model.encoder.layers.0.layer_norm1.bias", "linear_2.w_0": "visual.vision_model.encoder.layers.0.self_attn.k_proj.weight", "linear_2.b_0": "visual.vision_model.encoder.layers.0.self_attn.k_proj.bias", "linear_3.w_0": "visual.vision_model.encoder.layers.0.self_attn.v_proj.weight", "linear_3.b_0": "visual.vision_model.encoder.layers.0.self_attn.v_proj.bias", "linear_4.w_0": "visual.vision_model.encoder.layers.0.self_attn.q_proj.weight", "linear_4.b_0": "visual.vision_model.encoder.layers.0.self_attn.q_proj.bias", "linear_5.w_0": "visual.vision_model.encoder.layers.0.self_attn.out_proj.weight", "linear_5.b_0": "visual.vision_model.encoder.layers.0.self_attn.out_proj.bias", "layer_norm_2.w_0": "visual.vision_model.encoder.layers.0.layer_norm2.weight", "layer_norm_2.b_0": "visual.vision_model.encoder.layers.0.layer_norm2.bias", "linear_6.w_0": "visual.vision_model.encoder.layers.0.mlp.fc1.weight", "linear_6.b_0": "visual.vision_model.encoder.layers.0.mlp.fc1.bias", "linear_7.w_0": "visual.vision_model.encoder.layers.0.mlp.fc2.weight", "linear_7.b_0": "visual.vision_model.encoder.layers.0.mlp.fc2.bias", "layer_norm_3.w_0": "visual.vision_model.encoder.layers.1.layer_norm1.weight", "layer_norm_3.b_0": "visual.vision_model.encoder.layers.1.layer_norm1.bias", "linear_8.w_0": "visual.vision_model.encoder.layers.1.self_attn.k_proj.weight", "linear_8.b_0": "visual.vision_model.encoder.layers.1.self_attn.k_proj.bias", "linear_9.w_0": "visual.vision_model.encoder.layers.1.self_attn.v_proj.weight", "linear_9.b_0": "visual.vision_model.encoder.layers.1.self_attn.v_proj.bias", "linear_10.w_0": "visual.vision_model.encoder.layers.1.self_attn.q_proj.weight", "linear_10.b_0": "visual.vision_model.encoder.layers.1.self_attn.q_proj.bias", "linear_11.w_0": "visual.vision_model.encoder.layers.1.self_attn.out_proj.weight", "linear_11.b_0": "visual.vision_model.encoder.layers.1.self_attn.out_proj.bias", "layer_norm_4.w_0": "visual.vision_model.encoder.layers.1.layer_norm2.weight", "layer_norm_4.b_0": "visual.vision_model.encoder.layers.1.layer_norm2.bias", "linear_12.w_0": "visual.vision_model.encoder.layers.1.mlp.fc1.weight", "linear_12.b_0": "visual.vision_model.encoder.layers.1.mlp.fc1.bias", "linear_13.w_0": "visual.vision_model.encoder.layers.1.mlp.fc2.weight", "linear_13.b_0": "visual.vision_model.encoder.layers.1.mlp.fc2.bias", "layer_norm_5.w_0": "visual.vision_model.encoder.layers.2.layer_norm1.weight", "layer_norm_5.b_0": "visual.vision_model.encoder.layers.2.layer_norm1.bias", "linear_14.w_0": "visual.vision_model.encoder.layers.2.self_attn.k_proj.weight", "linear_14.b_0": "visual.vision_model.encoder.layers.2.self_attn.k_proj.bias", "linear_15.w_0": "visual.vision_model.encoder.layers.2.self_attn.v_proj.weight", "linear_15.b_0": "visual.vision_model.encoder.layers.2.self_attn.v_proj.bias", "linear_16.w_0": "visual.vision_model.encoder.layers.2.self_attn.q_proj.weight", "linear_16.b_0": "visual.vision_model.encoder.layers.2.self_attn.q_proj.bias", "linear_17.w_0": "visual.vision_model.encoder.layers.2.self_attn.out_proj.weight", "linear_17.b_0": "visual.vision_model.encoder.layers.2.self_attn.out_proj.bias", "layer_norm_6.w_0": "visual.vision_model.encoder.layers.2.layer_norm2.weight", "layer_norm_6.b_0": "visual.vision_model.encoder.layers.2.layer_norm2.bias", "linear_18.w_0": "visual.vision_model.encoder.layers.2.mlp.fc1.weight", "linear_18.b_0": "visual.vision_model.encoder.layers.2.mlp.fc1.bias", "linear_19.w_0": "visual.vision_model.encoder.layers.2.mlp.fc2.weight", "linear_19.b_0": "visual.vision_model.encoder.layers.2.mlp.fc2.bias", "layer_norm_7.w_0": "visual.vision_model.encoder.layers.3.layer_norm1.weight", "layer_norm_7.b_0": "visual.vision_model.encoder.layers.3.layer_norm1.bias", "linear_20.w_0": "visual.vision_model.encoder.layers.3.self_attn.k_proj.weight", "linear_20.b_0": "visual.vision_model.encoder.layers.3.self_attn.k_proj.bias", "linear_21.w_0": "visual.vision_model.encoder.layers.3.self_attn.v_proj.weight", "linear_21.b_0": "visual.vision_model.encoder.layers.3.self_attn.v_proj.bias", "linear_22.w_0": "visual.vision_model.encoder.layers.3.self_attn.q_proj.weight", "linear_22.b_0": "visual.vision_model.encoder.layers.3.self_attn.q_proj.bias", "linear_23.w_0": "visual.vision_model.encoder.layers.3.self_attn.out_proj.weight", "linear_23.b_0": "visual.vision_model.encoder.layers.3.self_attn.out_proj.bias", "layer_norm_8.w_0": "visual.vision_model.encoder.layers.3.layer_norm2.weight", "layer_norm_8.b_0": "visual.vision_model.encoder.layers.3.layer_norm2.bias", "linear_24.w_0": "visual.vision_model.encoder.layers.3.mlp.fc1.weight", "linear_24.b_0": "visual.vision_model.encoder.layers.3.mlp.fc1.bias", "linear_25.w_0": "visual.vision_model.encoder.layers.3.mlp.fc2.weight", "linear_25.b_0": "visual.vision_model.encoder.layers.3.mlp.fc2.bias", "layer_norm_9.w_0": "visual.vision_model.encoder.layers.4.layer_norm1.weight", "layer_norm_9.b_0": "visual.vision_model.encoder.layers.4.layer_norm1.bias", "linear_26.w_0": "visual.vision_model.encoder.layers.4.self_attn.k_proj.weight", "linear_26.b_0": "visual.vision_model.encoder.layers.4.self_attn.k_proj.bias", "linear_27.w_0": "visual.vision_model.encoder.layers.4.self_attn.v_proj.weight", "linear_27.b_0": "visual.vision_model.encoder.layers.4.self_attn.v_proj.bias", "linear_28.w_0": "visual.vision_model.encoder.layers.4.self_attn.q_proj.weight", "linear_28.b_0": "visual.vision_model.encoder.layers.4.self_attn.q_proj.bias", "linear_29.w_0": "visual.vision_model.encoder.layers.4.self_attn.out_proj.weight", "linear_29.b_0": "visual.vision_model.encoder.layers.4.self_attn.out_proj.bias", "layer_norm_10.w_0": "visual.vision_model.encoder.layers.4.layer_norm2.weight", "layer_norm_10.b_0": "visual.vision_model.encoder.layers.4.layer_norm2.bias", "linear_30.w_0": "visual.vision_model.encoder.layers.4.mlp.fc1.weight", "linear_30.b_0": "visual.vision_model.encoder.layers.4.mlp.fc1.bias", "linear_31.w_0": "visual.vision_model.encoder.layers.4.mlp.fc2.weight", "linear_31.b_0": "visual.vision_model.encoder.layers.4.mlp.fc2.bias", "layer_norm_11.w_0": "visual.vision_model.encoder.layers.5.layer_norm1.weight", "layer_norm_11.b_0": "visual.vision_model.encoder.layers.5.layer_norm1.bias", "linear_32.w_0": "visual.vision_model.encoder.layers.5.self_attn.k_proj.weight", "linear_32.b_0": "visual.vision_model.encoder.layers.5.self_attn.k_proj.bias", "linear_33.w_0": "visual.vision_model.encoder.layers.5.self_attn.v_proj.weight", "linear_33.b_0": "visual.vision_model.encoder.layers.5.self_attn.v_proj.bias", "linear_34.w_0": "visual.vision_model.encoder.layers.5.self_attn.q_proj.weight", "linear_34.b_0": "visual.vision_model.encoder.layers.5.self_attn.q_proj.bias", "linear_35.w_0": "visual.vision_model.encoder.layers.5.self_attn.out_proj.weight", "linear_35.b_0": "visual.vision_model.encoder.layers.5.self_attn.out_proj.bias", "layer_norm_12.w_0": "visual.vision_model.encoder.layers.5.layer_norm2.weight", "layer_norm_12.b_0": "visual.vision_model.encoder.layers.5.layer_norm2.bias", "linear_36.w_0": "visual.vision_model.encoder.layers.5.mlp.fc1.weight", "linear_36.b_0": "visual.vision_model.encoder.layers.5.mlp.fc1.bias", "linear_37.w_0": "visual.vision_model.encoder.layers.5.mlp.fc2.weight", "linear_37.b_0": "visual.vision_model.encoder.layers.5.mlp.fc2.bias", "layer_norm_13.w_0": "visual.vision_model.encoder.layers.6.layer_norm1.weight", "layer_norm_13.b_0": "visual.vision_model.encoder.layers.6.layer_norm1.bias", "linear_38.w_0": "visual.vision_model.encoder.layers.6.self_attn.k_proj.weight", "linear_38.b_0": "visual.vision_model.encoder.layers.6.self_attn.k_proj.bias", "linear_39.w_0": "visual.vision_model.encoder.layers.6.self_attn.v_proj.weight", "linear_39.b_0": "visual.vision_model.encoder.layers.6.self_attn.v_proj.bias", "linear_40.w_0": "visual.vision_model.encoder.layers.6.self_attn.q_proj.weight", "linear_40.b_0": "visual.vision_model.encoder.layers.6.self_attn.q_proj.bias", "linear_41.w_0": "visual.vision_model.encoder.layers.6.self_attn.out_proj.weight", "linear_41.b_0": "visual.vision_model.encoder.layers.6.self_attn.out_proj.bias", "layer_norm_14.w_0": "visual.vision_model.encoder.layers.6.layer_norm2.weight", "layer_norm_14.b_0": "visual.vision_model.encoder.layers.6.layer_norm2.bias", "linear_42.w_0": "visual.vision_model.encoder.layers.6.mlp.fc1.weight", "linear_42.b_0": "visual.vision_model.encoder.layers.6.mlp.fc1.bias", "linear_43.w_0": "visual.vision_model.encoder.layers.6.mlp.fc2.weight", "linear_43.b_0": "visual.vision_model.encoder.layers.6.mlp.fc2.bias", "layer_norm_15.w_0": "visual.vision_model.encoder.layers.7.layer_norm1.weight", "layer_norm_15.b_0": "visual.vision_model.encoder.layers.7.layer_norm1.bias", "linear_44.w_0": "visual.vision_model.encoder.layers.7.self_attn.k_proj.weight", "linear_44.b_0": "visual.vision_model.encoder.layers.7.self_attn.k_proj.bias", "linear_45.w_0": "visual.vision_model.encoder.layers.7.self_attn.v_proj.weight", "linear_45.b_0": "visual.vision_model.encoder.layers.7.self_attn.v_proj.bias", "linear_46.w_0": "visual.vision_model.encoder.layers.7.self_attn.q_proj.weight", "linear_46.b_0": "visual.vision_model.encoder.layers.7.self_attn.q_proj.bias", "linear_47.w_0": "visual.vision_model.encoder.layers.7.self_attn.out_proj.weight", "linear_47.b_0": "visual.vision_model.encoder.layers.7.self_attn.out_proj.bias", "layer_norm_16.w_0": "visual.vision_model.encoder.layers.7.layer_norm2.weight", "layer_norm_16.b_0": "visual.vision_model.encoder.layers.7.layer_norm2.bias", "linear_48.w_0": "visual.vision_model.encoder.layers.7.mlp.fc1.weight", "linear_48.b_0": "visual.vision_model.encoder.layers.7.mlp.fc1.bias", "linear_49.w_0": "visual.vision_model.encoder.layers.7.mlp.fc2.weight", "linear_49.b_0": "visual.vision_model.encoder.layers.7.mlp.fc2.bias", "layer_norm_17.w_0": "visual.vision_model.encoder.layers.8.layer_norm1.weight", "layer_norm_17.b_0": "visual.vision_model.encoder.layers.8.layer_norm1.bias", "linear_50.w_0": "visual.vision_model.encoder.layers.8.self_attn.k_proj.weight", "linear_50.b_0": "visual.vision_model.encoder.layers.8.self_attn.k_proj.bias", "linear_51.w_0": "visual.vision_model.encoder.layers.8.self_attn.v_proj.weight", "linear_51.b_0": "visual.vision_model.encoder.layers.8.self_attn.v_proj.bias", "linear_52.w_0": "visual.vision_model.encoder.layers.8.self_attn.q_proj.weight", "linear_52.b_0": "visual.vision_model.encoder.layers.8.self_attn.q_proj.bias", "linear_53.w_0": "visual.vision_model.encoder.layers.8.self_attn.out_proj.weight", "linear_53.b_0": "visual.vision_model.encoder.layers.8.self_attn.out_proj.bias", "layer_norm_18.w_0": "visual.vision_model.encoder.layers.8.layer_norm2.weight", "layer_norm_18.b_0": "visual.vision_model.encoder.layers.8.layer_norm2.bias", "linear_54.w_0": "visual.vision_model.encoder.layers.8.mlp.fc1.weight", "linear_54.b_0": "visual.vision_model.encoder.layers.8.mlp.fc1.bias", "linear_55.w_0": "visual.vision_model.encoder.layers.8.mlp.fc2.weight", "linear_55.b_0": "visual.vision_model.encoder.layers.8.mlp.fc2.bias", "layer_norm_19.w_0": "visual.vision_model.encoder.layers.9.layer_norm1.weight", "layer_norm_19.b_0": "visual.vision_model.encoder.layers.9.layer_norm1.bias", "linear_56.w_0": "visual.vision_model.encoder.layers.9.self_attn.k_proj.weight", "linear_56.b_0": "visual.vision_model.encoder.layers.9.self_attn.k_proj.bias", "linear_57.w_0": "visual.vision_model.encoder.layers.9.self_attn.v_proj.weight", "linear_57.b_0": "visual.vision_model.encoder.layers.9.self_attn.v_proj.bias", "linear_58.w_0": "visual.vision_model.encoder.layers.9.self_attn.q_proj.weight", "linear_58.b_0": "visual.vision_model.encoder.layers.9.self_attn.q_proj.bias", "linear_59.w_0": "visual.vision_model.encoder.layers.9.self_attn.out_proj.weight", "linear_59.b_0": "visual.vision_model.encoder.layers.9.self_attn.out_proj.bias", "layer_norm_20.w_0": "visual.vision_model.encoder.layers.9.layer_norm2.weight", "layer_norm_20.b_0": "visual.vision_model.encoder.layers.9.layer_norm2.bias", "linear_60.w_0": "visual.vision_model.encoder.layers.9.mlp.fc1.weight", "linear_60.b_0": "visual.vision_model.encoder.layers.9.mlp.fc1.bias", "linear_61.w_0": "visual.vision_model.encoder.layers.9.mlp.fc2.weight", "linear_61.b_0": "visual.vision_model.encoder.layers.9.mlp.fc2.bias", "layer_norm_21.w_0": "visual.vision_model.encoder.layers.10.layer_norm1.weight", "layer_norm_21.b_0": "visual.vision_model.encoder.layers.10.layer_norm1.bias", "linear_62.w_0": "visual.vision_model.encoder.layers.10.self_attn.k_proj.weight", "linear_62.b_0": "visual.vision_model.encoder.layers.10.self_attn.k_proj.bias", "linear_63.w_0": "visual.vision_model.encoder.layers.10.self_attn.v_proj.weight", "linear_63.b_0": "visual.vision_model.encoder.layers.10.self_attn.v_proj.bias", "linear_64.w_0": "visual.vision_model.encoder.layers.10.self_attn.q_proj.weight", "linear_64.b_0": "visual.vision_model.encoder.layers.10.self_attn.q_proj.bias", "linear_65.w_0": "visual.vision_model.encoder.layers.10.self_attn.out_proj.weight", "linear_65.b_0": "visual.vision_model.encoder.layers.10.self_attn.out_proj.bias", "layer_norm_22.w_0": "visual.vision_model.encoder.layers.10.layer_norm2.weight", "layer_norm_22.b_0": "visual.vision_model.encoder.layers.10.layer_norm2.bias", "linear_66.w_0": "visual.vision_model.encoder.layers.10.mlp.fc1.weight", "linear_66.b_0": "visual.vision_model.encoder.layers.10.mlp.fc1.bias", "linear_67.w_0": "visual.vision_model.encoder.layers.10.mlp.fc2.weight", "linear_67.b_0": "visual.vision_model.encoder.layers.10.mlp.fc2.bias", "layer_norm_23.w_0": "visual.vision_model.encoder.layers.11.layer_norm1.weight", "layer_norm_23.b_0": "visual.vision_model.encoder.layers.11.layer_norm1.bias", "linear_68.w_0": "visual.vision_model.encoder.layers.11.self_attn.k_proj.weight", "linear_68.b_0": "visual.vision_model.encoder.layers.11.self_attn.k_proj.bias", "linear_69.w_0": "visual.vision_model.encoder.layers.11.self_attn.v_proj.weight", "linear_69.b_0": "visual.vision_model.encoder.layers.11.self_attn.v_proj.bias", "linear_70.w_0": "visual.vision_model.encoder.layers.11.self_attn.q_proj.weight", "linear_70.b_0": "visual.vision_model.encoder.layers.11.self_attn.q_proj.bias", "linear_71.w_0": "visual.vision_model.encoder.layers.11.self_attn.out_proj.weight", "linear_71.b_0": "visual.vision_model.encoder.layers.11.self_attn.out_proj.bias", "layer_norm_24.w_0": "visual.vision_model.encoder.layers.11.layer_norm2.weight", "layer_norm_24.b_0": "visual.vision_model.encoder.layers.11.layer_norm2.bias", "linear_72.w_0": "visual.vision_model.encoder.layers.11.mlp.fc1.weight", "linear_72.b_0": "visual.vision_model.encoder.layers.11.mlp.fc1.bias", "linear_73.w_0": "visual.vision_model.encoder.layers.11.mlp.fc2.weight", "linear_73.b_0": "visual.vision_model.encoder.layers.11.mlp.fc2.bias", "layer_norm_25.w_0": "visual.vision_model.encoder.layers.12.layer_norm1.weight", "layer_norm_25.b_0": "visual.vision_model.encoder.layers.12.layer_norm1.bias", "linear_74.w_0": "visual.vision_model.encoder.layers.12.self_attn.k_proj.weight", "linear_74.b_0": "visual.vision_model.encoder.layers.12.self_attn.k_proj.bias", "linear_75.w_0": "visual.vision_model.encoder.layers.12.self_attn.v_proj.weight", "linear_75.b_0": "visual.vision_model.encoder.layers.12.self_attn.v_proj.bias", "linear_76.w_0": "visual.vision_model.encoder.layers.12.self_attn.q_proj.weight", "linear_76.b_0": "visual.vision_model.encoder.layers.12.self_attn.q_proj.bias", "linear_77.w_0": "visual.vision_model.encoder.layers.12.self_attn.out_proj.weight", "linear_77.b_0": "visual.vision_model.encoder.layers.12.self_attn.out_proj.bias", "layer_norm_26.w_0": "visual.vision_model.encoder.layers.12.layer_norm2.weight", "layer_norm_26.b_0": "visual.vision_model.encoder.layers.12.layer_norm2.bias", "linear_78.w_0": "visual.vision_model.encoder.layers.12.mlp.fc1.weight", "linear_78.b_0": "visual.vision_model.encoder.layers.12.mlp.fc1.bias", "linear_79.w_0": "visual.vision_model.encoder.layers.12.mlp.fc2.weight", "linear_79.b_0": "visual.vision_model.encoder.layers.12.mlp.fc2.bias", "layer_norm_27.w_0": "visual.vision_model.encoder.layers.13.layer_norm1.weight", "layer_norm_27.b_0": "visual.vision_model.encoder.layers.13.layer_norm1.bias", "linear_80.w_0": "visual.vision_model.encoder.layers.13.self_attn.k_proj.weight", "linear_80.b_0": "visual.vision_model.encoder.layers.13.self_attn.k_proj.bias", "linear_81.w_0": "visual.vision_model.encoder.layers.13.self_attn.v_proj.weight", "linear_81.b_0": "visual.vision_model.encoder.layers.13.self_attn.v_proj.bias", "linear_82.w_0": "visual.vision_model.encoder.layers.13.self_attn.q_proj.weight", "linear_82.b_0": "visual.vision_model.encoder.layers.13.self_attn.q_proj.bias", "linear_83.w_0": "visual.vision_model.encoder.layers.13.self_attn.out_proj.weight", "linear_83.b_0": "visual.vision_model.encoder.layers.13.self_attn.out_proj.bias", "layer_norm_28.w_0": "visual.vision_model.encoder.layers.13.layer_norm2.weight", "layer_norm_28.b_0": "visual.vision_model.encoder.layers.13.layer_norm2.bias", "linear_84.w_0": "visual.vision_model.encoder.layers.13.mlp.fc1.weight", "linear_84.b_0": "visual.vision_model.encoder.layers.13.mlp.fc1.bias", "linear_85.w_0": "visual.vision_model.encoder.layers.13.mlp.fc2.weight", "linear_85.b_0": "visual.vision_model.encoder.layers.13.mlp.fc2.bias", "layer_norm_29.w_0": "visual.vision_model.encoder.layers.14.layer_norm1.weight", "layer_norm_29.b_0": "visual.vision_model.encoder.layers.14.layer_norm1.bias", "linear_86.w_0": "visual.vision_model.encoder.layers.14.self_attn.k_proj.weight", "linear_86.b_0": "visual.vision_model.encoder.layers.14.self_attn.k_proj.bias", "linear_87.w_0": "visual.vision_model.encoder.layers.14.self_attn.v_proj.weight", "linear_87.b_0": "visual.vision_model.encoder.layers.14.self_attn.v_proj.bias", "linear_88.w_0": "visual.vision_model.encoder.layers.14.self_attn.q_proj.weight", "linear_88.b_0": "visual.vision_model.encoder.layers.14.self_attn.q_proj.bias", "linear_89.w_0": "visual.vision_model.encoder.layers.14.self_attn.out_proj.weight", "linear_89.b_0": "visual.vision_model.encoder.layers.14.self_attn.out_proj.bias", "layer_norm_30.w_0": "visual.vision_model.encoder.layers.14.layer_norm2.weight", "layer_norm_30.b_0": "visual.vision_model.encoder.layers.14.layer_norm2.bias", "linear_90.w_0": "visual.vision_model.encoder.layers.14.mlp.fc1.weight", "linear_90.b_0": "visual.vision_model.encoder.layers.14.mlp.fc1.bias", "linear_91.w_0": "visual.vision_model.encoder.layers.14.mlp.fc2.weight", "linear_91.b_0": "visual.vision_model.encoder.layers.14.mlp.fc2.bias", "layer_norm_31.w_0": "visual.vision_model.encoder.layers.15.layer_norm1.weight", "layer_norm_31.b_0": "visual.vision_model.encoder.layers.15.layer_norm1.bias", "linear_92.w_0": "visual.vision_model.encoder.layers.15.self_attn.k_proj.weight", "linear_92.b_0": "visual.vision_model.encoder.layers.15.self_attn.k_proj.bias", "linear_93.w_0": "visual.vision_model.encoder.layers.15.self_attn.v_proj.weight", "linear_93.b_0": "visual.vision_model.encoder.layers.15.self_attn.v_proj.bias", "linear_94.w_0": "visual.vision_model.encoder.layers.15.self_attn.q_proj.weight", "linear_94.b_0": "visual.vision_model.encoder.layers.15.self_attn.q_proj.bias", "linear_95.w_0": "visual.vision_model.encoder.layers.15.self_attn.out_proj.weight", "linear_95.b_0": "visual.vision_model.encoder.layers.15.self_attn.out_proj.bias", "layer_norm_32.w_0": "visual.vision_model.encoder.layers.15.layer_norm2.weight", "layer_norm_32.b_0": "visual.vision_model.encoder.layers.15.layer_norm2.bias", "linear_96.w_0": "visual.vision_model.encoder.layers.15.mlp.fc1.weight", "linear_96.b_0": "visual.vision_model.encoder.layers.15.mlp.fc1.bias", "linear_97.w_0": "visual.vision_model.encoder.layers.15.mlp.fc2.weight", "linear_97.b_0": "visual.vision_model.encoder.layers.15.mlp.fc2.bias", "layer_norm_33.w_0": "visual.vision_model.encoder.layers.16.layer_norm1.weight", "layer_norm_33.b_0": "visual.vision_model.encoder.layers.16.layer_norm1.bias", "linear_98.w_0": "visual.vision_model.encoder.layers.16.self_attn.k_proj.weight", "linear_98.b_0": "visual.vision_model.encoder.layers.16.self_attn.k_proj.bias", "linear_99.w_0": "visual.vision_model.encoder.layers.16.self_attn.v_proj.weight", "linear_99.b_0": "visual.vision_model.encoder.layers.16.self_attn.v_proj.bias", "linear_100.w_0": "visual.vision_model.encoder.layers.16.self_attn.q_proj.weight", "linear_100.b_0": "visual.vision_model.encoder.layers.16.self_attn.q_proj.bias", "linear_101.w_0": "visual.vision_model.encoder.layers.16.self_attn.out_proj.weight", "linear_101.b_0": "visual.vision_model.encoder.layers.16.self_attn.out_proj.bias", "layer_norm_34.w_0": "visual.vision_model.encoder.layers.16.layer_norm2.weight", "layer_norm_34.b_0": "visual.vision_model.encoder.layers.16.layer_norm2.bias", "linear_102.w_0": "visual.vision_model.encoder.layers.16.mlp.fc1.weight", "linear_102.b_0": "visual.vision_model.encoder.layers.16.mlp.fc1.bias", "linear_103.w_0": "visual.vision_model.encoder.layers.16.mlp.fc2.weight", "linear_103.b_0": "visual.vision_model.encoder.layers.16.mlp.fc2.bias", "layer_norm_35.w_0": "visual.vision_model.encoder.layers.17.layer_norm1.weight", "layer_norm_35.b_0": "visual.vision_model.encoder.layers.17.layer_norm1.bias", "linear_104.w_0": "visual.vision_model.encoder.layers.17.self_attn.k_proj.weight", "linear_104.b_0": "visual.vision_model.encoder.layers.17.self_attn.k_proj.bias", "linear_105.w_0": "visual.vision_model.encoder.layers.17.self_attn.v_proj.weight", "linear_105.b_0": "visual.vision_model.encoder.layers.17.self_attn.v_proj.bias", "linear_106.w_0": "visual.vision_model.encoder.layers.17.self_attn.q_proj.weight", "linear_106.b_0": "visual.vision_model.encoder.layers.17.self_attn.q_proj.bias", "linear_107.w_0": "visual.vision_model.encoder.layers.17.self_attn.out_proj.weight", "linear_107.b_0": "visual.vision_model.encoder.layers.17.self_attn.out_proj.bias", "layer_norm_36.w_0": "visual.vision_model.encoder.layers.17.layer_norm2.weight", "layer_norm_36.b_0": "visual.vision_model.encoder.layers.17.layer_norm2.bias", "linear_108.w_0": "visual.vision_model.encoder.layers.17.mlp.fc1.weight", "linear_108.b_0": "visual.vision_model.encoder.layers.17.mlp.fc1.bias", "linear_109.w_0": "visual.vision_model.encoder.layers.17.mlp.fc2.weight", "linear_109.b_0": "visual.vision_model.encoder.layers.17.mlp.fc2.bias", "layer_norm_37.w_0": "visual.vision_model.encoder.layers.18.layer_norm1.weight", "layer_norm_37.b_0": "visual.vision_model.encoder.layers.18.layer_norm1.bias", "linear_110.w_0": "visual.vision_model.encoder.layers.18.self_attn.k_proj.weight", "linear_110.b_0": "visual.vision_model.encoder.layers.18.self_attn.k_proj.bias", "linear_111.w_0": "visual.vision_model.encoder.layers.18.self_attn.v_proj.weight", "linear_111.b_0": "visual.vision_model.encoder.layers.18.self_attn.v_proj.bias", "linear_112.w_0": "visual.vision_model.encoder.layers.18.self_attn.q_proj.weight", "linear_112.b_0": "visual.vision_model.encoder.layers.18.self_attn.q_proj.bias", "linear_113.w_0": "visual.vision_model.encoder.layers.18.self_attn.out_proj.weight", "linear_113.b_0": "visual.vision_model.encoder.layers.18.self_attn.out_proj.bias", "layer_norm_38.w_0": "visual.vision_model.encoder.layers.18.layer_norm2.weight", "layer_norm_38.b_0": "visual.vision_model.encoder.layers.18.layer_norm2.bias", "linear_114.w_0": "visual.vision_model.encoder.layers.18.mlp.fc1.weight", "linear_114.b_0": "visual.vision_model.encoder.layers.18.mlp.fc1.bias", "linear_115.w_0": "visual.vision_model.encoder.layers.18.mlp.fc2.weight", "linear_115.b_0": "visual.vision_model.encoder.layers.18.mlp.fc2.bias", "layer_norm_39.w_0": "visual.vision_model.encoder.layers.19.layer_norm1.weight", "layer_norm_39.b_0": "visual.vision_model.encoder.layers.19.layer_norm1.bias", "linear_116.w_0": "visual.vision_model.encoder.layers.19.self_attn.k_proj.weight", "linear_116.b_0": "visual.vision_model.encoder.layers.19.self_attn.k_proj.bias", "linear_117.w_0": "visual.vision_model.encoder.layers.19.self_attn.v_proj.weight", "linear_117.b_0": "visual.vision_model.encoder.layers.19.self_attn.v_proj.bias", "linear_118.w_0": "visual.vision_model.encoder.layers.19.self_attn.q_proj.weight", "linear_118.b_0": "visual.vision_model.encoder.layers.19.self_attn.q_proj.bias", "linear_119.w_0": "visual.vision_model.encoder.layers.19.self_attn.out_proj.weight", "linear_119.b_0": "visual.vision_model.encoder.layers.19.self_attn.out_proj.bias", "layer_norm_40.w_0": "visual.vision_model.encoder.layers.19.layer_norm2.weight", "layer_norm_40.b_0": "visual.vision_model.encoder.layers.19.layer_norm2.bias", "linear_120.w_0": "visual.vision_model.encoder.layers.19.mlp.fc1.weight", "linear_120.b_0": "visual.vision_model.encoder.layers.19.mlp.fc1.bias", "linear_121.w_0": "visual.vision_model.encoder.layers.19.mlp.fc2.weight", "linear_121.b_0": "visual.vision_model.encoder.layers.19.mlp.fc2.bias", "layer_norm_41.w_0": "visual.vision_model.encoder.layers.20.layer_norm1.weight", "layer_norm_41.b_0": "visual.vision_model.encoder.layers.20.layer_norm1.bias", "linear_122.w_0": "visual.vision_model.encoder.layers.20.self_attn.k_proj.weight", "linear_122.b_0": "visual.vision_model.encoder.layers.20.self_attn.k_proj.bias", "linear_123.w_0": "visual.vision_model.encoder.layers.20.self_attn.v_proj.weight", "linear_123.b_0": "visual.vision_model.encoder.layers.20.self_attn.v_proj.bias", "linear_124.w_0": "visual.vision_model.encoder.layers.20.self_attn.q_proj.weight", "linear_124.b_0": "visual.vision_model.encoder.layers.20.self_attn.q_proj.bias", "linear_125.w_0": "visual.vision_model.encoder.layers.20.self_attn.out_proj.weight", "linear_125.b_0": "visual.vision_model.encoder.layers.20.self_attn.out_proj.bias", "layer_norm_42.w_0": "visual.vision_model.encoder.layers.20.layer_norm2.weight", "layer_norm_42.b_0": "visual.vision_model.encoder.layers.20.layer_norm2.bias", "linear_126.w_0": "visual.vision_model.encoder.layers.20.mlp.fc1.weight", "linear_126.b_0": "visual.vision_model.encoder.layers.20.mlp.fc1.bias", "linear_127.w_0": "visual.vision_model.encoder.layers.20.mlp.fc2.weight", "linear_127.b_0": "visual.vision_model.encoder.layers.20.mlp.fc2.bias", "layer_norm_43.w_0": "visual.vision_model.encoder.layers.21.layer_norm1.weight", "layer_norm_43.b_0": "visual.vision_model.encoder.layers.21.layer_norm1.bias", "linear_128.w_0": "visual.vision_model.encoder.layers.21.self_attn.k_proj.weight", "linear_128.b_0": "visual.vision_model.encoder.layers.21.self_attn.k_proj.bias", "linear_129.w_0": "visual.vision_model.encoder.layers.21.self_attn.v_proj.weight", "linear_129.b_0": "visual.vision_model.encoder.layers.21.self_attn.v_proj.bias", "linear_130.w_0": "visual.vision_model.encoder.layers.21.self_attn.q_proj.weight", "linear_130.b_0": "visual.vision_model.encoder.layers.21.self_attn.q_proj.bias", "linear_131.w_0": "visual.vision_model.encoder.layers.21.self_attn.out_proj.weight", "linear_131.b_0": "visual.vision_model.encoder.layers.21.self_attn.out_proj.bias", "layer_norm_44.w_0": "visual.vision_model.encoder.layers.21.layer_norm2.weight", "layer_norm_44.b_0": "visual.vision_model.encoder.layers.21.layer_norm2.bias", "linear_132.w_0": "visual.vision_model.encoder.layers.21.mlp.fc1.weight", "linear_132.b_0": "visual.vision_model.encoder.layers.21.mlp.fc1.bias", "linear_133.w_0": "visual.vision_model.encoder.layers.21.mlp.fc2.weight", "linear_133.b_0": "visual.vision_model.encoder.layers.21.mlp.fc2.bias", "layer_norm_45.w_0": "visual.vision_model.encoder.layers.22.layer_norm1.weight", "layer_norm_45.b_0": "visual.vision_model.encoder.layers.22.layer_norm1.bias", "linear_134.w_0": "visual.vision_model.encoder.layers.22.self_attn.k_proj.weight", "linear_134.b_0": "visual.vision_model.encoder.layers.22.self_attn.k_proj.bias", "linear_135.w_0": "visual.vision_model.encoder.layers.22.self_attn.v_proj.weight", "linear_135.b_0": "visual.vision_model.encoder.layers.22.self_attn.v_proj.bias", "linear_136.w_0": "visual.vision_model.encoder.layers.22.self_attn.q_proj.weight", "linear_136.b_0": "visual.vision_model.encoder.layers.22.self_attn.q_proj.bias", "linear_137.w_0": "visual.vision_model.encoder.layers.22.self_attn.out_proj.weight", "linear_137.b_0": "visual.vision_model.encoder.layers.22.self_attn.out_proj.bias", "layer_norm_46.w_0": "visual.vision_model.encoder.layers.22.layer_norm2.weight", "layer_norm_46.b_0": "visual.vision_model.encoder.layers.22.layer_norm2.bias", "linear_138.w_0": "visual.vision_model.encoder.layers.22.mlp.fc1.weight", "linear_138.b_0": "visual.vision_model.encoder.layers.22.mlp.fc1.bias", "linear_139.w_0": "visual.vision_model.encoder.layers.22.mlp.fc2.weight", "linear_139.b_0": "visual.vision_model.encoder.layers.22.mlp.fc2.bias", "layer_norm_47.w_0": "visual.vision_model.encoder.layers.23.layer_norm1.weight", "layer_norm_47.b_0": "visual.vision_model.encoder.layers.23.layer_norm1.bias", "linear_140.w_0": "visual.vision_model.encoder.layers.23.self_attn.k_proj.weight", "linear_140.b_0": "visual.vision_model.encoder.layers.23.self_attn.k_proj.bias", "linear_141.w_0": "visual.vision_model.encoder.layers.23.self_attn.v_proj.weight", "linear_141.b_0": "visual.vision_model.encoder.layers.23.self_attn.v_proj.bias", "linear_142.w_0": "visual.vision_model.encoder.layers.23.self_attn.q_proj.weight", "linear_142.b_0": "visual.vision_model.encoder.layers.23.self_attn.q_proj.bias", "linear_143.w_0": "visual.vision_model.encoder.layers.23.self_attn.out_proj.weight", "linear_143.b_0": "visual.vision_model.encoder.layers.23.self_attn.out_proj.bias", "layer_norm_48.w_0": "visual.vision_model.encoder.layers.23.layer_norm2.weight", "layer_norm_48.b_0": "visual.vision_model.encoder.layers.23.layer_norm2.bias", "linear_144.w_0": "visual.vision_model.encoder.layers.23.mlp.fc1.weight", "linear_144.b_0": "visual.vision_model.encoder.layers.23.mlp.fc1.bias", "linear_145.w_0": "visual.vision_model.encoder.layers.23.mlp.fc2.weight", "linear_145.b_0": "visual.vision_model.encoder.layers.23.mlp.fc2.bias", "layer_norm_49.w_0": "visual.vision_model.encoder.layers.24.layer_norm1.weight", "layer_norm_49.b_0": "visual.vision_model.encoder.layers.24.layer_norm1.bias", "linear_146.w_0": "visual.vision_model.encoder.layers.24.self_attn.k_proj.weight", "linear_146.b_0": "visual.vision_model.encoder.layers.24.self_attn.k_proj.bias", "linear_147.w_0": "visual.vision_model.encoder.layers.24.self_attn.v_proj.weight", "linear_147.b_0": "visual.vision_model.encoder.layers.24.self_attn.v_proj.bias", "linear_148.w_0": "visual.vision_model.encoder.layers.24.self_attn.q_proj.weight", "linear_148.b_0": "visual.vision_model.encoder.layers.24.self_attn.q_proj.bias", "linear_149.w_0": "visual.vision_model.encoder.layers.24.self_attn.out_proj.weight", "linear_149.b_0": "visual.vision_model.encoder.layers.24.self_attn.out_proj.bias", "layer_norm_50.w_0": "visual.vision_model.encoder.layers.24.layer_norm2.weight", "layer_norm_50.b_0": "visual.vision_model.encoder.layers.24.layer_norm2.bias", "linear_150.w_0": "visual.vision_model.encoder.layers.24.mlp.fc1.weight", "linear_150.b_0": "visual.vision_model.encoder.layers.24.mlp.fc1.bias", "linear_151.w_0": "visual.vision_model.encoder.layers.24.mlp.fc2.weight", "linear_151.b_0": "visual.vision_model.encoder.layers.24.mlp.fc2.bias", "layer_norm_51.w_0": "visual.vision_model.encoder.layers.25.layer_norm1.weight", "layer_norm_51.b_0": "visual.vision_model.encoder.layers.25.layer_norm1.bias", "linear_152.w_0": "visual.vision_model.encoder.layers.25.self_attn.k_proj.weight", "linear_152.b_0": "visual.vision_model.encoder.layers.25.self_attn.k_proj.bias", "linear_153.w_0": "visual.vision_model.encoder.layers.25.self_attn.v_proj.weight", "linear_153.b_0": "visual.vision_model.encoder.layers.25.self_attn.v_proj.bias", "linear_154.w_0": "visual.vision_model.encoder.layers.25.self_attn.q_proj.weight", "linear_154.b_0": "visual.vision_model.encoder.layers.25.self_attn.q_proj.bias", "linear_155.w_0": "visual.vision_model.encoder.layers.25.self_attn.out_proj.weight", "linear_155.b_0": "visual.vision_model.encoder.layers.25.self_attn.out_proj.bias", "layer_norm_52.w_0": "visual.vision_model.encoder.layers.25.layer_norm2.weight", "layer_norm_52.b_0": "visual.vision_model.encoder.layers.25.layer_norm2.bias", "linear_156.w_0": "visual.vision_model.encoder.layers.25.mlp.fc1.weight", "linear_156.b_0": "visual.vision_model.encoder.layers.25.mlp.fc1.bias", "linear_157.w_0": "visual.vision_model.encoder.layers.25.mlp.fc2.weight", "linear_157.b_0": "visual.vision_model.encoder.layers.25.mlp.fc2.bias", "layer_norm_53.w_0": "visual.vision_model.encoder.layers.26.layer_norm1.weight", "layer_norm_53.b_0": "visual.vision_model.encoder.layers.26.layer_norm1.bias", "linear_158.w_0": "visual.vision_model.encoder.layers.26.self_attn.k_proj.weight", "linear_158.b_0": "visual.vision_model.encoder.layers.26.self_attn.k_proj.bias", "linear_159.w_0": "visual.vision_model.encoder.layers.26.self_attn.v_proj.weight", "linear_159.b_0": "visual.vision_model.encoder.layers.26.self_attn.v_proj.bias", "linear_160.w_0": "visual.vision_model.encoder.layers.26.self_attn.q_proj.weight", "linear_160.b_0": "visual.vision_model.encoder.layers.26.self_attn.q_proj.bias", "linear_161.w_0": "visual.vision_model.encoder.layers.26.self_attn.out_proj.weight", "linear_161.b_0": "visual.vision_model.encoder.layers.26.self_attn.out_proj.bias", "layer_norm_54.w_0": "visual.vision_model.encoder.layers.26.layer_norm2.weight", "layer_norm_54.b_0": "visual.vision_model.encoder.layers.26.layer_norm2.bias", "linear_162.w_0": "visual.vision_model.encoder.layers.26.mlp.fc1.weight", "linear_162.b_0": "visual.vision_model.encoder.layers.26.mlp.fc1.bias", "linear_163.w_0": "visual.vision_model.encoder.layers.26.mlp.fc2.weight", "linear_163.b_0": "visual.vision_model.encoder.layers.26.mlp.fc2.bias", "layer_norm_55.w_0": "visual.vision_model.post_layernorm.weight", "layer_norm_55.b_0": "visual.vision_model.post_layernorm.bias", "siglip_multihead_attention_pooling_head_0.w_0": "visual.vision_model.head.probe", "multi_head_attention_0.w_0": "visual.vision_model.head.attention.in_proj_weight", "multi_head_attention_0.w_1": "visual.vision_model.head.attention.in_proj_bias", "linear_164.w_0": "visual.vision_model.head.attention.out_proj.weight", "linear_164.b_0": "visual.vision_model.head.attention.out_proj.bias", "layer_norm_56.w_0": "visual.vision_model.head.layernorm.weight", "layer_norm_56.b_0": "visual.vision_model.head.layernorm.bias", "linear_165.w_0": "visual.vision_model.head.mlp.fc1.weight", "linear_165.b_0": "visual.vision_model.head.mlp.fc1.bias", "linear_166.w_0": "visual.vision_model.head.mlp.fc2.weight", "linear_166.b_0": "visual.vision_model.head.mlp.fc2.bias", "embedding_2.w_0": "model.embed_tokens.weight", "linear_167.w_0": "model.layers.0.self_attn.q_proj.weight", "linear_168.w_0": "model.layers.0.self_attn.k_proj.weight", "linear_169.w_0": "model.layers.0.self_attn.v_proj.weight", "linear_170.w_0": "model.layers.0.self_attn.o_proj.weight", "linear_171.w_0": "model.layers.0.mlp.gate_proj.weight", "linear_172.w_0": "model.layers.0.mlp.up_proj.weight", "linear_173.w_0": "model.layers.0.mlp.down_proj.weight", "create_parameter_0.w_0": "model.layers.0.input_layernorm.weight", "create_parameter_1.w_0": "model.layers.0.post_attention_layernorm.weight", "linear_174.w_0": "model.layers.1.self_attn.q_proj.weight", "linear_175.w_0": "model.layers.1.self_attn.k_proj.weight", "linear_176.w_0": "model.layers.1.self_attn.v_proj.weight", "linear_177.w_0": "model.layers.1.self_attn.o_proj.weight", "linear_178.w_0": "model.layers.1.mlp.gate_proj.weight", "linear_179.w_0": "model.layers.1.mlp.up_proj.weight", "linear_180.w_0": "model.layers.1.mlp.down_proj.weight", "create_parameter_2.w_0": "model.layers.1.input_layernorm.weight", "create_parameter_3.w_0": "model.layers.1.post_attention_layernorm.weight", "linear_181.w_0": "model.layers.2.self_attn.q_proj.weight", "linear_182.w_0": "model.layers.2.self_attn.k_proj.weight", "linear_183.w_0": "model.layers.2.self_attn.v_proj.weight", "linear_184.w_0": "model.layers.2.self_attn.o_proj.weight", "linear_185.w_0": "model.layers.2.mlp.gate_proj.weight", "linear_186.w_0": "model.layers.2.mlp.up_proj.weight", "linear_187.w_0": "model.layers.2.mlp.down_proj.weight", "create_parameter_4.w_0": "model.layers.2.input_layernorm.weight", "create_parameter_5.w_0": "model.layers.2.post_attention_layernorm.weight", "linear_188.w_0": "model.layers.3.self_attn.q_proj.weight", "linear_189.w_0": "model.layers.3.self_attn.k_proj.weight", "linear_190.w_0": "model.layers.3.self_attn.v_proj.weight", "linear_191.w_0": "model.layers.3.self_attn.o_proj.weight", "linear_192.w_0": "model.layers.3.mlp.gate_proj.weight", "linear_193.w_0": "model.layers.3.mlp.up_proj.weight", "linear_194.w_0": "model.layers.3.mlp.down_proj.weight", "create_parameter_6.w_0": "model.layers.3.input_layernorm.weight", "create_parameter_7.w_0": "model.layers.3.post_attention_layernorm.weight", "linear_195.w_0": "model.layers.4.self_attn.q_proj.weight", "linear_196.w_0": "model.layers.4.self_attn.k_proj.weight", "linear_197.w_0": "model.layers.4.self_attn.v_proj.weight", "linear_198.w_0": "model.layers.4.self_attn.o_proj.weight", "linear_199.w_0": "model.layers.4.mlp.gate_proj.weight", "linear_200.w_0": "model.layers.4.mlp.up_proj.weight", "linear_201.w_0": "model.layers.4.mlp.down_proj.weight", "create_parameter_8.w_0": "model.layers.4.input_layernorm.weight", "create_parameter_9.w_0": "model.layers.4.post_attention_layernorm.weight", "linear_202.w_0": "model.layers.5.self_attn.q_proj.weight", "linear_203.w_0": "model.layers.5.self_attn.k_proj.weight", "linear_204.w_0": "model.layers.5.self_attn.v_proj.weight", "linear_205.w_0": "model.layers.5.self_attn.o_proj.weight", "linear_206.w_0": "model.layers.5.mlp.gate_proj.weight", "linear_207.w_0": "model.layers.5.mlp.up_proj.weight", "linear_208.w_0": "model.layers.5.mlp.down_proj.weight", "create_parameter_10.w_0": "model.layers.5.input_layernorm.weight", "create_parameter_11.w_0": "model.layers.5.post_attention_layernorm.weight", "linear_209.w_0": "model.layers.6.self_attn.q_proj.weight", "linear_210.w_0": "model.layers.6.self_attn.k_proj.weight", "linear_211.w_0": "model.layers.6.self_attn.v_proj.weight", "linear_212.w_0": "model.layers.6.self_attn.o_proj.weight", "linear_213.w_0": "model.layers.6.mlp.gate_proj.weight", "linear_214.w_0": "model.layers.6.mlp.up_proj.weight", "linear_215.w_0": "model.layers.6.mlp.down_proj.weight", "create_parameter_12.w_0": "model.layers.6.input_layernorm.weight", "create_parameter_13.w_0": "model.layers.6.post_attention_layernorm.weight", "linear_216.w_0": "model.layers.7.self_attn.q_proj.weight", "linear_217.w_0": "model.layers.7.self_attn.k_proj.weight", "linear_218.w_0": "model.layers.7.self_attn.v_proj.weight", "linear_219.w_0": "model.layers.7.self_attn.o_proj.weight", "linear_220.w_0": "model.layers.7.mlp.gate_proj.weight", "linear_221.w_0": "model.layers.7.mlp.up_proj.weight", "linear_222.w_0": "model.layers.7.mlp.down_proj.weight", "create_parameter_14.w_0": "model.layers.7.input_layernorm.weight", "create_parameter_15.w_0": "model.layers.7.post_attention_layernorm.weight", "linear_223.w_0": "model.layers.8.self_attn.q_proj.weight", "linear_224.w_0": "model.layers.8.self_attn.k_proj.weight", "linear_225.w_0": "model.layers.8.self_attn.v_proj.weight", "linear_226.w_0": "model.layers.8.self_attn.o_proj.weight", "linear_227.w_0": "model.layers.8.mlp.gate_proj.weight", "linear_228.w_0": "model.layers.8.mlp.up_proj.weight", "linear_229.w_0": "model.layers.8.mlp.down_proj.weight", "create_parameter_16.w_0": "model.layers.8.input_layernorm.weight", "create_parameter_17.w_0": "model.layers.8.post_attention_layernorm.weight", "linear_230.w_0": "model.layers.9.self_attn.q_proj.weight", "linear_231.w_0": "model.layers.9.self_attn.k_proj.weight", "linear_232.w_0": "model.layers.9.self_attn.v_proj.weight", "linear_233.w_0": "model.layers.9.self_attn.o_proj.weight", "linear_234.w_0": "model.layers.9.mlp.gate_proj.weight", "linear_235.w_0": "model.layers.9.mlp.up_proj.weight", "linear_236.w_0": "model.layers.9.mlp.down_proj.weight", "create_parameter_18.w_0": "model.layers.9.input_layernorm.weight", "create_parameter_19.w_0": "model.layers.9.post_attention_layernorm.weight", "linear_237.w_0": "model.layers.10.self_attn.q_proj.weight", "linear_238.w_0": "model.layers.10.self_attn.k_proj.weight", "linear_239.w_0": "model.layers.10.self_attn.v_proj.weight", "linear_240.w_0": "model.layers.10.self_attn.o_proj.weight", "linear_241.w_0": "model.layers.10.mlp.gate_proj.weight", "linear_242.w_0": "model.layers.10.mlp.up_proj.weight", "linear_243.w_0": "model.layers.10.mlp.down_proj.weight", "create_parameter_20.w_0": "model.layers.10.input_layernorm.weight", "create_parameter_21.w_0": "model.layers.10.post_attention_layernorm.weight", "linear_244.w_0": "model.layers.11.self_attn.q_proj.weight", "linear_245.w_0": "model.layers.11.self_attn.k_proj.weight", "linear_246.w_0": "model.layers.11.self_attn.v_proj.weight", "linear_247.w_0": "model.layers.11.self_attn.o_proj.weight", "linear_248.w_0": "model.layers.11.mlp.gate_proj.weight", "linear_249.w_0": "model.layers.11.mlp.up_proj.weight", "linear_250.w_0": "model.layers.11.mlp.down_proj.weight", "create_parameter_22.w_0": "model.layers.11.input_layernorm.weight", "create_parameter_23.w_0": "model.layers.11.post_attention_layernorm.weight", "linear_251.w_0": "model.layers.12.self_attn.q_proj.weight", "linear_252.w_0": "model.layers.12.self_attn.k_proj.weight", "linear_253.w_0": "model.layers.12.self_attn.v_proj.weight", "linear_254.w_0": "model.layers.12.self_attn.o_proj.weight", "linear_255.w_0": "model.layers.12.mlp.gate_proj.weight", "linear_256.w_0": "model.layers.12.mlp.up_proj.weight", "linear_257.w_0": "model.layers.12.mlp.down_proj.weight", "create_parameter_24.w_0": "model.layers.12.input_layernorm.weight", "create_parameter_25.w_0": "model.layers.12.post_attention_layernorm.weight", "linear_258.w_0": "model.layers.13.self_attn.q_proj.weight", "linear_259.w_0": "model.layers.13.self_attn.k_proj.weight", "linear_260.w_0": "model.layers.13.self_attn.v_proj.weight", "linear_261.w_0": "model.layers.13.self_attn.o_proj.weight", "linear_262.w_0": "model.layers.13.mlp.gate_proj.weight", "linear_263.w_0": "model.layers.13.mlp.up_proj.weight", "linear_264.w_0": "model.layers.13.mlp.down_proj.weight", "create_parameter_26.w_0": "model.layers.13.input_layernorm.weight", "create_parameter_27.w_0": "model.layers.13.post_attention_layernorm.weight", "linear_265.w_0": "model.layers.14.self_attn.q_proj.weight", "linear_266.w_0": "model.layers.14.self_attn.k_proj.weight", "linear_267.w_0": "model.layers.14.self_attn.v_proj.weight", "linear_268.w_0": "model.layers.14.self_attn.o_proj.weight", "linear_269.w_0": "model.layers.14.mlp.gate_proj.weight", "linear_270.w_0": "model.layers.14.mlp.up_proj.weight", "linear_271.w_0": "model.layers.14.mlp.down_proj.weight", "create_parameter_28.w_0": "model.layers.14.input_layernorm.weight", "create_parameter_29.w_0": "model.layers.14.post_attention_layernorm.weight", "linear_272.w_0": "model.layers.15.self_attn.q_proj.weight", "linear_273.w_0": "model.layers.15.self_attn.k_proj.weight", "linear_274.w_0": "model.layers.15.self_attn.v_proj.weight", "linear_275.w_0": "model.layers.15.self_attn.o_proj.weight", "linear_276.w_0": "model.layers.15.mlp.gate_proj.weight", "linear_277.w_0": "model.layers.15.mlp.up_proj.weight", "linear_278.w_0": "model.layers.15.mlp.down_proj.weight", "create_parameter_30.w_0": "model.layers.15.input_layernorm.weight", "create_parameter_31.w_0": "model.layers.15.post_attention_layernorm.weight", "linear_279.w_0": "model.layers.16.self_attn.q_proj.weight", "linear_280.w_0": "model.layers.16.self_attn.k_proj.weight", "linear_281.w_0": "model.layers.16.self_attn.v_proj.weight", "linear_282.w_0": "model.layers.16.self_attn.o_proj.weight", "linear_283.w_0": "model.layers.16.mlp.gate_proj.weight", "linear_284.w_0": "model.layers.16.mlp.up_proj.weight", "linear_285.w_0": "model.layers.16.mlp.down_proj.weight", "create_parameter_32.w_0": "model.layers.16.input_layernorm.weight", "create_parameter_33.w_0": "model.layers.16.post_attention_layernorm.weight", "linear_286.w_0": "model.layers.17.self_attn.q_proj.weight", "linear_287.w_0": "model.layers.17.self_attn.k_proj.weight", "linear_288.w_0": "model.layers.17.self_attn.v_proj.weight", "linear_289.w_0": "model.layers.17.self_attn.o_proj.weight", "linear_290.w_0": "model.layers.17.mlp.gate_proj.weight", "linear_291.w_0": "model.layers.17.mlp.up_proj.weight", "linear_292.w_0": "model.layers.17.mlp.down_proj.weight", "create_parameter_34.w_0": "model.layers.17.input_layernorm.weight", "create_parameter_35.w_0": "model.layers.17.post_attention_layernorm.weight", "create_parameter_36.w_0": "model.norm.weight", "linear_293.w_0": "lm_head.weight"}
checkpoint-336/tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34ef7db83df785924fb83d7b887b6e822a031c56e15cff40aaf9b982988180df
3
+ size 1614363
checkpoint-336/tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-336/trainer_state.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "consumed_samples": 0,
5
+ "epoch": 0.7980997624703088,
6
+ "global_step": 336,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [],
10
+ "max_steps": 421,
11
+ "num_train_epochs": 9223372036854775807,
12
+ "total_flos": 0,
13
+ "trial_name": null,
14
+ "trial_params": null
15
+ }
checkpoint-336/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0888f0d3f65facdf04c57f4bd4b2036e3fec9976c17b7eb281a443129536c8e7
3
+ size 6425
checkpoint-378/added_tokens.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"<unk>": 0, "<s>": 1, "</s>": 2, "0": 3, "1": 4, "2": 5, "3": 6, "4": 7, "5": 8, "6": 9, "7": 10, "8": 11, "9": 12, "<|end_of_sentence|>": 100272, "<|begin_of_sentence|>": 100273, "<mask:1>": 100274, "<mask:4>": 100277, "<mask:5>": 100278, "<mask:6>": 100279, "<mask:7>": 100280, "<|IMAGE_PLACEHOLDER|>": 100295, "<|AUDIO_PLACEHOLDER|>": 100296, "<|LOC_0|>": 100297, "<|LOC_1|>": 100298, "<|LOC_2|>": 100299, "<|LOC_3|>": 100300, "<|LOC_4|>": 100301, "<|LOC_5|>": 100302, "<|LOC_6|>": 100303, "<|LOC_7|>": 100304, "<|LOC_8|>": 100305, "<|LOC_9|>": 100306, "<|LOC_10|>": 100307, "<|LOC_11|>": 100308, "<|LOC_12|>": 100309, "<|LOC_13|>": 100310, "<|LOC_14|>": 100311, "<|LOC_15|>": 100312, "<|LOC_16|>": 100313, "<|LOC_17|>": 100314, "<|LOC_18|>": 100315, "<|LOC_19|>": 100316, "<|LOC_20|>": 100317, "<|LOC_21|>": 100318, "<|LOC_22|>": 100319, "<|LOC_23|>": 100320, "<|LOC_24|>": 100321, "<|LOC_25|>": 100322, "<|LOC_26|>": 100323, "<|LOC_27|>": 100324, "<|LOC_28|>": 100325, "<|LOC_29|>": 100326, "<|LOC_30|>": 100327, "<|LOC_31|>": 100328, "<|LOC_32|>": 100329, "<|LOC_33|>": 100330, "<|LOC_34|>": 100331, "<|LOC_35|>": 100332, "<|LOC_36|>": 100333, "<|LOC_37|>": 100334, "<|LOC_38|>": 100335, "<|LOC_39|>": 100336, "<|LOC_40|>": 100337, "<|LOC_41|>": 100338, "<|LOC_42|>": 100339, "<|LOC_43|>": 100340, "<|LOC_44|>": 100341, "<|LOC_45|>": 100342, "<|LOC_46|>": 100343, "<|LOC_47|>": 100344, "<|LOC_48|>": 100345, "<|LOC_49|>": 100346, "<|LOC_50|>": 100347, "<|LOC_51|>": 100348, "<|LOC_52|>": 100349, "<|LOC_53|>": 100350, "<|LOC_54|>": 100351, "<|LOC_55|>": 100352, "<|LOC_56|>": 100353, "<|LOC_57|>": 100354, "<|LOC_58|>": 100355, "<|LOC_59|>": 100356, "<|LOC_60|>": 100357, "<|LOC_61|>": 100358, "<|LOC_62|>": 100359, "<|LOC_63|>": 100360, "<|LOC_64|>": 100361, "<|LOC_65|>": 100362, "<|LOC_66|>": 100363, "<|LOC_67|>": 100364, "<|LOC_68|>": 100365, "<|LOC_69|>": 100366, "<|LOC_70|>": 100367, "<|LOC_71|>": 100368, "<|LOC_72|>": 100369, "<|LOC_73|>": 100370, "<|LOC_74|>": 100371, "<|LOC_75|>": 100372, "<|LOC_76|>": 100373, "<|LOC_77|>": 100374, "<|LOC_78|>": 100375, "<|LOC_79|>": 100376, "<|LOC_80|>": 100377, "<|LOC_81|>": 100378, "<|LOC_82|>": 100379, "<|LOC_83|>": 100380, "<|LOC_84|>": 100381, "<|LOC_85|>": 100382, "<|LOC_86|>": 100383, "<|LOC_87|>": 100384, "<|LOC_88|>": 100385, "<|LOC_89|>": 100386, "<|LOC_90|>": 100387, "<|LOC_91|>": 100388, "<|LOC_92|>": 100389, "<|LOC_93|>": 100390, "<|LOC_94|>": 100391, "<|LOC_95|>": 100392, "<|LOC_96|>": 100393, "<|LOC_97|>": 100394, "<|LOC_98|>": 100395, "<|LOC_99|>": 100396, "<|LOC_100|>": 100397, "<|LOC_101|>": 100398, "<|LOC_102|>": 100399, "<|LOC_103|>": 100400, "<|LOC_104|>": 100401, "<|LOC_105|>": 100402, "<|LOC_106|>": 100403, "<|LOC_107|>": 100404, "<|LOC_108|>": 100405, "<|LOC_109|>": 100406, "<|LOC_110|>": 100407, "<|LOC_111|>": 100408, "<|LOC_112|>": 100409, "<|LOC_113|>": 100410, "<|LOC_114|>": 100411, "<|LOC_115|>": 100412, "<|LOC_116|>": 100413, "<|LOC_117|>": 100414, "<|LOC_118|>": 100415, "<|LOC_119|>": 100416, "<|LOC_120|>": 100417, "<|LOC_121|>": 100418, "<|LOC_122|>": 100419, "<|LOC_123|>": 100420, "<|LOC_124|>": 100421, "<|LOC_125|>": 100422, "<|LOC_126|>": 100423, "<|LOC_127|>": 100424, "<|LOC_128|>": 100425, "<|LOC_129|>": 100426, "<|LOC_130|>": 100427, "<|LOC_131|>": 100428, "<|LOC_132|>": 100429, "<|LOC_133|>": 100430, "<|LOC_134|>": 100431, "<|LOC_135|>": 100432, "<|LOC_136|>": 100433, "<|LOC_137|>": 100434, "<|LOC_138|>": 100435, "<|LOC_139|>": 100436, "<|LOC_140|>": 100437, "<|LOC_141|>": 100438, "<|LOC_142|>": 100439, "<|LOC_143|>": 100440, "<|LOC_144|>": 100441, "<|LOC_145|>": 100442, "<|LOC_146|>": 100443, "<|LOC_147|>": 100444, "<|LOC_148|>": 100445, "<|LOC_149|>": 100446, "<|LOC_150|>": 100447, "<|LOC_151|>": 100448, "<|LOC_152|>": 100449, "<|LOC_153|>": 100450, "<|LOC_154|>": 100451, "<|LOC_155|>": 100452, "<|LOC_156|>": 100453, "<|LOC_157|>": 100454, "<|LOC_158|>": 100455, "<|LOC_159|>": 100456, "<|LOC_160|>": 100457, "<|LOC_161|>": 100458, "<|LOC_162|>": 100459, "<|LOC_163|>": 100460, "<|LOC_164|>": 100461, "<|LOC_165|>": 100462, "<|LOC_166|>": 100463, "<|LOC_167|>": 100464, "<|LOC_168|>": 100465, "<|LOC_169|>": 100466, "<|LOC_170|>": 100467, "<|LOC_171|>": 100468, "<|LOC_172|>": 100469, "<|LOC_173|>": 100470, "<|LOC_174|>": 100471, "<|LOC_175|>": 100472, "<|LOC_176|>": 100473, "<|LOC_177|>": 100474, "<|LOC_178|>": 100475, "<|LOC_179|>": 100476, "<|LOC_180|>": 100477, "<|LOC_181|>": 100478, "<|LOC_182|>": 100479, "<|LOC_183|>": 100480, "<|LOC_184|>": 100481, "<|LOC_185|>": 100482, "<|LOC_186|>": 100483, "<|LOC_187|>": 100484, "<|LOC_188|>": 100485, "<|LOC_189|>": 100486, "<|LOC_190|>": 100487, "<|LOC_191|>": 100488, "<|LOC_192|>": 100489, "<|LOC_193|>": 100490, "<|LOC_194|>": 100491, "<|LOC_195|>": 100492, "<|LOC_196|>": 100493, "<|LOC_197|>": 100494, "<|LOC_198|>": 100495, "<|LOC_199|>": 100496, "<|LOC_200|>": 100497, "<|LOC_201|>": 100498, "<|LOC_202|>": 100499, "<|LOC_203|>": 100500, "<|LOC_204|>": 100501, "<|LOC_205|>": 100502, "<|LOC_206|>": 100503, "<|LOC_207|>": 100504, "<|LOC_208|>": 100505, "<|LOC_209|>": 100506, "<|LOC_210|>": 100507, "<|LOC_211|>": 100508, "<|LOC_212|>": 100509, "<|LOC_213|>": 100510, "<|LOC_214|>": 100511, "<|LOC_215|>": 100512, "<|LOC_216|>": 100513, "<|LOC_217|>": 100514, "<|LOC_218|>": 100515, "<|LOC_219|>": 100516, "<|LOC_220|>": 100517, "<|LOC_221|>": 100518, "<|LOC_222|>": 100519, "<|LOC_223|>": 100520, "<|LOC_224|>": 100521, "<|LOC_225|>": 100522, "<|LOC_226|>": 100523, "<|LOC_227|>": 100524, "<|LOC_228|>": 100525, "<|LOC_229|>": 100526, "<|LOC_230|>": 100527, "<|LOC_231|>": 100528, "<|LOC_232|>": 100529, "<|LOC_233|>": 100530, "<|LOC_234|>": 100531, "<|LOC_235|>": 100532, "<|LOC_236|>": 100533, "<|LOC_237|>": 100534, "<|LOC_238|>": 100535, "<|LOC_239|>": 100536, "<|LOC_240|>": 100537, "<|LOC_241|>": 100538, "<|LOC_242|>": 100539, "<|LOC_243|>": 100540, "<|LOC_244|>": 100541, "<|LOC_245|>": 100542, "<|LOC_246|>": 100543, "<|LOC_247|>": 100544, "<|LOC_248|>": 100545, "<|LOC_249|>": 100546, "<|LOC_250|>": 100547, "<|LOC_251|>": 100548, "<|LOC_252|>": 100549, "<|LOC_253|>": 100550, "<|LOC_254|>": 100551, "<|LOC_255|>": 100552, "<|LOC_256|>": 100553, "<|LOC_257|>": 100554, "<|LOC_258|>": 100555, "<|LOC_259|>": 100556, "<|LOC_260|>": 100557, "<|LOC_261|>": 100558, "<|LOC_262|>": 100559, "<|LOC_263|>": 100560, "<|LOC_264|>": 100561, "<|LOC_265|>": 100562, "<|LOC_266|>": 100563, "<|LOC_267|>": 100564, "<|LOC_268|>": 100565, "<|LOC_269|>": 100566, "<|LOC_270|>": 100567, "<|LOC_271|>": 100568, "<|LOC_272|>": 100569, "<|LOC_273|>": 100570, "<|LOC_274|>": 100571, "<|LOC_275|>": 100572, "<|LOC_276|>": 100573, "<|LOC_277|>": 100574, "<|LOC_278|>": 100575, "<|LOC_279|>": 100576, "<|LOC_280|>": 100577, "<|LOC_281|>": 100578, "<|LOC_282|>": 100579, "<|LOC_283|>": 100580, "<|LOC_284|>": 100581, "<|LOC_285|>": 100582, "<|LOC_286|>": 100583, "<|LOC_287|>": 100584, "<|LOC_288|>": 100585, "<|LOC_289|>": 100586, "<|LOC_290|>": 100587, "<|LOC_291|>": 100588, "<|LOC_292|>": 100589, "<|LOC_293|>": 100590, "<|LOC_294|>": 100591, "<|LOC_295|>": 100592, "<|LOC_296|>": 100593, "<|LOC_297|>": 100594, "<|LOC_298|>": 100595, "<|LOC_299|>": 100596, "<|LOC_300|>": 100597, "<|LOC_301|>": 100598, "<|LOC_302|>": 100599, "<|LOC_303|>": 100600, "<|LOC_304|>": 100601, "<|LOC_305|>": 100602, "<|LOC_306|>": 100603, "<|LOC_307|>": 100604, "<|LOC_308|>": 100605, "<|LOC_309|>": 100606, "<|LOC_310|>": 100607, "<|LOC_311|>": 100608, "<|LOC_312|>": 100609, "<|LOC_313|>": 100610, "<|LOC_314|>": 100611, "<|LOC_315|>": 100612, "<|LOC_316|>": 100613, "<|LOC_317|>": 100614, "<|LOC_318|>": 100615, "<|LOC_319|>": 100616, "<|LOC_320|>": 100617, "<|LOC_321|>": 100618, "<|LOC_322|>": 100619, "<|LOC_323|>": 100620, "<|LOC_324|>": 100621, "<|LOC_325|>": 100622, "<|LOC_326|>": 100623, "<|LOC_327|>": 100624, "<|LOC_328|>": 100625, "<|LOC_329|>": 100626, "<|LOC_330|>": 100627, "<|LOC_331|>": 100628, "<|LOC_332|>": 100629, "<|LOC_333|>": 100630, "<|LOC_334|>": 100631, "<|LOC_335|>": 100632, "<|LOC_336|>": 100633, "<|LOC_337|>": 100634, "<|LOC_338|>": 100635, "<|LOC_339|>": 100636, "<|LOC_340|>": 100637, "<|LOC_341|>": 100638, "<|LOC_342|>": 100639, "<|LOC_343|>": 100640, "<|LOC_344|>": 100641, "<|LOC_345|>": 100642, "<|LOC_346|>": 100643, "<|LOC_347|>": 100644, "<|LOC_348|>": 100645, "<|LOC_349|>": 100646, "<|LOC_350|>": 100647, "<|LOC_351|>": 100648, "<|LOC_352|>": 100649, "<|LOC_353|>": 100650, "<|LOC_354|>": 100651, "<|LOC_355|>": 100652, "<|LOC_356|>": 100653, "<|LOC_357|>": 100654, "<|LOC_358|>": 100655, "<|LOC_359|>": 100656, "<|LOC_360|>": 100657, "<|LOC_361|>": 100658, "<|LOC_362|>": 100659, "<|LOC_363|>": 100660, "<|LOC_364|>": 100661, "<|LOC_365|>": 100662, "<|LOC_366|>": 100663, "<|LOC_367|>": 100664, "<|LOC_368|>": 100665, "<|LOC_369|>": 100666, "<|LOC_370|>": 100667, "<|LOC_371|>": 100668, "<|LOC_372|>": 100669, "<|LOC_373|>": 100670, "<|LOC_374|>": 100671, "<|LOC_375|>": 100672, "<|LOC_376|>": 100673, "<|LOC_377|>": 100674, "<|LOC_378|>": 100675, "<|LOC_379|>": 100676, "<|LOC_380|>": 100677, "<|LOC_381|>": 100678, "<|LOC_382|>": 100679, "<|LOC_383|>": 100680, "<|LOC_384|>": 100681, "<|LOC_385|>": 100682, "<|LOC_386|>": 100683, "<|LOC_387|>": 100684, "<|LOC_388|>": 100685, "<|LOC_389|>": 100686, "<|LOC_390|>": 100687, "<|LOC_391|>": 100688, "<|LOC_392|>": 100689, "<|LOC_393|>": 100690, "<|LOC_394|>": 100691, "<|LOC_395|>": 100692, "<|LOC_396|>": 100693, "<|LOC_397|>": 100694, "<|LOC_398|>": 100695, "<|LOC_399|>": 100696, "<|LOC_400|>": 100697, "<|LOC_401|>": 100698, "<|LOC_402|>": 100699, "<|LOC_403|>": 100700, "<|LOC_404|>": 100701, "<|LOC_405|>": 100702, "<|LOC_406|>": 100703, "<|LOC_407|>": 100704, "<|LOC_408|>": 100705, "<|LOC_409|>": 100706, "<|LOC_410|>": 100707, "<|LOC_411|>": 100708, "<|LOC_412|>": 100709, "<|LOC_413|>": 100710, "<|LOC_414|>": 100711, "<|LOC_415|>": 100712, "<|LOC_416|>": 100713, "<|LOC_417|>": 100714, "<|LOC_418|>": 100715, "<|LOC_419|>": 100716, "<|LOC_420|>": 100717, "<|LOC_421|>": 100718, "<|LOC_422|>": 100719, "<|LOC_423|>": 100720, "<|LOC_424|>": 100721, "<|LOC_425|>": 100722, "<|LOC_426|>": 100723, "<|LOC_427|>": 100724, "<|LOC_428|>": 100725, "<|LOC_429|>": 100726, "<|LOC_430|>": 100727, "<|LOC_431|>": 100728, "<|LOC_432|>": 100729, "<|LOC_433|>": 100730, "<|LOC_434|>": 100731, "<|LOC_435|>": 100732, "<|LOC_436|>": 100733, "<|LOC_437|>": 100734, "<|LOC_438|>": 100735, "<|LOC_439|>": 100736, "<|LOC_440|>": 100737, "<|LOC_441|>": 100738, "<|LOC_442|>": 100739, "<|LOC_443|>": 100740, "<|LOC_444|>": 100741, "<|LOC_445|>": 100742, "<|LOC_446|>": 100743, "<|LOC_447|>": 100744, "<|LOC_448|>": 100745, "<|LOC_449|>": 100746, "<|LOC_450|>": 100747, "<|LOC_451|>": 100748, "<|LOC_452|>": 100749, "<|LOC_453|>": 100750, "<|LOC_454|>": 100751, "<|LOC_455|>": 100752, "<|LOC_456|>": 100753, "<|LOC_457|>": 100754, "<|LOC_458|>": 100755, "<|LOC_459|>": 100756, "<|LOC_460|>": 100757, "<|LOC_461|>": 100758, "<|LOC_462|>": 100759, "<|LOC_463|>": 100760, "<|LOC_464|>": 100761, "<|LOC_465|>": 100762, "<|LOC_466|>": 100763, "<|LOC_467|>": 100764, "<|LOC_468|>": 100765, "<|LOC_469|>": 100766, "<|LOC_470|>": 100767, "<|LOC_471|>": 100768, "<|LOC_472|>": 100769, "<|LOC_473|>": 100770, "<|LOC_474|>": 100771, "<|LOC_475|>": 100772, "<|LOC_476|>": 100773, "<|LOC_477|>": 100774, "<|LOC_478|>": 100775, "<|LOC_479|>": 100776, "<|LOC_480|>": 100777, "<|LOC_481|>": 100778, "<|LOC_482|>": 100779, "<|LOC_483|>": 100780, "<|LOC_484|>": 100781, "<|LOC_485|>": 100782, "<|LOC_486|>": 100783, "<|LOC_487|>": 100784, "<|LOC_488|>": 100785, "<|LOC_489|>": 100786, "<|LOC_490|>": 100787, "<|LOC_491|>": 100788, "<|LOC_492|>": 100789, "<|LOC_493|>": 100790, "<|LOC_494|>": 100791, "<|LOC_495|>": 100792, "<|LOC_496|>": 100793, "<|LOC_497|>": 100794, "<|LOC_498|>": 100795, "<|LOC_499|>": 100796, "<|LOC_500|>": 100797, "<|LOC_501|>": 100798, "<|LOC_502|>": 100799, "<|LOC_503|>": 100800, "<|LOC_504|>": 100801, "<|LOC_505|>": 100802, "<|LOC_506|>": 100803, "<|LOC_507|>": 100804, "<|LOC_508|>": 100805, "<|LOC_509|>": 100806, "<|LOC_510|>": 100807, "<|LOC_511|>": 100808, "<|LOC_512|>": 100809, "<|LOC_513|>": 100810, "<|LOC_514|>": 100811, "<|LOC_515|>": 100812, "<|LOC_516|>": 100813, "<|LOC_517|>": 100814, "<|LOC_518|>": 100815, "<|LOC_519|>": 100816, "<|LOC_520|>": 100817, "<|LOC_521|>": 100818, "<|LOC_522|>": 100819, "<|LOC_523|>": 100820, "<|LOC_524|>": 100821, "<|LOC_525|>": 100822, "<|LOC_526|>": 100823, "<|LOC_527|>": 100824, "<|LOC_528|>": 100825, "<|LOC_529|>": 100826, "<|LOC_530|>": 100827, "<|LOC_531|>": 100828, "<|LOC_532|>": 100829, "<|LOC_533|>": 100830, "<|LOC_534|>": 100831, "<|LOC_535|>": 100832, "<|LOC_536|>": 100833, "<|LOC_537|>": 100834, "<|LOC_538|>": 100835, "<|LOC_539|>": 100836, "<|LOC_540|>": 100837, "<|LOC_541|>": 100838, "<|LOC_542|>": 100839, "<|LOC_543|>": 100840, "<|LOC_544|>": 100841, "<|LOC_545|>": 100842, "<|LOC_546|>": 100843, "<|LOC_547|>": 100844, "<|LOC_548|>": 100845, "<|LOC_549|>": 100846, "<|LOC_550|>": 100847, "<|LOC_551|>": 100848, "<|LOC_552|>": 100849, "<|LOC_553|>": 100850, "<|LOC_554|>": 100851, "<|LOC_555|>": 100852, "<|LOC_556|>": 100853, "<|LOC_557|>": 100854, "<|LOC_558|>": 100855, "<|LOC_559|>": 100856, "<|LOC_560|>": 100857, "<|LOC_561|>": 100858, "<|LOC_562|>": 100859, "<|LOC_563|>": 100860, "<|LOC_564|>": 100861, "<|LOC_565|>": 100862, "<|LOC_566|>": 100863, "<|LOC_567|>": 100864, "<|LOC_568|>": 100865, "<|LOC_569|>": 100866, "<|LOC_570|>": 100867, "<|LOC_571|>": 100868, "<|LOC_572|>": 100869, "<|LOC_573|>": 100870, "<|LOC_574|>": 100871, "<|LOC_575|>": 100872, "<|LOC_576|>": 100873, "<|LOC_577|>": 100874, "<|LOC_578|>": 100875, "<|LOC_579|>": 100876, "<|LOC_580|>": 100877, "<|LOC_581|>": 100878, "<|LOC_582|>": 100879, "<|LOC_583|>": 100880, "<|LOC_584|>": 100881, "<|LOC_585|>": 100882, "<|LOC_586|>": 100883, "<|LOC_587|>": 100884, "<|LOC_588|>": 100885, "<|LOC_589|>": 100886, "<|LOC_590|>": 100887, "<|LOC_591|>": 100888, "<|LOC_592|>": 100889, "<|LOC_593|>": 100890, "<|LOC_594|>": 100891, "<|LOC_595|>": 100892, "<|LOC_596|>": 100893, "<|LOC_597|>": 100894, "<|LOC_598|>": 100895, "<|LOC_599|>": 100896, "<|LOC_600|>": 100897, "<|LOC_601|>": 100898, "<|LOC_602|>": 100899, "<|LOC_603|>": 100900, "<|LOC_604|>": 100901, "<|LOC_605|>": 100902, "<|LOC_606|>": 100903, "<|LOC_607|>": 100904, "<|LOC_608|>": 100905, "<|LOC_609|>": 100906, "<|LOC_610|>": 100907, "<|LOC_611|>": 100908, "<|LOC_612|>": 100909, "<|LOC_613|>": 100910, "<|LOC_614|>": 100911, "<|LOC_615|>": 100912, "<|LOC_616|>": 100913, "<|LOC_617|>": 100914, "<|LOC_618|>": 100915, "<|LOC_619|>": 100916, "<|LOC_620|>": 100917, "<|LOC_621|>": 100918, "<|LOC_622|>": 100919, "<|LOC_623|>": 100920, "<|LOC_624|>": 100921, "<|LOC_625|>": 100922, "<|LOC_626|>": 100923, "<|LOC_627|>": 100924, "<|LOC_628|>": 100925, "<|LOC_629|>": 100926, "<|LOC_630|>": 100927, "<|LOC_631|>": 100928, "<|LOC_632|>": 100929, "<|LOC_633|>": 100930, "<|LOC_634|>": 100931, "<|LOC_635|>": 100932, "<|LOC_636|>": 100933, "<|LOC_637|>": 100934, "<|LOC_638|>": 100935, "<|LOC_639|>": 100936, "<|LOC_640|>": 100937, "<|LOC_641|>": 100938, "<|LOC_642|>": 100939, "<|LOC_643|>": 100940, "<|LOC_644|>": 100941, "<|LOC_645|>": 100942, "<|LOC_646|>": 100943, "<|LOC_647|>": 100944, "<|LOC_648|>": 100945, "<|LOC_649|>": 100946, "<|LOC_650|>": 100947, "<|LOC_651|>": 100948, "<|LOC_652|>": 100949, "<|LOC_653|>": 100950, "<|LOC_654|>": 100951, "<|LOC_655|>": 100952, "<|LOC_656|>": 100953, "<|LOC_657|>": 100954, "<|LOC_658|>": 100955, "<|LOC_659|>": 100956, "<|LOC_660|>": 100957, "<|LOC_661|>": 100958, "<|LOC_662|>": 100959, "<|LOC_663|>": 100960, "<|LOC_664|>": 100961, "<|LOC_665|>": 100962, "<|LOC_666|>": 100963, "<|LOC_667|>": 100964, "<|LOC_668|>": 100965, "<|LOC_669|>": 100966, "<|LOC_670|>": 100967, "<|LOC_671|>": 100968, "<|LOC_672|>": 100969, "<|LOC_673|>": 100970, "<|LOC_674|>": 100971, "<|LOC_675|>": 100972, "<|LOC_676|>": 100973, "<|LOC_677|>": 100974, "<|LOC_678|>": 100975, "<|LOC_679|>": 100976, "<|LOC_680|>": 100977, "<|LOC_681|>": 100978, "<|LOC_682|>": 100979, "<|LOC_683|>": 100980, "<|LOC_684|>": 100981, "<|LOC_685|>": 100982, "<|LOC_686|>": 100983, "<|LOC_687|>": 100984, "<|LOC_688|>": 100985, "<|LOC_689|>": 100986, "<|LOC_690|>": 100987, "<|LOC_691|>": 100988, "<|LOC_692|>": 100989, "<|LOC_693|>": 100990, "<|LOC_694|>": 100991, "<|LOC_695|>": 100992, "<|LOC_696|>": 100993, "<|LOC_697|>": 100994, "<|LOC_698|>": 100995, "<|LOC_699|>": 100996, "<|LOC_700|>": 100997, "<|LOC_701|>": 100998, "<|LOC_702|>": 100999, "<|LOC_703|>": 101000, "<|LOC_704|>": 101001, "<|LOC_705|>": 101002, "<|LOC_706|>": 101003, "<|LOC_707|>": 101004, "<|LOC_708|>": 101005, "<|LOC_709|>": 101006, "<|LOC_710|>": 101007, "<|LOC_711|>": 101008, "<|LOC_712|>": 101009, "<|LOC_713|>": 101010, "<|LOC_714|>": 101011, "<|LOC_715|>": 101012, "<|LOC_716|>": 101013, "<|LOC_717|>": 101014, "<|LOC_718|>": 101015, "<|LOC_719|>": 101016, "<|LOC_720|>": 101017, "<|LOC_721|>": 101018, "<|LOC_722|>": 101019, "<|LOC_723|>": 101020, "<|LOC_724|>": 101021, "<|LOC_725|>": 101022, "<|LOC_726|>": 101023, "<|LOC_727|>": 101024, "<|LOC_728|>": 101025, "<|LOC_729|>": 101026, "<|LOC_730|>": 101027, "<|LOC_731|>": 101028, "<|LOC_732|>": 101029, "<|LOC_733|>": 101030, "<|LOC_734|>": 101031, "<|LOC_735|>": 101032, "<|LOC_736|>": 101033, "<|LOC_737|>": 101034, "<|LOC_738|>": 101035, "<|LOC_739|>": 101036, "<|LOC_740|>": 101037, "<|LOC_741|>": 101038, "<|LOC_742|>": 101039, "<|LOC_743|>": 101040, "<|LOC_744|>": 101041, "<|LOC_745|>": 101042, "<|LOC_746|>": 101043, "<|LOC_747|>": 101044, "<|LOC_748|>": 101045, "<|LOC_749|>": 101046, "<|LOC_750|>": 101047, "<|LOC_751|>": 101048, "<|LOC_752|>": 101049, "<|LOC_753|>": 101050, "<|LOC_754|>": 101051, "<|LOC_755|>": 101052, "<|LOC_756|>": 101053, "<|LOC_757|>": 101054, "<|LOC_758|>": 101055, "<|LOC_759|>": 101056, "<|LOC_760|>": 101057, "<|LOC_761|>": 101058, "<|LOC_762|>": 101059, "<|LOC_763|>": 101060, "<|LOC_764|>": 101061, "<|LOC_765|>": 101062, "<|LOC_766|>": 101063, "<|LOC_767|>": 101064, "<|LOC_768|>": 101065, "<|LOC_769|>": 101066, "<|LOC_770|>": 101067, "<|LOC_771|>": 101068, "<|LOC_772|>": 101069, "<|LOC_773|>": 101070, "<|LOC_774|>": 101071, "<|LOC_775|>": 101072, "<|LOC_776|>": 101073, "<|LOC_777|>": 101074, "<|LOC_778|>": 101075, "<|LOC_779|>": 101076, "<|LOC_780|>": 101077, "<|LOC_781|>": 101078, "<|LOC_782|>": 101079, "<|LOC_783|>": 101080, "<|LOC_784|>": 101081, "<|LOC_785|>": 101082, "<|LOC_786|>": 101083, "<|LOC_787|>": 101084, "<|LOC_788|>": 101085, "<|LOC_789|>": 101086, "<|LOC_790|>": 101087, "<|LOC_791|>": 101088, "<|LOC_792|>": 101089, "<|LOC_793|>": 101090, "<|LOC_794|>": 101091, "<|LOC_795|>": 101092, "<|LOC_796|>": 101093, "<|LOC_797|>": 101094, "<|LOC_798|>": 101095, "<|LOC_799|>": 101096, "<|LOC_800|>": 101097, "<|LOC_801|>": 101098, "<|LOC_802|>": 101099, "<|LOC_803|>": 101100, "<|LOC_804|>": 101101, "<|LOC_805|>": 101102, "<|LOC_806|>": 101103, "<|LOC_807|>": 101104, "<|LOC_808|>": 101105, "<|LOC_809|>": 101106, "<|LOC_810|>": 101107, "<|LOC_811|>": 101108, "<|LOC_812|>": 101109, "<|LOC_813|>": 101110, "<|LOC_814|>": 101111, "<|LOC_815|>": 101112, "<|LOC_816|>": 101113, "<|LOC_817|>": 101114, "<|LOC_818|>": 101115, "<|LOC_819|>": 101116, "<|LOC_820|>": 101117, "<|LOC_821|>": 101118, "<|LOC_822|>": 101119, "<|LOC_823|>": 101120, "<|LOC_824|>": 101121, "<|LOC_825|>": 101122, "<|LOC_826|>": 101123, "<|LOC_827|>": 101124, "<|LOC_828|>": 101125, "<|LOC_829|>": 101126, "<|LOC_830|>": 101127, "<|LOC_831|>": 101128, "<|LOC_832|>": 101129, "<|LOC_833|>": 101130, "<|LOC_834|>": 101131, "<|LOC_835|>": 101132, "<|LOC_836|>": 101133, "<|LOC_837|>": 101134, "<|LOC_838|>": 101135, "<|LOC_839|>": 101136, "<|LOC_840|>": 101137, "<|LOC_841|>": 101138, "<|LOC_842|>": 101139, "<|LOC_843|>": 101140, "<|LOC_844|>": 101141, "<|LOC_845|>": 101142, "<|LOC_846|>": 101143, "<|LOC_847|>": 101144, "<|LOC_848|>": 101145, "<|LOC_849|>": 101146, "<|LOC_850|>": 101147, "<|LOC_851|>": 101148, "<|LOC_852|>": 101149, "<|LOC_853|>": 101150, "<|LOC_854|>": 101151, "<|LOC_855|>": 101152, "<|LOC_856|>": 101153, "<|LOC_857|>": 101154, "<|LOC_858|>": 101155, "<|LOC_859|>": 101156, "<|LOC_860|>": 101157, "<|LOC_861|>": 101158, "<|LOC_862|>": 101159, "<|LOC_863|>": 101160, "<|LOC_864|>": 101161, "<|LOC_865|>": 101162, "<|LOC_866|>": 101163, "<|LOC_867|>": 101164, "<|LOC_868|>": 101165, "<|LOC_869|>": 101166, "<|LOC_870|>": 101167, "<|LOC_871|>": 101168, "<|LOC_872|>": 101169, "<|LOC_873|>": 101170, "<|LOC_874|>": 101171, "<|LOC_875|>": 101172, "<|LOC_876|>": 101173, "<|LOC_877|>": 101174, "<|LOC_878|>": 101175, "<|LOC_879|>": 101176, "<|LOC_880|>": 101177, "<|LOC_881|>": 101178, "<|LOC_882|>": 101179, "<|LOC_883|>": 101180, "<|LOC_884|>": 101181, "<|LOC_885|>": 101182, "<|LOC_886|>": 101183, "<|LOC_887|>": 101184, "<|LOC_888|>": 101185, "<|LOC_889|>": 101186, "<|LOC_890|>": 101187, "<|LOC_891|>": 101188, "<|LOC_892|>": 101189, "<|LOC_893|>": 101190, "<|LOC_894|>": 101191, "<|LOC_895|>": 101192, "<|LOC_896|>": 101193, "<|LOC_897|>": 101194, "<|LOC_898|>": 101195, "<|LOC_899|>": 101196, "<|LOC_900|>": 101197, "<|LOC_901|>": 101198, "<|LOC_902|>": 101199, "<|LOC_903|>": 101200, "<|LOC_904|>": 101201, "<|LOC_905|>": 101202, "<|LOC_906|>": 101203, "<|LOC_907|>": 101204, "<|LOC_908|>": 101205, "<|LOC_909|>": 101206, "<|LOC_910|>": 101207, "<|LOC_911|>": 101208, "<|LOC_912|>": 101209, "<|LOC_913|>": 101210, "<|LOC_914|>": 101211, "<|LOC_915|>": 101212, "<|LOC_916|>": 101213, "<|LOC_917|>": 101214, "<|LOC_918|>": 101215, "<|LOC_919|>": 101216, "<|LOC_920|>": 101217, "<|LOC_921|>": 101218, "<|LOC_922|>": 101219, "<|LOC_923|>": 101220, "<|LOC_924|>": 101221, "<|LOC_925|>": 101222, "<|LOC_926|>": 101223, "<|LOC_927|>": 101224, "<|LOC_928|>": 101225, "<|LOC_929|>": 101226, "<|LOC_930|>": 101227, "<|LOC_931|>": 101228, "<|LOC_932|>": 101229, "<|LOC_933|>": 101230, "<|LOC_934|>": 101231, "<|LOC_935|>": 101232, "<|LOC_936|>": 101233, "<|LOC_937|>": 101234, "<|LOC_938|>": 101235, "<|LOC_939|>": 101236, "<|LOC_940|>": 101237, "<|LOC_941|>": 101238, "<|LOC_942|>": 101239, "<|LOC_943|>": 101240, "<|LOC_944|>": 101241, "<|LOC_945|>": 101242, "<|LOC_946|>": 101243, "<|LOC_947|>": 101244, "<|LOC_948|>": 101245, "<|LOC_949|>": 101246, "<|LOC_950|>": 101247, "<|LOC_951|>": 101248, "<|LOC_952|>": 101249, "<|LOC_953|>": 101250, "<|LOC_954|>": 101251, "<|LOC_955|>": 101252, "<|LOC_956|>": 101253, "<|LOC_957|>": 101254, "<|LOC_958|>": 101255, "<|LOC_959|>": 101256, "<|LOC_960|>": 101257, "<|LOC_961|>": 101258, "<|LOC_962|>": 101259, "<|LOC_963|>": 101260, "<|LOC_964|>": 101261, "<|LOC_965|>": 101262, "<|LOC_966|>": 101263, "<|LOC_967|>": 101264, "<|LOC_968|>": 101265, "<|LOC_969|>": 101266, "<|LOC_970|>": 101267, "<|LOC_971|>": 101268, "<|LOC_972|>": 101269, "<|LOC_973|>": 101270, "<|LOC_974|>": 101271, "<|LOC_975|>": 101272, "<|LOC_976|>": 101273, "<|LOC_977|>": 101274, "<|LOC_978|>": 101275, "<|LOC_979|>": 101276, "<|LOC_980|>": 101277, "<|LOC_981|>": 101278, "<|LOC_982|>": 101279, "<|LOC_983|>": 101280, "<|LOC_984|>": 101281, "<|LOC_985|>": 101282, "<|LOC_986|>": 101283, "<|LOC_987|>": 101284, "<|LOC_988|>": 101285, "<|LOC_989|>": 101286, "<|LOC_990|>": 101287, "<|LOC_991|>": 101288, "<|LOC_992|>": 101289, "<|LOC_993|>": 101290, "<|LOC_994|>": 101291, "<|LOC_995|>": 101292, "<|LOC_996|>": 101293, "<|LOC_997|>": 101294, "<|LOC_998|>": 101295, "<|LOC_999|>": 101296, "<|LOC_1000|>": 101297, "<|LOC_BEGIN|>": 101298, "<|LOC_END|>": 101299, "<|LOC_SEP|>": 101300, "<|CROP_COL_SEP|>": 101301, "<|CROP_ROW_SEP|>": 101302, "<|IMAGE_SEP|>": 101303, "<|image_pad|>": 101304, "<|IMAGE_START|>": 101305, "<|IMAGE_END|>": 101306, "<|video_pad|>": 101307, "<ecel>": 101308, "<fcel>": 101309, "<xcel>": 101310, "<lcel>": 101311, "<ucel>": 101312, "<nl>": 101313}
checkpoint-378/config.json ADDED
@@ -0,0 +1,223 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "PaddleOCRVLForConditionalGeneration"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.0,
6
+ "auto_map": {
7
+ "AutoConfig": "configuration_paddleocr_vl.PaddleOCRVLConfig",
8
+ "AutoModel": "modeling_paddleocr_vl.PaddleOCRVLForConditionalGeneration",
9
+ "AutoModelForCausalLM": "modeling_paddleocr_vl.PaddleOCRVLForConditionalGeneration"
10
+ },
11
+ "compression_ratio": 1.0,
12
+ "disable_pipeline_warmup": false,
13
+ "enable_mtp_magic_send": false,
14
+ "fp16_opt_level": "O2",
15
+ "freq_allocation": 20,
16
+ "fuse_ln": false,
17
+ "fuse_rms_norm": true,
18
+ "head_dim": 128,
19
+ "hidden_act": "silu",
20
+ "hidden_dropout_prob": 0.0,
21
+ "hidden_size": 1024,
22
+ "ignored_index": -100,
23
+ "im_patch_id": 100295,
24
+ "image_token_id": 100295,
25
+ "intermediate_size": 3072,
26
+ "max_position_embeddings": 131072,
27
+ "max_text_id": 100295,
28
+ "model_type": "paddleocr_vl",
29
+ "moe_dropout_prob": 0.0,
30
+ "moe_multimodal_dispatch_use_allgather": "v2-alltoall-unpad",
31
+ "num_attention_heads": 16,
32
+ "num_hidden_layers": 18,
33
+ "num_key_value_heads": 2,
34
+ "paddleformers_version": "0.4.0",
35
+ "pixel_hidden_size": 1152,
36
+ "rms_norm_eps": 1e-05,
37
+ "rope_is_neox_style": true,
38
+ "rope_scaling": {
39
+ "mrope_section": [
40
+ 16,
41
+ 24,
42
+ 24
43
+ ],
44
+ "rope_type": "default",
45
+ "type": "default"
46
+ },
47
+ "rope_theta": 500000,
48
+ "scale_qk_coeff": 1.0,
49
+ "seqlen": 16384,
50
+ "sliding_window": null,
51
+ "tie_word_embeddings": false,
52
+ "token_balance_loss": false,
53
+ "token_balance_seqlen": 16384,
54
+ "torch_dtype": "bfloat16",
55
+ "use_3d_rope": true,
56
+ "use_bias": false,
57
+ "use_flash_attn_with_mask": true,
58
+ "use_fp8": false,
59
+ "use_mem_eff_attn": true,
60
+ "use_recompute_moe": false,
61
+ "use_rmsnorm": true,
62
+ "video_token_id": 101307,
63
+ "vision_config": {
64
+ "_attn_implementation": "eager",
65
+ "_name_or_path": "",
66
+ "_save_to_hf": false,
67
+ "add_cross_attention": false,
68
+ "add_tail_layers": 0,
69
+ "architectures": [
70
+ "PaddleOCRVisionModel"
71
+ ],
72
+ "attention_dropout": 0.0,
73
+ "auto_map": {
74
+ "AutoConfig": "configuration_paddleocr_vl.PaddleOCRVLConfig",
75
+ "AutoModel": "modeling_paddleocr_vl.PaddleOCRVisionModel"
76
+ },
77
+ "bad_words_ids": null,
78
+ "begin_suppress_tokens": null,
79
+ "bos_token_id": null,
80
+ "chunk_size_feed_forward": 0,
81
+ "classifier_dropout": null,
82
+ "context_parallel_degree": 1,
83
+ "cross_attention_hidden_size": null,
84
+ "decoder_start_token_id": null,
85
+ "diversity_penalty": 0.0,
86
+ "do_sample": false,
87
+ "dpo_config": null,
88
+ "dtype": "bfloat16",
89
+ "early_stopping": false,
90
+ "encoder_no_repeat_ngram_size": 0,
91
+ "eos_token_id": null,
92
+ "ep_communication_type": "deepep",
93
+ "exponential_decay_length_penalty": null,
94
+ "finetuning_task": null,
95
+ "forced_bos_token_id": null,
96
+ "forced_eos_token_id": null,
97
+ "fuse_attention_ffn": false,
98
+ "fuse_attention_qkv": false,
99
+ "fuse_linear": false,
100
+ "fuse_rope": false,
101
+ "fuse_sequence_parallel_allreduce": false,
102
+ "fuse_swiglu": false,
103
+ "hidden_act": "gelu_new",
104
+ "hidden_size": 1152,
105
+ "id2label": {
106
+ "0": "LABEL_0",
107
+ "1": "LABEL_1"
108
+ },
109
+ "image_size": 384,
110
+ "intermediate_size": 4304,
111
+ "is_decoder": false,
112
+ "is_encoder_decoder": false,
113
+ "kto_config": null,
114
+ "label2id": {
115
+ "LABEL_0": 0,
116
+ "LABEL_1": 1
117
+ },
118
+ "layer_norm_eps": 1e-06,
119
+ "length_penalty": 1.0,
120
+ "loss_subbatch_sequence_length": -1,
121
+ "max_length": 20,
122
+ "min_length": 0,
123
+ "model_type": "paddleocr_vl",
124
+ "moe_subbatch_token_num": 0,
125
+ "no_recompute_layers": null,
126
+ "no_repeat_ngram_size": 0,
127
+ "num_attention_heads": 16,
128
+ "num_beam_groups": 1,
129
+ "num_beams": 1,
130
+ "num_channels": 3,
131
+ "num_choices": null,
132
+ "num_hidden_layers": 27,
133
+ "num_nextn_predict_layers": 0,
134
+ "num_return_sequences": 1,
135
+ "offload_recompute_inputs": false,
136
+ "output_attentions": false,
137
+ "output_hidden_states": false,
138
+ "output_scores": false,
139
+ "pad_token_id": 0,
140
+ "patch_size": 14,
141
+ "pipeline_parallel_degree": 1,
142
+ "pp_recompute_interval": 1,
143
+ "prefix": null,
144
+ "problem_type": null,
145
+ "pruned_heads": {},
146
+ "quantization_config": {
147
+ "act_quant_method": "abs_max",
148
+ "activation_scheme": null,
149
+ "actscale_moving_rate": 0.01,
150
+ "apply_hadamard": false,
151
+ "apply_online_actscale_step": 200,
152
+ "dense_quant_type": "",
153
+ "dtype": null,
154
+ "fmt": null,
155
+ "fp8_format_type": "hybrid",
156
+ "group_size": -1,
157
+ "hadamard_block_size": 32,
158
+ "ignore_modules": null,
159
+ "llm_int8_threshold": 6.0,
160
+ "moe_quant_type": "",
161
+ "qlora_weight_blocksize": 64,
162
+ "qlora_weight_double_quant": false,
163
+ "qlora_weight_double_quant_block_size": 256,
164
+ "quant_input_grad": false,
165
+ "quant_method": null,
166
+ "quant_round_type": 0,
167
+ "quant_type": null,
168
+ "quant_weight_grad": false,
169
+ "quantization": "",
170
+ "scale_epsilon": 1e-08,
171
+ "shift": false,
172
+ "shift_smooth_all_linears": false,
173
+ "smooth": false,
174
+ "weight_block_size": null,
175
+ "weight_quant_method": "abs_max_channel_wise",
176
+ "weight_quantize_algo": null
177
+ },
178
+ "recompute": true,
179
+ "recompute_granularity": "full",
180
+ "recompute_use_reentrant": false,
181
+ "refined_recompute": "",
182
+ "remove_invalid_values": false,
183
+ "repetition_penalty": 1.0,
184
+ "return_dict": false,
185
+ "return_dict_in_generate": false,
186
+ "sep_parallel_degree": 1,
187
+ "sep_token_id": null,
188
+ "sequence_parallel": false,
189
+ "spatial_merge_size": 2,
190
+ "suppress_tokens": null,
191
+ "task_specific_params": null,
192
+ "temperature": 1.0,
193
+ "temporal_patch_size": 2,
194
+ "tensor_parallel_degree": 1,
195
+ "tensor_parallel_output": true,
196
+ "tensor_parallel_rank": 0,
197
+ "tie_encoder_decoder": false,
198
+ "tie_word_embeddings": true,
199
+ "tokenizer_class": null,
200
+ "tokens_per_second": 2,
201
+ "top_k": 50,
202
+ "top_p": 1.0,
203
+ "typical_p": 1.0,
204
+ "use_cache": false,
205
+ "use_filtered_label_loss": false,
206
+ "use_flash_attention": true,
207
+ "use_fused_dropout_add": false,
208
+ "use_fused_head_and_loss_fn": false,
209
+ "use_fused_linear": false,
210
+ "use_fused_linear_cross_entropy": false,
211
+ "use_fused_rms_norm": false,
212
+ "use_fused_rope": false,
213
+ "use_sparse_flash_attn": true,
214
+ "use_sparse_head_and_loss_fn": false,
215
+ "use_unified_moe": false,
216
+ "using_fake_gate": false,
217
+ "virtual_pp_degree": 1
218
+ },
219
+ "vision_end_token_id": 101306,
220
+ "vision_start_token_id": 101305,
221
+ "vocab_size": 103424,
222
+ "weight_share_add_bias": true
223
+ }
checkpoint-378/generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "eos_token_id": 2,
4
+ "pad_token_id": 0,
5
+ "use_cache": false
6
+ }
checkpoint-378/master_weights-00001-of-00001.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd1c110b76954659e6dac8e032e963ae04742fa40d976974604898de01f72a71
3
+ size 3622483352
checkpoint-378/master_weights.safetensors.index.json ADDED
@@ -0,0 +1,615 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 3622406592
4
+ },
5
+ "weight_map": {
6
+ "mlp_AR.pre_norm.weight": "master_weights-00001-of-00001.safetensors",
7
+ "mlp_AR.pre_norm.bias": "master_weights-00001-of-00001.safetensors",
8
+ "mlp_AR.linear_1.weight": "master_weights-00001-of-00001.safetensors",
9
+ "mlp_AR.linear_1.bias": "master_weights-00001-of-00001.safetensors",
10
+ "mlp_AR.linear_2.weight": "master_weights-00001-of-00001.safetensors",
11
+ "mlp_AR.linear_2.bias": "master_weights-00001-of-00001.safetensors",
12
+ "visual.vision_model.embeddings.patch_embedding.weight": "master_weights-00001-of-00001.safetensors",
13
+ "visual.vision_model.embeddings.patch_embedding.bias": "master_weights-00001-of-00001.safetensors",
14
+ "visual.vision_model.embeddings.position_embedding.weight": "master_weights-00001-of-00001.safetensors",
15
+ "visual.vision_model.encoder.layers.0.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
16
+ "visual.vision_model.encoder.layers.0.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
17
+ "visual.vision_model.encoder.layers.0.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
18
+ "visual.vision_model.encoder.layers.0.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
19
+ "visual.vision_model.encoder.layers.0.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
20
+ "visual.vision_model.encoder.layers.0.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
21
+ "visual.vision_model.encoder.layers.0.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
22
+ "visual.vision_model.encoder.layers.0.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
23
+ "visual.vision_model.encoder.layers.0.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
24
+ "visual.vision_model.encoder.layers.0.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
25
+ "visual.vision_model.encoder.layers.0.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
26
+ "visual.vision_model.encoder.layers.0.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
27
+ "visual.vision_model.encoder.layers.0.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
28
+ "visual.vision_model.encoder.layers.0.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
29
+ "visual.vision_model.encoder.layers.0.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
30
+ "visual.vision_model.encoder.layers.0.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
31
+ "visual.vision_model.encoder.layers.1.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
32
+ "visual.vision_model.encoder.layers.1.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
33
+ "visual.vision_model.encoder.layers.1.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
34
+ "visual.vision_model.encoder.layers.1.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
35
+ "visual.vision_model.encoder.layers.1.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
36
+ "visual.vision_model.encoder.layers.1.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
37
+ "visual.vision_model.encoder.layers.1.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
38
+ "visual.vision_model.encoder.layers.1.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
39
+ "visual.vision_model.encoder.layers.1.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
40
+ "visual.vision_model.encoder.layers.1.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
41
+ "visual.vision_model.encoder.layers.1.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
42
+ "visual.vision_model.encoder.layers.1.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
43
+ "visual.vision_model.encoder.layers.1.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
44
+ "visual.vision_model.encoder.layers.1.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
45
+ "visual.vision_model.encoder.layers.1.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
46
+ "visual.vision_model.encoder.layers.1.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
47
+ "visual.vision_model.encoder.layers.2.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
48
+ "visual.vision_model.encoder.layers.2.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
49
+ "visual.vision_model.encoder.layers.2.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
50
+ "visual.vision_model.encoder.layers.2.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
51
+ "visual.vision_model.encoder.layers.2.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
52
+ "visual.vision_model.encoder.layers.2.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
53
+ "visual.vision_model.encoder.layers.2.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
54
+ "visual.vision_model.encoder.layers.2.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
55
+ "visual.vision_model.encoder.layers.2.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
56
+ "visual.vision_model.encoder.layers.2.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
57
+ "visual.vision_model.encoder.layers.2.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
58
+ "visual.vision_model.encoder.layers.2.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
59
+ "visual.vision_model.encoder.layers.2.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
60
+ "visual.vision_model.encoder.layers.2.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
61
+ "visual.vision_model.encoder.layers.2.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
62
+ "visual.vision_model.encoder.layers.2.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
63
+ "visual.vision_model.encoder.layers.3.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
64
+ "visual.vision_model.encoder.layers.3.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
65
+ "visual.vision_model.encoder.layers.3.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
66
+ "visual.vision_model.encoder.layers.3.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
67
+ "visual.vision_model.encoder.layers.3.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
68
+ "visual.vision_model.encoder.layers.3.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
69
+ "visual.vision_model.encoder.layers.3.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
70
+ "visual.vision_model.encoder.layers.3.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
71
+ "visual.vision_model.encoder.layers.3.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
72
+ "visual.vision_model.encoder.layers.3.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
73
+ "visual.vision_model.encoder.layers.3.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
74
+ "visual.vision_model.encoder.layers.3.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
75
+ "visual.vision_model.encoder.layers.3.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
76
+ "visual.vision_model.encoder.layers.3.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
77
+ "visual.vision_model.encoder.layers.3.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
78
+ "visual.vision_model.encoder.layers.3.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
79
+ "visual.vision_model.encoder.layers.4.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
80
+ "visual.vision_model.encoder.layers.4.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
81
+ "visual.vision_model.encoder.layers.4.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
82
+ "visual.vision_model.encoder.layers.4.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
83
+ "visual.vision_model.encoder.layers.4.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
84
+ "visual.vision_model.encoder.layers.4.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
85
+ "visual.vision_model.encoder.layers.4.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
86
+ "visual.vision_model.encoder.layers.4.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
87
+ "visual.vision_model.encoder.layers.4.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
88
+ "visual.vision_model.encoder.layers.4.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
89
+ "visual.vision_model.encoder.layers.4.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
90
+ "visual.vision_model.encoder.layers.4.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
91
+ "visual.vision_model.encoder.layers.4.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
92
+ "visual.vision_model.encoder.layers.4.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
93
+ "visual.vision_model.encoder.layers.4.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
94
+ "visual.vision_model.encoder.layers.4.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
95
+ "visual.vision_model.encoder.layers.5.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
96
+ "visual.vision_model.encoder.layers.5.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
97
+ "visual.vision_model.encoder.layers.5.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
98
+ "visual.vision_model.encoder.layers.5.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
99
+ "visual.vision_model.encoder.layers.5.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
100
+ "visual.vision_model.encoder.layers.5.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
101
+ "visual.vision_model.encoder.layers.5.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
102
+ "visual.vision_model.encoder.layers.5.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
103
+ "visual.vision_model.encoder.layers.5.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
104
+ "visual.vision_model.encoder.layers.5.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
105
+ "visual.vision_model.encoder.layers.5.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
106
+ "visual.vision_model.encoder.layers.5.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
107
+ "visual.vision_model.encoder.layers.5.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
108
+ "visual.vision_model.encoder.layers.5.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
109
+ "visual.vision_model.encoder.layers.5.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
110
+ "visual.vision_model.encoder.layers.5.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
111
+ "visual.vision_model.encoder.layers.6.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
112
+ "visual.vision_model.encoder.layers.6.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
113
+ "visual.vision_model.encoder.layers.6.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
114
+ "visual.vision_model.encoder.layers.6.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
115
+ "visual.vision_model.encoder.layers.6.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
116
+ "visual.vision_model.encoder.layers.6.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
117
+ "visual.vision_model.encoder.layers.6.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
118
+ "visual.vision_model.encoder.layers.6.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
119
+ "visual.vision_model.encoder.layers.6.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
120
+ "visual.vision_model.encoder.layers.6.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
121
+ "visual.vision_model.encoder.layers.6.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
122
+ "visual.vision_model.encoder.layers.6.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
123
+ "visual.vision_model.encoder.layers.6.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
124
+ "visual.vision_model.encoder.layers.6.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
125
+ "visual.vision_model.encoder.layers.6.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
126
+ "visual.vision_model.encoder.layers.6.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
127
+ "visual.vision_model.encoder.layers.7.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
128
+ "visual.vision_model.encoder.layers.7.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
129
+ "visual.vision_model.encoder.layers.7.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
130
+ "visual.vision_model.encoder.layers.7.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
131
+ "visual.vision_model.encoder.layers.7.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
132
+ "visual.vision_model.encoder.layers.7.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
133
+ "visual.vision_model.encoder.layers.7.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
134
+ "visual.vision_model.encoder.layers.7.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
135
+ "visual.vision_model.encoder.layers.7.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
136
+ "visual.vision_model.encoder.layers.7.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
137
+ "visual.vision_model.encoder.layers.7.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
138
+ "visual.vision_model.encoder.layers.7.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
139
+ "visual.vision_model.encoder.layers.7.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
140
+ "visual.vision_model.encoder.layers.7.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
141
+ "visual.vision_model.encoder.layers.7.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
142
+ "visual.vision_model.encoder.layers.7.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
143
+ "visual.vision_model.encoder.layers.8.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
144
+ "visual.vision_model.encoder.layers.8.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
145
+ "visual.vision_model.encoder.layers.8.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
146
+ "visual.vision_model.encoder.layers.8.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
147
+ "visual.vision_model.encoder.layers.8.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
148
+ "visual.vision_model.encoder.layers.8.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
149
+ "visual.vision_model.encoder.layers.8.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
150
+ "visual.vision_model.encoder.layers.8.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
151
+ "visual.vision_model.encoder.layers.8.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
152
+ "visual.vision_model.encoder.layers.8.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
153
+ "visual.vision_model.encoder.layers.8.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
154
+ "visual.vision_model.encoder.layers.8.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
155
+ "visual.vision_model.encoder.layers.8.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
156
+ "visual.vision_model.encoder.layers.8.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
157
+ "visual.vision_model.encoder.layers.8.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
158
+ "visual.vision_model.encoder.layers.8.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
159
+ "visual.vision_model.encoder.layers.9.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
160
+ "visual.vision_model.encoder.layers.9.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
161
+ "visual.vision_model.encoder.layers.9.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
162
+ "visual.vision_model.encoder.layers.9.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
163
+ "visual.vision_model.encoder.layers.9.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
164
+ "visual.vision_model.encoder.layers.9.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
165
+ "visual.vision_model.encoder.layers.9.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
166
+ "visual.vision_model.encoder.layers.9.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
167
+ "visual.vision_model.encoder.layers.9.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
168
+ "visual.vision_model.encoder.layers.9.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
169
+ "visual.vision_model.encoder.layers.9.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
170
+ "visual.vision_model.encoder.layers.9.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
171
+ "visual.vision_model.encoder.layers.9.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
172
+ "visual.vision_model.encoder.layers.9.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
173
+ "visual.vision_model.encoder.layers.9.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
174
+ "visual.vision_model.encoder.layers.9.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
175
+ "visual.vision_model.encoder.layers.10.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
176
+ "visual.vision_model.encoder.layers.10.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
177
+ "visual.vision_model.encoder.layers.10.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
178
+ "visual.vision_model.encoder.layers.10.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
179
+ "visual.vision_model.encoder.layers.10.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
180
+ "visual.vision_model.encoder.layers.10.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
181
+ "visual.vision_model.encoder.layers.10.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
182
+ "visual.vision_model.encoder.layers.10.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
183
+ "visual.vision_model.encoder.layers.10.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
184
+ "visual.vision_model.encoder.layers.10.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
185
+ "visual.vision_model.encoder.layers.10.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
186
+ "visual.vision_model.encoder.layers.10.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
187
+ "visual.vision_model.encoder.layers.10.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
188
+ "visual.vision_model.encoder.layers.10.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
189
+ "visual.vision_model.encoder.layers.10.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
190
+ "visual.vision_model.encoder.layers.10.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
191
+ "visual.vision_model.encoder.layers.11.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
192
+ "visual.vision_model.encoder.layers.11.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
193
+ "visual.vision_model.encoder.layers.11.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
194
+ "visual.vision_model.encoder.layers.11.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
195
+ "visual.vision_model.encoder.layers.11.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
196
+ "visual.vision_model.encoder.layers.11.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
197
+ "visual.vision_model.encoder.layers.11.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
198
+ "visual.vision_model.encoder.layers.11.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
199
+ "visual.vision_model.encoder.layers.11.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
200
+ "visual.vision_model.encoder.layers.11.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
201
+ "visual.vision_model.encoder.layers.11.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
202
+ "visual.vision_model.encoder.layers.11.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
203
+ "visual.vision_model.encoder.layers.11.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
204
+ "visual.vision_model.encoder.layers.11.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
205
+ "visual.vision_model.encoder.layers.11.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
206
+ "visual.vision_model.encoder.layers.11.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
207
+ "visual.vision_model.encoder.layers.12.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
208
+ "visual.vision_model.encoder.layers.12.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
209
+ "visual.vision_model.encoder.layers.12.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
210
+ "visual.vision_model.encoder.layers.12.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
211
+ "visual.vision_model.encoder.layers.12.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
212
+ "visual.vision_model.encoder.layers.12.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
213
+ "visual.vision_model.encoder.layers.12.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
214
+ "visual.vision_model.encoder.layers.12.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
215
+ "visual.vision_model.encoder.layers.12.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
216
+ "visual.vision_model.encoder.layers.12.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
217
+ "visual.vision_model.encoder.layers.12.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
218
+ "visual.vision_model.encoder.layers.12.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
219
+ "visual.vision_model.encoder.layers.12.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
220
+ "visual.vision_model.encoder.layers.12.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
221
+ "visual.vision_model.encoder.layers.12.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
222
+ "visual.vision_model.encoder.layers.12.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
223
+ "visual.vision_model.encoder.layers.13.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
224
+ "visual.vision_model.encoder.layers.13.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
225
+ "visual.vision_model.encoder.layers.13.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
226
+ "visual.vision_model.encoder.layers.13.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
227
+ "visual.vision_model.encoder.layers.13.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
228
+ "visual.vision_model.encoder.layers.13.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
229
+ "visual.vision_model.encoder.layers.13.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
230
+ "visual.vision_model.encoder.layers.13.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
231
+ "visual.vision_model.encoder.layers.13.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
232
+ "visual.vision_model.encoder.layers.13.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
233
+ "visual.vision_model.encoder.layers.13.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
234
+ "visual.vision_model.encoder.layers.13.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
235
+ "visual.vision_model.encoder.layers.13.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
236
+ "visual.vision_model.encoder.layers.13.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
237
+ "visual.vision_model.encoder.layers.13.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
238
+ "visual.vision_model.encoder.layers.13.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
239
+ "visual.vision_model.encoder.layers.14.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
240
+ "visual.vision_model.encoder.layers.14.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
241
+ "visual.vision_model.encoder.layers.14.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
242
+ "visual.vision_model.encoder.layers.14.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
243
+ "visual.vision_model.encoder.layers.14.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
244
+ "visual.vision_model.encoder.layers.14.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
245
+ "visual.vision_model.encoder.layers.14.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
246
+ "visual.vision_model.encoder.layers.14.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
247
+ "visual.vision_model.encoder.layers.14.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
248
+ "visual.vision_model.encoder.layers.14.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
249
+ "visual.vision_model.encoder.layers.14.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
250
+ "visual.vision_model.encoder.layers.14.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
251
+ "visual.vision_model.encoder.layers.14.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
252
+ "visual.vision_model.encoder.layers.14.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
253
+ "visual.vision_model.encoder.layers.14.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
254
+ "visual.vision_model.encoder.layers.14.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
255
+ "visual.vision_model.encoder.layers.15.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
256
+ "visual.vision_model.encoder.layers.15.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
257
+ "visual.vision_model.encoder.layers.15.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
258
+ "visual.vision_model.encoder.layers.15.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
259
+ "visual.vision_model.encoder.layers.15.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
260
+ "visual.vision_model.encoder.layers.15.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
261
+ "visual.vision_model.encoder.layers.15.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
262
+ "visual.vision_model.encoder.layers.15.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
263
+ "visual.vision_model.encoder.layers.15.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
264
+ "visual.vision_model.encoder.layers.15.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
265
+ "visual.vision_model.encoder.layers.15.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
266
+ "visual.vision_model.encoder.layers.15.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
267
+ "visual.vision_model.encoder.layers.15.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
268
+ "visual.vision_model.encoder.layers.15.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
269
+ "visual.vision_model.encoder.layers.15.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
270
+ "visual.vision_model.encoder.layers.15.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
271
+ "visual.vision_model.encoder.layers.16.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
272
+ "visual.vision_model.encoder.layers.16.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
273
+ "visual.vision_model.encoder.layers.16.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
274
+ "visual.vision_model.encoder.layers.16.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
275
+ "visual.vision_model.encoder.layers.16.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
276
+ "visual.vision_model.encoder.layers.16.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
277
+ "visual.vision_model.encoder.layers.16.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
278
+ "visual.vision_model.encoder.layers.16.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
279
+ "visual.vision_model.encoder.layers.16.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
280
+ "visual.vision_model.encoder.layers.16.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
281
+ "visual.vision_model.encoder.layers.16.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
282
+ "visual.vision_model.encoder.layers.16.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
283
+ "visual.vision_model.encoder.layers.16.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
284
+ "visual.vision_model.encoder.layers.16.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
285
+ "visual.vision_model.encoder.layers.16.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
286
+ "visual.vision_model.encoder.layers.16.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
287
+ "visual.vision_model.encoder.layers.17.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
288
+ "visual.vision_model.encoder.layers.17.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
289
+ "visual.vision_model.encoder.layers.17.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
290
+ "visual.vision_model.encoder.layers.17.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
291
+ "visual.vision_model.encoder.layers.17.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
292
+ "visual.vision_model.encoder.layers.17.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
293
+ "visual.vision_model.encoder.layers.17.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
294
+ "visual.vision_model.encoder.layers.17.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
295
+ "visual.vision_model.encoder.layers.17.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
296
+ "visual.vision_model.encoder.layers.17.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
297
+ "visual.vision_model.encoder.layers.17.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
298
+ "visual.vision_model.encoder.layers.17.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
299
+ "visual.vision_model.encoder.layers.17.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
300
+ "visual.vision_model.encoder.layers.17.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
301
+ "visual.vision_model.encoder.layers.17.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
302
+ "visual.vision_model.encoder.layers.17.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
303
+ "visual.vision_model.encoder.layers.18.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
304
+ "visual.vision_model.encoder.layers.18.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
305
+ "visual.vision_model.encoder.layers.18.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
306
+ "visual.vision_model.encoder.layers.18.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
307
+ "visual.vision_model.encoder.layers.18.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
308
+ "visual.vision_model.encoder.layers.18.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
309
+ "visual.vision_model.encoder.layers.18.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
310
+ "visual.vision_model.encoder.layers.18.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
311
+ "visual.vision_model.encoder.layers.18.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
312
+ "visual.vision_model.encoder.layers.18.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
313
+ "visual.vision_model.encoder.layers.18.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
314
+ "visual.vision_model.encoder.layers.18.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
315
+ "visual.vision_model.encoder.layers.18.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
316
+ "visual.vision_model.encoder.layers.18.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
317
+ "visual.vision_model.encoder.layers.18.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
318
+ "visual.vision_model.encoder.layers.18.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
319
+ "visual.vision_model.encoder.layers.19.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
320
+ "visual.vision_model.encoder.layers.19.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
321
+ "visual.vision_model.encoder.layers.19.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
322
+ "visual.vision_model.encoder.layers.19.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
323
+ "visual.vision_model.encoder.layers.19.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
324
+ "visual.vision_model.encoder.layers.19.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
325
+ "visual.vision_model.encoder.layers.19.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
326
+ "visual.vision_model.encoder.layers.19.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
327
+ "visual.vision_model.encoder.layers.19.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
328
+ "visual.vision_model.encoder.layers.19.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
329
+ "visual.vision_model.encoder.layers.19.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
330
+ "visual.vision_model.encoder.layers.19.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
331
+ "visual.vision_model.encoder.layers.19.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
332
+ "visual.vision_model.encoder.layers.19.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
333
+ "visual.vision_model.encoder.layers.19.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
334
+ "visual.vision_model.encoder.layers.19.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
335
+ "visual.vision_model.encoder.layers.20.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
336
+ "visual.vision_model.encoder.layers.20.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
337
+ "visual.vision_model.encoder.layers.20.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
338
+ "visual.vision_model.encoder.layers.20.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
339
+ "visual.vision_model.encoder.layers.20.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
340
+ "visual.vision_model.encoder.layers.20.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
341
+ "visual.vision_model.encoder.layers.20.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
342
+ "visual.vision_model.encoder.layers.20.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
343
+ "visual.vision_model.encoder.layers.20.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
344
+ "visual.vision_model.encoder.layers.20.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
345
+ "visual.vision_model.encoder.layers.20.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
346
+ "visual.vision_model.encoder.layers.20.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
347
+ "visual.vision_model.encoder.layers.20.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
348
+ "visual.vision_model.encoder.layers.20.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
349
+ "visual.vision_model.encoder.layers.20.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
350
+ "visual.vision_model.encoder.layers.20.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
351
+ "visual.vision_model.encoder.layers.21.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
352
+ "visual.vision_model.encoder.layers.21.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
353
+ "visual.vision_model.encoder.layers.21.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
354
+ "visual.vision_model.encoder.layers.21.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
355
+ "visual.vision_model.encoder.layers.21.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
356
+ "visual.vision_model.encoder.layers.21.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
357
+ "visual.vision_model.encoder.layers.21.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
358
+ "visual.vision_model.encoder.layers.21.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
359
+ "visual.vision_model.encoder.layers.21.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
360
+ "visual.vision_model.encoder.layers.21.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
361
+ "visual.vision_model.encoder.layers.21.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
362
+ "visual.vision_model.encoder.layers.21.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
363
+ "visual.vision_model.encoder.layers.21.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
364
+ "visual.vision_model.encoder.layers.21.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
365
+ "visual.vision_model.encoder.layers.21.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
366
+ "visual.vision_model.encoder.layers.21.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
367
+ "visual.vision_model.encoder.layers.22.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
368
+ "visual.vision_model.encoder.layers.22.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
369
+ "visual.vision_model.encoder.layers.22.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
370
+ "visual.vision_model.encoder.layers.22.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
371
+ "visual.vision_model.encoder.layers.22.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
372
+ "visual.vision_model.encoder.layers.22.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
373
+ "visual.vision_model.encoder.layers.22.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
374
+ "visual.vision_model.encoder.layers.22.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
375
+ "visual.vision_model.encoder.layers.22.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
376
+ "visual.vision_model.encoder.layers.22.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
377
+ "visual.vision_model.encoder.layers.22.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
378
+ "visual.vision_model.encoder.layers.22.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
379
+ "visual.vision_model.encoder.layers.22.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
380
+ "visual.vision_model.encoder.layers.22.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
381
+ "visual.vision_model.encoder.layers.22.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
382
+ "visual.vision_model.encoder.layers.22.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
383
+ "visual.vision_model.encoder.layers.23.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
384
+ "visual.vision_model.encoder.layers.23.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
385
+ "visual.vision_model.encoder.layers.23.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
386
+ "visual.vision_model.encoder.layers.23.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
387
+ "visual.vision_model.encoder.layers.23.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
388
+ "visual.vision_model.encoder.layers.23.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
389
+ "visual.vision_model.encoder.layers.23.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
390
+ "visual.vision_model.encoder.layers.23.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
391
+ "visual.vision_model.encoder.layers.23.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
392
+ "visual.vision_model.encoder.layers.23.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
393
+ "visual.vision_model.encoder.layers.23.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
394
+ "visual.vision_model.encoder.layers.23.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
395
+ "visual.vision_model.encoder.layers.23.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
396
+ "visual.vision_model.encoder.layers.23.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
397
+ "visual.vision_model.encoder.layers.23.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
398
+ "visual.vision_model.encoder.layers.23.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
399
+ "visual.vision_model.encoder.layers.24.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
400
+ "visual.vision_model.encoder.layers.24.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
401
+ "visual.vision_model.encoder.layers.24.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
402
+ "visual.vision_model.encoder.layers.24.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
403
+ "visual.vision_model.encoder.layers.24.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
404
+ "visual.vision_model.encoder.layers.24.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
405
+ "visual.vision_model.encoder.layers.24.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
406
+ "visual.vision_model.encoder.layers.24.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
407
+ "visual.vision_model.encoder.layers.24.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
408
+ "visual.vision_model.encoder.layers.24.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
409
+ "visual.vision_model.encoder.layers.24.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
410
+ "visual.vision_model.encoder.layers.24.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
411
+ "visual.vision_model.encoder.layers.24.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
412
+ "visual.vision_model.encoder.layers.24.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
413
+ "visual.vision_model.encoder.layers.24.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
414
+ "visual.vision_model.encoder.layers.24.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
415
+ "visual.vision_model.encoder.layers.25.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
416
+ "visual.vision_model.encoder.layers.25.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
417
+ "visual.vision_model.encoder.layers.25.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
418
+ "visual.vision_model.encoder.layers.25.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
419
+ "visual.vision_model.encoder.layers.25.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
420
+ "visual.vision_model.encoder.layers.25.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
421
+ "visual.vision_model.encoder.layers.25.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
422
+ "visual.vision_model.encoder.layers.25.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
423
+ "visual.vision_model.encoder.layers.25.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
424
+ "visual.vision_model.encoder.layers.25.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
425
+ "visual.vision_model.encoder.layers.25.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
426
+ "visual.vision_model.encoder.layers.25.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
427
+ "visual.vision_model.encoder.layers.25.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
428
+ "visual.vision_model.encoder.layers.25.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
429
+ "visual.vision_model.encoder.layers.25.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
430
+ "visual.vision_model.encoder.layers.25.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
431
+ "visual.vision_model.encoder.layers.26.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
432
+ "visual.vision_model.encoder.layers.26.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
433
+ "visual.vision_model.encoder.layers.26.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
434
+ "visual.vision_model.encoder.layers.26.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
435
+ "visual.vision_model.encoder.layers.26.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
436
+ "visual.vision_model.encoder.layers.26.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
437
+ "visual.vision_model.encoder.layers.26.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
438
+ "visual.vision_model.encoder.layers.26.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
439
+ "visual.vision_model.encoder.layers.26.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
440
+ "visual.vision_model.encoder.layers.26.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
441
+ "visual.vision_model.encoder.layers.26.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
442
+ "visual.vision_model.encoder.layers.26.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
443
+ "visual.vision_model.encoder.layers.26.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
444
+ "visual.vision_model.encoder.layers.26.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
445
+ "visual.vision_model.encoder.layers.26.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
446
+ "visual.vision_model.encoder.layers.26.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
447
+ "visual.vision_model.post_layernorm.weight": "master_weights-00001-of-00001.safetensors",
448
+ "visual.vision_model.post_layernorm.bias": "master_weights-00001-of-00001.safetensors",
449
+ "model.embed_tokens.weight": "master_weights-00001-of-00001.safetensors",
450
+ "model.layers.0.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
451
+ "model.layers.0.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
452
+ "model.layers.0.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
453
+ "model.layers.0.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
454
+ "model.layers.0.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
455
+ "model.layers.0.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
456
+ "model.layers.0.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
457
+ "model.layers.0.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
458
+ "model.layers.0.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
459
+ "model.layers.1.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
460
+ "model.layers.1.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
461
+ "model.layers.1.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
462
+ "model.layers.1.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
463
+ "model.layers.1.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
464
+ "model.layers.1.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
465
+ "model.layers.1.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
466
+ "model.layers.1.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
467
+ "model.layers.1.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
468
+ "model.layers.2.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
469
+ "model.layers.2.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
470
+ "model.layers.2.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
471
+ "model.layers.2.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
472
+ "model.layers.2.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
473
+ "model.layers.2.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
474
+ "model.layers.2.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
475
+ "model.layers.2.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
476
+ "model.layers.2.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
477
+ "model.layers.3.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
478
+ "model.layers.3.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
479
+ "model.layers.3.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
480
+ "model.layers.3.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
481
+ "model.layers.3.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
482
+ "model.layers.3.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
483
+ "model.layers.3.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
484
+ "model.layers.3.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
485
+ "model.layers.3.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
486
+ "model.layers.4.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
487
+ "model.layers.4.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
488
+ "model.layers.4.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
489
+ "model.layers.4.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
490
+ "model.layers.4.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
491
+ "model.layers.4.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
492
+ "model.layers.4.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
493
+ "model.layers.4.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
494
+ "model.layers.4.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
495
+ "model.layers.5.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
496
+ "model.layers.5.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
497
+ "model.layers.5.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
498
+ "model.layers.5.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
499
+ "model.layers.5.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
500
+ "model.layers.5.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
501
+ "model.layers.5.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
502
+ "model.layers.5.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
503
+ "model.layers.5.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
504
+ "model.layers.6.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
505
+ "model.layers.6.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
506
+ "model.layers.6.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
507
+ "model.layers.6.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
508
+ "model.layers.6.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
509
+ "model.layers.6.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
510
+ "model.layers.6.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
511
+ "model.layers.6.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
512
+ "model.layers.6.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
513
+ "model.layers.7.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
514
+ "model.layers.7.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
515
+ "model.layers.7.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
516
+ "model.layers.7.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
517
+ "model.layers.7.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
518
+ "model.layers.7.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
519
+ "model.layers.7.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
520
+ "model.layers.7.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
521
+ "model.layers.7.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
522
+ "model.layers.8.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
523
+ "model.layers.8.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
524
+ "model.layers.8.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
525
+ "model.layers.8.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
526
+ "model.layers.8.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
527
+ "model.layers.8.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
528
+ "model.layers.8.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
529
+ "model.layers.8.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
530
+ "model.layers.8.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
531
+ "model.layers.9.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
532
+ "model.layers.9.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
533
+ "model.layers.9.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
534
+ "model.layers.9.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
535
+ "model.layers.9.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
536
+ "model.layers.9.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
537
+ "model.layers.9.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
538
+ "model.layers.9.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
539
+ "model.layers.9.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
540
+ "model.layers.10.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
541
+ "model.layers.10.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
542
+ "model.layers.10.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
543
+ "model.layers.10.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
544
+ "model.layers.10.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
545
+ "model.layers.10.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
546
+ "model.layers.10.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
547
+ "model.layers.10.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
548
+ "model.layers.10.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
549
+ "model.layers.11.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
550
+ "model.layers.11.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
551
+ "model.layers.11.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
552
+ "model.layers.11.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
553
+ "model.layers.11.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
554
+ "model.layers.11.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
555
+ "model.layers.11.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
556
+ "model.layers.11.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
557
+ "model.layers.11.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
558
+ "model.layers.12.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
559
+ "model.layers.12.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
560
+ "model.layers.12.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
561
+ "model.layers.12.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
562
+ "model.layers.12.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
563
+ "model.layers.12.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
564
+ "model.layers.12.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
565
+ "model.layers.12.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
566
+ "model.layers.12.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
567
+ "model.layers.13.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
568
+ "model.layers.13.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
569
+ "model.layers.13.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
570
+ "model.layers.13.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
571
+ "model.layers.13.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
572
+ "model.layers.13.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
573
+ "model.layers.13.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
574
+ "model.layers.13.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
575
+ "model.layers.13.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
576
+ "model.layers.14.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
577
+ "model.layers.14.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
578
+ "model.layers.14.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
579
+ "model.layers.14.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
580
+ "model.layers.14.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
581
+ "model.layers.14.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
582
+ "model.layers.14.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
583
+ "model.layers.14.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
584
+ "model.layers.14.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
585
+ "model.layers.15.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
586
+ "model.layers.15.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
587
+ "model.layers.15.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
588
+ "model.layers.15.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
589
+ "model.layers.15.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
590
+ "model.layers.15.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
591
+ "model.layers.15.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
592
+ "model.layers.15.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
593
+ "model.layers.15.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
594
+ "model.layers.16.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
595
+ "model.layers.16.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
596
+ "model.layers.16.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
597
+ "model.layers.16.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
598
+ "model.layers.16.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
599
+ "model.layers.16.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
600
+ "model.layers.16.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
601
+ "model.layers.16.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
602
+ "model.layers.16.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
603
+ "model.layers.17.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
604
+ "model.layers.17.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
605
+ "model.layers.17.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
606
+ "model.layers.17.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
607
+ "model.layers.17.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
608
+ "model.layers.17.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
609
+ "model.layers.17.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
610
+ "model.layers.17.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
611
+ "model.layers.17.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
612
+ "model.norm.weight": "master_weights-00001-of-00001.safetensors",
613
+ "lm_head.weight": "master_weights-00001-of-00001.safetensors"
614
+ }
615
+ }
checkpoint-378/model-00001-of-00001.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1daec2a5420a9168e38e1ee07bc43febae885c45dbb49f81d30ab572eb25686
3
+ size 1917255968
checkpoint-378/model.safetensors.index.json ADDED
@@ -0,0 +1,627 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 1917177472
4
+ },
5
+ "weight_map": {
6
+ "mlp_AR.pre_norm.weight": "model-00001-of-00001.safetensors",
7
+ "mlp_AR.pre_norm.bias": "model-00001-of-00001.safetensors",
8
+ "mlp_AR.linear_1.weight": "model-00001-of-00001.safetensors",
9
+ "mlp_AR.linear_1.bias": "model-00001-of-00001.safetensors",
10
+ "mlp_AR.linear_2.weight": "model-00001-of-00001.safetensors",
11
+ "mlp_AR.linear_2.bias": "model-00001-of-00001.safetensors",
12
+ "visual.vision_model.embeddings.patch_embedding.weight": "model-00001-of-00001.safetensors",
13
+ "visual.vision_model.embeddings.patch_embedding.bias": "model-00001-of-00001.safetensors",
14
+ "visual.vision_model.embeddings.position_embedding.weight": "model-00001-of-00001.safetensors",
15
+ "visual.vision_model.embeddings.packing_position_embedding.weight": "model-00001-of-00001.safetensors",
16
+ "visual.vision_model.encoder.layers.0.layer_norm1.weight": "model-00001-of-00001.safetensors",
17
+ "visual.vision_model.encoder.layers.0.layer_norm1.bias": "model-00001-of-00001.safetensors",
18
+ "visual.vision_model.encoder.layers.0.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
19
+ "visual.vision_model.encoder.layers.0.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
20
+ "visual.vision_model.encoder.layers.0.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
21
+ "visual.vision_model.encoder.layers.0.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
22
+ "visual.vision_model.encoder.layers.0.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
23
+ "visual.vision_model.encoder.layers.0.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
24
+ "visual.vision_model.encoder.layers.0.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
25
+ "visual.vision_model.encoder.layers.0.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
26
+ "visual.vision_model.encoder.layers.0.layer_norm2.weight": "model-00001-of-00001.safetensors",
27
+ "visual.vision_model.encoder.layers.0.layer_norm2.bias": "model-00001-of-00001.safetensors",
28
+ "visual.vision_model.encoder.layers.0.mlp.fc1.weight": "model-00001-of-00001.safetensors",
29
+ "visual.vision_model.encoder.layers.0.mlp.fc1.bias": "model-00001-of-00001.safetensors",
30
+ "visual.vision_model.encoder.layers.0.mlp.fc2.weight": "model-00001-of-00001.safetensors",
31
+ "visual.vision_model.encoder.layers.0.mlp.fc2.bias": "model-00001-of-00001.safetensors",
32
+ "visual.vision_model.encoder.layers.1.layer_norm1.weight": "model-00001-of-00001.safetensors",
33
+ "visual.vision_model.encoder.layers.1.layer_norm1.bias": "model-00001-of-00001.safetensors",
34
+ "visual.vision_model.encoder.layers.1.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
35
+ "visual.vision_model.encoder.layers.1.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
36
+ "visual.vision_model.encoder.layers.1.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
37
+ "visual.vision_model.encoder.layers.1.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
38
+ "visual.vision_model.encoder.layers.1.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
39
+ "visual.vision_model.encoder.layers.1.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
40
+ "visual.vision_model.encoder.layers.1.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
41
+ "visual.vision_model.encoder.layers.1.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
42
+ "visual.vision_model.encoder.layers.1.layer_norm2.weight": "model-00001-of-00001.safetensors",
43
+ "visual.vision_model.encoder.layers.1.layer_norm2.bias": "model-00001-of-00001.safetensors",
44
+ "visual.vision_model.encoder.layers.1.mlp.fc1.weight": "model-00001-of-00001.safetensors",
45
+ "visual.vision_model.encoder.layers.1.mlp.fc1.bias": "model-00001-of-00001.safetensors",
46
+ "visual.vision_model.encoder.layers.1.mlp.fc2.weight": "model-00001-of-00001.safetensors",
47
+ "visual.vision_model.encoder.layers.1.mlp.fc2.bias": "model-00001-of-00001.safetensors",
48
+ "visual.vision_model.encoder.layers.2.layer_norm1.weight": "model-00001-of-00001.safetensors",
49
+ "visual.vision_model.encoder.layers.2.layer_norm1.bias": "model-00001-of-00001.safetensors",
50
+ "visual.vision_model.encoder.layers.2.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
51
+ "visual.vision_model.encoder.layers.2.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
52
+ "visual.vision_model.encoder.layers.2.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
53
+ "visual.vision_model.encoder.layers.2.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
54
+ "visual.vision_model.encoder.layers.2.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
55
+ "visual.vision_model.encoder.layers.2.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
56
+ "visual.vision_model.encoder.layers.2.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
57
+ "visual.vision_model.encoder.layers.2.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
58
+ "visual.vision_model.encoder.layers.2.layer_norm2.weight": "model-00001-of-00001.safetensors",
59
+ "visual.vision_model.encoder.layers.2.layer_norm2.bias": "model-00001-of-00001.safetensors",
60
+ "visual.vision_model.encoder.layers.2.mlp.fc1.weight": "model-00001-of-00001.safetensors",
61
+ "visual.vision_model.encoder.layers.2.mlp.fc1.bias": "model-00001-of-00001.safetensors",
62
+ "visual.vision_model.encoder.layers.2.mlp.fc2.weight": "model-00001-of-00001.safetensors",
63
+ "visual.vision_model.encoder.layers.2.mlp.fc2.bias": "model-00001-of-00001.safetensors",
64
+ "visual.vision_model.encoder.layers.3.layer_norm1.weight": "model-00001-of-00001.safetensors",
65
+ "visual.vision_model.encoder.layers.3.layer_norm1.bias": "model-00001-of-00001.safetensors",
66
+ "visual.vision_model.encoder.layers.3.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
67
+ "visual.vision_model.encoder.layers.3.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
68
+ "visual.vision_model.encoder.layers.3.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
69
+ "visual.vision_model.encoder.layers.3.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
70
+ "visual.vision_model.encoder.layers.3.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
71
+ "visual.vision_model.encoder.layers.3.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
72
+ "visual.vision_model.encoder.layers.3.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
73
+ "visual.vision_model.encoder.layers.3.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
74
+ "visual.vision_model.encoder.layers.3.layer_norm2.weight": "model-00001-of-00001.safetensors",
75
+ "visual.vision_model.encoder.layers.3.layer_norm2.bias": "model-00001-of-00001.safetensors",
76
+ "visual.vision_model.encoder.layers.3.mlp.fc1.weight": "model-00001-of-00001.safetensors",
77
+ "visual.vision_model.encoder.layers.3.mlp.fc1.bias": "model-00001-of-00001.safetensors",
78
+ "visual.vision_model.encoder.layers.3.mlp.fc2.weight": "model-00001-of-00001.safetensors",
79
+ "visual.vision_model.encoder.layers.3.mlp.fc2.bias": "model-00001-of-00001.safetensors",
80
+ "visual.vision_model.encoder.layers.4.layer_norm1.weight": "model-00001-of-00001.safetensors",
81
+ "visual.vision_model.encoder.layers.4.layer_norm1.bias": "model-00001-of-00001.safetensors",
82
+ "visual.vision_model.encoder.layers.4.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
83
+ "visual.vision_model.encoder.layers.4.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
84
+ "visual.vision_model.encoder.layers.4.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
85
+ "visual.vision_model.encoder.layers.4.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
86
+ "visual.vision_model.encoder.layers.4.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
87
+ "visual.vision_model.encoder.layers.4.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
88
+ "visual.vision_model.encoder.layers.4.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
89
+ "visual.vision_model.encoder.layers.4.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
90
+ "visual.vision_model.encoder.layers.4.layer_norm2.weight": "model-00001-of-00001.safetensors",
91
+ "visual.vision_model.encoder.layers.4.layer_norm2.bias": "model-00001-of-00001.safetensors",
92
+ "visual.vision_model.encoder.layers.4.mlp.fc1.weight": "model-00001-of-00001.safetensors",
93
+ "visual.vision_model.encoder.layers.4.mlp.fc1.bias": "model-00001-of-00001.safetensors",
94
+ "visual.vision_model.encoder.layers.4.mlp.fc2.weight": "model-00001-of-00001.safetensors",
95
+ "visual.vision_model.encoder.layers.4.mlp.fc2.bias": "model-00001-of-00001.safetensors",
96
+ "visual.vision_model.encoder.layers.5.layer_norm1.weight": "model-00001-of-00001.safetensors",
97
+ "visual.vision_model.encoder.layers.5.layer_norm1.bias": "model-00001-of-00001.safetensors",
98
+ "visual.vision_model.encoder.layers.5.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
99
+ "visual.vision_model.encoder.layers.5.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
100
+ "visual.vision_model.encoder.layers.5.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
101
+ "visual.vision_model.encoder.layers.5.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
102
+ "visual.vision_model.encoder.layers.5.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
103
+ "visual.vision_model.encoder.layers.5.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
104
+ "visual.vision_model.encoder.layers.5.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
105
+ "visual.vision_model.encoder.layers.5.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
106
+ "visual.vision_model.encoder.layers.5.layer_norm2.weight": "model-00001-of-00001.safetensors",
107
+ "visual.vision_model.encoder.layers.5.layer_norm2.bias": "model-00001-of-00001.safetensors",
108
+ "visual.vision_model.encoder.layers.5.mlp.fc1.weight": "model-00001-of-00001.safetensors",
109
+ "visual.vision_model.encoder.layers.5.mlp.fc1.bias": "model-00001-of-00001.safetensors",
110
+ "visual.vision_model.encoder.layers.5.mlp.fc2.weight": "model-00001-of-00001.safetensors",
111
+ "visual.vision_model.encoder.layers.5.mlp.fc2.bias": "model-00001-of-00001.safetensors",
112
+ "visual.vision_model.encoder.layers.6.layer_norm1.weight": "model-00001-of-00001.safetensors",
113
+ "visual.vision_model.encoder.layers.6.layer_norm1.bias": "model-00001-of-00001.safetensors",
114
+ "visual.vision_model.encoder.layers.6.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
115
+ "visual.vision_model.encoder.layers.6.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
116
+ "visual.vision_model.encoder.layers.6.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
117
+ "visual.vision_model.encoder.layers.6.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
118
+ "visual.vision_model.encoder.layers.6.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
119
+ "visual.vision_model.encoder.layers.6.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
120
+ "visual.vision_model.encoder.layers.6.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
121
+ "visual.vision_model.encoder.layers.6.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
122
+ "visual.vision_model.encoder.layers.6.layer_norm2.weight": "model-00001-of-00001.safetensors",
123
+ "visual.vision_model.encoder.layers.6.layer_norm2.bias": "model-00001-of-00001.safetensors",
124
+ "visual.vision_model.encoder.layers.6.mlp.fc1.weight": "model-00001-of-00001.safetensors",
125
+ "visual.vision_model.encoder.layers.6.mlp.fc1.bias": "model-00001-of-00001.safetensors",
126
+ "visual.vision_model.encoder.layers.6.mlp.fc2.weight": "model-00001-of-00001.safetensors",
127
+ "visual.vision_model.encoder.layers.6.mlp.fc2.bias": "model-00001-of-00001.safetensors",
128
+ "visual.vision_model.encoder.layers.7.layer_norm1.weight": "model-00001-of-00001.safetensors",
129
+ "visual.vision_model.encoder.layers.7.layer_norm1.bias": "model-00001-of-00001.safetensors",
130
+ "visual.vision_model.encoder.layers.7.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
131
+ "visual.vision_model.encoder.layers.7.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
132
+ "visual.vision_model.encoder.layers.7.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
133
+ "visual.vision_model.encoder.layers.7.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
134
+ "visual.vision_model.encoder.layers.7.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
135
+ "visual.vision_model.encoder.layers.7.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
136
+ "visual.vision_model.encoder.layers.7.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
137
+ "visual.vision_model.encoder.layers.7.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
138
+ "visual.vision_model.encoder.layers.7.layer_norm2.weight": "model-00001-of-00001.safetensors",
139
+ "visual.vision_model.encoder.layers.7.layer_norm2.bias": "model-00001-of-00001.safetensors",
140
+ "visual.vision_model.encoder.layers.7.mlp.fc1.weight": "model-00001-of-00001.safetensors",
141
+ "visual.vision_model.encoder.layers.7.mlp.fc1.bias": "model-00001-of-00001.safetensors",
142
+ "visual.vision_model.encoder.layers.7.mlp.fc2.weight": "model-00001-of-00001.safetensors",
143
+ "visual.vision_model.encoder.layers.7.mlp.fc2.bias": "model-00001-of-00001.safetensors",
144
+ "visual.vision_model.encoder.layers.8.layer_norm1.weight": "model-00001-of-00001.safetensors",
145
+ "visual.vision_model.encoder.layers.8.layer_norm1.bias": "model-00001-of-00001.safetensors",
146
+ "visual.vision_model.encoder.layers.8.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
147
+ "visual.vision_model.encoder.layers.8.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
148
+ "visual.vision_model.encoder.layers.8.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
149
+ "visual.vision_model.encoder.layers.8.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
150
+ "visual.vision_model.encoder.layers.8.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
151
+ "visual.vision_model.encoder.layers.8.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
152
+ "visual.vision_model.encoder.layers.8.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
153
+ "visual.vision_model.encoder.layers.8.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
154
+ "visual.vision_model.encoder.layers.8.layer_norm2.weight": "model-00001-of-00001.safetensors",
155
+ "visual.vision_model.encoder.layers.8.layer_norm2.bias": "model-00001-of-00001.safetensors",
156
+ "visual.vision_model.encoder.layers.8.mlp.fc1.weight": "model-00001-of-00001.safetensors",
157
+ "visual.vision_model.encoder.layers.8.mlp.fc1.bias": "model-00001-of-00001.safetensors",
158
+ "visual.vision_model.encoder.layers.8.mlp.fc2.weight": "model-00001-of-00001.safetensors",
159
+ "visual.vision_model.encoder.layers.8.mlp.fc2.bias": "model-00001-of-00001.safetensors",
160
+ "visual.vision_model.encoder.layers.9.layer_norm1.weight": "model-00001-of-00001.safetensors",
161
+ "visual.vision_model.encoder.layers.9.layer_norm1.bias": "model-00001-of-00001.safetensors",
162
+ "visual.vision_model.encoder.layers.9.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
163
+ "visual.vision_model.encoder.layers.9.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
164
+ "visual.vision_model.encoder.layers.9.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
165
+ "visual.vision_model.encoder.layers.9.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
166
+ "visual.vision_model.encoder.layers.9.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
167
+ "visual.vision_model.encoder.layers.9.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
168
+ "visual.vision_model.encoder.layers.9.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
169
+ "visual.vision_model.encoder.layers.9.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
170
+ "visual.vision_model.encoder.layers.9.layer_norm2.weight": "model-00001-of-00001.safetensors",
171
+ "visual.vision_model.encoder.layers.9.layer_norm2.bias": "model-00001-of-00001.safetensors",
172
+ "visual.vision_model.encoder.layers.9.mlp.fc1.weight": "model-00001-of-00001.safetensors",
173
+ "visual.vision_model.encoder.layers.9.mlp.fc1.bias": "model-00001-of-00001.safetensors",
174
+ "visual.vision_model.encoder.layers.9.mlp.fc2.weight": "model-00001-of-00001.safetensors",
175
+ "visual.vision_model.encoder.layers.9.mlp.fc2.bias": "model-00001-of-00001.safetensors",
176
+ "visual.vision_model.encoder.layers.10.layer_norm1.weight": "model-00001-of-00001.safetensors",
177
+ "visual.vision_model.encoder.layers.10.layer_norm1.bias": "model-00001-of-00001.safetensors",
178
+ "visual.vision_model.encoder.layers.10.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
179
+ "visual.vision_model.encoder.layers.10.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
180
+ "visual.vision_model.encoder.layers.10.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
181
+ "visual.vision_model.encoder.layers.10.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
182
+ "visual.vision_model.encoder.layers.10.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
183
+ "visual.vision_model.encoder.layers.10.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
184
+ "visual.vision_model.encoder.layers.10.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
185
+ "visual.vision_model.encoder.layers.10.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
186
+ "visual.vision_model.encoder.layers.10.layer_norm2.weight": "model-00001-of-00001.safetensors",
187
+ "visual.vision_model.encoder.layers.10.layer_norm2.bias": "model-00001-of-00001.safetensors",
188
+ "visual.vision_model.encoder.layers.10.mlp.fc1.weight": "model-00001-of-00001.safetensors",
189
+ "visual.vision_model.encoder.layers.10.mlp.fc1.bias": "model-00001-of-00001.safetensors",
190
+ "visual.vision_model.encoder.layers.10.mlp.fc2.weight": "model-00001-of-00001.safetensors",
191
+ "visual.vision_model.encoder.layers.10.mlp.fc2.bias": "model-00001-of-00001.safetensors",
192
+ "visual.vision_model.encoder.layers.11.layer_norm1.weight": "model-00001-of-00001.safetensors",
193
+ "visual.vision_model.encoder.layers.11.layer_norm1.bias": "model-00001-of-00001.safetensors",
194
+ "visual.vision_model.encoder.layers.11.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
195
+ "visual.vision_model.encoder.layers.11.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
196
+ "visual.vision_model.encoder.layers.11.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
197
+ "visual.vision_model.encoder.layers.11.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
198
+ "visual.vision_model.encoder.layers.11.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
199
+ "visual.vision_model.encoder.layers.11.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
200
+ "visual.vision_model.encoder.layers.11.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
201
+ "visual.vision_model.encoder.layers.11.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
202
+ "visual.vision_model.encoder.layers.11.layer_norm2.weight": "model-00001-of-00001.safetensors",
203
+ "visual.vision_model.encoder.layers.11.layer_norm2.bias": "model-00001-of-00001.safetensors",
204
+ "visual.vision_model.encoder.layers.11.mlp.fc1.weight": "model-00001-of-00001.safetensors",
205
+ "visual.vision_model.encoder.layers.11.mlp.fc1.bias": "model-00001-of-00001.safetensors",
206
+ "visual.vision_model.encoder.layers.11.mlp.fc2.weight": "model-00001-of-00001.safetensors",
207
+ "visual.vision_model.encoder.layers.11.mlp.fc2.bias": "model-00001-of-00001.safetensors",
208
+ "visual.vision_model.encoder.layers.12.layer_norm1.weight": "model-00001-of-00001.safetensors",
209
+ "visual.vision_model.encoder.layers.12.layer_norm1.bias": "model-00001-of-00001.safetensors",
210
+ "visual.vision_model.encoder.layers.12.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
211
+ "visual.vision_model.encoder.layers.12.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
212
+ "visual.vision_model.encoder.layers.12.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
213
+ "visual.vision_model.encoder.layers.12.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
214
+ "visual.vision_model.encoder.layers.12.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
215
+ "visual.vision_model.encoder.layers.12.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
216
+ "visual.vision_model.encoder.layers.12.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
217
+ "visual.vision_model.encoder.layers.12.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
218
+ "visual.vision_model.encoder.layers.12.layer_norm2.weight": "model-00001-of-00001.safetensors",
219
+ "visual.vision_model.encoder.layers.12.layer_norm2.bias": "model-00001-of-00001.safetensors",
220
+ "visual.vision_model.encoder.layers.12.mlp.fc1.weight": "model-00001-of-00001.safetensors",
221
+ "visual.vision_model.encoder.layers.12.mlp.fc1.bias": "model-00001-of-00001.safetensors",
222
+ "visual.vision_model.encoder.layers.12.mlp.fc2.weight": "model-00001-of-00001.safetensors",
223
+ "visual.vision_model.encoder.layers.12.mlp.fc2.bias": "model-00001-of-00001.safetensors",
224
+ "visual.vision_model.encoder.layers.13.layer_norm1.weight": "model-00001-of-00001.safetensors",
225
+ "visual.vision_model.encoder.layers.13.layer_norm1.bias": "model-00001-of-00001.safetensors",
226
+ "visual.vision_model.encoder.layers.13.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
227
+ "visual.vision_model.encoder.layers.13.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
228
+ "visual.vision_model.encoder.layers.13.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
229
+ "visual.vision_model.encoder.layers.13.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
230
+ "visual.vision_model.encoder.layers.13.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
231
+ "visual.vision_model.encoder.layers.13.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
232
+ "visual.vision_model.encoder.layers.13.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
233
+ "visual.vision_model.encoder.layers.13.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
234
+ "visual.vision_model.encoder.layers.13.layer_norm2.weight": "model-00001-of-00001.safetensors",
235
+ "visual.vision_model.encoder.layers.13.layer_norm2.bias": "model-00001-of-00001.safetensors",
236
+ "visual.vision_model.encoder.layers.13.mlp.fc1.weight": "model-00001-of-00001.safetensors",
237
+ "visual.vision_model.encoder.layers.13.mlp.fc1.bias": "model-00001-of-00001.safetensors",
238
+ "visual.vision_model.encoder.layers.13.mlp.fc2.weight": "model-00001-of-00001.safetensors",
239
+ "visual.vision_model.encoder.layers.13.mlp.fc2.bias": "model-00001-of-00001.safetensors",
240
+ "visual.vision_model.encoder.layers.14.layer_norm1.weight": "model-00001-of-00001.safetensors",
241
+ "visual.vision_model.encoder.layers.14.layer_norm1.bias": "model-00001-of-00001.safetensors",
242
+ "visual.vision_model.encoder.layers.14.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
243
+ "visual.vision_model.encoder.layers.14.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
244
+ "visual.vision_model.encoder.layers.14.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
245
+ "visual.vision_model.encoder.layers.14.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
246
+ "visual.vision_model.encoder.layers.14.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
247
+ "visual.vision_model.encoder.layers.14.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
248
+ "visual.vision_model.encoder.layers.14.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
249
+ "visual.vision_model.encoder.layers.14.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
250
+ "visual.vision_model.encoder.layers.14.layer_norm2.weight": "model-00001-of-00001.safetensors",
251
+ "visual.vision_model.encoder.layers.14.layer_norm2.bias": "model-00001-of-00001.safetensors",
252
+ "visual.vision_model.encoder.layers.14.mlp.fc1.weight": "model-00001-of-00001.safetensors",
253
+ "visual.vision_model.encoder.layers.14.mlp.fc1.bias": "model-00001-of-00001.safetensors",
254
+ "visual.vision_model.encoder.layers.14.mlp.fc2.weight": "model-00001-of-00001.safetensors",
255
+ "visual.vision_model.encoder.layers.14.mlp.fc2.bias": "model-00001-of-00001.safetensors",
256
+ "visual.vision_model.encoder.layers.15.layer_norm1.weight": "model-00001-of-00001.safetensors",
257
+ "visual.vision_model.encoder.layers.15.layer_norm1.bias": "model-00001-of-00001.safetensors",
258
+ "visual.vision_model.encoder.layers.15.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
259
+ "visual.vision_model.encoder.layers.15.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
260
+ "visual.vision_model.encoder.layers.15.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
261
+ "visual.vision_model.encoder.layers.15.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
262
+ "visual.vision_model.encoder.layers.15.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
263
+ "visual.vision_model.encoder.layers.15.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
264
+ "visual.vision_model.encoder.layers.15.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
265
+ "visual.vision_model.encoder.layers.15.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
266
+ "visual.vision_model.encoder.layers.15.layer_norm2.weight": "model-00001-of-00001.safetensors",
267
+ "visual.vision_model.encoder.layers.15.layer_norm2.bias": "model-00001-of-00001.safetensors",
268
+ "visual.vision_model.encoder.layers.15.mlp.fc1.weight": "model-00001-of-00001.safetensors",
269
+ "visual.vision_model.encoder.layers.15.mlp.fc1.bias": "model-00001-of-00001.safetensors",
270
+ "visual.vision_model.encoder.layers.15.mlp.fc2.weight": "model-00001-of-00001.safetensors",
271
+ "visual.vision_model.encoder.layers.15.mlp.fc2.bias": "model-00001-of-00001.safetensors",
272
+ "visual.vision_model.encoder.layers.16.layer_norm1.weight": "model-00001-of-00001.safetensors",
273
+ "visual.vision_model.encoder.layers.16.layer_norm1.bias": "model-00001-of-00001.safetensors",
274
+ "visual.vision_model.encoder.layers.16.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
275
+ "visual.vision_model.encoder.layers.16.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
276
+ "visual.vision_model.encoder.layers.16.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
277
+ "visual.vision_model.encoder.layers.16.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
278
+ "visual.vision_model.encoder.layers.16.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
279
+ "visual.vision_model.encoder.layers.16.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
280
+ "visual.vision_model.encoder.layers.16.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
281
+ "visual.vision_model.encoder.layers.16.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
282
+ "visual.vision_model.encoder.layers.16.layer_norm2.weight": "model-00001-of-00001.safetensors",
283
+ "visual.vision_model.encoder.layers.16.layer_norm2.bias": "model-00001-of-00001.safetensors",
284
+ "visual.vision_model.encoder.layers.16.mlp.fc1.weight": "model-00001-of-00001.safetensors",
285
+ "visual.vision_model.encoder.layers.16.mlp.fc1.bias": "model-00001-of-00001.safetensors",
286
+ "visual.vision_model.encoder.layers.16.mlp.fc2.weight": "model-00001-of-00001.safetensors",
287
+ "visual.vision_model.encoder.layers.16.mlp.fc2.bias": "model-00001-of-00001.safetensors",
288
+ "visual.vision_model.encoder.layers.17.layer_norm1.weight": "model-00001-of-00001.safetensors",
289
+ "visual.vision_model.encoder.layers.17.layer_norm1.bias": "model-00001-of-00001.safetensors",
290
+ "visual.vision_model.encoder.layers.17.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
291
+ "visual.vision_model.encoder.layers.17.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
292
+ "visual.vision_model.encoder.layers.17.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
293
+ "visual.vision_model.encoder.layers.17.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
294
+ "visual.vision_model.encoder.layers.17.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
295
+ "visual.vision_model.encoder.layers.17.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
296
+ "visual.vision_model.encoder.layers.17.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
297
+ "visual.vision_model.encoder.layers.17.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
298
+ "visual.vision_model.encoder.layers.17.layer_norm2.weight": "model-00001-of-00001.safetensors",
299
+ "visual.vision_model.encoder.layers.17.layer_norm2.bias": "model-00001-of-00001.safetensors",
300
+ "visual.vision_model.encoder.layers.17.mlp.fc1.weight": "model-00001-of-00001.safetensors",
301
+ "visual.vision_model.encoder.layers.17.mlp.fc1.bias": "model-00001-of-00001.safetensors",
302
+ "visual.vision_model.encoder.layers.17.mlp.fc2.weight": "model-00001-of-00001.safetensors",
303
+ "visual.vision_model.encoder.layers.17.mlp.fc2.bias": "model-00001-of-00001.safetensors",
304
+ "visual.vision_model.encoder.layers.18.layer_norm1.weight": "model-00001-of-00001.safetensors",
305
+ "visual.vision_model.encoder.layers.18.layer_norm1.bias": "model-00001-of-00001.safetensors",
306
+ "visual.vision_model.encoder.layers.18.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
307
+ "visual.vision_model.encoder.layers.18.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
308
+ "visual.vision_model.encoder.layers.18.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
309
+ "visual.vision_model.encoder.layers.18.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
310
+ "visual.vision_model.encoder.layers.18.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
311
+ "visual.vision_model.encoder.layers.18.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
312
+ "visual.vision_model.encoder.layers.18.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
313
+ "visual.vision_model.encoder.layers.18.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
314
+ "visual.vision_model.encoder.layers.18.layer_norm2.weight": "model-00001-of-00001.safetensors",
315
+ "visual.vision_model.encoder.layers.18.layer_norm2.bias": "model-00001-of-00001.safetensors",
316
+ "visual.vision_model.encoder.layers.18.mlp.fc1.weight": "model-00001-of-00001.safetensors",
317
+ "visual.vision_model.encoder.layers.18.mlp.fc1.bias": "model-00001-of-00001.safetensors",
318
+ "visual.vision_model.encoder.layers.18.mlp.fc2.weight": "model-00001-of-00001.safetensors",
319
+ "visual.vision_model.encoder.layers.18.mlp.fc2.bias": "model-00001-of-00001.safetensors",
320
+ "visual.vision_model.encoder.layers.19.layer_norm1.weight": "model-00001-of-00001.safetensors",
321
+ "visual.vision_model.encoder.layers.19.layer_norm1.bias": "model-00001-of-00001.safetensors",
322
+ "visual.vision_model.encoder.layers.19.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
323
+ "visual.vision_model.encoder.layers.19.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
324
+ "visual.vision_model.encoder.layers.19.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
325
+ "visual.vision_model.encoder.layers.19.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
326
+ "visual.vision_model.encoder.layers.19.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
327
+ "visual.vision_model.encoder.layers.19.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
328
+ "visual.vision_model.encoder.layers.19.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
329
+ "visual.vision_model.encoder.layers.19.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
330
+ "visual.vision_model.encoder.layers.19.layer_norm2.weight": "model-00001-of-00001.safetensors",
331
+ "visual.vision_model.encoder.layers.19.layer_norm2.bias": "model-00001-of-00001.safetensors",
332
+ "visual.vision_model.encoder.layers.19.mlp.fc1.weight": "model-00001-of-00001.safetensors",
333
+ "visual.vision_model.encoder.layers.19.mlp.fc1.bias": "model-00001-of-00001.safetensors",
334
+ "visual.vision_model.encoder.layers.19.mlp.fc2.weight": "model-00001-of-00001.safetensors",
335
+ "visual.vision_model.encoder.layers.19.mlp.fc2.bias": "model-00001-of-00001.safetensors",
336
+ "visual.vision_model.encoder.layers.20.layer_norm1.weight": "model-00001-of-00001.safetensors",
337
+ "visual.vision_model.encoder.layers.20.layer_norm1.bias": "model-00001-of-00001.safetensors",
338
+ "visual.vision_model.encoder.layers.20.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
339
+ "visual.vision_model.encoder.layers.20.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
340
+ "visual.vision_model.encoder.layers.20.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
341
+ "visual.vision_model.encoder.layers.20.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
342
+ "visual.vision_model.encoder.layers.20.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
343
+ "visual.vision_model.encoder.layers.20.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
344
+ "visual.vision_model.encoder.layers.20.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
345
+ "visual.vision_model.encoder.layers.20.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
346
+ "visual.vision_model.encoder.layers.20.layer_norm2.weight": "model-00001-of-00001.safetensors",
347
+ "visual.vision_model.encoder.layers.20.layer_norm2.bias": "model-00001-of-00001.safetensors",
348
+ "visual.vision_model.encoder.layers.20.mlp.fc1.weight": "model-00001-of-00001.safetensors",
349
+ "visual.vision_model.encoder.layers.20.mlp.fc1.bias": "model-00001-of-00001.safetensors",
350
+ "visual.vision_model.encoder.layers.20.mlp.fc2.weight": "model-00001-of-00001.safetensors",
351
+ "visual.vision_model.encoder.layers.20.mlp.fc2.bias": "model-00001-of-00001.safetensors",
352
+ "visual.vision_model.encoder.layers.21.layer_norm1.weight": "model-00001-of-00001.safetensors",
353
+ "visual.vision_model.encoder.layers.21.layer_norm1.bias": "model-00001-of-00001.safetensors",
354
+ "visual.vision_model.encoder.layers.21.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
355
+ "visual.vision_model.encoder.layers.21.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
356
+ "visual.vision_model.encoder.layers.21.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
357
+ "visual.vision_model.encoder.layers.21.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
358
+ "visual.vision_model.encoder.layers.21.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
359
+ "visual.vision_model.encoder.layers.21.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
360
+ "visual.vision_model.encoder.layers.21.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
361
+ "visual.vision_model.encoder.layers.21.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
362
+ "visual.vision_model.encoder.layers.21.layer_norm2.weight": "model-00001-of-00001.safetensors",
363
+ "visual.vision_model.encoder.layers.21.layer_norm2.bias": "model-00001-of-00001.safetensors",
364
+ "visual.vision_model.encoder.layers.21.mlp.fc1.weight": "model-00001-of-00001.safetensors",
365
+ "visual.vision_model.encoder.layers.21.mlp.fc1.bias": "model-00001-of-00001.safetensors",
366
+ "visual.vision_model.encoder.layers.21.mlp.fc2.weight": "model-00001-of-00001.safetensors",
367
+ "visual.vision_model.encoder.layers.21.mlp.fc2.bias": "model-00001-of-00001.safetensors",
368
+ "visual.vision_model.encoder.layers.22.layer_norm1.weight": "model-00001-of-00001.safetensors",
369
+ "visual.vision_model.encoder.layers.22.layer_norm1.bias": "model-00001-of-00001.safetensors",
370
+ "visual.vision_model.encoder.layers.22.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
371
+ "visual.vision_model.encoder.layers.22.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
372
+ "visual.vision_model.encoder.layers.22.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
373
+ "visual.vision_model.encoder.layers.22.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
374
+ "visual.vision_model.encoder.layers.22.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
375
+ "visual.vision_model.encoder.layers.22.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
376
+ "visual.vision_model.encoder.layers.22.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
377
+ "visual.vision_model.encoder.layers.22.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
378
+ "visual.vision_model.encoder.layers.22.layer_norm2.weight": "model-00001-of-00001.safetensors",
379
+ "visual.vision_model.encoder.layers.22.layer_norm2.bias": "model-00001-of-00001.safetensors",
380
+ "visual.vision_model.encoder.layers.22.mlp.fc1.weight": "model-00001-of-00001.safetensors",
381
+ "visual.vision_model.encoder.layers.22.mlp.fc1.bias": "model-00001-of-00001.safetensors",
382
+ "visual.vision_model.encoder.layers.22.mlp.fc2.weight": "model-00001-of-00001.safetensors",
383
+ "visual.vision_model.encoder.layers.22.mlp.fc2.bias": "model-00001-of-00001.safetensors",
384
+ "visual.vision_model.encoder.layers.23.layer_norm1.weight": "model-00001-of-00001.safetensors",
385
+ "visual.vision_model.encoder.layers.23.layer_norm1.bias": "model-00001-of-00001.safetensors",
386
+ "visual.vision_model.encoder.layers.23.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
387
+ "visual.vision_model.encoder.layers.23.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
388
+ "visual.vision_model.encoder.layers.23.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
389
+ "visual.vision_model.encoder.layers.23.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
390
+ "visual.vision_model.encoder.layers.23.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
391
+ "visual.vision_model.encoder.layers.23.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
392
+ "visual.vision_model.encoder.layers.23.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
393
+ "visual.vision_model.encoder.layers.23.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
394
+ "visual.vision_model.encoder.layers.23.layer_norm2.weight": "model-00001-of-00001.safetensors",
395
+ "visual.vision_model.encoder.layers.23.layer_norm2.bias": "model-00001-of-00001.safetensors",
396
+ "visual.vision_model.encoder.layers.23.mlp.fc1.weight": "model-00001-of-00001.safetensors",
397
+ "visual.vision_model.encoder.layers.23.mlp.fc1.bias": "model-00001-of-00001.safetensors",
398
+ "visual.vision_model.encoder.layers.23.mlp.fc2.weight": "model-00001-of-00001.safetensors",
399
+ "visual.vision_model.encoder.layers.23.mlp.fc2.bias": "model-00001-of-00001.safetensors",
400
+ "visual.vision_model.encoder.layers.24.layer_norm1.weight": "model-00001-of-00001.safetensors",
401
+ "visual.vision_model.encoder.layers.24.layer_norm1.bias": "model-00001-of-00001.safetensors",
402
+ "visual.vision_model.encoder.layers.24.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
403
+ "visual.vision_model.encoder.layers.24.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
404
+ "visual.vision_model.encoder.layers.24.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
405
+ "visual.vision_model.encoder.layers.24.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
406
+ "visual.vision_model.encoder.layers.24.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
407
+ "visual.vision_model.encoder.layers.24.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
408
+ "visual.vision_model.encoder.layers.24.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
409
+ "visual.vision_model.encoder.layers.24.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
410
+ "visual.vision_model.encoder.layers.24.layer_norm2.weight": "model-00001-of-00001.safetensors",
411
+ "visual.vision_model.encoder.layers.24.layer_norm2.bias": "model-00001-of-00001.safetensors",
412
+ "visual.vision_model.encoder.layers.24.mlp.fc1.weight": "model-00001-of-00001.safetensors",
413
+ "visual.vision_model.encoder.layers.24.mlp.fc1.bias": "model-00001-of-00001.safetensors",
414
+ "visual.vision_model.encoder.layers.24.mlp.fc2.weight": "model-00001-of-00001.safetensors",
415
+ "visual.vision_model.encoder.layers.24.mlp.fc2.bias": "model-00001-of-00001.safetensors",
416
+ "visual.vision_model.encoder.layers.25.layer_norm1.weight": "model-00001-of-00001.safetensors",
417
+ "visual.vision_model.encoder.layers.25.layer_norm1.bias": "model-00001-of-00001.safetensors",
418
+ "visual.vision_model.encoder.layers.25.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
419
+ "visual.vision_model.encoder.layers.25.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
420
+ "visual.vision_model.encoder.layers.25.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
421
+ "visual.vision_model.encoder.layers.25.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
422
+ "visual.vision_model.encoder.layers.25.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
423
+ "visual.vision_model.encoder.layers.25.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
424
+ "visual.vision_model.encoder.layers.25.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
425
+ "visual.vision_model.encoder.layers.25.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
426
+ "visual.vision_model.encoder.layers.25.layer_norm2.weight": "model-00001-of-00001.safetensors",
427
+ "visual.vision_model.encoder.layers.25.layer_norm2.bias": "model-00001-of-00001.safetensors",
428
+ "visual.vision_model.encoder.layers.25.mlp.fc1.weight": "model-00001-of-00001.safetensors",
429
+ "visual.vision_model.encoder.layers.25.mlp.fc1.bias": "model-00001-of-00001.safetensors",
430
+ "visual.vision_model.encoder.layers.25.mlp.fc2.weight": "model-00001-of-00001.safetensors",
431
+ "visual.vision_model.encoder.layers.25.mlp.fc2.bias": "model-00001-of-00001.safetensors",
432
+ "visual.vision_model.encoder.layers.26.layer_norm1.weight": "model-00001-of-00001.safetensors",
433
+ "visual.vision_model.encoder.layers.26.layer_norm1.bias": "model-00001-of-00001.safetensors",
434
+ "visual.vision_model.encoder.layers.26.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
435
+ "visual.vision_model.encoder.layers.26.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
436
+ "visual.vision_model.encoder.layers.26.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
437
+ "visual.vision_model.encoder.layers.26.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
438
+ "visual.vision_model.encoder.layers.26.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
439
+ "visual.vision_model.encoder.layers.26.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
440
+ "visual.vision_model.encoder.layers.26.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
441
+ "visual.vision_model.encoder.layers.26.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
442
+ "visual.vision_model.encoder.layers.26.layer_norm2.weight": "model-00001-of-00001.safetensors",
443
+ "visual.vision_model.encoder.layers.26.layer_norm2.bias": "model-00001-of-00001.safetensors",
444
+ "visual.vision_model.encoder.layers.26.mlp.fc1.weight": "model-00001-of-00001.safetensors",
445
+ "visual.vision_model.encoder.layers.26.mlp.fc1.bias": "model-00001-of-00001.safetensors",
446
+ "visual.vision_model.encoder.layers.26.mlp.fc2.weight": "model-00001-of-00001.safetensors",
447
+ "visual.vision_model.encoder.layers.26.mlp.fc2.bias": "model-00001-of-00001.safetensors",
448
+ "visual.vision_model.post_layernorm.weight": "model-00001-of-00001.safetensors",
449
+ "visual.vision_model.post_layernorm.bias": "model-00001-of-00001.safetensors",
450
+ "visual.vision_model.head.probe": "model-00001-of-00001.safetensors",
451
+ "visual.vision_model.head.attention.in_proj_weight": "model-00001-of-00001.safetensors",
452
+ "visual.vision_model.head.attention.in_proj_bias": "model-00001-of-00001.safetensors",
453
+ "visual.vision_model.head.attention.out_proj.weight": "model-00001-of-00001.safetensors",
454
+ "visual.vision_model.head.attention.out_proj.bias": "model-00001-of-00001.safetensors",
455
+ "visual.vision_model.head.layernorm.weight": "model-00001-of-00001.safetensors",
456
+ "visual.vision_model.head.layernorm.bias": "model-00001-of-00001.safetensors",
457
+ "visual.vision_model.head.mlp.fc1.weight": "model-00001-of-00001.safetensors",
458
+ "visual.vision_model.head.mlp.fc1.bias": "model-00001-of-00001.safetensors",
459
+ "visual.vision_model.head.mlp.fc2.weight": "model-00001-of-00001.safetensors",
460
+ "visual.vision_model.head.mlp.fc2.bias": "model-00001-of-00001.safetensors",
461
+ "model.embed_tokens.weight": "model-00001-of-00001.safetensors",
462
+ "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
463
+ "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
464
+ "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
465
+ "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
466
+ "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
467
+ "model.layers.0.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
468
+ "model.layers.0.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
469
+ "model.layers.0.input_layernorm.weight": "model-00001-of-00001.safetensors",
470
+ "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
471
+ "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
472
+ "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
473
+ "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
474
+ "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
475
+ "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
476
+ "model.layers.1.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
477
+ "model.layers.1.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
478
+ "model.layers.1.input_layernorm.weight": "model-00001-of-00001.safetensors",
479
+ "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
480
+ "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
481
+ "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
482
+ "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
483
+ "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
484
+ "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
485
+ "model.layers.2.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
486
+ "model.layers.2.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
487
+ "model.layers.2.input_layernorm.weight": "model-00001-of-00001.safetensors",
488
+ "model.layers.2.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
489
+ "model.layers.3.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
490
+ "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
491
+ "model.layers.3.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
492
+ "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
493
+ "model.layers.3.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
494
+ "model.layers.3.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
495
+ "model.layers.3.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
496
+ "model.layers.3.input_layernorm.weight": "model-00001-of-00001.safetensors",
497
+ "model.layers.3.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
498
+ "model.layers.4.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
499
+ "model.layers.4.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
500
+ "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
501
+ "model.layers.4.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
502
+ "model.layers.4.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
503
+ "model.layers.4.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
504
+ "model.layers.4.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
505
+ "model.layers.4.input_layernorm.weight": "model-00001-of-00001.safetensors",
506
+ "model.layers.4.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
507
+ "model.layers.5.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
508
+ "model.layers.5.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
509
+ "model.layers.5.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
510
+ "model.layers.5.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
511
+ "model.layers.5.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
512
+ "model.layers.5.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
513
+ "model.layers.5.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
514
+ "model.layers.5.input_layernorm.weight": "model-00001-of-00001.safetensors",
515
+ "model.layers.5.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
516
+ "model.layers.6.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
517
+ "model.layers.6.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
518
+ "model.layers.6.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
519
+ "model.layers.6.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
520
+ "model.layers.6.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
521
+ "model.layers.6.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
522
+ "model.layers.6.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
523
+ "model.layers.6.input_layernorm.weight": "model-00001-of-00001.safetensors",
524
+ "model.layers.6.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
525
+ "model.layers.7.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
526
+ "model.layers.7.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
527
+ "model.layers.7.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
528
+ "model.layers.7.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
529
+ "model.layers.7.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
530
+ "model.layers.7.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
531
+ "model.layers.7.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
532
+ "model.layers.7.input_layernorm.weight": "model-00001-of-00001.safetensors",
533
+ "model.layers.7.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
534
+ "model.layers.8.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
535
+ "model.layers.8.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
536
+ "model.layers.8.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
537
+ "model.layers.8.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
538
+ "model.layers.8.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
539
+ "model.layers.8.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
540
+ "model.layers.8.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
541
+ "model.layers.8.input_layernorm.weight": "model-00001-of-00001.safetensors",
542
+ "model.layers.8.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
543
+ "model.layers.9.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
544
+ "model.layers.9.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
545
+ "model.layers.9.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
546
+ "model.layers.9.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
547
+ "model.layers.9.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
548
+ "model.layers.9.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
549
+ "model.layers.9.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
550
+ "model.layers.9.input_layernorm.weight": "model-00001-of-00001.safetensors",
551
+ "model.layers.9.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
552
+ "model.layers.10.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
553
+ "model.layers.10.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
554
+ "model.layers.10.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
555
+ "model.layers.10.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
556
+ "model.layers.10.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
557
+ "model.layers.10.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
558
+ "model.layers.10.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
559
+ "model.layers.10.input_layernorm.weight": "model-00001-of-00001.safetensors",
560
+ "model.layers.10.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
561
+ "model.layers.11.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
562
+ "model.layers.11.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
563
+ "model.layers.11.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
564
+ "model.layers.11.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
565
+ "model.layers.11.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
566
+ "model.layers.11.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
567
+ "model.layers.11.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
568
+ "model.layers.11.input_layernorm.weight": "model-00001-of-00001.safetensors",
569
+ "model.layers.11.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
570
+ "model.layers.12.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
571
+ "model.layers.12.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
572
+ "model.layers.12.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
573
+ "model.layers.12.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
574
+ "model.layers.12.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
575
+ "model.layers.12.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
576
+ "model.layers.12.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
577
+ "model.layers.12.input_layernorm.weight": "model-00001-of-00001.safetensors",
578
+ "model.layers.12.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
579
+ "model.layers.13.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
580
+ "model.layers.13.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
581
+ "model.layers.13.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
582
+ "model.layers.13.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
583
+ "model.layers.13.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
584
+ "model.layers.13.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
585
+ "model.layers.13.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
586
+ "model.layers.13.input_layernorm.weight": "model-00001-of-00001.safetensors",
587
+ "model.layers.13.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
588
+ "model.layers.14.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
589
+ "model.layers.14.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
590
+ "model.layers.14.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
591
+ "model.layers.14.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
592
+ "model.layers.14.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
593
+ "model.layers.14.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
594
+ "model.layers.14.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
595
+ "model.layers.14.input_layernorm.weight": "model-00001-of-00001.safetensors",
596
+ "model.layers.14.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
597
+ "model.layers.15.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
598
+ "model.layers.15.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
599
+ "model.layers.15.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
600
+ "model.layers.15.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
601
+ "model.layers.15.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
602
+ "model.layers.15.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
603
+ "model.layers.15.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
604
+ "model.layers.15.input_layernorm.weight": "model-00001-of-00001.safetensors",
605
+ "model.layers.15.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
606
+ "model.layers.16.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
607
+ "model.layers.16.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
608
+ "model.layers.16.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
609
+ "model.layers.16.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
610
+ "model.layers.16.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
611
+ "model.layers.16.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
612
+ "model.layers.16.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
613
+ "model.layers.16.input_layernorm.weight": "model-00001-of-00001.safetensors",
614
+ "model.layers.16.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
615
+ "model.layers.17.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
616
+ "model.layers.17.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
617
+ "model.layers.17.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
618
+ "model.layers.17.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
619
+ "model.layers.17.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
620
+ "model.layers.17.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
621
+ "model.layers.17.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
622
+ "model.layers.17.input_layernorm.weight": "model-00001-of-00001.safetensors",
623
+ "model.layers.17.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
624
+ "model.norm.weight": "model-00001-of-00001.safetensors",
625
+ "lm_head.weight": "model-00001-of-00001.safetensors"
626
+ }
627
+ }
checkpoint-378/optimizer-00001-of-00001.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d15e7372d8ff5e0050feda539c5dee7463c8bcc5ec1b4e3bc4d2f3e64b83df6b
3
+ size 7245150152
checkpoint-378/optimizer.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-378/preprocessor_config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_convert_rgb": true,
3
+ "do_normalize": true,
4
+ "do_rescale": true,
5
+ "do_resize": true,
6
+ "image_mean": [
7
+ 0.5,
8
+ 0.5,
9
+ 0.5
10
+ ],
11
+ "image_processor_type": "SiglipImageProcessor",
12
+ "image_std": [
13
+ 0.5,
14
+ 0.5,
15
+ 0.5
16
+ ],
17
+ "max_pixels": 2822400,
18
+ "merge_size": 2,
19
+ "min_pixels": 147384,
20
+ "patch_size": 14,
21
+ "resample": 3,
22
+ "rescale_factor": 0.00392156862745098,
23
+ "size": {
24
+ "max_pixels": 2822400,
25
+ "min_pixels": 147384
26
+ },
27
+ "temporal_conv_size": 1,
28
+ "temporal_patch_size": 1
29
+ }
checkpoint-378/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:29cde6478b49d2682dfa8c033ed7da4f33ab76ccd1941cfe950453d3923dfec3
3
+ size 32008
checkpoint-378/scheduler.pdparams ADDED
Binary file (51 Bytes). View file
 
checkpoint-378/special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": false, "special": false}, "eos_token": "</s>", "unk_token": {"content": "<unk>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": false, "special": false}, "sep_token": "<|end_of_sentence|>", "pad_token": {"content": "<unk>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": false, "special": false}, "cls_token": "<|begin_of_sentence|>", "mask_token": {"content": "<mask:1>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": false, "special": false}, "additional_special_tokens": ["<|IMAGE_PLACEHOLDER|>", "<|image_pad|>", "<|IMAGE_START|>", "<|IMAGE_END|>", "<|video_pad|>"]}
checkpoint-378/static_name_to_dyg_name.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"layer_norm_0.w_0": "mlp_AR.pre_norm.weight", "layer_norm_0.b_0": "mlp_AR.pre_norm.bias", "linear_0.w_0": "mlp_AR.linear_1.weight", "linear_0.b_0": "mlp_AR.linear_1.bias", "linear_1.w_0": "mlp_AR.linear_2.weight", "linear_1.b_0": "mlp_AR.linear_2.bias", "conv2d_0.w_0": "visual.vision_model.embeddings.patch_embedding.weight", "conv2d_0.b_0": "visual.vision_model.embeddings.patch_embedding.bias", "embedding_0.w_0": "visual.vision_model.embeddings.position_embedding.weight", "embedding_1.w_0": "visual.vision_model.embeddings.packing_position_embedding.weight", "layer_norm_1.w_0": "visual.vision_model.encoder.layers.0.layer_norm1.weight", "layer_norm_1.b_0": "visual.vision_model.encoder.layers.0.layer_norm1.bias", "linear_2.w_0": "visual.vision_model.encoder.layers.0.self_attn.k_proj.weight", "linear_2.b_0": "visual.vision_model.encoder.layers.0.self_attn.k_proj.bias", "linear_3.w_0": "visual.vision_model.encoder.layers.0.self_attn.v_proj.weight", "linear_3.b_0": "visual.vision_model.encoder.layers.0.self_attn.v_proj.bias", "linear_4.w_0": "visual.vision_model.encoder.layers.0.self_attn.q_proj.weight", "linear_4.b_0": "visual.vision_model.encoder.layers.0.self_attn.q_proj.bias", "linear_5.w_0": "visual.vision_model.encoder.layers.0.self_attn.out_proj.weight", "linear_5.b_0": "visual.vision_model.encoder.layers.0.self_attn.out_proj.bias", "layer_norm_2.w_0": "visual.vision_model.encoder.layers.0.layer_norm2.weight", "layer_norm_2.b_0": "visual.vision_model.encoder.layers.0.layer_norm2.bias", "linear_6.w_0": "visual.vision_model.encoder.layers.0.mlp.fc1.weight", "linear_6.b_0": "visual.vision_model.encoder.layers.0.mlp.fc1.bias", "linear_7.w_0": "visual.vision_model.encoder.layers.0.mlp.fc2.weight", "linear_7.b_0": "visual.vision_model.encoder.layers.0.mlp.fc2.bias", "layer_norm_3.w_0": "visual.vision_model.encoder.layers.1.layer_norm1.weight", "layer_norm_3.b_0": "visual.vision_model.encoder.layers.1.layer_norm1.bias", "linear_8.w_0": "visual.vision_model.encoder.layers.1.self_attn.k_proj.weight", "linear_8.b_0": "visual.vision_model.encoder.layers.1.self_attn.k_proj.bias", "linear_9.w_0": "visual.vision_model.encoder.layers.1.self_attn.v_proj.weight", "linear_9.b_0": "visual.vision_model.encoder.layers.1.self_attn.v_proj.bias", "linear_10.w_0": "visual.vision_model.encoder.layers.1.self_attn.q_proj.weight", "linear_10.b_0": "visual.vision_model.encoder.layers.1.self_attn.q_proj.bias", "linear_11.w_0": "visual.vision_model.encoder.layers.1.self_attn.out_proj.weight", "linear_11.b_0": "visual.vision_model.encoder.layers.1.self_attn.out_proj.bias", "layer_norm_4.w_0": "visual.vision_model.encoder.layers.1.layer_norm2.weight", "layer_norm_4.b_0": "visual.vision_model.encoder.layers.1.layer_norm2.bias", "linear_12.w_0": "visual.vision_model.encoder.layers.1.mlp.fc1.weight", "linear_12.b_0": "visual.vision_model.encoder.layers.1.mlp.fc1.bias", "linear_13.w_0": "visual.vision_model.encoder.layers.1.mlp.fc2.weight", "linear_13.b_0": "visual.vision_model.encoder.layers.1.mlp.fc2.bias", "layer_norm_5.w_0": "visual.vision_model.encoder.layers.2.layer_norm1.weight", "layer_norm_5.b_0": "visual.vision_model.encoder.layers.2.layer_norm1.bias", "linear_14.w_0": "visual.vision_model.encoder.layers.2.self_attn.k_proj.weight", "linear_14.b_0": "visual.vision_model.encoder.layers.2.self_attn.k_proj.bias", "linear_15.w_0": "visual.vision_model.encoder.layers.2.self_attn.v_proj.weight", "linear_15.b_0": "visual.vision_model.encoder.layers.2.self_attn.v_proj.bias", "linear_16.w_0": "visual.vision_model.encoder.layers.2.self_attn.q_proj.weight", "linear_16.b_0": "visual.vision_model.encoder.layers.2.self_attn.q_proj.bias", "linear_17.w_0": "visual.vision_model.encoder.layers.2.self_attn.out_proj.weight", "linear_17.b_0": "visual.vision_model.encoder.layers.2.self_attn.out_proj.bias", "layer_norm_6.w_0": "visual.vision_model.encoder.layers.2.layer_norm2.weight", "layer_norm_6.b_0": "visual.vision_model.encoder.layers.2.layer_norm2.bias", "linear_18.w_0": "visual.vision_model.encoder.layers.2.mlp.fc1.weight", "linear_18.b_0": "visual.vision_model.encoder.layers.2.mlp.fc1.bias", "linear_19.w_0": "visual.vision_model.encoder.layers.2.mlp.fc2.weight", "linear_19.b_0": "visual.vision_model.encoder.layers.2.mlp.fc2.bias", "layer_norm_7.w_0": "visual.vision_model.encoder.layers.3.layer_norm1.weight", "layer_norm_7.b_0": "visual.vision_model.encoder.layers.3.layer_norm1.bias", "linear_20.w_0": "visual.vision_model.encoder.layers.3.self_attn.k_proj.weight", "linear_20.b_0": "visual.vision_model.encoder.layers.3.self_attn.k_proj.bias", "linear_21.w_0": "visual.vision_model.encoder.layers.3.self_attn.v_proj.weight", "linear_21.b_0": "visual.vision_model.encoder.layers.3.self_attn.v_proj.bias", "linear_22.w_0": "visual.vision_model.encoder.layers.3.self_attn.q_proj.weight", "linear_22.b_0": "visual.vision_model.encoder.layers.3.self_attn.q_proj.bias", "linear_23.w_0": "visual.vision_model.encoder.layers.3.self_attn.out_proj.weight", "linear_23.b_0": "visual.vision_model.encoder.layers.3.self_attn.out_proj.bias", "layer_norm_8.w_0": "visual.vision_model.encoder.layers.3.layer_norm2.weight", "layer_norm_8.b_0": "visual.vision_model.encoder.layers.3.layer_norm2.bias", "linear_24.w_0": "visual.vision_model.encoder.layers.3.mlp.fc1.weight", "linear_24.b_0": "visual.vision_model.encoder.layers.3.mlp.fc1.bias", "linear_25.w_0": "visual.vision_model.encoder.layers.3.mlp.fc2.weight", "linear_25.b_0": "visual.vision_model.encoder.layers.3.mlp.fc2.bias", "layer_norm_9.w_0": "visual.vision_model.encoder.layers.4.layer_norm1.weight", "layer_norm_9.b_0": "visual.vision_model.encoder.layers.4.layer_norm1.bias", "linear_26.w_0": "visual.vision_model.encoder.layers.4.self_attn.k_proj.weight", "linear_26.b_0": "visual.vision_model.encoder.layers.4.self_attn.k_proj.bias", "linear_27.w_0": "visual.vision_model.encoder.layers.4.self_attn.v_proj.weight", "linear_27.b_0": "visual.vision_model.encoder.layers.4.self_attn.v_proj.bias", "linear_28.w_0": "visual.vision_model.encoder.layers.4.self_attn.q_proj.weight", "linear_28.b_0": "visual.vision_model.encoder.layers.4.self_attn.q_proj.bias", "linear_29.w_0": "visual.vision_model.encoder.layers.4.self_attn.out_proj.weight", "linear_29.b_0": "visual.vision_model.encoder.layers.4.self_attn.out_proj.bias", "layer_norm_10.w_0": "visual.vision_model.encoder.layers.4.layer_norm2.weight", "layer_norm_10.b_0": "visual.vision_model.encoder.layers.4.layer_norm2.bias", "linear_30.w_0": "visual.vision_model.encoder.layers.4.mlp.fc1.weight", "linear_30.b_0": "visual.vision_model.encoder.layers.4.mlp.fc1.bias", "linear_31.w_0": "visual.vision_model.encoder.layers.4.mlp.fc2.weight", "linear_31.b_0": "visual.vision_model.encoder.layers.4.mlp.fc2.bias", "layer_norm_11.w_0": "visual.vision_model.encoder.layers.5.layer_norm1.weight", "layer_norm_11.b_0": "visual.vision_model.encoder.layers.5.layer_norm1.bias", "linear_32.w_0": "visual.vision_model.encoder.layers.5.self_attn.k_proj.weight", "linear_32.b_0": "visual.vision_model.encoder.layers.5.self_attn.k_proj.bias", "linear_33.w_0": "visual.vision_model.encoder.layers.5.self_attn.v_proj.weight", "linear_33.b_0": "visual.vision_model.encoder.layers.5.self_attn.v_proj.bias", "linear_34.w_0": "visual.vision_model.encoder.layers.5.self_attn.q_proj.weight", "linear_34.b_0": "visual.vision_model.encoder.layers.5.self_attn.q_proj.bias", "linear_35.w_0": "visual.vision_model.encoder.layers.5.self_attn.out_proj.weight", "linear_35.b_0": "visual.vision_model.encoder.layers.5.self_attn.out_proj.bias", "layer_norm_12.w_0": "visual.vision_model.encoder.layers.5.layer_norm2.weight", "layer_norm_12.b_0": "visual.vision_model.encoder.layers.5.layer_norm2.bias", "linear_36.w_0": "visual.vision_model.encoder.layers.5.mlp.fc1.weight", "linear_36.b_0": "visual.vision_model.encoder.layers.5.mlp.fc1.bias", "linear_37.w_0": "visual.vision_model.encoder.layers.5.mlp.fc2.weight", "linear_37.b_0": "visual.vision_model.encoder.layers.5.mlp.fc2.bias", "layer_norm_13.w_0": "visual.vision_model.encoder.layers.6.layer_norm1.weight", "layer_norm_13.b_0": "visual.vision_model.encoder.layers.6.layer_norm1.bias", "linear_38.w_0": "visual.vision_model.encoder.layers.6.self_attn.k_proj.weight", "linear_38.b_0": "visual.vision_model.encoder.layers.6.self_attn.k_proj.bias", "linear_39.w_0": "visual.vision_model.encoder.layers.6.self_attn.v_proj.weight", "linear_39.b_0": "visual.vision_model.encoder.layers.6.self_attn.v_proj.bias", "linear_40.w_0": "visual.vision_model.encoder.layers.6.self_attn.q_proj.weight", "linear_40.b_0": "visual.vision_model.encoder.layers.6.self_attn.q_proj.bias", "linear_41.w_0": "visual.vision_model.encoder.layers.6.self_attn.out_proj.weight", "linear_41.b_0": "visual.vision_model.encoder.layers.6.self_attn.out_proj.bias", "layer_norm_14.w_0": "visual.vision_model.encoder.layers.6.layer_norm2.weight", "layer_norm_14.b_0": "visual.vision_model.encoder.layers.6.layer_norm2.bias", "linear_42.w_0": "visual.vision_model.encoder.layers.6.mlp.fc1.weight", "linear_42.b_0": "visual.vision_model.encoder.layers.6.mlp.fc1.bias", "linear_43.w_0": "visual.vision_model.encoder.layers.6.mlp.fc2.weight", "linear_43.b_0": "visual.vision_model.encoder.layers.6.mlp.fc2.bias", "layer_norm_15.w_0": "visual.vision_model.encoder.layers.7.layer_norm1.weight", "layer_norm_15.b_0": "visual.vision_model.encoder.layers.7.layer_norm1.bias", "linear_44.w_0": "visual.vision_model.encoder.layers.7.self_attn.k_proj.weight", "linear_44.b_0": "visual.vision_model.encoder.layers.7.self_attn.k_proj.bias", "linear_45.w_0": "visual.vision_model.encoder.layers.7.self_attn.v_proj.weight", "linear_45.b_0": "visual.vision_model.encoder.layers.7.self_attn.v_proj.bias", "linear_46.w_0": "visual.vision_model.encoder.layers.7.self_attn.q_proj.weight", "linear_46.b_0": "visual.vision_model.encoder.layers.7.self_attn.q_proj.bias", "linear_47.w_0": "visual.vision_model.encoder.layers.7.self_attn.out_proj.weight", "linear_47.b_0": "visual.vision_model.encoder.layers.7.self_attn.out_proj.bias", "layer_norm_16.w_0": "visual.vision_model.encoder.layers.7.layer_norm2.weight", "layer_norm_16.b_0": "visual.vision_model.encoder.layers.7.layer_norm2.bias", "linear_48.w_0": "visual.vision_model.encoder.layers.7.mlp.fc1.weight", "linear_48.b_0": "visual.vision_model.encoder.layers.7.mlp.fc1.bias", "linear_49.w_0": "visual.vision_model.encoder.layers.7.mlp.fc2.weight", "linear_49.b_0": "visual.vision_model.encoder.layers.7.mlp.fc2.bias", "layer_norm_17.w_0": "visual.vision_model.encoder.layers.8.layer_norm1.weight", "layer_norm_17.b_0": "visual.vision_model.encoder.layers.8.layer_norm1.bias", "linear_50.w_0": "visual.vision_model.encoder.layers.8.self_attn.k_proj.weight", "linear_50.b_0": "visual.vision_model.encoder.layers.8.self_attn.k_proj.bias", "linear_51.w_0": "visual.vision_model.encoder.layers.8.self_attn.v_proj.weight", "linear_51.b_0": "visual.vision_model.encoder.layers.8.self_attn.v_proj.bias", "linear_52.w_0": "visual.vision_model.encoder.layers.8.self_attn.q_proj.weight", "linear_52.b_0": "visual.vision_model.encoder.layers.8.self_attn.q_proj.bias", "linear_53.w_0": "visual.vision_model.encoder.layers.8.self_attn.out_proj.weight", "linear_53.b_0": "visual.vision_model.encoder.layers.8.self_attn.out_proj.bias", "layer_norm_18.w_0": "visual.vision_model.encoder.layers.8.layer_norm2.weight", "layer_norm_18.b_0": "visual.vision_model.encoder.layers.8.layer_norm2.bias", "linear_54.w_0": "visual.vision_model.encoder.layers.8.mlp.fc1.weight", "linear_54.b_0": "visual.vision_model.encoder.layers.8.mlp.fc1.bias", "linear_55.w_0": "visual.vision_model.encoder.layers.8.mlp.fc2.weight", "linear_55.b_0": "visual.vision_model.encoder.layers.8.mlp.fc2.bias", "layer_norm_19.w_0": "visual.vision_model.encoder.layers.9.layer_norm1.weight", "layer_norm_19.b_0": "visual.vision_model.encoder.layers.9.layer_norm1.bias", "linear_56.w_0": "visual.vision_model.encoder.layers.9.self_attn.k_proj.weight", "linear_56.b_0": "visual.vision_model.encoder.layers.9.self_attn.k_proj.bias", "linear_57.w_0": "visual.vision_model.encoder.layers.9.self_attn.v_proj.weight", "linear_57.b_0": "visual.vision_model.encoder.layers.9.self_attn.v_proj.bias", "linear_58.w_0": "visual.vision_model.encoder.layers.9.self_attn.q_proj.weight", "linear_58.b_0": "visual.vision_model.encoder.layers.9.self_attn.q_proj.bias", "linear_59.w_0": "visual.vision_model.encoder.layers.9.self_attn.out_proj.weight", "linear_59.b_0": "visual.vision_model.encoder.layers.9.self_attn.out_proj.bias", "layer_norm_20.w_0": "visual.vision_model.encoder.layers.9.layer_norm2.weight", "layer_norm_20.b_0": "visual.vision_model.encoder.layers.9.layer_norm2.bias", "linear_60.w_0": "visual.vision_model.encoder.layers.9.mlp.fc1.weight", "linear_60.b_0": "visual.vision_model.encoder.layers.9.mlp.fc1.bias", "linear_61.w_0": "visual.vision_model.encoder.layers.9.mlp.fc2.weight", "linear_61.b_0": "visual.vision_model.encoder.layers.9.mlp.fc2.bias", "layer_norm_21.w_0": "visual.vision_model.encoder.layers.10.layer_norm1.weight", "layer_norm_21.b_0": "visual.vision_model.encoder.layers.10.layer_norm1.bias", "linear_62.w_0": "visual.vision_model.encoder.layers.10.self_attn.k_proj.weight", "linear_62.b_0": "visual.vision_model.encoder.layers.10.self_attn.k_proj.bias", "linear_63.w_0": "visual.vision_model.encoder.layers.10.self_attn.v_proj.weight", "linear_63.b_0": "visual.vision_model.encoder.layers.10.self_attn.v_proj.bias", "linear_64.w_0": "visual.vision_model.encoder.layers.10.self_attn.q_proj.weight", "linear_64.b_0": "visual.vision_model.encoder.layers.10.self_attn.q_proj.bias", "linear_65.w_0": "visual.vision_model.encoder.layers.10.self_attn.out_proj.weight", "linear_65.b_0": "visual.vision_model.encoder.layers.10.self_attn.out_proj.bias", "layer_norm_22.w_0": "visual.vision_model.encoder.layers.10.layer_norm2.weight", "layer_norm_22.b_0": "visual.vision_model.encoder.layers.10.layer_norm2.bias", "linear_66.w_0": "visual.vision_model.encoder.layers.10.mlp.fc1.weight", "linear_66.b_0": "visual.vision_model.encoder.layers.10.mlp.fc1.bias", "linear_67.w_0": "visual.vision_model.encoder.layers.10.mlp.fc2.weight", "linear_67.b_0": "visual.vision_model.encoder.layers.10.mlp.fc2.bias", "layer_norm_23.w_0": "visual.vision_model.encoder.layers.11.layer_norm1.weight", "layer_norm_23.b_0": "visual.vision_model.encoder.layers.11.layer_norm1.bias", "linear_68.w_0": "visual.vision_model.encoder.layers.11.self_attn.k_proj.weight", "linear_68.b_0": "visual.vision_model.encoder.layers.11.self_attn.k_proj.bias", "linear_69.w_0": "visual.vision_model.encoder.layers.11.self_attn.v_proj.weight", "linear_69.b_0": "visual.vision_model.encoder.layers.11.self_attn.v_proj.bias", "linear_70.w_0": "visual.vision_model.encoder.layers.11.self_attn.q_proj.weight", "linear_70.b_0": "visual.vision_model.encoder.layers.11.self_attn.q_proj.bias", "linear_71.w_0": "visual.vision_model.encoder.layers.11.self_attn.out_proj.weight", "linear_71.b_0": "visual.vision_model.encoder.layers.11.self_attn.out_proj.bias", "layer_norm_24.w_0": "visual.vision_model.encoder.layers.11.layer_norm2.weight", "layer_norm_24.b_0": "visual.vision_model.encoder.layers.11.layer_norm2.bias", "linear_72.w_0": "visual.vision_model.encoder.layers.11.mlp.fc1.weight", "linear_72.b_0": "visual.vision_model.encoder.layers.11.mlp.fc1.bias", "linear_73.w_0": "visual.vision_model.encoder.layers.11.mlp.fc2.weight", "linear_73.b_0": "visual.vision_model.encoder.layers.11.mlp.fc2.bias", "layer_norm_25.w_0": "visual.vision_model.encoder.layers.12.layer_norm1.weight", "layer_norm_25.b_0": "visual.vision_model.encoder.layers.12.layer_norm1.bias", "linear_74.w_0": "visual.vision_model.encoder.layers.12.self_attn.k_proj.weight", "linear_74.b_0": "visual.vision_model.encoder.layers.12.self_attn.k_proj.bias", "linear_75.w_0": "visual.vision_model.encoder.layers.12.self_attn.v_proj.weight", "linear_75.b_0": "visual.vision_model.encoder.layers.12.self_attn.v_proj.bias", "linear_76.w_0": "visual.vision_model.encoder.layers.12.self_attn.q_proj.weight", "linear_76.b_0": "visual.vision_model.encoder.layers.12.self_attn.q_proj.bias", "linear_77.w_0": "visual.vision_model.encoder.layers.12.self_attn.out_proj.weight", "linear_77.b_0": "visual.vision_model.encoder.layers.12.self_attn.out_proj.bias", "layer_norm_26.w_0": "visual.vision_model.encoder.layers.12.layer_norm2.weight", "layer_norm_26.b_0": "visual.vision_model.encoder.layers.12.layer_norm2.bias", "linear_78.w_0": "visual.vision_model.encoder.layers.12.mlp.fc1.weight", "linear_78.b_0": "visual.vision_model.encoder.layers.12.mlp.fc1.bias", "linear_79.w_0": "visual.vision_model.encoder.layers.12.mlp.fc2.weight", "linear_79.b_0": "visual.vision_model.encoder.layers.12.mlp.fc2.bias", "layer_norm_27.w_0": "visual.vision_model.encoder.layers.13.layer_norm1.weight", "layer_norm_27.b_0": "visual.vision_model.encoder.layers.13.layer_norm1.bias", "linear_80.w_0": "visual.vision_model.encoder.layers.13.self_attn.k_proj.weight", "linear_80.b_0": "visual.vision_model.encoder.layers.13.self_attn.k_proj.bias", "linear_81.w_0": "visual.vision_model.encoder.layers.13.self_attn.v_proj.weight", "linear_81.b_0": "visual.vision_model.encoder.layers.13.self_attn.v_proj.bias", "linear_82.w_0": "visual.vision_model.encoder.layers.13.self_attn.q_proj.weight", "linear_82.b_0": "visual.vision_model.encoder.layers.13.self_attn.q_proj.bias", "linear_83.w_0": "visual.vision_model.encoder.layers.13.self_attn.out_proj.weight", "linear_83.b_0": "visual.vision_model.encoder.layers.13.self_attn.out_proj.bias", "layer_norm_28.w_0": "visual.vision_model.encoder.layers.13.layer_norm2.weight", "layer_norm_28.b_0": "visual.vision_model.encoder.layers.13.layer_norm2.bias", "linear_84.w_0": "visual.vision_model.encoder.layers.13.mlp.fc1.weight", "linear_84.b_0": "visual.vision_model.encoder.layers.13.mlp.fc1.bias", "linear_85.w_0": "visual.vision_model.encoder.layers.13.mlp.fc2.weight", "linear_85.b_0": "visual.vision_model.encoder.layers.13.mlp.fc2.bias", "layer_norm_29.w_0": "visual.vision_model.encoder.layers.14.layer_norm1.weight", "layer_norm_29.b_0": "visual.vision_model.encoder.layers.14.layer_norm1.bias", "linear_86.w_0": "visual.vision_model.encoder.layers.14.self_attn.k_proj.weight", "linear_86.b_0": "visual.vision_model.encoder.layers.14.self_attn.k_proj.bias", "linear_87.w_0": "visual.vision_model.encoder.layers.14.self_attn.v_proj.weight", "linear_87.b_0": "visual.vision_model.encoder.layers.14.self_attn.v_proj.bias", "linear_88.w_0": "visual.vision_model.encoder.layers.14.self_attn.q_proj.weight", "linear_88.b_0": "visual.vision_model.encoder.layers.14.self_attn.q_proj.bias", "linear_89.w_0": "visual.vision_model.encoder.layers.14.self_attn.out_proj.weight", "linear_89.b_0": "visual.vision_model.encoder.layers.14.self_attn.out_proj.bias", "layer_norm_30.w_0": "visual.vision_model.encoder.layers.14.layer_norm2.weight", "layer_norm_30.b_0": "visual.vision_model.encoder.layers.14.layer_norm2.bias", "linear_90.w_0": "visual.vision_model.encoder.layers.14.mlp.fc1.weight", "linear_90.b_0": "visual.vision_model.encoder.layers.14.mlp.fc1.bias", "linear_91.w_0": "visual.vision_model.encoder.layers.14.mlp.fc2.weight", "linear_91.b_0": "visual.vision_model.encoder.layers.14.mlp.fc2.bias", "layer_norm_31.w_0": "visual.vision_model.encoder.layers.15.layer_norm1.weight", "layer_norm_31.b_0": "visual.vision_model.encoder.layers.15.layer_norm1.bias", "linear_92.w_0": "visual.vision_model.encoder.layers.15.self_attn.k_proj.weight", "linear_92.b_0": "visual.vision_model.encoder.layers.15.self_attn.k_proj.bias", "linear_93.w_0": "visual.vision_model.encoder.layers.15.self_attn.v_proj.weight", "linear_93.b_0": "visual.vision_model.encoder.layers.15.self_attn.v_proj.bias", "linear_94.w_0": "visual.vision_model.encoder.layers.15.self_attn.q_proj.weight", "linear_94.b_0": "visual.vision_model.encoder.layers.15.self_attn.q_proj.bias", "linear_95.w_0": "visual.vision_model.encoder.layers.15.self_attn.out_proj.weight", "linear_95.b_0": "visual.vision_model.encoder.layers.15.self_attn.out_proj.bias", "layer_norm_32.w_0": "visual.vision_model.encoder.layers.15.layer_norm2.weight", "layer_norm_32.b_0": "visual.vision_model.encoder.layers.15.layer_norm2.bias", "linear_96.w_0": "visual.vision_model.encoder.layers.15.mlp.fc1.weight", "linear_96.b_0": "visual.vision_model.encoder.layers.15.mlp.fc1.bias", "linear_97.w_0": "visual.vision_model.encoder.layers.15.mlp.fc2.weight", "linear_97.b_0": "visual.vision_model.encoder.layers.15.mlp.fc2.bias", "layer_norm_33.w_0": "visual.vision_model.encoder.layers.16.layer_norm1.weight", "layer_norm_33.b_0": "visual.vision_model.encoder.layers.16.layer_norm1.bias", "linear_98.w_0": "visual.vision_model.encoder.layers.16.self_attn.k_proj.weight", "linear_98.b_0": "visual.vision_model.encoder.layers.16.self_attn.k_proj.bias", "linear_99.w_0": "visual.vision_model.encoder.layers.16.self_attn.v_proj.weight", "linear_99.b_0": "visual.vision_model.encoder.layers.16.self_attn.v_proj.bias", "linear_100.w_0": "visual.vision_model.encoder.layers.16.self_attn.q_proj.weight", "linear_100.b_0": "visual.vision_model.encoder.layers.16.self_attn.q_proj.bias", "linear_101.w_0": "visual.vision_model.encoder.layers.16.self_attn.out_proj.weight", "linear_101.b_0": "visual.vision_model.encoder.layers.16.self_attn.out_proj.bias", "layer_norm_34.w_0": "visual.vision_model.encoder.layers.16.layer_norm2.weight", "layer_norm_34.b_0": "visual.vision_model.encoder.layers.16.layer_norm2.bias", "linear_102.w_0": "visual.vision_model.encoder.layers.16.mlp.fc1.weight", "linear_102.b_0": "visual.vision_model.encoder.layers.16.mlp.fc1.bias", "linear_103.w_0": "visual.vision_model.encoder.layers.16.mlp.fc2.weight", "linear_103.b_0": "visual.vision_model.encoder.layers.16.mlp.fc2.bias", "layer_norm_35.w_0": "visual.vision_model.encoder.layers.17.layer_norm1.weight", "layer_norm_35.b_0": "visual.vision_model.encoder.layers.17.layer_norm1.bias", "linear_104.w_0": "visual.vision_model.encoder.layers.17.self_attn.k_proj.weight", "linear_104.b_0": "visual.vision_model.encoder.layers.17.self_attn.k_proj.bias", "linear_105.w_0": "visual.vision_model.encoder.layers.17.self_attn.v_proj.weight", "linear_105.b_0": "visual.vision_model.encoder.layers.17.self_attn.v_proj.bias", "linear_106.w_0": "visual.vision_model.encoder.layers.17.self_attn.q_proj.weight", "linear_106.b_0": "visual.vision_model.encoder.layers.17.self_attn.q_proj.bias", "linear_107.w_0": "visual.vision_model.encoder.layers.17.self_attn.out_proj.weight", "linear_107.b_0": "visual.vision_model.encoder.layers.17.self_attn.out_proj.bias", "layer_norm_36.w_0": "visual.vision_model.encoder.layers.17.layer_norm2.weight", "layer_norm_36.b_0": "visual.vision_model.encoder.layers.17.layer_norm2.bias", "linear_108.w_0": "visual.vision_model.encoder.layers.17.mlp.fc1.weight", "linear_108.b_0": "visual.vision_model.encoder.layers.17.mlp.fc1.bias", "linear_109.w_0": "visual.vision_model.encoder.layers.17.mlp.fc2.weight", "linear_109.b_0": "visual.vision_model.encoder.layers.17.mlp.fc2.bias", "layer_norm_37.w_0": "visual.vision_model.encoder.layers.18.layer_norm1.weight", "layer_norm_37.b_0": "visual.vision_model.encoder.layers.18.layer_norm1.bias", "linear_110.w_0": "visual.vision_model.encoder.layers.18.self_attn.k_proj.weight", "linear_110.b_0": "visual.vision_model.encoder.layers.18.self_attn.k_proj.bias", "linear_111.w_0": "visual.vision_model.encoder.layers.18.self_attn.v_proj.weight", "linear_111.b_0": "visual.vision_model.encoder.layers.18.self_attn.v_proj.bias", "linear_112.w_0": "visual.vision_model.encoder.layers.18.self_attn.q_proj.weight", "linear_112.b_0": "visual.vision_model.encoder.layers.18.self_attn.q_proj.bias", "linear_113.w_0": "visual.vision_model.encoder.layers.18.self_attn.out_proj.weight", "linear_113.b_0": "visual.vision_model.encoder.layers.18.self_attn.out_proj.bias", "layer_norm_38.w_0": "visual.vision_model.encoder.layers.18.layer_norm2.weight", "layer_norm_38.b_0": "visual.vision_model.encoder.layers.18.layer_norm2.bias", "linear_114.w_0": "visual.vision_model.encoder.layers.18.mlp.fc1.weight", "linear_114.b_0": "visual.vision_model.encoder.layers.18.mlp.fc1.bias", "linear_115.w_0": "visual.vision_model.encoder.layers.18.mlp.fc2.weight", "linear_115.b_0": "visual.vision_model.encoder.layers.18.mlp.fc2.bias", "layer_norm_39.w_0": "visual.vision_model.encoder.layers.19.layer_norm1.weight", "layer_norm_39.b_0": "visual.vision_model.encoder.layers.19.layer_norm1.bias", "linear_116.w_0": "visual.vision_model.encoder.layers.19.self_attn.k_proj.weight", "linear_116.b_0": "visual.vision_model.encoder.layers.19.self_attn.k_proj.bias", "linear_117.w_0": "visual.vision_model.encoder.layers.19.self_attn.v_proj.weight", "linear_117.b_0": "visual.vision_model.encoder.layers.19.self_attn.v_proj.bias", "linear_118.w_0": "visual.vision_model.encoder.layers.19.self_attn.q_proj.weight", "linear_118.b_0": "visual.vision_model.encoder.layers.19.self_attn.q_proj.bias", "linear_119.w_0": "visual.vision_model.encoder.layers.19.self_attn.out_proj.weight", "linear_119.b_0": "visual.vision_model.encoder.layers.19.self_attn.out_proj.bias", "layer_norm_40.w_0": "visual.vision_model.encoder.layers.19.layer_norm2.weight", "layer_norm_40.b_0": "visual.vision_model.encoder.layers.19.layer_norm2.bias", "linear_120.w_0": "visual.vision_model.encoder.layers.19.mlp.fc1.weight", "linear_120.b_0": "visual.vision_model.encoder.layers.19.mlp.fc1.bias", "linear_121.w_0": "visual.vision_model.encoder.layers.19.mlp.fc2.weight", "linear_121.b_0": "visual.vision_model.encoder.layers.19.mlp.fc2.bias", "layer_norm_41.w_0": "visual.vision_model.encoder.layers.20.layer_norm1.weight", "layer_norm_41.b_0": "visual.vision_model.encoder.layers.20.layer_norm1.bias", "linear_122.w_0": "visual.vision_model.encoder.layers.20.self_attn.k_proj.weight", "linear_122.b_0": "visual.vision_model.encoder.layers.20.self_attn.k_proj.bias", "linear_123.w_0": "visual.vision_model.encoder.layers.20.self_attn.v_proj.weight", "linear_123.b_0": "visual.vision_model.encoder.layers.20.self_attn.v_proj.bias", "linear_124.w_0": "visual.vision_model.encoder.layers.20.self_attn.q_proj.weight", "linear_124.b_0": "visual.vision_model.encoder.layers.20.self_attn.q_proj.bias", "linear_125.w_0": "visual.vision_model.encoder.layers.20.self_attn.out_proj.weight", "linear_125.b_0": "visual.vision_model.encoder.layers.20.self_attn.out_proj.bias", "layer_norm_42.w_0": "visual.vision_model.encoder.layers.20.layer_norm2.weight", "layer_norm_42.b_0": "visual.vision_model.encoder.layers.20.layer_norm2.bias", "linear_126.w_0": "visual.vision_model.encoder.layers.20.mlp.fc1.weight", "linear_126.b_0": "visual.vision_model.encoder.layers.20.mlp.fc1.bias", "linear_127.w_0": "visual.vision_model.encoder.layers.20.mlp.fc2.weight", "linear_127.b_0": "visual.vision_model.encoder.layers.20.mlp.fc2.bias", "layer_norm_43.w_0": "visual.vision_model.encoder.layers.21.layer_norm1.weight", "layer_norm_43.b_0": "visual.vision_model.encoder.layers.21.layer_norm1.bias", "linear_128.w_0": "visual.vision_model.encoder.layers.21.self_attn.k_proj.weight", "linear_128.b_0": "visual.vision_model.encoder.layers.21.self_attn.k_proj.bias", "linear_129.w_0": "visual.vision_model.encoder.layers.21.self_attn.v_proj.weight", "linear_129.b_0": "visual.vision_model.encoder.layers.21.self_attn.v_proj.bias", "linear_130.w_0": "visual.vision_model.encoder.layers.21.self_attn.q_proj.weight", "linear_130.b_0": "visual.vision_model.encoder.layers.21.self_attn.q_proj.bias", "linear_131.w_0": "visual.vision_model.encoder.layers.21.self_attn.out_proj.weight", "linear_131.b_0": "visual.vision_model.encoder.layers.21.self_attn.out_proj.bias", "layer_norm_44.w_0": "visual.vision_model.encoder.layers.21.layer_norm2.weight", "layer_norm_44.b_0": "visual.vision_model.encoder.layers.21.layer_norm2.bias", "linear_132.w_0": "visual.vision_model.encoder.layers.21.mlp.fc1.weight", "linear_132.b_0": "visual.vision_model.encoder.layers.21.mlp.fc1.bias", "linear_133.w_0": "visual.vision_model.encoder.layers.21.mlp.fc2.weight", "linear_133.b_0": "visual.vision_model.encoder.layers.21.mlp.fc2.bias", "layer_norm_45.w_0": "visual.vision_model.encoder.layers.22.layer_norm1.weight", "layer_norm_45.b_0": "visual.vision_model.encoder.layers.22.layer_norm1.bias", "linear_134.w_0": "visual.vision_model.encoder.layers.22.self_attn.k_proj.weight", "linear_134.b_0": "visual.vision_model.encoder.layers.22.self_attn.k_proj.bias", "linear_135.w_0": "visual.vision_model.encoder.layers.22.self_attn.v_proj.weight", "linear_135.b_0": "visual.vision_model.encoder.layers.22.self_attn.v_proj.bias", "linear_136.w_0": "visual.vision_model.encoder.layers.22.self_attn.q_proj.weight", "linear_136.b_0": "visual.vision_model.encoder.layers.22.self_attn.q_proj.bias", "linear_137.w_0": "visual.vision_model.encoder.layers.22.self_attn.out_proj.weight", "linear_137.b_0": "visual.vision_model.encoder.layers.22.self_attn.out_proj.bias", "layer_norm_46.w_0": "visual.vision_model.encoder.layers.22.layer_norm2.weight", "layer_norm_46.b_0": "visual.vision_model.encoder.layers.22.layer_norm2.bias", "linear_138.w_0": "visual.vision_model.encoder.layers.22.mlp.fc1.weight", "linear_138.b_0": "visual.vision_model.encoder.layers.22.mlp.fc1.bias", "linear_139.w_0": "visual.vision_model.encoder.layers.22.mlp.fc2.weight", "linear_139.b_0": "visual.vision_model.encoder.layers.22.mlp.fc2.bias", "layer_norm_47.w_0": "visual.vision_model.encoder.layers.23.layer_norm1.weight", "layer_norm_47.b_0": "visual.vision_model.encoder.layers.23.layer_norm1.bias", "linear_140.w_0": "visual.vision_model.encoder.layers.23.self_attn.k_proj.weight", "linear_140.b_0": "visual.vision_model.encoder.layers.23.self_attn.k_proj.bias", "linear_141.w_0": "visual.vision_model.encoder.layers.23.self_attn.v_proj.weight", "linear_141.b_0": "visual.vision_model.encoder.layers.23.self_attn.v_proj.bias", "linear_142.w_0": "visual.vision_model.encoder.layers.23.self_attn.q_proj.weight", "linear_142.b_0": "visual.vision_model.encoder.layers.23.self_attn.q_proj.bias", "linear_143.w_0": "visual.vision_model.encoder.layers.23.self_attn.out_proj.weight", "linear_143.b_0": "visual.vision_model.encoder.layers.23.self_attn.out_proj.bias", "layer_norm_48.w_0": "visual.vision_model.encoder.layers.23.layer_norm2.weight", "layer_norm_48.b_0": "visual.vision_model.encoder.layers.23.layer_norm2.bias", "linear_144.w_0": "visual.vision_model.encoder.layers.23.mlp.fc1.weight", "linear_144.b_0": "visual.vision_model.encoder.layers.23.mlp.fc1.bias", "linear_145.w_0": "visual.vision_model.encoder.layers.23.mlp.fc2.weight", "linear_145.b_0": "visual.vision_model.encoder.layers.23.mlp.fc2.bias", "layer_norm_49.w_0": "visual.vision_model.encoder.layers.24.layer_norm1.weight", "layer_norm_49.b_0": "visual.vision_model.encoder.layers.24.layer_norm1.bias", "linear_146.w_0": "visual.vision_model.encoder.layers.24.self_attn.k_proj.weight", "linear_146.b_0": "visual.vision_model.encoder.layers.24.self_attn.k_proj.bias", "linear_147.w_0": "visual.vision_model.encoder.layers.24.self_attn.v_proj.weight", "linear_147.b_0": "visual.vision_model.encoder.layers.24.self_attn.v_proj.bias", "linear_148.w_0": "visual.vision_model.encoder.layers.24.self_attn.q_proj.weight", "linear_148.b_0": "visual.vision_model.encoder.layers.24.self_attn.q_proj.bias", "linear_149.w_0": "visual.vision_model.encoder.layers.24.self_attn.out_proj.weight", "linear_149.b_0": "visual.vision_model.encoder.layers.24.self_attn.out_proj.bias", "layer_norm_50.w_0": "visual.vision_model.encoder.layers.24.layer_norm2.weight", "layer_norm_50.b_0": "visual.vision_model.encoder.layers.24.layer_norm2.bias", "linear_150.w_0": "visual.vision_model.encoder.layers.24.mlp.fc1.weight", "linear_150.b_0": "visual.vision_model.encoder.layers.24.mlp.fc1.bias", "linear_151.w_0": "visual.vision_model.encoder.layers.24.mlp.fc2.weight", "linear_151.b_0": "visual.vision_model.encoder.layers.24.mlp.fc2.bias", "layer_norm_51.w_0": "visual.vision_model.encoder.layers.25.layer_norm1.weight", "layer_norm_51.b_0": "visual.vision_model.encoder.layers.25.layer_norm1.bias", "linear_152.w_0": "visual.vision_model.encoder.layers.25.self_attn.k_proj.weight", "linear_152.b_0": "visual.vision_model.encoder.layers.25.self_attn.k_proj.bias", "linear_153.w_0": "visual.vision_model.encoder.layers.25.self_attn.v_proj.weight", "linear_153.b_0": "visual.vision_model.encoder.layers.25.self_attn.v_proj.bias", "linear_154.w_0": "visual.vision_model.encoder.layers.25.self_attn.q_proj.weight", "linear_154.b_0": "visual.vision_model.encoder.layers.25.self_attn.q_proj.bias", "linear_155.w_0": "visual.vision_model.encoder.layers.25.self_attn.out_proj.weight", "linear_155.b_0": "visual.vision_model.encoder.layers.25.self_attn.out_proj.bias", "layer_norm_52.w_0": "visual.vision_model.encoder.layers.25.layer_norm2.weight", "layer_norm_52.b_0": "visual.vision_model.encoder.layers.25.layer_norm2.bias", "linear_156.w_0": "visual.vision_model.encoder.layers.25.mlp.fc1.weight", "linear_156.b_0": "visual.vision_model.encoder.layers.25.mlp.fc1.bias", "linear_157.w_0": "visual.vision_model.encoder.layers.25.mlp.fc2.weight", "linear_157.b_0": "visual.vision_model.encoder.layers.25.mlp.fc2.bias", "layer_norm_53.w_0": "visual.vision_model.encoder.layers.26.layer_norm1.weight", "layer_norm_53.b_0": "visual.vision_model.encoder.layers.26.layer_norm1.bias", "linear_158.w_0": "visual.vision_model.encoder.layers.26.self_attn.k_proj.weight", "linear_158.b_0": "visual.vision_model.encoder.layers.26.self_attn.k_proj.bias", "linear_159.w_0": "visual.vision_model.encoder.layers.26.self_attn.v_proj.weight", "linear_159.b_0": "visual.vision_model.encoder.layers.26.self_attn.v_proj.bias", "linear_160.w_0": "visual.vision_model.encoder.layers.26.self_attn.q_proj.weight", "linear_160.b_0": "visual.vision_model.encoder.layers.26.self_attn.q_proj.bias", "linear_161.w_0": "visual.vision_model.encoder.layers.26.self_attn.out_proj.weight", "linear_161.b_0": "visual.vision_model.encoder.layers.26.self_attn.out_proj.bias", "layer_norm_54.w_0": "visual.vision_model.encoder.layers.26.layer_norm2.weight", "layer_norm_54.b_0": "visual.vision_model.encoder.layers.26.layer_norm2.bias", "linear_162.w_0": "visual.vision_model.encoder.layers.26.mlp.fc1.weight", "linear_162.b_0": "visual.vision_model.encoder.layers.26.mlp.fc1.bias", "linear_163.w_0": "visual.vision_model.encoder.layers.26.mlp.fc2.weight", "linear_163.b_0": "visual.vision_model.encoder.layers.26.mlp.fc2.bias", "layer_norm_55.w_0": "visual.vision_model.post_layernorm.weight", "layer_norm_55.b_0": "visual.vision_model.post_layernorm.bias", "siglip_multihead_attention_pooling_head_0.w_0": "visual.vision_model.head.probe", "multi_head_attention_0.w_0": "visual.vision_model.head.attention.in_proj_weight", "multi_head_attention_0.w_1": "visual.vision_model.head.attention.in_proj_bias", "linear_164.w_0": "visual.vision_model.head.attention.out_proj.weight", "linear_164.b_0": "visual.vision_model.head.attention.out_proj.bias", "layer_norm_56.w_0": "visual.vision_model.head.layernorm.weight", "layer_norm_56.b_0": "visual.vision_model.head.layernorm.bias", "linear_165.w_0": "visual.vision_model.head.mlp.fc1.weight", "linear_165.b_0": "visual.vision_model.head.mlp.fc1.bias", "linear_166.w_0": "visual.vision_model.head.mlp.fc2.weight", "linear_166.b_0": "visual.vision_model.head.mlp.fc2.bias", "embedding_2.w_0": "model.embed_tokens.weight", "linear_167.w_0": "model.layers.0.self_attn.q_proj.weight", "linear_168.w_0": "model.layers.0.self_attn.k_proj.weight", "linear_169.w_0": "model.layers.0.self_attn.v_proj.weight", "linear_170.w_0": "model.layers.0.self_attn.o_proj.weight", "linear_171.w_0": "model.layers.0.mlp.gate_proj.weight", "linear_172.w_0": "model.layers.0.mlp.up_proj.weight", "linear_173.w_0": "model.layers.0.mlp.down_proj.weight", "create_parameter_0.w_0": "model.layers.0.input_layernorm.weight", "create_parameter_1.w_0": "model.layers.0.post_attention_layernorm.weight", "linear_174.w_0": "model.layers.1.self_attn.q_proj.weight", "linear_175.w_0": "model.layers.1.self_attn.k_proj.weight", "linear_176.w_0": "model.layers.1.self_attn.v_proj.weight", "linear_177.w_0": "model.layers.1.self_attn.o_proj.weight", "linear_178.w_0": "model.layers.1.mlp.gate_proj.weight", "linear_179.w_0": "model.layers.1.mlp.up_proj.weight", "linear_180.w_0": "model.layers.1.mlp.down_proj.weight", "create_parameter_2.w_0": "model.layers.1.input_layernorm.weight", "create_parameter_3.w_0": "model.layers.1.post_attention_layernorm.weight", "linear_181.w_0": "model.layers.2.self_attn.q_proj.weight", "linear_182.w_0": "model.layers.2.self_attn.k_proj.weight", "linear_183.w_0": "model.layers.2.self_attn.v_proj.weight", "linear_184.w_0": "model.layers.2.self_attn.o_proj.weight", "linear_185.w_0": "model.layers.2.mlp.gate_proj.weight", "linear_186.w_0": "model.layers.2.mlp.up_proj.weight", "linear_187.w_0": "model.layers.2.mlp.down_proj.weight", "create_parameter_4.w_0": "model.layers.2.input_layernorm.weight", "create_parameter_5.w_0": "model.layers.2.post_attention_layernorm.weight", "linear_188.w_0": "model.layers.3.self_attn.q_proj.weight", "linear_189.w_0": "model.layers.3.self_attn.k_proj.weight", "linear_190.w_0": "model.layers.3.self_attn.v_proj.weight", "linear_191.w_0": "model.layers.3.self_attn.o_proj.weight", "linear_192.w_0": "model.layers.3.mlp.gate_proj.weight", "linear_193.w_0": "model.layers.3.mlp.up_proj.weight", "linear_194.w_0": "model.layers.3.mlp.down_proj.weight", "create_parameter_6.w_0": "model.layers.3.input_layernorm.weight", "create_parameter_7.w_0": "model.layers.3.post_attention_layernorm.weight", "linear_195.w_0": "model.layers.4.self_attn.q_proj.weight", "linear_196.w_0": "model.layers.4.self_attn.k_proj.weight", "linear_197.w_0": "model.layers.4.self_attn.v_proj.weight", "linear_198.w_0": "model.layers.4.self_attn.o_proj.weight", "linear_199.w_0": "model.layers.4.mlp.gate_proj.weight", "linear_200.w_0": "model.layers.4.mlp.up_proj.weight", "linear_201.w_0": "model.layers.4.mlp.down_proj.weight", "create_parameter_8.w_0": "model.layers.4.input_layernorm.weight", "create_parameter_9.w_0": "model.layers.4.post_attention_layernorm.weight", "linear_202.w_0": "model.layers.5.self_attn.q_proj.weight", "linear_203.w_0": "model.layers.5.self_attn.k_proj.weight", "linear_204.w_0": "model.layers.5.self_attn.v_proj.weight", "linear_205.w_0": "model.layers.5.self_attn.o_proj.weight", "linear_206.w_0": "model.layers.5.mlp.gate_proj.weight", "linear_207.w_0": "model.layers.5.mlp.up_proj.weight", "linear_208.w_0": "model.layers.5.mlp.down_proj.weight", "create_parameter_10.w_0": "model.layers.5.input_layernorm.weight", "create_parameter_11.w_0": "model.layers.5.post_attention_layernorm.weight", "linear_209.w_0": "model.layers.6.self_attn.q_proj.weight", "linear_210.w_0": "model.layers.6.self_attn.k_proj.weight", "linear_211.w_0": "model.layers.6.self_attn.v_proj.weight", "linear_212.w_0": "model.layers.6.self_attn.o_proj.weight", "linear_213.w_0": "model.layers.6.mlp.gate_proj.weight", "linear_214.w_0": "model.layers.6.mlp.up_proj.weight", "linear_215.w_0": "model.layers.6.mlp.down_proj.weight", "create_parameter_12.w_0": "model.layers.6.input_layernorm.weight", "create_parameter_13.w_0": "model.layers.6.post_attention_layernorm.weight", "linear_216.w_0": "model.layers.7.self_attn.q_proj.weight", "linear_217.w_0": "model.layers.7.self_attn.k_proj.weight", "linear_218.w_0": "model.layers.7.self_attn.v_proj.weight", "linear_219.w_0": "model.layers.7.self_attn.o_proj.weight", "linear_220.w_0": "model.layers.7.mlp.gate_proj.weight", "linear_221.w_0": "model.layers.7.mlp.up_proj.weight", "linear_222.w_0": "model.layers.7.mlp.down_proj.weight", "create_parameter_14.w_0": "model.layers.7.input_layernorm.weight", "create_parameter_15.w_0": "model.layers.7.post_attention_layernorm.weight", "linear_223.w_0": "model.layers.8.self_attn.q_proj.weight", "linear_224.w_0": "model.layers.8.self_attn.k_proj.weight", "linear_225.w_0": "model.layers.8.self_attn.v_proj.weight", "linear_226.w_0": "model.layers.8.self_attn.o_proj.weight", "linear_227.w_0": "model.layers.8.mlp.gate_proj.weight", "linear_228.w_0": "model.layers.8.mlp.up_proj.weight", "linear_229.w_0": "model.layers.8.mlp.down_proj.weight", "create_parameter_16.w_0": "model.layers.8.input_layernorm.weight", "create_parameter_17.w_0": "model.layers.8.post_attention_layernorm.weight", "linear_230.w_0": "model.layers.9.self_attn.q_proj.weight", "linear_231.w_0": "model.layers.9.self_attn.k_proj.weight", "linear_232.w_0": "model.layers.9.self_attn.v_proj.weight", "linear_233.w_0": "model.layers.9.self_attn.o_proj.weight", "linear_234.w_0": "model.layers.9.mlp.gate_proj.weight", "linear_235.w_0": "model.layers.9.mlp.up_proj.weight", "linear_236.w_0": "model.layers.9.mlp.down_proj.weight", "create_parameter_18.w_0": "model.layers.9.input_layernorm.weight", "create_parameter_19.w_0": "model.layers.9.post_attention_layernorm.weight", "linear_237.w_0": "model.layers.10.self_attn.q_proj.weight", "linear_238.w_0": "model.layers.10.self_attn.k_proj.weight", "linear_239.w_0": "model.layers.10.self_attn.v_proj.weight", "linear_240.w_0": "model.layers.10.self_attn.o_proj.weight", "linear_241.w_0": "model.layers.10.mlp.gate_proj.weight", "linear_242.w_0": "model.layers.10.mlp.up_proj.weight", "linear_243.w_0": "model.layers.10.mlp.down_proj.weight", "create_parameter_20.w_0": "model.layers.10.input_layernorm.weight", "create_parameter_21.w_0": "model.layers.10.post_attention_layernorm.weight", "linear_244.w_0": "model.layers.11.self_attn.q_proj.weight", "linear_245.w_0": "model.layers.11.self_attn.k_proj.weight", "linear_246.w_0": "model.layers.11.self_attn.v_proj.weight", "linear_247.w_0": "model.layers.11.self_attn.o_proj.weight", "linear_248.w_0": "model.layers.11.mlp.gate_proj.weight", "linear_249.w_0": "model.layers.11.mlp.up_proj.weight", "linear_250.w_0": "model.layers.11.mlp.down_proj.weight", "create_parameter_22.w_0": "model.layers.11.input_layernorm.weight", "create_parameter_23.w_0": "model.layers.11.post_attention_layernorm.weight", "linear_251.w_0": "model.layers.12.self_attn.q_proj.weight", "linear_252.w_0": "model.layers.12.self_attn.k_proj.weight", "linear_253.w_0": "model.layers.12.self_attn.v_proj.weight", "linear_254.w_0": "model.layers.12.self_attn.o_proj.weight", "linear_255.w_0": "model.layers.12.mlp.gate_proj.weight", "linear_256.w_0": "model.layers.12.mlp.up_proj.weight", "linear_257.w_0": "model.layers.12.mlp.down_proj.weight", "create_parameter_24.w_0": "model.layers.12.input_layernorm.weight", "create_parameter_25.w_0": "model.layers.12.post_attention_layernorm.weight", "linear_258.w_0": "model.layers.13.self_attn.q_proj.weight", "linear_259.w_0": "model.layers.13.self_attn.k_proj.weight", "linear_260.w_0": "model.layers.13.self_attn.v_proj.weight", "linear_261.w_0": "model.layers.13.self_attn.o_proj.weight", "linear_262.w_0": "model.layers.13.mlp.gate_proj.weight", "linear_263.w_0": "model.layers.13.mlp.up_proj.weight", "linear_264.w_0": "model.layers.13.mlp.down_proj.weight", "create_parameter_26.w_0": "model.layers.13.input_layernorm.weight", "create_parameter_27.w_0": "model.layers.13.post_attention_layernorm.weight", "linear_265.w_0": "model.layers.14.self_attn.q_proj.weight", "linear_266.w_0": "model.layers.14.self_attn.k_proj.weight", "linear_267.w_0": "model.layers.14.self_attn.v_proj.weight", "linear_268.w_0": "model.layers.14.self_attn.o_proj.weight", "linear_269.w_0": "model.layers.14.mlp.gate_proj.weight", "linear_270.w_0": "model.layers.14.mlp.up_proj.weight", "linear_271.w_0": "model.layers.14.mlp.down_proj.weight", "create_parameter_28.w_0": "model.layers.14.input_layernorm.weight", "create_parameter_29.w_0": "model.layers.14.post_attention_layernorm.weight", "linear_272.w_0": "model.layers.15.self_attn.q_proj.weight", "linear_273.w_0": "model.layers.15.self_attn.k_proj.weight", "linear_274.w_0": "model.layers.15.self_attn.v_proj.weight", "linear_275.w_0": "model.layers.15.self_attn.o_proj.weight", "linear_276.w_0": "model.layers.15.mlp.gate_proj.weight", "linear_277.w_0": "model.layers.15.mlp.up_proj.weight", "linear_278.w_0": "model.layers.15.mlp.down_proj.weight", "create_parameter_30.w_0": "model.layers.15.input_layernorm.weight", "create_parameter_31.w_0": "model.layers.15.post_attention_layernorm.weight", "linear_279.w_0": "model.layers.16.self_attn.q_proj.weight", "linear_280.w_0": "model.layers.16.self_attn.k_proj.weight", "linear_281.w_0": "model.layers.16.self_attn.v_proj.weight", "linear_282.w_0": "model.layers.16.self_attn.o_proj.weight", "linear_283.w_0": "model.layers.16.mlp.gate_proj.weight", "linear_284.w_0": "model.layers.16.mlp.up_proj.weight", "linear_285.w_0": "model.layers.16.mlp.down_proj.weight", "create_parameter_32.w_0": "model.layers.16.input_layernorm.weight", "create_parameter_33.w_0": "model.layers.16.post_attention_layernorm.weight", "linear_286.w_0": "model.layers.17.self_attn.q_proj.weight", "linear_287.w_0": "model.layers.17.self_attn.k_proj.weight", "linear_288.w_0": "model.layers.17.self_attn.v_proj.weight", "linear_289.w_0": "model.layers.17.self_attn.o_proj.weight", "linear_290.w_0": "model.layers.17.mlp.gate_proj.weight", "linear_291.w_0": "model.layers.17.mlp.up_proj.weight", "linear_292.w_0": "model.layers.17.mlp.down_proj.weight", "create_parameter_34.w_0": "model.layers.17.input_layernorm.weight", "create_parameter_35.w_0": "model.layers.17.post_attention_layernorm.weight", "create_parameter_36.w_0": "model.norm.weight", "linear_293.w_0": "lm_head.weight"}
checkpoint-378/tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34ef7db83df785924fb83d7b887b6e822a031c56e15cff40aaf9b982988180df
3
+ size 1614363
checkpoint-378/tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-378/trainer_state.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "consumed_samples": 0,
5
+ "epoch": 0.8978622327790974,
6
+ "global_step": 378,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [],
10
+ "max_steps": 421,
11
+ "num_train_epochs": 9223372036854775807,
12
+ "total_flos": 0,
13
+ "trial_name": null,
14
+ "trial_params": null
15
+ }
checkpoint-378/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0888f0d3f65facdf04c57f4bd4b2036e3fec9976c17b7eb281a443129536c8e7
3
+ size 6425
checkpoint-420/added_tokens.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"<unk>": 0, "<s>": 1, "</s>": 2, "0": 3, "1": 4, "2": 5, "3": 6, "4": 7, "5": 8, "6": 9, "7": 10, "8": 11, "9": 12, "<|end_of_sentence|>": 100272, "<|begin_of_sentence|>": 100273, "<mask:1>": 100274, "<mask:4>": 100277, "<mask:5>": 100278, "<mask:6>": 100279, "<mask:7>": 100280, "<|IMAGE_PLACEHOLDER|>": 100295, "<|AUDIO_PLACEHOLDER|>": 100296, "<|LOC_0|>": 100297, "<|LOC_1|>": 100298, "<|LOC_2|>": 100299, "<|LOC_3|>": 100300, "<|LOC_4|>": 100301, "<|LOC_5|>": 100302, "<|LOC_6|>": 100303, "<|LOC_7|>": 100304, "<|LOC_8|>": 100305, "<|LOC_9|>": 100306, "<|LOC_10|>": 100307, "<|LOC_11|>": 100308, "<|LOC_12|>": 100309, "<|LOC_13|>": 100310, "<|LOC_14|>": 100311, "<|LOC_15|>": 100312, "<|LOC_16|>": 100313, "<|LOC_17|>": 100314, "<|LOC_18|>": 100315, "<|LOC_19|>": 100316, "<|LOC_20|>": 100317, "<|LOC_21|>": 100318, "<|LOC_22|>": 100319, "<|LOC_23|>": 100320, "<|LOC_24|>": 100321, "<|LOC_25|>": 100322, "<|LOC_26|>": 100323, "<|LOC_27|>": 100324, "<|LOC_28|>": 100325, "<|LOC_29|>": 100326, "<|LOC_30|>": 100327, "<|LOC_31|>": 100328, "<|LOC_32|>": 100329, "<|LOC_33|>": 100330, "<|LOC_34|>": 100331, "<|LOC_35|>": 100332, "<|LOC_36|>": 100333, "<|LOC_37|>": 100334, "<|LOC_38|>": 100335, "<|LOC_39|>": 100336, "<|LOC_40|>": 100337, "<|LOC_41|>": 100338, "<|LOC_42|>": 100339, "<|LOC_43|>": 100340, "<|LOC_44|>": 100341, "<|LOC_45|>": 100342, "<|LOC_46|>": 100343, "<|LOC_47|>": 100344, "<|LOC_48|>": 100345, "<|LOC_49|>": 100346, "<|LOC_50|>": 100347, "<|LOC_51|>": 100348, "<|LOC_52|>": 100349, "<|LOC_53|>": 100350, "<|LOC_54|>": 100351, "<|LOC_55|>": 100352, "<|LOC_56|>": 100353, "<|LOC_57|>": 100354, "<|LOC_58|>": 100355, "<|LOC_59|>": 100356, "<|LOC_60|>": 100357, "<|LOC_61|>": 100358, "<|LOC_62|>": 100359, "<|LOC_63|>": 100360, "<|LOC_64|>": 100361, "<|LOC_65|>": 100362, "<|LOC_66|>": 100363, "<|LOC_67|>": 100364, "<|LOC_68|>": 100365, "<|LOC_69|>": 100366, "<|LOC_70|>": 100367, "<|LOC_71|>": 100368, "<|LOC_72|>": 100369, "<|LOC_73|>": 100370, "<|LOC_74|>": 100371, "<|LOC_75|>": 100372, "<|LOC_76|>": 100373, "<|LOC_77|>": 100374, "<|LOC_78|>": 100375, "<|LOC_79|>": 100376, "<|LOC_80|>": 100377, "<|LOC_81|>": 100378, "<|LOC_82|>": 100379, "<|LOC_83|>": 100380, "<|LOC_84|>": 100381, "<|LOC_85|>": 100382, "<|LOC_86|>": 100383, "<|LOC_87|>": 100384, "<|LOC_88|>": 100385, "<|LOC_89|>": 100386, "<|LOC_90|>": 100387, "<|LOC_91|>": 100388, "<|LOC_92|>": 100389, "<|LOC_93|>": 100390, "<|LOC_94|>": 100391, "<|LOC_95|>": 100392, "<|LOC_96|>": 100393, "<|LOC_97|>": 100394, "<|LOC_98|>": 100395, "<|LOC_99|>": 100396, "<|LOC_100|>": 100397, "<|LOC_101|>": 100398, "<|LOC_102|>": 100399, "<|LOC_103|>": 100400, "<|LOC_104|>": 100401, "<|LOC_105|>": 100402, "<|LOC_106|>": 100403, "<|LOC_107|>": 100404, "<|LOC_108|>": 100405, "<|LOC_109|>": 100406, "<|LOC_110|>": 100407, "<|LOC_111|>": 100408, "<|LOC_112|>": 100409, "<|LOC_113|>": 100410, "<|LOC_114|>": 100411, "<|LOC_115|>": 100412, "<|LOC_116|>": 100413, "<|LOC_117|>": 100414, "<|LOC_118|>": 100415, "<|LOC_119|>": 100416, "<|LOC_120|>": 100417, "<|LOC_121|>": 100418, "<|LOC_122|>": 100419, "<|LOC_123|>": 100420, "<|LOC_124|>": 100421, "<|LOC_125|>": 100422, "<|LOC_126|>": 100423, "<|LOC_127|>": 100424, "<|LOC_128|>": 100425, "<|LOC_129|>": 100426, "<|LOC_130|>": 100427, "<|LOC_131|>": 100428, "<|LOC_132|>": 100429, "<|LOC_133|>": 100430, "<|LOC_134|>": 100431, "<|LOC_135|>": 100432, "<|LOC_136|>": 100433, "<|LOC_137|>": 100434, "<|LOC_138|>": 100435, "<|LOC_139|>": 100436, "<|LOC_140|>": 100437, "<|LOC_141|>": 100438, "<|LOC_142|>": 100439, "<|LOC_143|>": 100440, "<|LOC_144|>": 100441, "<|LOC_145|>": 100442, "<|LOC_146|>": 100443, "<|LOC_147|>": 100444, "<|LOC_148|>": 100445, "<|LOC_149|>": 100446, "<|LOC_150|>": 100447, "<|LOC_151|>": 100448, "<|LOC_152|>": 100449, "<|LOC_153|>": 100450, "<|LOC_154|>": 100451, "<|LOC_155|>": 100452, "<|LOC_156|>": 100453, "<|LOC_157|>": 100454, "<|LOC_158|>": 100455, "<|LOC_159|>": 100456, "<|LOC_160|>": 100457, "<|LOC_161|>": 100458, "<|LOC_162|>": 100459, "<|LOC_163|>": 100460, "<|LOC_164|>": 100461, "<|LOC_165|>": 100462, "<|LOC_166|>": 100463, "<|LOC_167|>": 100464, "<|LOC_168|>": 100465, "<|LOC_169|>": 100466, "<|LOC_170|>": 100467, "<|LOC_171|>": 100468, "<|LOC_172|>": 100469, "<|LOC_173|>": 100470, "<|LOC_174|>": 100471, "<|LOC_175|>": 100472, "<|LOC_176|>": 100473, "<|LOC_177|>": 100474, "<|LOC_178|>": 100475, "<|LOC_179|>": 100476, "<|LOC_180|>": 100477, "<|LOC_181|>": 100478, "<|LOC_182|>": 100479, "<|LOC_183|>": 100480, "<|LOC_184|>": 100481, "<|LOC_185|>": 100482, "<|LOC_186|>": 100483, "<|LOC_187|>": 100484, "<|LOC_188|>": 100485, "<|LOC_189|>": 100486, "<|LOC_190|>": 100487, "<|LOC_191|>": 100488, "<|LOC_192|>": 100489, "<|LOC_193|>": 100490, "<|LOC_194|>": 100491, "<|LOC_195|>": 100492, "<|LOC_196|>": 100493, "<|LOC_197|>": 100494, "<|LOC_198|>": 100495, "<|LOC_199|>": 100496, "<|LOC_200|>": 100497, "<|LOC_201|>": 100498, "<|LOC_202|>": 100499, "<|LOC_203|>": 100500, "<|LOC_204|>": 100501, "<|LOC_205|>": 100502, "<|LOC_206|>": 100503, "<|LOC_207|>": 100504, "<|LOC_208|>": 100505, "<|LOC_209|>": 100506, "<|LOC_210|>": 100507, "<|LOC_211|>": 100508, "<|LOC_212|>": 100509, "<|LOC_213|>": 100510, "<|LOC_214|>": 100511, "<|LOC_215|>": 100512, "<|LOC_216|>": 100513, "<|LOC_217|>": 100514, "<|LOC_218|>": 100515, "<|LOC_219|>": 100516, "<|LOC_220|>": 100517, "<|LOC_221|>": 100518, "<|LOC_222|>": 100519, "<|LOC_223|>": 100520, "<|LOC_224|>": 100521, "<|LOC_225|>": 100522, "<|LOC_226|>": 100523, "<|LOC_227|>": 100524, "<|LOC_228|>": 100525, "<|LOC_229|>": 100526, "<|LOC_230|>": 100527, "<|LOC_231|>": 100528, "<|LOC_232|>": 100529, "<|LOC_233|>": 100530, "<|LOC_234|>": 100531, "<|LOC_235|>": 100532, "<|LOC_236|>": 100533, "<|LOC_237|>": 100534, "<|LOC_238|>": 100535, "<|LOC_239|>": 100536, "<|LOC_240|>": 100537, "<|LOC_241|>": 100538, "<|LOC_242|>": 100539, "<|LOC_243|>": 100540, "<|LOC_244|>": 100541, "<|LOC_245|>": 100542, "<|LOC_246|>": 100543, "<|LOC_247|>": 100544, "<|LOC_248|>": 100545, "<|LOC_249|>": 100546, "<|LOC_250|>": 100547, "<|LOC_251|>": 100548, "<|LOC_252|>": 100549, "<|LOC_253|>": 100550, "<|LOC_254|>": 100551, "<|LOC_255|>": 100552, "<|LOC_256|>": 100553, "<|LOC_257|>": 100554, "<|LOC_258|>": 100555, "<|LOC_259|>": 100556, "<|LOC_260|>": 100557, "<|LOC_261|>": 100558, "<|LOC_262|>": 100559, "<|LOC_263|>": 100560, "<|LOC_264|>": 100561, "<|LOC_265|>": 100562, "<|LOC_266|>": 100563, "<|LOC_267|>": 100564, "<|LOC_268|>": 100565, "<|LOC_269|>": 100566, "<|LOC_270|>": 100567, "<|LOC_271|>": 100568, "<|LOC_272|>": 100569, "<|LOC_273|>": 100570, "<|LOC_274|>": 100571, "<|LOC_275|>": 100572, "<|LOC_276|>": 100573, "<|LOC_277|>": 100574, "<|LOC_278|>": 100575, "<|LOC_279|>": 100576, "<|LOC_280|>": 100577, "<|LOC_281|>": 100578, "<|LOC_282|>": 100579, "<|LOC_283|>": 100580, "<|LOC_284|>": 100581, "<|LOC_285|>": 100582, "<|LOC_286|>": 100583, "<|LOC_287|>": 100584, "<|LOC_288|>": 100585, "<|LOC_289|>": 100586, "<|LOC_290|>": 100587, "<|LOC_291|>": 100588, "<|LOC_292|>": 100589, "<|LOC_293|>": 100590, "<|LOC_294|>": 100591, "<|LOC_295|>": 100592, "<|LOC_296|>": 100593, "<|LOC_297|>": 100594, "<|LOC_298|>": 100595, "<|LOC_299|>": 100596, "<|LOC_300|>": 100597, "<|LOC_301|>": 100598, "<|LOC_302|>": 100599, "<|LOC_303|>": 100600, "<|LOC_304|>": 100601, "<|LOC_305|>": 100602, "<|LOC_306|>": 100603, "<|LOC_307|>": 100604, "<|LOC_308|>": 100605, "<|LOC_309|>": 100606, "<|LOC_310|>": 100607, "<|LOC_311|>": 100608, "<|LOC_312|>": 100609, "<|LOC_313|>": 100610, "<|LOC_314|>": 100611, "<|LOC_315|>": 100612, "<|LOC_316|>": 100613, "<|LOC_317|>": 100614, "<|LOC_318|>": 100615, "<|LOC_319|>": 100616, "<|LOC_320|>": 100617, "<|LOC_321|>": 100618, "<|LOC_322|>": 100619, "<|LOC_323|>": 100620, "<|LOC_324|>": 100621, "<|LOC_325|>": 100622, "<|LOC_326|>": 100623, "<|LOC_327|>": 100624, "<|LOC_328|>": 100625, "<|LOC_329|>": 100626, "<|LOC_330|>": 100627, "<|LOC_331|>": 100628, "<|LOC_332|>": 100629, "<|LOC_333|>": 100630, "<|LOC_334|>": 100631, "<|LOC_335|>": 100632, "<|LOC_336|>": 100633, "<|LOC_337|>": 100634, "<|LOC_338|>": 100635, "<|LOC_339|>": 100636, "<|LOC_340|>": 100637, "<|LOC_341|>": 100638, "<|LOC_342|>": 100639, "<|LOC_343|>": 100640, "<|LOC_344|>": 100641, "<|LOC_345|>": 100642, "<|LOC_346|>": 100643, "<|LOC_347|>": 100644, "<|LOC_348|>": 100645, "<|LOC_349|>": 100646, "<|LOC_350|>": 100647, "<|LOC_351|>": 100648, "<|LOC_352|>": 100649, "<|LOC_353|>": 100650, "<|LOC_354|>": 100651, "<|LOC_355|>": 100652, "<|LOC_356|>": 100653, "<|LOC_357|>": 100654, "<|LOC_358|>": 100655, "<|LOC_359|>": 100656, "<|LOC_360|>": 100657, "<|LOC_361|>": 100658, "<|LOC_362|>": 100659, "<|LOC_363|>": 100660, "<|LOC_364|>": 100661, "<|LOC_365|>": 100662, "<|LOC_366|>": 100663, "<|LOC_367|>": 100664, "<|LOC_368|>": 100665, "<|LOC_369|>": 100666, "<|LOC_370|>": 100667, "<|LOC_371|>": 100668, "<|LOC_372|>": 100669, "<|LOC_373|>": 100670, "<|LOC_374|>": 100671, "<|LOC_375|>": 100672, "<|LOC_376|>": 100673, "<|LOC_377|>": 100674, "<|LOC_378|>": 100675, "<|LOC_379|>": 100676, "<|LOC_380|>": 100677, "<|LOC_381|>": 100678, "<|LOC_382|>": 100679, "<|LOC_383|>": 100680, "<|LOC_384|>": 100681, "<|LOC_385|>": 100682, "<|LOC_386|>": 100683, "<|LOC_387|>": 100684, "<|LOC_388|>": 100685, "<|LOC_389|>": 100686, "<|LOC_390|>": 100687, "<|LOC_391|>": 100688, "<|LOC_392|>": 100689, "<|LOC_393|>": 100690, "<|LOC_394|>": 100691, "<|LOC_395|>": 100692, "<|LOC_396|>": 100693, "<|LOC_397|>": 100694, "<|LOC_398|>": 100695, "<|LOC_399|>": 100696, "<|LOC_400|>": 100697, "<|LOC_401|>": 100698, "<|LOC_402|>": 100699, "<|LOC_403|>": 100700, "<|LOC_404|>": 100701, "<|LOC_405|>": 100702, "<|LOC_406|>": 100703, "<|LOC_407|>": 100704, "<|LOC_408|>": 100705, "<|LOC_409|>": 100706, "<|LOC_410|>": 100707, "<|LOC_411|>": 100708, "<|LOC_412|>": 100709, "<|LOC_413|>": 100710, "<|LOC_414|>": 100711, "<|LOC_415|>": 100712, "<|LOC_416|>": 100713, "<|LOC_417|>": 100714, "<|LOC_418|>": 100715, "<|LOC_419|>": 100716, "<|LOC_420|>": 100717, "<|LOC_421|>": 100718, "<|LOC_422|>": 100719, "<|LOC_423|>": 100720, "<|LOC_424|>": 100721, "<|LOC_425|>": 100722, "<|LOC_426|>": 100723, "<|LOC_427|>": 100724, "<|LOC_428|>": 100725, "<|LOC_429|>": 100726, "<|LOC_430|>": 100727, "<|LOC_431|>": 100728, "<|LOC_432|>": 100729, "<|LOC_433|>": 100730, "<|LOC_434|>": 100731, "<|LOC_435|>": 100732, "<|LOC_436|>": 100733, "<|LOC_437|>": 100734, "<|LOC_438|>": 100735, "<|LOC_439|>": 100736, "<|LOC_440|>": 100737, "<|LOC_441|>": 100738, "<|LOC_442|>": 100739, "<|LOC_443|>": 100740, "<|LOC_444|>": 100741, "<|LOC_445|>": 100742, "<|LOC_446|>": 100743, "<|LOC_447|>": 100744, "<|LOC_448|>": 100745, "<|LOC_449|>": 100746, "<|LOC_450|>": 100747, "<|LOC_451|>": 100748, "<|LOC_452|>": 100749, "<|LOC_453|>": 100750, "<|LOC_454|>": 100751, "<|LOC_455|>": 100752, "<|LOC_456|>": 100753, "<|LOC_457|>": 100754, "<|LOC_458|>": 100755, "<|LOC_459|>": 100756, "<|LOC_460|>": 100757, "<|LOC_461|>": 100758, "<|LOC_462|>": 100759, "<|LOC_463|>": 100760, "<|LOC_464|>": 100761, "<|LOC_465|>": 100762, "<|LOC_466|>": 100763, "<|LOC_467|>": 100764, "<|LOC_468|>": 100765, "<|LOC_469|>": 100766, "<|LOC_470|>": 100767, "<|LOC_471|>": 100768, "<|LOC_472|>": 100769, "<|LOC_473|>": 100770, "<|LOC_474|>": 100771, "<|LOC_475|>": 100772, "<|LOC_476|>": 100773, "<|LOC_477|>": 100774, "<|LOC_478|>": 100775, "<|LOC_479|>": 100776, "<|LOC_480|>": 100777, "<|LOC_481|>": 100778, "<|LOC_482|>": 100779, "<|LOC_483|>": 100780, "<|LOC_484|>": 100781, "<|LOC_485|>": 100782, "<|LOC_486|>": 100783, "<|LOC_487|>": 100784, "<|LOC_488|>": 100785, "<|LOC_489|>": 100786, "<|LOC_490|>": 100787, "<|LOC_491|>": 100788, "<|LOC_492|>": 100789, "<|LOC_493|>": 100790, "<|LOC_494|>": 100791, "<|LOC_495|>": 100792, "<|LOC_496|>": 100793, "<|LOC_497|>": 100794, "<|LOC_498|>": 100795, "<|LOC_499|>": 100796, "<|LOC_500|>": 100797, "<|LOC_501|>": 100798, "<|LOC_502|>": 100799, "<|LOC_503|>": 100800, "<|LOC_504|>": 100801, "<|LOC_505|>": 100802, "<|LOC_506|>": 100803, "<|LOC_507|>": 100804, "<|LOC_508|>": 100805, "<|LOC_509|>": 100806, "<|LOC_510|>": 100807, "<|LOC_511|>": 100808, "<|LOC_512|>": 100809, "<|LOC_513|>": 100810, "<|LOC_514|>": 100811, "<|LOC_515|>": 100812, "<|LOC_516|>": 100813, "<|LOC_517|>": 100814, "<|LOC_518|>": 100815, "<|LOC_519|>": 100816, "<|LOC_520|>": 100817, "<|LOC_521|>": 100818, "<|LOC_522|>": 100819, "<|LOC_523|>": 100820, "<|LOC_524|>": 100821, "<|LOC_525|>": 100822, "<|LOC_526|>": 100823, "<|LOC_527|>": 100824, "<|LOC_528|>": 100825, "<|LOC_529|>": 100826, "<|LOC_530|>": 100827, "<|LOC_531|>": 100828, "<|LOC_532|>": 100829, "<|LOC_533|>": 100830, "<|LOC_534|>": 100831, "<|LOC_535|>": 100832, "<|LOC_536|>": 100833, "<|LOC_537|>": 100834, "<|LOC_538|>": 100835, "<|LOC_539|>": 100836, "<|LOC_540|>": 100837, "<|LOC_541|>": 100838, "<|LOC_542|>": 100839, "<|LOC_543|>": 100840, "<|LOC_544|>": 100841, "<|LOC_545|>": 100842, "<|LOC_546|>": 100843, "<|LOC_547|>": 100844, "<|LOC_548|>": 100845, "<|LOC_549|>": 100846, "<|LOC_550|>": 100847, "<|LOC_551|>": 100848, "<|LOC_552|>": 100849, "<|LOC_553|>": 100850, "<|LOC_554|>": 100851, "<|LOC_555|>": 100852, "<|LOC_556|>": 100853, "<|LOC_557|>": 100854, "<|LOC_558|>": 100855, "<|LOC_559|>": 100856, "<|LOC_560|>": 100857, "<|LOC_561|>": 100858, "<|LOC_562|>": 100859, "<|LOC_563|>": 100860, "<|LOC_564|>": 100861, "<|LOC_565|>": 100862, "<|LOC_566|>": 100863, "<|LOC_567|>": 100864, "<|LOC_568|>": 100865, "<|LOC_569|>": 100866, "<|LOC_570|>": 100867, "<|LOC_571|>": 100868, "<|LOC_572|>": 100869, "<|LOC_573|>": 100870, "<|LOC_574|>": 100871, "<|LOC_575|>": 100872, "<|LOC_576|>": 100873, "<|LOC_577|>": 100874, "<|LOC_578|>": 100875, "<|LOC_579|>": 100876, "<|LOC_580|>": 100877, "<|LOC_581|>": 100878, "<|LOC_582|>": 100879, "<|LOC_583|>": 100880, "<|LOC_584|>": 100881, "<|LOC_585|>": 100882, "<|LOC_586|>": 100883, "<|LOC_587|>": 100884, "<|LOC_588|>": 100885, "<|LOC_589|>": 100886, "<|LOC_590|>": 100887, "<|LOC_591|>": 100888, "<|LOC_592|>": 100889, "<|LOC_593|>": 100890, "<|LOC_594|>": 100891, "<|LOC_595|>": 100892, "<|LOC_596|>": 100893, "<|LOC_597|>": 100894, "<|LOC_598|>": 100895, "<|LOC_599|>": 100896, "<|LOC_600|>": 100897, "<|LOC_601|>": 100898, "<|LOC_602|>": 100899, "<|LOC_603|>": 100900, "<|LOC_604|>": 100901, "<|LOC_605|>": 100902, "<|LOC_606|>": 100903, "<|LOC_607|>": 100904, "<|LOC_608|>": 100905, "<|LOC_609|>": 100906, "<|LOC_610|>": 100907, "<|LOC_611|>": 100908, "<|LOC_612|>": 100909, "<|LOC_613|>": 100910, "<|LOC_614|>": 100911, "<|LOC_615|>": 100912, "<|LOC_616|>": 100913, "<|LOC_617|>": 100914, "<|LOC_618|>": 100915, "<|LOC_619|>": 100916, "<|LOC_620|>": 100917, "<|LOC_621|>": 100918, "<|LOC_622|>": 100919, "<|LOC_623|>": 100920, "<|LOC_624|>": 100921, "<|LOC_625|>": 100922, "<|LOC_626|>": 100923, "<|LOC_627|>": 100924, "<|LOC_628|>": 100925, "<|LOC_629|>": 100926, "<|LOC_630|>": 100927, "<|LOC_631|>": 100928, "<|LOC_632|>": 100929, "<|LOC_633|>": 100930, "<|LOC_634|>": 100931, "<|LOC_635|>": 100932, "<|LOC_636|>": 100933, "<|LOC_637|>": 100934, "<|LOC_638|>": 100935, "<|LOC_639|>": 100936, "<|LOC_640|>": 100937, "<|LOC_641|>": 100938, "<|LOC_642|>": 100939, "<|LOC_643|>": 100940, "<|LOC_644|>": 100941, "<|LOC_645|>": 100942, "<|LOC_646|>": 100943, "<|LOC_647|>": 100944, "<|LOC_648|>": 100945, "<|LOC_649|>": 100946, "<|LOC_650|>": 100947, "<|LOC_651|>": 100948, "<|LOC_652|>": 100949, "<|LOC_653|>": 100950, "<|LOC_654|>": 100951, "<|LOC_655|>": 100952, "<|LOC_656|>": 100953, "<|LOC_657|>": 100954, "<|LOC_658|>": 100955, "<|LOC_659|>": 100956, "<|LOC_660|>": 100957, "<|LOC_661|>": 100958, "<|LOC_662|>": 100959, "<|LOC_663|>": 100960, "<|LOC_664|>": 100961, "<|LOC_665|>": 100962, "<|LOC_666|>": 100963, "<|LOC_667|>": 100964, "<|LOC_668|>": 100965, "<|LOC_669|>": 100966, "<|LOC_670|>": 100967, "<|LOC_671|>": 100968, "<|LOC_672|>": 100969, "<|LOC_673|>": 100970, "<|LOC_674|>": 100971, "<|LOC_675|>": 100972, "<|LOC_676|>": 100973, "<|LOC_677|>": 100974, "<|LOC_678|>": 100975, "<|LOC_679|>": 100976, "<|LOC_680|>": 100977, "<|LOC_681|>": 100978, "<|LOC_682|>": 100979, "<|LOC_683|>": 100980, "<|LOC_684|>": 100981, "<|LOC_685|>": 100982, "<|LOC_686|>": 100983, "<|LOC_687|>": 100984, "<|LOC_688|>": 100985, "<|LOC_689|>": 100986, "<|LOC_690|>": 100987, "<|LOC_691|>": 100988, "<|LOC_692|>": 100989, "<|LOC_693|>": 100990, "<|LOC_694|>": 100991, "<|LOC_695|>": 100992, "<|LOC_696|>": 100993, "<|LOC_697|>": 100994, "<|LOC_698|>": 100995, "<|LOC_699|>": 100996, "<|LOC_700|>": 100997, "<|LOC_701|>": 100998, "<|LOC_702|>": 100999, "<|LOC_703|>": 101000, "<|LOC_704|>": 101001, "<|LOC_705|>": 101002, "<|LOC_706|>": 101003, "<|LOC_707|>": 101004, "<|LOC_708|>": 101005, "<|LOC_709|>": 101006, "<|LOC_710|>": 101007, "<|LOC_711|>": 101008, "<|LOC_712|>": 101009, "<|LOC_713|>": 101010, "<|LOC_714|>": 101011, "<|LOC_715|>": 101012, "<|LOC_716|>": 101013, "<|LOC_717|>": 101014, "<|LOC_718|>": 101015, "<|LOC_719|>": 101016, "<|LOC_720|>": 101017, "<|LOC_721|>": 101018, "<|LOC_722|>": 101019, "<|LOC_723|>": 101020, "<|LOC_724|>": 101021, "<|LOC_725|>": 101022, "<|LOC_726|>": 101023, "<|LOC_727|>": 101024, "<|LOC_728|>": 101025, "<|LOC_729|>": 101026, "<|LOC_730|>": 101027, "<|LOC_731|>": 101028, "<|LOC_732|>": 101029, "<|LOC_733|>": 101030, "<|LOC_734|>": 101031, "<|LOC_735|>": 101032, "<|LOC_736|>": 101033, "<|LOC_737|>": 101034, "<|LOC_738|>": 101035, "<|LOC_739|>": 101036, "<|LOC_740|>": 101037, "<|LOC_741|>": 101038, "<|LOC_742|>": 101039, "<|LOC_743|>": 101040, "<|LOC_744|>": 101041, "<|LOC_745|>": 101042, "<|LOC_746|>": 101043, "<|LOC_747|>": 101044, "<|LOC_748|>": 101045, "<|LOC_749|>": 101046, "<|LOC_750|>": 101047, "<|LOC_751|>": 101048, "<|LOC_752|>": 101049, "<|LOC_753|>": 101050, "<|LOC_754|>": 101051, "<|LOC_755|>": 101052, "<|LOC_756|>": 101053, "<|LOC_757|>": 101054, "<|LOC_758|>": 101055, "<|LOC_759|>": 101056, "<|LOC_760|>": 101057, "<|LOC_761|>": 101058, "<|LOC_762|>": 101059, "<|LOC_763|>": 101060, "<|LOC_764|>": 101061, "<|LOC_765|>": 101062, "<|LOC_766|>": 101063, "<|LOC_767|>": 101064, "<|LOC_768|>": 101065, "<|LOC_769|>": 101066, "<|LOC_770|>": 101067, "<|LOC_771|>": 101068, "<|LOC_772|>": 101069, "<|LOC_773|>": 101070, "<|LOC_774|>": 101071, "<|LOC_775|>": 101072, "<|LOC_776|>": 101073, "<|LOC_777|>": 101074, "<|LOC_778|>": 101075, "<|LOC_779|>": 101076, "<|LOC_780|>": 101077, "<|LOC_781|>": 101078, "<|LOC_782|>": 101079, "<|LOC_783|>": 101080, "<|LOC_784|>": 101081, "<|LOC_785|>": 101082, "<|LOC_786|>": 101083, "<|LOC_787|>": 101084, "<|LOC_788|>": 101085, "<|LOC_789|>": 101086, "<|LOC_790|>": 101087, "<|LOC_791|>": 101088, "<|LOC_792|>": 101089, "<|LOC_793|>": 101090, "<|LOC_794|>": 101091, "<|LOC_795|>": 101092, "<|LOC_796|>": 101093, "<|LOC_797|>": 101094, "<|LOC_798|>": 101095, "<|LOC_799|>": 101096, "<|LOC_800|>": 101097, "<|LOC_801|>": 101098, "<|LOC_802|>": 101099, "<|LOC_803|>": 101100, "<|LOC_804|>": 101101, "<|LOC_805|>": 101102, "<|LOC_806|>": 101103, "<|LOC_807|>": 101104, "<|LOC_808|>": 101105, "<|LOC_809|>": 101106, "<|LOC_810|>": 101107, "<|LOC_811|>": 101108, "<|LOC_812|>": 101109, "<|LOC_813|>": 101110, "<|LOC_814|>": 101111, "<|LOC_815|>": 101112, "<|LOC_816|>": 101113, "<|LOC_817|>": 101114, "<|LOC_818|>": 101115, "<|LOC_819|>": 101116, "<|LOC_820|>": 101117, "<|LOC_821|>": 101118, "<|LOC_822|>": 101119, "<|LOC_823|>": 101120, "<|LOC_824|>": 101121, "<|LOC_825|>": 101122, "<|LOC_826|>": 101123, "<|LOC_827|>": 101124, "<|LOC_828|>": 101125, "<|LOC_829|>": 101126, "<|LOC_830|>": 101127, "<|LOC_831|>": 101128, "<|LOC_832|>": 101129, "<|LOC_833|>": 101130, "<|LOC_834|>": 101131, "<|LOC_835|>": 101132, "<|LOC_836|>": 101133, "<|LOC_837|>": 101134, "<|LOC_838|>": 101135, "<|LOC_839|>": 101136, "<|LOC_840|>": 101137, "<|LOC_841|>": 101138, "<|LOC_842|>": 101139, "<|LOC_843|>": 101140, "<|LOC_844|>": 101141, "<|LOC_845|>": 101142, "<|LOC_846|>": 101143, "<|LOC_847|>": 101144, "<|LOC_848|>": 101145, "<|LOC_849|>": 101146, "<|LOC_850|>": 101147, "<|LOC_851|>": 101148, "<|LOC_852|>": 101149, "<|LOC_853|>": 101150, "<|LOC_854|>": 101151, "<|LOC_855|>": 101152, "<|LOC_856|>": 101153, "<|LOC_857|>": 101154, "<|LOC_858|>": 101155, "<|LOC_859|>": 101156, "<|LOC_860|>": 101157, "<|LOC_861|>": 101158, "<|LOC_862|>": 101159, "<|LOC_863|>": 101160, "<|LOC_864|>": 101161, "<|LOC_865|>": 101162, "<|LOC_866|>": 101163, "<|LOC_867|>": 101164, "<|LOC_868|>": 101165, "<|LOC_869|>": 101166, "<|LOC_870|>": 101167, "<|LOC_871|>": 101168, "<|LOC_872|>": 101169, "<|LOC_873|>": 101170, "<|LOC_874|>": 101171, "<|LOC_875|>": 101172, "<|LOC_876|>": 101173, "<|LOC_877|>": 101174, "<|LOC_878|>": 101175, "<|LOC_879|>": 101176, "<|LOC_880|>": 101177, "<|LOC_881|>": 101178, "<|LOC_882|>": 101179, "<|LOC_883|>": 101180, "<|LOC_884|>": 101181, "<|LOC_885|>": 101182, "<|LOC_886|>": 101183, "<|LOC_887|>": 101184, "<|LOC_888|>": 101185, "<|LOC_889|>": 101186, "<|LOC_890|>": 101187, "<|LOC_891|>": 101188, "<|LOC_892|>": 101189, "<|LOC_893|>": 101190, "<|LOC_894|>": 101191, "<|LOC_895|>": 101192, "<|LOC_896|>": 101193, "<|LOC_897|>": 101194, "<|LOC_898|>": 101195, "<|LOC_899|>": 101196, "<|LOC_900|>": 101197, "<|LOC_901|>": 101198, "<|LOC_902|>": 101199, "<|LOC_903|>": 101200, "<|LOC_904|>": 101201, "<|LOC_905|>": 101202, "<|LOC_906|>": 101203, "<|LOC_907|>": 101204, "<|LOC_908|>": 101205, "<|LOC_909|>": 101206, "<|LOC_910|>": 101207, "<|LOC_911|>": 101208, "<|LOC_912|>": 101209, "<|LOC_913|>": 101210, "<|LOC_914|>": 101211, "<|LOC_915|>": 101212, "<|LOC_916|>": 101213, "<|LOC_917|>": 101214, "<|LOC_918|>": 101215, "<|LOC_919|>": 101216, "<|LOC_920|>": 101217, "<|LOC_921|>": 101218, "<|LOC_922|>": 101219, "<|LOC_923|>": 101220, "<|LOC_924|>": 101221, "<|LOC_925|>": 101222, "<|LOC_926|>": 101223, "<|LOC_927|>": 101224, "<|LOC_928|>": 101225, "<|LOC_929|>": 101226, "<|LOC_930|>": 101227, "<|LOC_931|>": 101228, "<|LOC_932|>": 101229, "<|LOC_933|>": 101230, "<|LOC_934|>": 101231, "<|LOC_935|>": 101232, "<|LOC_936|>": 101233, "<|LOC_937|>": 101234, "<|LOC_938|>": 101235, "<|LOC_939|>": 101236, "<|LOC_940|>": 101237, "<|LOC_941|>": 101238, "<|LOC_942|>": 101239, "<|LOC_943|>": 101240, "<|LOC_944|>": 101241, "<|LOC_945|>": 101242, "<|LOC_946|>": 101243, "<|LOC_947|>": 101244, "<|LOC_948|>": 101245, "<|LOC_949|>": 101246, "<|LOC_950|>": 101247, "<|LOC_951|>": 101248, "<|LOC_952|>": 101249, "<|LOC_953|>": 101250, "<|LOC_954|>": 101251, "<|LOC_955|>": 101252, "<|LOC_956|>": 101253, "<|LOC_957|>": 101254, "<|LOC_958|>": 101255, "<|LOC_959|>": 101256, "<|LOC_960|>": 101257, "<|LOC_961|>": 101258, "<|LOC_962|>": 101259, "<|LOC_963|>": 101260, "<|LOC_964|>": 101261, "<|LOC_965|>": 101262, "<|LOC_966|>": 101263, "<|LOC_967|>": 101264, "<|LOC_968|>": 101265, "<|LOC_969|>": 101266, "<|LOC_970|>": 101267, "<|LOC_971|>": 101268, "<|LOC_972|>": 101269, "<|LOC_973|>": 101270, "<|LOC_974|>": 101271, "<|LOC_975|>": 101272, "<|LOC_976|>": 101273, "<|LOC_977|>": 101274, "<|LOC_978|>": 101275, "<|LOC_979|>": 101276, "<|LOC_980|>": 101277, "<|LOC_981|>": 101278, "<|LOC_982|>": 101279, "<|LOC_983|>": 101280, "<|LOC_984|>": 101281, "<|LOC_985|>": 101282, "<|LOC_986|>": 101283, "<|LOC_987|>": 101284, "<|LOC_988|>": 101285, "<|LOC_989|>": 101286, "<|LOC_990|>": 101287, "<|LOC_991|>": 101288, "<|LOC_992|>": 101289, "<|LOC_993|>": 101290, "<|LOC_994|>": 101291, "<|LOC_995|>": 101292, "<|LOC_996|>": 101293, "<|LOC_997|>": 101294, "<|LOC_998|>": 101295, "<|LOC_999|>": 101296, "<|LOC_1000|>": 101297, "<|LOC_BEGIN|>": 101298, "<|LOC_END|>": 101299, "<|LOC_SEP|>": 101300, "<|CROP_COL_SEP|>": 101301, "<|CROP_ROW_SEP|>": 101302, "<|IMAGE_SEP|>": 101303, "<|image_pad|>": 101304, "<|IMAGE_START|>": 101305, "<|IMAGE_END|>": 101306, "<|video_pad|>": 101307, "<ecel>": 101308, "<fcel>": 101309, "<xcel>": 101310, "<lcel>": 101311, "<ucel>": 101312, "<nl>": 101313}
checkpoint-420/config.json ADDED
@@ -0,0 +1,223 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "PaddleOCRVLForConditionalGeneration"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.0,
6
+ "auto_map": {
7
+ "AutoConfig": "configuration_paddleocr_vl.PaddleOCRVLConfig",
8
+ "AutoModel": "modeling_paddleocr_vl.PaddleOCRVLForConditionalGeneration",
9
+ "AutoModelForCausalLM": "modeling_paddleocr_vl.PaddleOCRVLForConditionalGeneration"
10
+ },
11
+ "compression_ratio": 1.0,
12
+ "disable_pipeline_warmup": false,
13
+ "enable_mtp_magic_send": false,
14
+ "fp16_opt_level": "O2",
15
+ "freq_allocation": 20,
16
+ "fuse_ln": false,
17
+ "fuse_rms_norm": true,
18
+ "head_dim": 128,
19
+ "hidden_act": "silu",
20
+ "hidden_dropout_prob": 0.0,
21
+ "hidden_size": 1024,
22
+ "ignored_index": -100,
23
+ "im_patch_id": 100295,
24
+ "image_token_id": 100295,
25
+ "intermediate_size": 3072,
26
+ "max_position_embeddings": 131072,
27
+ "max_text_id": 100295,
28
+ "model_type": "paddleocr_vl",
29
+ "moe_dropout_prob": 0.0,
30
+ "moe_multimodal_dispatch_use_allgather": "v2-alltoall-unpad",
31
+ "num_attention_heads": 16,
32
+ "num_hidden_layers": 18,
33
+ "num_key_value_heads": 2,
34
+ "paddleformers_version": "0.4.0",
35
+ "pixel_hidden_size": 1152,
36
+ "rms_norm_eps": 1e-05,
37
+ "rope_is_neox_style": true,
38
+ "rope_scaling": {
39
+ "mrope_section": [
40
+ 16,
41
+ 24,
42
+ 24
43
+ ],
44
+ "rope_type": "default",
45
+ "type": "default"
46
+ },
47
+ "rope_theta": 500000,
48
+ "scale_qk_coeff": 1.0,
49
+ "seqlen": 16384,
50
+ "sliding_window": null,
51
+ "tie_word_embeddings": false,
52
+ "token_balance_loss": false,
53
+ "token_balance_seqlen": 16384,
54
+ "torch_dtype": "bfloat16",
55
+ "use_3d_rope": true,
56
+ "use_bias": false,
57
+ "use_flash_attn_with_mask": true,
58
+ "use_fp8": false,
59
+ "use_mem_eff_attn": true,
60
+ "use_recompute_moe": false,
61
+ "use_rmsnorm": true,
62
+ "video_token_id": 101307,
63
+ "vision_config": {
64
+ "_attn_implementation": "eager",
65
+ "_name_or_path": "",
66
+ "_save_to_hf": false,
67
+ "add_cross_attention": false,
68
+ "add_tail_layers": 0,
69
+ "architectures": [
70
+ "PaddleOCRVisionModel"
71
+ ],
72
+ "attention_dropout": 0.0,
73
+ "auto_map": {
74
+ "AutoConfig": "configuration_paddleocr_vl.PaddleOCRVLConfig",
75
+ "AutoModel": "modeling_paddleocr_vl.PaddleOCRVisionModel"
76
+ },
77
+ "bad_words_ids": null,
78
+ "begin_suppress_tokens": null,
79
+ "bos_token_id": null,
80
+ "chunk_size_feed_forward": 0,
81
+ "classifier_dropout": null,
82
+ "context_parallel_degree": 1,
83
+ "cross_attention_hidden_size": null,
84
+ "decoder_start_token_id": null,
85
+ "diversity_penalty": 0.0,
86
+ "do_sample": false,
87
+ "dpo_config": null,
88
+ "dtype": "bfloat16",
89
+ "early_stopping": false,
90
+ "encoder_no_repeat_ngram_size": 0,
91
+ "eos_token_id": null,
92
+ "ep_communication_type": "deepep",
93
+ "exponential_decay_length_penalty": null,
94
+ "finetuning_task": null,
95
+ "forced_bos_token_id": null,
96
+ "forced_eos_token_id": null,
97
+ "fuse_attention_ffn": false,
98
+ "fuse_attention_qkv": false,
99
+ "fuse_linear": false,
100
+ "fuse_rope": false,
101
+ "fuse_sequence_parallel_allreduce": false,
102
+ "fuse_swiglu": false,
103
+ "hidden_act": "gelu_new",
104
+ "hidden_size": 1152,
105
+ "id2label": {
106
+ "0": "LABEL_0",
107
+ "1": "LABEL_1"
108
+ },
109
+ "image_size": 384,
110
+ "intermediate_size": 4304,
111
+ "is_decoder": false,
112
+ "is_encoder_decoder": false,
113
+ "kto_config": null,
114
+ "label2id": {
115
+ "LABEL_0": 0,
116
+ "LABEL_1": 1
117
+ },
118
+ "layer_norm_eps": 1e-06,
119
+ "length_penalty": 1.0,
120
+ "loss_subbatch_sequence_length": -1,
121
+ "max_length": 20,
122
+ "min_length": 0,
123
+ "model_type": "paddleocr_vl",
124
+ "moe_subbatch_token_num": 0,
125
+ "no_recompute_layers": null,
126
+ "no_repeat_ngram_size": 0,
127
+ "num_attention_heads": 16,
128
+ "num_beam_groups": 1,
129
+ "num_beams": 1,
130
+ "num_channels": 3,
131
+ "num_choices": null,
132
+ "num_hidden_layers": 27,
133
+ "num_nextn_predict_layers": 0,
134
+ "num_return_sequences": 1,
135
+ "offload_recompute_inputs": false,
136
+ "output_attentions": false,
137
+ "output_hidden_states": false,
138
+ "output_scores": false,
139
+ "pad_token_id": 0,
140
+ "patch_size": 14,
141
+ "pipeline_parallel_degree": 1,
142
+ "pp_recompute_interval": 1,
143
+ "prefix": null,
144
+ "problem_type": null,
145
+ "pruned_heads": {},
146
+ "quantization_config": {
147
+ "act_quant_method": "abs_max",
148
+ "activation_scheme": null,
149
+ "actscale_moving_rate": 0.01,
150
+ "apply_hadamard": false,
151
+ "apply_online_actscale_step": 200,
152
+ "dense_quant_type": "",
153
+ "dtype": null,
154
+ "fmt": null,
155
+ "fp8_format_type": "hybrid",
156
+ "group_size": -1,
157
+ "hadamard_block_size": 32,
158
+ "ignore_modules": null,
159
+ "llm_int8_threshold": 6.0,
160
+ "moe_quant_type": "",
161
+ "qlora_weight_blocksize": 64,
162
+ "qlora_weight_double_quant": false,
163
+ "qlora_weight_double_quant_block_size": 256,
164
+ "quant_input_grad": false,
165
+ "quant_method": null,
166
+ "quant_round_type": 0,
167
+ "quant_type": null,
168
+ "quant_weight_grad": false,
169
+ "quantization": "",
170
+ "scale_epsilon": 1e-08,
171
+ "shift": false,
172
+ "shift_smooth_all_linears": false,
173
+ "smooth": false,
174
+ "weight_block_size": null,
175
+ "weight_quant_method": "abs_max_channel_wise",
176
+ "weight_quantize_algo": null
177
+ },
178
+ "recompute": true,
179
+ "recompute_granularity": "full",
180
+ "recompute_use_reentrant": false,
181
+ "refined_recompute": "",
182
+ "remove_invalid_values": false,
183
+ "repetition_penalty": 1.0,
184
+ "return_dict": false,
185
+ "return_dict_in_generate": false,
186
+ "sep_parallel_degree": 1,
187
+ "sep_token_id": null,
188
+ "sequence_parallel": false,
189
+ "spatial_merge_size": 2,
190
+ "suppress_tokens": null,
191
+ "task_specific_params": null,
192
+ "temperature": 1.0,
193
+ "temporal_patch_size": 2,
194
+ "tensor_parallel_degree": 1,
195
+ "tensor_parallel_output": true,
196
+ "tensor_parallel_rank": 0,
197
+ "tie_encoder_decoder": false,
198
+ "tie_word_embeddings": true,
199
+ "tokenizer_class": null,
200
+ "tokens_per_second": 2,
201
+ "top_k": 50,
202
+ "top_p": 1.0,
203
+ "typical_p": 1.0,
204
+ "use_cache": false,
205
+ "use_filtered_label_loss": false,
206
+ "use_flash_attention": true,
207
+ "use_fused_dropout_add": false,
208
+ "use_fused_head_and_loss_fn": false,
209
+ "use_fused_linear": false,
210
+ "use_fused_linear_cross_entropy": false,
211
+ "use_fused_rms_norm": false,
212
+ "use_fused_rope": false,
213
+ "use_sparse_flash_attn": true,
214
+ "use_sparse_head_and_loss_fn": false,
215
+ "use_unified_moe": false,
216
+ "using_fake_gate": false,
217
+ "virtual_pp_degree": 1
218
+ },
219
+ "vision_end_token_id": 101306,
220
+ "vision_start_token_id": 101305,
221
+ "vocab_size": 103424,
222
+ "weight_share_add_bias": true
223
+ }
checkpoint-420/generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "eos_token_id": 2,
4
+ "pad_token_id": 0,
5
+ "use_cache": false
6
+ }
checkpoint-420/master_weights-00001-of-00001.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c939d1ffb8d46bfb1b37af1fa66b6781f66ce0dbeda32eafb408fb1a114ae62
3
+ size 3622483352
checkpoint-420/master_weights.safetensors.index.json ADDED
@@ -0,0 +1,615 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 3622406592
4
+ },
5
+ "weight_map": {
6
+ "mlp_AR.pre_norm.weight": "master_weights-00001-of-00001.safetensors",
7
+ "mlp_AR.pre_norm.bias": "master_weights-00001-of-00001.safetensors",
8
+ "mlp_AR.linear_1.weight": "master_weights-00001-of-00001.safetensors",
9
+ "mlp_AR.linear_1.bias": "master_weights-00001-of-00001.safetensors",
10
+ "mlp_AR.linear_2.weight": "master_weights-00001-of-00001.safetensors",
11
+ "mlp_AR.linear_2.bias": "master_weights-00001-of-00001.safetensors",
12
+ "visual.vision_model.embeddings.patch_embedding.weight": "master_weights-00001-of-00001.safetensors",
13
+ "visual.vision_model.embeddings.patch_embedding.bias": "master_weights-00001-of-00001.safetensors",
14
+ "visual.vision_model.embeddings.position_embedding.weight": "master_weights-00001-of-00001.safetensors",
15
+ "visual.vision_model.encoder.layers.0.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
16
+ "visual.vision_model.encoder.layers.0.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
17
+ "visual.vision_model.encoder.layers.0.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
18
+ "visual.vision_model.encoder.layers.0.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
19
+ "visual.vision_model.encoder.layers.0.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
20
+ "visual.vision_model.encoder.layers.0.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
21
+ "visual.vision_model.encoder.layers.0.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
22
+ "visual.vision_model.encoder.layers.0.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
23
+ "visual.vision_model.encoder.layers.0.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
24
+ "visual.vision_model.encoder.layers.0.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
25
+ "visual.vision_model.encoder.layers.0.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
26
+ "visual.vision_model.encoder.layers.0.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
27
+ "visual.vision_model.encoder.layers.0.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
28
+ "visual.vision_model.encoder.layers.0.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
29
+ "visual.vision_model.encoder.layers.0.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
30
+ "visual.vision_model.encoder.layers.0.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
31
+ "visual.vision_model.encoder.layers.1.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
32
+ "visual.vision_model.encoder.layers.1.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
33
+ "visual.vision_model.encoder.layers.1.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
34
+ "visual.vision_model.encoder.layers.1.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
35
+ "visual.vision_model.encoder.layers.1.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
36
+ "visual.vision_model.encoder.layers.1.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
37
+ "visual.vision_model.encoder.layers.1.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
38
+ "visual.vision_model.encoder.layers.1.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
39
+ "visual.vision_model.encoder.layers.1.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
40
+ "visual.vision_model.encoder.layers.1.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
41
+ "visual.vision_model.encoder.layers.1.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
42
+ "visual.vision_model.encoder.layers.1.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
43
+ "visual.vision_model.encoder.layers.1.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
44
+ "visual.vision_model.encoder.layers.1.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
45
+ "visual.vision_model.encoder.layers.1.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
46
+ "visual.vision_model.encoder.layers.1.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
47
+ "visual.vision_model.encoder.layers.2.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
48
+ "visual.vision_model.encoder.layers.2.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
49
+ "visual.vision_model.encoder.layers.2.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
50
+ "visual.vision_model.encoder.layers.2.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
51
+ "visual.vision_model.encoder.layers.2.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
52
+ "visual.vision_model.encoder.layers.2.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
53
+ "visual.vision_model.encoder.layers.2.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
54
+ "visual.vision_model.encoder.layers.2.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
55
+ "visual.vision_model.encoder.layers.2.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
56
+ "visual.vision_model.encoder.layers.2.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
57
+ "visual.vision_model.encoder.layers.2.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
58
+ "visual.vision_model.encoder.layers.2.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
59
+ "visual.vision_model.encoder.layers.2.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
60
+ "visual.vision_model.encoder.layers.2.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
61
+ "visual.vision_model.encoder.layers.2.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
62
+ "visual.vision_model.encoder.layers.2.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
63
+ "visual.vision_model.encoder.layers.3.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
64
+ "visual.vision_model.encoder.layers.3.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
65
+ "visual.vision_model.encoder.layers.3.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
66
+ "visual.vision_model.encoder.layers.3.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
67
+ "visual.vision_model.encoder.layers.3.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
68
+ "visual.vision_model.encoder.layers.3.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
69
+ "visual.vision_model.encoder.layers.3.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
70
+ "visual.vision_model.encoder.layers.3.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
71
+ "visual.vision_model.encoder.layers.3.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
72
+ "visual.vision_model.encoder.layers.3.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
73
+ "visual.vision_model.encoder.layers.3.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
74
+ "visual.vision_model.encoder.layers.3.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
75
+ "visual.vision_model.encoder.layers.3.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
76
+ "visual.vision_model.encoder.layers.3.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
77
+ "visual.vision_model.encoder.layers.3.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
78
+ "visual.vision_model.encoder.layers.3.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
79
+ "visual.vision_model.encoder.layers.4.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
80
+ "visual.vision_model.encoder.layers.4.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
81
+ "visual.vision_model.encoder.layers.4.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
82
+ "visual.vision_model.encoder.layers.4.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
83
+ "visual.vision_model.encoder.layers.4.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
84
+ "visual.vision_model.encoder.layers.4.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
85
+ "visual.vision_model.encoder.layers.4.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
86
+ "visual.vision_model.encoder.layers.4.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
87
+ "visual.vision_model.encoder.layers.4.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
88
+ "visual.vision_model.encoder.layers.4.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
89
+ "visual.vision_model.encoder.layers.4.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
90
+ "visual.vision_model.encoder.layers.4.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
91
+ "visual.vision_model.encoder.layers.4.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
92
+ "visual.vision_model.encoder.layers.4.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
93
+ "visual.vision_model.encoder.layers.4.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
94
+ "visual.vision_model.encoder.layers.4.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
95
+ "visual.vision_model.encoder.layers.5.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
96
+ "visual.vision_model.encoder.layers.5.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
97
+ "visual.vision_model.encoder.layers.5.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
98
+ "visual.vision_model.encoder.layers.5.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
99
+ "visual.vision_model.encoder.layers.5.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
100
+ "visual.vision_model.encoder.layers.5.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
101
+ "visual.vision_model.encoder.layers.5.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
102
+ "visual.vision_model.encoder.layers.5.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
103
+ "visual.vision_model.encoder.layers.5.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
104
+ "visual.vision_model.encoder.layers.5.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
105
+ "visual.vision_model.encoder.layers.5.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
106
+ "visual.vision_model.encoder.layers.5.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
107
+ "visual.vision_model.encoder.layers.5.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
108
+ "visual.vision_model.encoder.layers.5.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
109
+ "visual.vision_model.encoder.layers.5.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
110
+ "visual.vision_model.encoder.layers.5.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
111
+ "visual.vision_model.encoder.layers.6.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
112
+ "visual.vision_model.encoder.layers.6.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
113
+ "visual.vision_model.encoder.layers.6.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
114
+ "visual.vision_model.encoder.layers.6.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
115
+ "visual.vision_model.encoder.layers.6.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
116
+ "visual.vision_model.encoder.layers.6.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
117
+ "visual.vision_model.encoder.layers.6.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
118
+ "visual.vision_model.encoder.layers.6.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
119
+ "visual.vision_model.encoder.layers.6.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
120
+ "visual.vision_model.encoder.layers.6.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
121
+ "visual.vision_model.encoder.layers.6.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
122
+ "visual.vision_model.encoder.layers.6.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
123
+ "visual.vision_model.encoder.layers.6.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
124
+ "visual.vision_model.encoder.layers.6.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
125
+ "visual.vision_model.encoder.layers.6.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
126
+ "visual.vision_model.encoder.layers.6.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
127
+ "visual.vision_model.encoder.layers.7.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
128
+ "visual.vision_model.encoder.layers.7.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
129
+ "visual.vision_model.encoder.layers.7.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
130
+ "visual.vision_model.encoder.layers.7.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
131
+ "visual.vision_model.encoder.layers.7.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
132
+ "visual.vision_model.encoder.layers.7.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
133
+ "visual.vision_model.encoder.layers.7.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
134
+ "visual.vision_model.encoder.layers.7.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
135
+ "visual.vision_model.encoder.layers.7.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
136
+ "visual.vision_model.encoder.layers.7.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
137
+ "visual.vision_model.encoder.layers.7.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
138
+ "visual.vision_model.encoder.layers.7.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
139
+ "visual.vision_model.encoder.layers.7.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
140
+ "visual.vision_model.encoder.layers.7.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
141
+ "visual.vision_model.encoder.layers.7.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
142
+ "visual.vision_model.encoder.layers.7.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
143
+ "visual.vision_model.encoder.layers.8.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
144
+ "visual.vision_model.encoder.layers.8.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
145
+ "visual.vision_model.encoder.layers.8.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
146
+ "visual.vision_model.encoder.layers.8.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
147
+ "visual.vision_model.encoder.layers.8.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
148
+ "visual.vision_model.encoder.layers.8.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
149
+ "visual.vision_model.encoder.layers.8.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
150
+ "visual.vision_model.encoder.layers.8.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
151
+ "visual.vision_model.encoder.layers.8.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
152
+ "visual.vision_model.encoder.layers.8.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
153
+ "visual.vision_model.encoder.layers.8.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
154
+ "visual.vision_model.encoder.layers.8.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
155
+ "visual.vision_model.encoder.layers.8.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
156
+ "visual.vision_model.encoder.layers.8.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
157
+ "visual.vision_model.encoder.layers.8.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
158
+ "visual.vision_model.encoder.layers.8.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
159
+ "visual.vision_model.encoder.layers.9.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
160
+ "visual.vision_model.encoder.layers.9.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
161
+ "visual.vision_model.encoder.layers.9.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
162
+ "visual.vision_model.encoder.layers.9.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
163
+ "visual.vision_model.encoder.layers.9.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
164
+ "visual.vision_model.encoder.layers.9.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
165
+ "visual.vision_model.encoder.layers.9.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
166
+ "visual.vision_model.encoder.layers.9.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
167
+ "visual.vision_model.encoder.layers.9.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
168
+ "visual.vision_model.encoder.layers.9.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
169
+ "visual.vision_model.encoder.layers.9.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
170
+ "visual.vision_model.encoder.layers.9.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
171
+ "visual.vision_model.encoder.layers.9.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
172
+ "visual.vision_model.encoder.layers.9.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
173
+ "visual.vision_model.encoder.layers.9.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
174
+ "visual.vision_model.encoder.layers.9.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
175
+ "visual.vision_model.encoder.layers.10.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
176
+ "visual.vision_model.encoder.layers.10.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
177
+ "visual.vision_model.encoder.layers.10.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
178
+ "visual.vision_model.encoder.layers.10.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
179
+ "visual.vision_model.encoder.layers.10.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
180
+ "visual.vision_model.encoder.layers.10.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
181
+ "visual.vision_model.encoder.layers.10.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
182
+ "visual.vision_model.encoder.layers.10.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
183
+ "visual.vision_model.encoder.layers.10.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
184
+ "visual.vision_model.encoder.layers.10.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
185
+ "visual.vision_model.encoder.layers.10.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
186
+ "visual.vision_model.encoder.layers.10.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
187
+ "visual.vision_model.encoder.layers.10.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
188
+ "visual.vision_model.encoder.layers.10.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
189
+ "visual.vision_model.encoder.layers.10.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
190
+ "visual.vision_model.encoder.layers.10.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
191
+ "visual.vision_model.encoder.layers.11.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
192
+ "visual.vision_model.encoder.layers.11.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
193
+ "visual.vision_model.encoder.layers.11.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
194
+ "visual.vision_model.encoder.layers.11.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
195
+ "visual.vision_model.encoder.layers.11.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
196
+ "visual.vision_model.encoder.layers.11.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
197
+ "visual.vision_model.encoder.layers.11.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
198
+ "visual.vision_model.encoder.layers.11.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
199
+ "visual.vision_model.encoder.layers.11.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
200
+ "visual.vision_model.encoder.layers.11.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
201
+ "visual.vision_model.encoder.layers.11.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
202
+ "visual.vision_model.encoder.layers.11.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
203
+ "visual.vision_model.encoder.layers.11.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
204
+ "visual.vision_model.encoder.layers.11.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
205
+ "visual.vision_model.encoder.layers.11.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
206
+ "visual.vision_model.encoder.layers.11.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
207
+ "visual.vision_model.encoder.layers.12.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
208
+ "visual.vision_model.encoder.layers.12.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
209
+ "visual.vision_model.encoder.layers.12.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
210
+ "visual.vision_model.encoder.layers.12.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
211
+ "visual.vision_model.encoder.layers.12.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
212
+ "visual.vision_model.encoder.layers.12.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
213
+ "visual.vision_model.encoder.layers.12.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
214
+ "visual.vision_model.encoder.layers.12.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
215
+ "visual.vision_model.encoder.layers.12.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
216
+ "visual.vision_model.encoder.layers.12.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
217
+ "visual.vision_model.encoder.layers.12.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
218
+ "visual.vision_model.encoder.layers.12.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
219
+ "visual.vision_model.encoder.layers.12.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
220
+ "visual.vision_model.encoder.layers.12.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
221
+ "visual.vision_model.encoder.layers.12.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
222
+ "visual.vision_model.encoder.layers.12.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
223
+ "visual.vision_model.encoder.layers.13.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
224
+ "visual.vision_model.encoder.layers.13.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
225
+ "visual.vision_model.encoder.layers.13.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
226
+ "visual.vision_model.encoder.layers.13.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
227
+ "visual.vision_model.encoder.layers.13.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
228
+ "visual.vision_model.encoder.layers.13.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
229
+ "visual.vision_model.encoder.layers.13.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
230
+ "visual.vision_model.encoder.layers.13.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
231
+ "visual.vision_model.encoder.layers.13.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
232
+ "visual.vision_model.encoder.layers.13.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
233
+ "visual.vision_model.encoder.layers.13.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
234
+ "visual.vision_model.encoder.layers.13.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
235
+ "visual.vision_model.encoder.layers.13.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
236
+ "visual.vision_model.encoder.layers.13.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
237
+ "visual.vision_model.encoder.layers.13.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
238
+ "visual.vision_model.encoder.layers.13.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
239
+ "visual.vision_model.encoder.layers.14.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
240
+ "visual.vision_model.encoder.layers.14.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
241
+ "visual.vision_model.encoder.layers.14.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
242
+ "visual.vision_model.encoder.layers.14.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
243
+ "visual.vision_model.encoder.layers.14.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
244
+ "visual.vision_model.encoder.layers.14.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
245
+ "visual.vision_model.encoder.layers.14.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
246
+ "visual.vision_model.encoder.layers.14.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
247
+ "visual.vision_model.encoder.layers.14.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
248
+ "visual.vision_model.encoder.layers.14.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
249
+ "visual.vision_model.encoder.layers.14.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
250
+ "visual.vision_model.encoder.layers.14.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
251
+ "visual.vision_model.encoder.layers.14.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
252
+ "visual.vision_model.encoder.layers.14.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
253
+ "visual.vision_model.encoder.layers.14.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
254
+ "visual.vision_model.encoder.layers.14.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
255
+ "visual.vision_model.encoder.layers.15.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
256
+ "visual.vision_model.encoder.layers.15.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
257
+ "visual.vision_model.encoder.layers.15.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
258
+ "visual.vision_model.encoder.layers.15.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
259
+ "visual.vision_model.encoder.layers.15.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
260
+ "visual.vision_model.encoder.layers.15.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
261
+ "visual.vision_model.encoder.layers.15.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
262
+ "visual.vision_model.encoder.layers.15.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
263
+ "visual.vision_model.encoder.layers.15.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
264
+ "visual.vision_model.encoder.layers.15.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
265
+ "visual.vision_model.encoder.layers.15.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
266
+ "visual.vision_model.encoder.layers.15.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
267
+ "visual.vision_model.encoder.layers.15.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
268
+ "visual.vision_model.encoder.layers.15.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
269
+ "visual.vision_model.encoder.layers.15.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
270
+ "visual.vision_model.encoder.layers.15.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
271
+ "visual.vision_model.encoder.layers.16.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
272
+ "visual.vision_model.encoder.layers.16.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
273
+ "visual.vision_model.encoder.layers.16.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
274
+ "visual.vision_model.encoder.layers.16.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
275
+ "visual.vision_model.encoder.layers.16.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
276
+ "visual.vision_model.encoder.layers.16.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
277
+ "visual.vision_model.encoder.layers.16.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
278
+ "visual.vision_model.encoder.layers.16.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
279
+ "visual.vision_model.encoder.layers.16.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
280
+ "visual.vision_model.encoder.layers.16.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
281
+ "visual.vision_model.encoder.layers.16.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
282
+ "visual.vision_model.encoder.layers.16.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
283
+ "visual.vision_model.encoder.layers.16.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
284
+ "visual.vision_model.encoder.layers.16.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
285
+ "visual.vision_model.encoder.layers.16.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
286
+ "visual.vision_model.encoder.layers.16.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
287
+ "visual.vision_model.encoder.layers.17.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
288
+ "visual.vision_model.encoder.layers.17.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
289
+ "visual.vision_model.encoder.layers.17.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
290
+ "visual.vision_model.encoder.layers.17.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
291
+ "visual.vision_model.encoder.layers.17.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
292
+ "visual.vision_model.encoder.layers.17.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
293
+ "visual.vision_model.encoder.layers.17.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
294
+ "visual.vision_model.encoder.layers.17.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
295
+ "visual.vision_model.encoder.layers.17.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
296
+ "visual.vision_model.encoder.layers.17.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
297
+ "visual.vision_model.encoder.layers.17.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
298
+ "visual.vision_model.encoder.layers.17.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
299
+ "visual.vision_model.encoder.layers.17.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
300
+ "visual.vision_model.encoder.layers.17.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
301
+ "visual.vision_model.encoder.layers.17.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
302
+ "visual.vision_model.encoder.layers.17.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
303
+ "visual.vision_model.encoder.layers.18.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
304
+ "visual.vision_model.encoder.layers.18.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
305
+ "visual.vision_model.encoder.layers.18.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
306
+ "visual.vision_model.encoder.layers.18.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
307
+ "visual.vision_model.encoder.layers.18.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
308
+ "visual.vision_model.encoder.layers.18.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
309
+ "visual.vision_model.encoder.layers.18.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
310
+ "visual.vision_model.encoder.layers.18.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
311
+ "visual.vision_model.encoder.layers.18.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
312
+ "visual.vision_model.encoder.layers.18.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
313
+ "visual.vision_model.encoder.layers.18.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
314
+ "visual.vision_model.encoder.layers.18.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
315
+ "visual.vision_model.encoder.layers.18.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
316
+ "visual.vision_model.encoder.layers.18.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
317
+ "visual.vision_model.encoder.layers.18.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
318
+ "visual.vision_model.encoder.layers.18.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
319
+ "visual.vision_model.encoder.layers.19.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
320
+ "visual.vision_model.encoder.layers.19.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
321
+ "visual.vision_model.encoder.layers.19.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
322
+ "visual.vision_model.encoder.layers.19.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
323
+ "visual.vision_model.encoder.layers.19.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
324
+ "visual.vision_model.encoder.layers.19.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
325
+ "visual.vision_model.encoder.layers.19.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
326
+ "visual.vision_model.encoder.layers.19.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
327
+ "visual.vision_model.encoder.layers.19.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
328
+ "visual.vision_model.encoder.layers.19.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
329
+ "visual.vision_model.encoder.layers.19.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
330
+ "visual.vision_model.encoder.layers.19.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
331
+ "visual.vision_model.encoder.layers.19.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
332
+ "visual.vision_model.encoder.layers.19.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
333
+ "visual.vision_model.encoder.layers.19.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
334
+ "visual.vision_model.encoder.layers.19.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
335
+ "visual.vision_model.encoder.layers.20.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
336
+ "visual.vision_model.encoder.layers.20.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
337
+ "visual.vision_model.encoder.layers.20.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
338
+ "visual.vision_model.encoder.layers.20.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
339
+ "visual.vision_model.encoder.layers.20.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
340
+ "visual.vision_model.encoder.layers.20.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
341
+ "visual.vision_model.encoder.layers.20.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
342
+ "visual.vision_model.encoder.layers.20.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
343
+ "visual.vision_model.encoder.layers.20.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
344
+ "visual.vision_model.encoder.layers.20.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
345
+ "visual.vision_model.encoder.layers.20.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
346
+ "visual.vision_model.encoder.layers.20.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
347
+ "visual.vision_model.encoder.layers.20.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
348
+ "visual.vision_model.encoder.layers.20.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
349
+ "visual.vision_model.encoder.layers.20.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
350
+ "visual.vision_model.encoder.layers.20.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
351
+ "visual.vision_model.encoder.layers.21.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
352
+ "visual.vision_model.encoder.layers.21.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
353
+ "visual.vision_model.encoder.layers.21.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
354
+ "visual.vision_model.encoder.layers.21.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
355
+ "visual.vision_model.encoder.layers.21.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
356
+ "visual.vision_model.encoder.layers.21.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
357
+ "visual.vision_model.encoder.layers.21.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
358
+ "visual.vision_model.encoder.layers.21.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
359
+ "visual.vision_model.encoder.layers.21.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
360
+ "visual.vision_model.encoder.layers.21.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
361
+ "visual.vision_model.encoder.layers.21.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
362
+ "visual.vision_model.encoder.layers.21.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
363
+ "visual.vision_model.encoder.layers.21.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
364
+ "visual.vision_model.encoder.layers.21.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
365
+ "visual.vision_model.encoder.layers.21.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
366
+ "visual.vision_model.encoder.layers.21.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
367
+ "visual.vision_model.encoder.layers.22.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
368
+ "visual.vision_model.encoder.layers.22.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
369
+ "visual.vision_model.encoder.layers.22.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
370
+ "visual.vision_model.encoder.layers.22.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
371
+ "visual.vision_model.encoder.layers.22.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
372
+ "visual.vision_model.encoder.layers.22.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
373
+ "visual.vision_model.encoder.layers.22.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
374
+ "visual.vision_model.encoder.layers.22.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
375
+ "visual.vision_model.encoder.layers.22.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
376
+ "visual.vision_model.encoder.layers.22.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
377
+ "visual.vision_model.encoder.layers.22.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
378
+ "visual.vision_model.encoder.layers.22.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
379
+ "visual.vision_model.encoder.layers.22.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
380
+ "visual.vision_model.encoder.layers.22.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
381
+ "visual.vision_model.encoder.layers.22.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
382
+ "visual.vision_model.encoder.layers.22.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
383
+ "visual.vision_model.encoder.layers.23.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
384
+ "visual.vision_model.encoder.layers.23.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
385
+ "visual.vision_model.encoder.layers.23.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
386
+ "visual.vision_model.encoder.layers.23.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
387
+ "visual.vision_model.encoder.layers.23.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
388
+ "visual.vision_model.encoder.layers.23.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
389
+ "visual.vision_model.encoder.layers.23.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
390
+ "visual.vision_model.encoder.layers.23.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
391
+ "visual.vision_model.encoder.layers.23.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
392
+ "visual.vision_model.encoder.layers.23.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
393
+ "visual.vision_model.encoder.layers.23.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
394
+ "visual.vision_model.encoder.layers.23.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
395
+ "visual.vision_model.encoder.layers.23.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
396
+ "visual.vision_model.encoder.layers.23.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
397
+ "visual.vision_model.encoder.layers.23.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
398
+ "visual.vision_model.encoder.layers.23.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
399
+ "visual.vision_model.encoder.layers.24.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
400
+ "visual.vision_model.encoder.layers.24.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
401
+ "visual.vision_model.encoder.layers.24.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
402
+ "visual.vision_model.encoder.layers.24.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
403
+ "visual.vision_model.encoder.layers.24.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
404
+ "visual.vision_model.encoder.layers.24.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
405
+ "visual.vision_model.encoder.layers.24.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
406
+ "visual.vision_model.encoder.layers.24.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
407
+ "visual.vision_model.encoder.layers.24.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
408
+ "visual.vision_model.encoder.layers.24.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
409
+ "visual.vision_model.encoder.layers.24.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
410
+ "visual.vision_model.encoder.layers.24.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
411
+ "visual.vision_model.encoder.layers.24.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
412
+ "visual.vision_model.encoder.layers.24.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
413
+ "visual.vision_model.encoder.layers.24.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
414
+ "visual.vision_model.encoder.layers.24.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
415
+ "visual.vision_model.encoder.layers.25.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
416
+ "visual.vision_model.encoder.layers.25.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
417
+ "visual.vision_model.encoder.layers.25.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
418
+ "visual.vision_model.encoder.layers.25.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
419
+ "visual.vision_model.encoder.layers.25.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
420
+ "visual.vision_model.encoder.layers.25.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
421
+ "visual.vision_model.encoder.layers.25.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
422
+ "visual.vision_model.encoder.layers.25.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
423
+ "visual.vision_model.encoder.layers.25.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
424
+ "visual.vision_model.encoder.layers.25.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
425
+ "visual.vision_model.encoder.layers.25.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
426
+ "visual.vision_model.encoder.layers.25.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
427
+ "visual.vision_model.encoder.layers.25.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
428
+ "visual.vision_model.encoder.layers.25.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
429
+ "visual.vision_model.encoder.layers.25.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
430
+ "visual.vision_model.encoder.layers.25.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
431
+ "visual.vision_model.encoder.layers.26.layer_norm1.weight": "master_weights-00001-of-00001.safetensors",
432
+ "visual.vision_model.encoder.layers.26.layer_norm1.bias": "master_weights-00001-of-00001.safetensors",
433
+ "visual.vision_model.encoder.layers.26.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
434
+ "visual.vision_model.encoder.layers.26.self_attn.k_proj.bias": "master_weights-00001-of-00001.safetensors",
435
+ "visual.vision_model.encoder.layers.26.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
436
+ "visual.vision_model.encoder.layers.26.self_attn.v_proj.bias": "master_weights-00001-of-00001.safetensors",
437
+ "visual.vision_model.encoder.layers.26.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
438
+ "visual.vision_model.encoder.layers.26.self_attn.q_proj.bias": "master_weights-00001-of-00001.safetensors",
439
+ "visual.vision_model.encoder.layers.26.self_attn.out_proj.weight": "master_weights-00001-of-00001.safetensors",
440
+ "visual.vision_model.encoder.layers.26.self_attn.out_proj.bias": "master_weights-00001-of-00001.safetensors",
441
+ "visual.vision_model.encoder.layers.26.layer_norm2.weight": "master_weights-00001-of-00001.safetensors",
442
+ "visual.vision_model.encoder.layers.26.layer_norm2.bias": "master_weights-00001-of-00001.safetensors",
443
+ "visual.vision_model.encoder.layers.26.mlp.fc1.weight": "master_weights-00001-of-00001.safetensors",
444
+ "visual.vision_model.encoder.layers.26.mlp.fc1.bias": "master_weights-00001-of-00001.safetensors",
445
+ "visual.vision_model.encoder.layers.26.mlp.fc2.weight": "master_weights-00001-of-00001.safetensors",
446
+ "visual.vision_model.encoder.layers.26.mlp.fc2.bias": "master_weights-00001-of-00001.safetensors",
447
+ "visual.vision_model.post_layernorm.weight": "master_weights-00001-of-00001.safetensors",
448
+ "visual.vision_model.post_layernorm.bias": "master_weights-00001-of-00001.safetensors",
449
+ "model.embed_tokens.weight": "master_weights-00001-of-00001.safetensors",
450
+ "model.layers.0.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
451
+ "model.layers.0.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
452
+ "model.layers.0.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
453
+ "model.layers.0.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
454
+ "model.layers.0.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
455
+ "model.layers.0.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
456
+ "model.layers.0.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
457
+ "model.layers.0.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
458
+ "model.layers.0.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
459
+ "model.layers.1.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
460
+ "model.layers.1.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
461
+ "model.layers.1.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
462
+ "model.layers.1.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
463
+ "model.layers.1.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
464
+ "model.layers.1.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
465
+ "model.layers.1.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
466
+ "model.layers.1.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
467
+ "model.layers.1.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
468
+ "model.layers.2.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
469
+ "model.layers.2.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
470
+ "model.layers.2.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
471
+ "model.layers.2.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
472
+ "model.layers.2.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
473
+ "model.layers.2.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
474
+ "model.layers.2.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
475
+ "model.layers.2.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
476
+ "model.layers.2.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
477
+ "model.layers.3.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
478
+ "model.layers.3.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
479
+ "model.layers.3.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
480
+ "model.layers.3.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
481
+ "model.layers.3.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
482
+ "model.layers.3.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
483
+ "model.layers.3.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
484
+ "model.layers.3.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
485
+ "model.layers.3.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
486
+ "model.layers.4.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
487
+ "model.layers.4.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
488
+ "model.layers.4.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
489
+ "model.layers.4.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
490
+ "model.layers.4.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
491
+ "model.layers.4.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
492
+ "model.layers.4.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
493
+ "model.layers.4.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
494
+ "model.layers.4.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
495
+ "model.layers.5.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
496
+ "model.layers.5.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
497
+ "model.layers.5.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
498
+ "model.layers.5.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
499
+ "model.layers.5.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
500
+ "model.layers.5.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
501
+ "model.layers.5.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
502
+ "model.layers.5.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
503
+ "model.layers.5.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
504
+ "model.layers.6.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
505
+ "model.layers.6.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
506
+ "model.layers.6.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
507
+ "model.layers.6.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
508
+ "model.layers.6.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
509
+ "model.layers.6.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
510
+ "model.layers.6.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
511
+ "model.layers.6.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
512
+ "model.layers.6.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
513
+ "model.layers.7.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
514
+ "model.layers.7.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
515
+ "model.layers.7.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
516
+ "model.layers.7.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
517
+ "model.layers.7.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
518
+ "model.layers.7.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
519
+ "model.layers.7.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
520
+ "model.layers.7.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
521
+ "model.layers.7.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
522
+ "model.layers.8.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
523
+ "model.layers.8.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
524
+ "model.layers.8.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
525
+ "model.layers.8.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
526
+ "model.layers.8.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
527
+ "model.layers.8.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
528
+ "model.layers.8.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
529
+ "model.layers.8.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
530
+ "model.layers.8.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
531
+ "model.layers.9.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
532
+ "model.layers.9.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
533
+ "model.layers.9.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
534
+ "model.layers.9.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
535
+ "model.layers.9.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
536
+ "model.layers.9.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
537
+ "model.layers.9.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
538
+ "model.layers.9.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
539
+ "model.layers.9.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
540
+ "model.layers.10.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
541
+ "model.layers.10.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
542
+ "model.layers.10.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
543
+ "model.layers.10.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
544
+ "model.layers.10.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
545
+ "model.layers.10.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
546
+ "model.layers.10.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
547
+ "model.layers.10.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
548
+ "model.layers.10.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
549
+ "model.layers.11.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
550
+ "model.layers.11.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
551
+ "model.layers.11.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
552
+ "model.layers.11.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
553
+ "model.layers.11.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
554
+ "model.layers.11.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
555
+ "model.layers.11.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
556
+ "model.layers.11.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
557
+ "model.layers.11.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
558
+ "model.layers.12.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
559
+ "model.layers.12.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
560
+ "model.layers.12.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
561
+ "model.layers.12.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
562
+ "model.layers.12.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
563
+ "model.layers.12.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
564
+ "model.layers.12.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
565
+ "model.layers.12.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
566
+ "model.layers.12.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
567
+ "model.layers.13.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
568
+ "model.layers.13.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
569
+ "model.layers.13.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
570
+ "model.layers.13.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
571
+ "model.layers.13.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
572
+ "model.layers.13.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
573
+ "model.layers.13.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
574
+ "model.layers.13.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
575
+ "model.layers.13.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
576
+ "model.layers.14.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
577
+ "model.layers.14.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
578
+ "model.layers.14.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
579
+ "model.layers.14.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
580
+ "model.layers.14.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
581
+ "model.layers.14.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
582
+ "model.layers.14.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
583
+ "model.layers.14.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
584
+ "model.layers.14.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
585
+ "model.layers.15.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
586
+ "model.layers.15.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
587
+ "model.layers.15.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
588
+ "model.layers.15.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
589
+ "model.layers.15.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
590
+ "model.layers.15.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
591
+ "model.layers.15.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
592
+ "model.layers.15.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
593
+ "model.layers.15.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
594
+ "model.layers.16.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
595
+ "model.layers.16.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
596
+ "model.layers.16.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
597
+ "model.layers.16.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
598
+ "model.layers.16.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
599
+ "model.layers.16.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
600
+ "model.layers.16.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
601
+ "model.layers.16.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
602
+ "model.layers.16.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
603
+ "model.layers.17.self_attn.q_proj.weight": "master_weights-00001-of-00001.safetensors",
604
+ "model.layers.17.self_attn.k_proj.weight": "master_weights-00001-of-00001.safetensors",
605
+ "model.layers.17.self_attn.v_proj.weight": "master_weights-00001-of-00001.safetensors",
606
+ "model.layers.17.self_attn.o_proj.weight": "master_weights-00001-of-00001.safetensors",
607
+ "model.layers.17.mlp.gate_proj.weight": "master_weights-00001-of-00001.safetensors",
608
+ "model.layers.17.mlp.up_proj.weight": "master_weights-00001-of-00001.safetensors",
609
+ "model.layers.17.mlp.down_proj.weight": "master_weights-00001-of-00001.safetensors",
610
+ "model.layers.17.input_layernorm.weight": "master_weights-00001-of-00001.safetensors",
611
+ "model.layers.17.post_attention_layernorm.weight": "master_weights-00001-of-00001.safetensors",
612
+ "model.norm.weight": "master_weights-00001-of-00001.safetensors",
613
+ "lm_head.weight": "master_weights-00001-of-00001.safetensors"
614
+ }
615
+ }
checkpoint-420/model-00001-of-00001.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c94d703a0352affef9bd2203aee6cc51e4dfb33b7b294542eb6ae19def2237d
3
+ size 1917255968
checkpoint-420/model.safetensors.index.json ADDED
@@ -0,0 +1,627 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 1917177472
4
+ },
5
+ "weight_map": {
6
+ "mlp_AR.pre_norm.weight": "model-00001-of-00001.safetensors",
7
+ "mlp_AR.pre_norm.bias": "model-00001-of-00001.safetensors",
8
+ "mlp_AR.linear_1.weight": "model-00001-of-00001.safetensors",
9
+ "mlp_AR.linear_1.bias": "model-00001-of-00001.safetensors",
10
+ "mlp_AR.linear_2.weight": "model-00001-of-00001.safetensors",
11
+ "mlp_AR.linear_2.bias": "model-00001-of-00001.safetensors",
12
+ "visual.vision_model.embeddings.patch_embedding.weight": "model-00001-of-00001.safetensors",
13
+ "visual.vision_model.embeddings.patch_embedding.bias": "model-00001-of-00001.safetensors",
14
+ "visual.vision_model.embeddings.position_embedding.weight": "model-00001-of-00001.safetensors",
15
+ "visual.vision_model.embeddings.packing_position_embedding.weight": "model-00001-of-00001.safetensors",
16
+ "visual.vision_model.encoder.layers.0.layer_norm1.weight": "model-00001-of-00001.safetensors",
17
+ "visual.vision_model.encoder.layers.0.layer_norm1.bias": "model-00001-of-00001.safetensors",
18
+ "visual.vision_model.encoder.layers.0.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
19
+ "visual.vision_model.encoder.layers.0.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
20
+ "visual.vision_model.encoder.layers.0.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
21
+ "visual.vision_model.encoder.layers.0.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
22
+ "visual.vision_model.encoder.layers.0.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
23
+ "visual.vision_model.encoder.layers.0.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
24
+ "visual.vision_model.encoder.layers.0.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
25
+ "visual.vision_model.encoder.layers.0.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
26
+ "visual.vision_model.encoder.layers.0.layer_norm2.weight": "model-00001-of-00001.safetensors",
27
+ "visual.vision_model.encoder.layers.0.layer_norm2.bias": "model-00001-of-00001.safetensors",
28
+ "visual.vision_model.encoder.layers.0.mlp.fc1.weight": "model-00001-of-00001.safetensors",
29
+ "visual.vision_model.encoder.layers.0.mlp.fc1.bias": "model-00001-of-00001.safetensors",
30
+ "visual.vision_model.encoder.layers.0.mlp.fc2.weight": "model-00001-of-00001.safetensors",
31
+ "visual.vision_model.encoder.layers.0.mlp.fc2.bias": "model-00001-of-00001.safetensors",
32
+ "visual.vision_model.encoder.layers.1.layer_norm1.weight": "model-00001-of-00001.safetensors",
33
+ "visual.vision_model.encoder.layers.1.layer_norm1.bias": "model-00001-of-00001.safetensors",
34
+ "visual.vision_model.encoder.layers.1.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
35
+ "visual.vision_model.encoder.layers.1.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
36
+ "visual.vision_model.encoder.layers.1.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
37
+ "visual.vision_model.encoder.layers.1.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
38
+ "visual.vision_model.encoder.layers.1.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
39
+ "visual.vision_model.encoder.layers.1.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
40
+ "visual.vision_model.encoder.layers.1.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
41
+ "visual.vision_model.encoder.layers.1.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
42
+ "visual.vision_model.encoder.layers.1.layer_norm2.weight": "model-00001-of-00001.safetensors",
43
+ "visual.vision_model.encoder.layers.1.layer_norm2.bias": "model-00001-of-00001.safetensors",
44
+ "visual.vision_model.encoder.layers.1.mlp.fc1.weight": "model-00001-of-00001.safetensors",
45
+ "visual.vision_model.encoder.layers.1.mlp.fc1.bias": "model-00001-of-00001.safetensors",
46
+ "visual.vision_model.encoder.layers.1.mlp.fc2.weight": "model-00001-of-00001.safetensors",
47
+ "visual.vision_model.encoder.layers.1.mlp.fc2.bias": "model-00001-of-00001.safetensors",
48
+ "visual.vision_model.encoder.layers.2.layer_norm1.weight": "model-00001-of-00001.safetensors",
49
+ "visual.vision_model.encoder.layers.2.layer_norm1.bias": "model-00001-of-00001.safetensors",
50
+ "visual.vision_model.encoder.layers.2.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
51
+ "visual.vision_model.encoder.layers.2.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
52
+ "visual.vision_model.encoder.layers.2.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
53
+ "visual.vision_model.encoder.layers.2.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
54
+ "visual.vision_model.encoder.layers.2.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
55
+ "visual.vision_model.encoder.layers.2.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
56
+ "visual.vision_model.encoder.layers.2.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
57
+ "visual.vision_model.encoder.layers.2.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
58
+ "visual.vision_model.encoder.layers.2.layer_norm2.weight": "model-00001-of-00001.safetensors",
59
+ "visual.vision_model.encoder.layers.2.layer_norm2.bias": "model-00001-of-00001.safetensors",
60
+ "visual.vision_model.encoder.layers.2.mlp.fc1.weight": "model-00001-of-00001.safetensors",
61
+ "visual.vision_model.encoder.layers.2.mlp.fc1.bias": "model-00001-of-00001.safetensors",
62
+ "visual.vision_model.encoder.layers.2.mlp.fc2.weight": "model-00001-of-00001.safetensors",
63
+ "visual.vision_model.encoder.layers.2.mlp.fc2.bias": "model-00001-of-00001.safetensors",
64
+ "visual.vision_model.encoder.layers.3.layer_norm1.weight": "model-00001-of-00001.safetensors",
65
+ "visual.vision_model.encoder.layers.3.layer_norm1.bias": "model-00001-of-00001.safetensors",
66
+ "visual.vision_model.encoder.layers.3.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
67
+ "visual.vision_model.encoder.layers.3.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
68
+ "visual.vision_model.encoder.layers.3.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
69
+ "visual.vision_model.encoder.layers.3.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
70
+ "visual.vision_model.encoder.layers.3.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
71
+ "visual.vision_model.encoder.layers.3.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
72
+ "visual.vision_model.encoder.layers.3.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
73
+ "visual.vision_model.encoder.layers.3.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
74
+ "visual.vision_model.encoder.layers.3.layer_norm2.weight": "model-00001-of-00001.safetensors",
75
+ "visual.vision_model.encoder.layers.3.layer_norm2.bias": "model-00001-of-00001.safetensors",
76
+ "visual.vision_model.encoder.layers.3.mlp.fc1.weight": "model-00001-of-00001.safetensors",
77
+ "visual.vision_model.encoder.layers.3.mlp.fc1.bias": "model-00001-of-00001.safetensors",
78
+ "visual.vision_model.encoder.layers.3.mlp.fc2.weight": "model-00001-of-00001.safetensors",
79
+ "visual.vision_model.encoder.layers.3.mlp.fc2.bias": "model-00001-of-00001.safetensors",
80
+ "visual.vision_model.encoder.layers.4.layer_norm1.weight": "model-00001-of-00001.safetensors",
81
+ "visual.vision_model.encoder.layers.4.layer_norm1.bias": "model-00001-of-00001.safetensors",
82
+ "visual.vision_model.encoder.layers.4.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
83
+ "visual.vision_model.encoder.layers.4.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
84
+ "visual.vision_model.encoder.layers.4.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
85
+ "visual.vision_model.encoder.layers.4.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
86
+ "visual.vision_model.encoder.layers.4.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
87
+ "visual.vision_model.encoder.layers.4.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
88
+ "visual.vision_model.encoder.layers.4.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
89
+ "visual.vision_model.encoder.layers.4.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
90
+ "visual.vision_model.encoder.layers.4.layer_norm2.weight": "model-00001-of-00001.safetensors",
91
+ "visual.vision_model.encoder.layers.4.layer_norm2.bias": "model-00001-of-00001.safetensors",
92
+ "visual.vision_model.encoder.layers.4.mlp.fc1.weight": "model-00001-of-00001.safetensors",
93
+ "visual.vision_model.encoder.layers.4.mlp.fc1.bias": "model-00001-of-00001.safetensors",
94
+ "visual.vision_model.encoder.layers.4.mlp.fc2.weight": "model-00001-of-00001.safetensors",
95
+ "visual.vision_model.encoder.layers.4.mlp.fc2.bias": "model-00001-of-00001.safetensors",
96
+ "visual.vision_model.encoder.layers.5.layer_norm1.weight": "model-00001-of-00001.safetensors",
97
+ "visual.vision_model.encoder.layers.5.layer_norm1.bias": "model-00001-of-00001.safetensors",
98
+ "visual.vision_model.encoder.layers.5.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
99
+ "visual.vision_model.encoder.layers.5.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
100
+ "visual.vision_model.encoder.layers.5.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
101
+ "visual.vision_model.encoder.layers.5.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
102
+ "visual.vision_model.encoder.layers.5.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
103
+ "visual.vision_model.encoder.layers.5.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
104
+ "visual.vision_model.encoder.layers.5.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
105
+ "visual.vision_model.encoder.layers.5.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
106
+ "visual.vision_model.encoder.layers.5.layer_norm2.weight": "model-00001-of-00001.safetensors",
107
+ "visual.vision_model.encoder.layers.5.layer_norm2.bias": "model-00001-of-00001.safetensors",
108
+ "visual.vision_model.encoder.layers.5.mlp.fc1.weight": "model-00001-of-00001.safetensors",
109
+ "visual.vision_model.encoder.layers.5.mlp.fc1.bias": "model-00001-of-00001.safetensors",
110
+ "visual.vision_model.encoder.layers.5.mlp.fc2.weight": "model-00001-of-00001.safetensors",
111
+ "visual.vision_model.encoder.layers.5.mlp.fc2.bias": "model-00001-of-00001.safetensors",
112
+ "visual.vision_model.encoder.layers.6.layer_norm1.weight": "model-00001-of-00001.safetensors",
113
+ "visual.vision_model.encoder.layers.6.layer_norm1.bias": "model-00001-of-00001.safetensors",
114
+ "visual.vision_model.encoder.layers.6.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
115
+ "visual.vision_model.encoder.layers.6.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
116
+ "visual.vision_model.encoder.layers.6.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
117
+ "visual.vision_model.encoder.layers.6.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
118
+ "visual.vision_model.encoder.layers.6.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
119
+ "visual.vision_model.encoder.layers.6.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
120
+ "visual.vision_model.encoder.layers.6.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
121
+ "visual.vision_model.encoder.layers.6.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
122
+ "visual.vision_model.encoder.layers.6.layer_norm2.weight": "model-00001-of-00001.safetensors",
123
+ "visual.vision_model.encoder.layers.6.layer_norm2.bias": "model-00001-of-00001.safetensors",
124
+ "visual.vision_model.encoder.layers.6.mlp.fc1.weight": "model-00001-of-00001.safetensors",
125
+ "visual.vision_model.encoder.layers.6.mlp.fc1.bias": "model-00001-of-00001.safetensors",
126
+ "visual.vision_model.encoder.layers.6.mlp.fc2.weight": "model-00001-of-00001.safetensors",
127
+ "visual.vision_model.encoder.layers.6.mlp.fc2.bias": "model-00001-of-00001.safetensors",
128
+ "visual.vision_model.encoder.layers.7.layer_norm1.weight": "model-00001-of-00001.safetensors",
129
+ "visual.vision_model.encoder.layers.7.layer_norm1.bias": "model-00001-of-00001.safetensors",
130
+ "visual.vision_model.encoder.layers.7.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
131
+ "visual.vision_model.encoder.layers.7.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
132
+ "visual.vision_model.encoder.layers.7.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
133
+ "visual.vision_model.encoder.layers.7.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
134
+ "visual.vision_model.encoder.layers.7.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
135
+ "visual.vision_model.encoder.layers.7.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
136
+ "visual.vision_model.encoder.layers.7.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
137
+ "visual.vision_model.encoder.layers.7.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
138
+ "visual.vision_model.encoder.layers.7.layer_norm2.weight": "model-00001-of-00001.safetensors",
139
+ "visual.vision_model.encoder.layers.7.layer_norm2.bias": "model-00001-of-00001.safetensors",
140
+ "visual.vision_model.encoder.layers.7.mlp.fc1.weight": "model-00001-of-00001.safetensors",
141
+ "visual.vision_model.encoder.layers.7.mlp.fc1.bias": "model-00001-of-00001.safetensors",
142
+ "visual.vision_model.encoder.layers.7.mlp.fc2.weight": "model-00001-of-00001.safetensors",
143
+ "visual.vision_model.encoder.layers.7.mlp.fc2.bias": "model-00001-of-00001.safetensors",
144
+ "visual.vision_model.encoder.layers.8.layer_norm1.weight": "model-00001-of-00001.safetensors",
145
+ "visual.vision_model.encoder.layers.8.layer_norm1.bias": "model-00001-of-00001.safetensors",
146
+ "visual.vision_model.encoder.layers.8.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
147
+ "visual.vision_model.encoder.layers.8.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
148
+ "visual.vision_model.encoder.layers.8.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
149
+ "visual.vision_model.encoder.layers.8.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
150
+ "visual.vision_model.encoder.layers.8.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
151
+ "visual.vision_model.encoder.layers.8.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
152
+ "visual.vision_model.encoder.layers.8.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
153
+ "visual.vision_model.encoder.layers.8.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
154
+ "visual.vision_model.encoder.layers.8.layer_norm2.weight": "model-00001-of-00001.safetensors",
155
+ "visual.vision_model.encoder.layers.8.layer_norm2.bias": "model-00001-of-00001.safetensors",
156
+ "visual.vision_model.encoder.layers.8.mlp.fc1.weight": "model-00001-of-00001.safetensors",
157
+ "visual.vision_model.encoder.layers.8.mlp.fc1.bias": "model-00001-of-00001.safetensors",
158
+ "visual.vision_model.encoder.layers.8.mlp.fc2.weight": "model-00001-of-00001.safetensors",
159
+ "visual.vision_model.encoder.layers.8.mlp.fc2.bias": "model-00001-of-00001.safetensors",
160
+ "visual.vision_model.encoder.layers.9.layer_norm1.weight": "model-00001-of-00001.safetensors",
161
+ "visual.vision_model.encoder.layers.9.layer_norm1.bias": "model-00001-of-00001.safetensors",
162
+ "visual.vision_model.encoder.layers.9.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
163
+ "visual.vision_model.encoder.layers.9.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
164
+ "visual.vision_model.encoder.layers.9.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
165
+ "visual.vision_model.encoder.layers.9.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
166
+ "visual.vision_model.encoder.layers.9.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
167
+ "visual.vision_model.encoder.layers.9.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
168
+ "visual.vision_model.encoder.layers.9.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
169
+ "visual.vision_model.encoder.layers.9.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
170
+ "visual.vision_model.encoder.layers.9.layer_norm2.weight": "model-00001-of-00001.safetensors",
171
+ "visual.vision_model.encoder.layers.9.layer_norm2.bias": "model-00001-of-00001.safetensors",
172
+ "visual.vision_model.encoder.layers.9.mlp.fc1.weight": "model-00001-of-00001.safetensors",
173
+ "visual.vision_model.encoder.layers.9.mlp.fc1.bias": "model-00001-of-00001.safetensors",
174
+ "visual.vision_model.encoder.layers.9.mlp.fc2.weight": "model-00001-of-00001.safetensors",
175
+ "visual.vision_model.encoder.layers.9.mlp.fc2.bias": "model-00001-of-00001.safetensors",
176
+ "visual.vision_model.encoder.layers.10.layer_norm1.weight": "model-00001-of-00001.safetensors",
177
+ "visual.vision_model.encoder.layers.10.layer_norm1.bias": "model-00001-of-00001.safetensors",
178
+ "visual.vision_model.encoder.layers.10.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
179
+ "visual.vision_model.encoder.layers.10.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
180
+ "visual.vision_model.encoder.layers.10.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
181
+ "visual.vision_model.encoder.layers.10.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
182
+ "visual.vision_model.encoder.layers.10.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
183
+ "visual.vision_model.encoder.layers.10.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
184
+ "visual.vision_model.encoder.layers.10.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
185
+ "visual.vision_model.encoder.layers.10.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
186
+ "visual.vision_model.encoder.layers.10.layer_norm2.weight": "model-00001-of-00001.safetensors",
187
+ "visual.vision_model.encoder.layers.10.layer_norm2.bias": "model-00001-of-00001.safetensors",
188
+ "visual.vision_model.encoder.layers.10.mlp.fc1.weight": "model-00001-of-00001.safetensors",
189
+ "visual.vision_model.encoder.layers.10.mlp.fc1.bias": "model-00001-of-00001.safetensors",
190
+ "visual.vision_model.encoder.layers.10.mlp.fc2.weight": "model-00001-of-00001.safetensors",
191
+ "visual.vision_model.encoder.layers.10.mlp.fc2.bias": "model-00001-of-00001.safetensors",
192
+ "visual.vision_model.encoder.layers.11.layer_norm1.weight": "model-00001-of-00001.safetensors",
193
+ "visual.vision_model.encoder.layers.11.layer_norm1.bias": "model-00001-of-00001.safetensors",
194
+ "visual.vision_model.encoder.layers.11.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
195
+ "visual.vision_model.encoder.layers.11.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
196
+ "visual.vision_model.encoder.layers.11.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
197
+ "visual.vision_model.encoder.layers.11.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
198
+ "visual.vision_model.encoder.layers.11.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
199
+ "visual.vision_model.encoder.layers.11.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
200
+ "visual.vision_model.encoder.layers.11.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
201
+ "visual.vision_model.encoder.layers.11.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
202
+ "visual.vision_model.encoder.layers.11.layer_norm2.weight": "model-00001-of-00001.safetensors",
203
+ "visual.vision_model.encoder.layers.11.layer_norm2.bias": "model-00001-of-00001.safetensors",
204
+ "visual.vision_model.encoder.layers.11.mlp.fc1.weight": "model-00001-of-00001.safetensors",
205
+ "visual.vision_model.encoder.layers.11.mlp.fc1.bias": "model-00001-of-00001.safetensors",
206
+ "visual.vision_model.encoder.layers.11.mlp.fc2.weight": "model-00001-of-00001.safetensors",
207
+ "visual.vision_model.encoder.layers.11.mlp.fc2.bias": "model-00001-of-00001.safetensors",
208
+ "visual.vision_model.encoder.layers.12.layer_norm1.weight": "model-00001-of-00001.safetensors",
209
+ "visual.vision_model.encoder.layers.12.layer_norm1.bias": "model-00001-of-00001.safetensors",
210
+ "visual.vision_model.encoder.layers.12.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
211
+ "visual.vision_model.encoder.layers.12.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
212
+ "visual.vision_model.encoder.layers.12.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
213
+ "visual.vision_model.encoder.layers.12.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
214
+ "visual.vision_model.encoder.layers.12.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
215
+ "visual.vision_model.encoder.layers.12.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
216
+ "visual.vision_model.encoder.layers.12.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
217
+ "visual.vision_model.encoder.layers.12.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
218
+ "visual.vision_model.encoder.layers.12.layer_norm2.weight": "model-00001-of-00001.safetensors",
219
+ "visual.vision_model.encoder.layers.12.layer_norm2.bias": "model-00001-of-00001.safetensors",
220
+ "visual.vision_model.encoder.layers.12.mlp.fc1.weight": "model-00001-of-00001.safetensors",
221
+ "visual.vision_model.encoder.layers.12.mlp.fc1.bias": "model-00001-of-00001.safetensors",
222
+ "visual.vision_model.encoder.layers.12.mlp.fc2.weight": "model-00001-of-00001.safetensors",
223
+ "visual.vision_model.encoder.layers.12.mlp.fc2.bias": "model-00001-of-00001.safetensors",
224
+ "visual.vision_model.encoder.layers.13.layer_norm1.weight": "model-00001-of-00001.safetensors",
225
+ "visual.vision_model.encoder.layers.13.layer_norm1.bias": "model-00001-of-00001.safetensors",
226
+ "visual.vision_model.encoder.layers.13.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
227
+ "visual.vision_model.encoder.layers.13.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
228
+ "visual.vision_model.encoder.layers.13.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
229
+ "visual.vision_model.encoder.layers.13.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
230
+ "visual.vision_model.encoder.layers.13.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
231
+ "visual.vision_model.encoder.layers.13.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
232
+ "visual.vision_model.encoder.layers.13.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
233
+ "visual.vision_model.encoder.layers.13.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
234
+ "visual.vision_model.encoder.layers.13.layer_norm2.weight": "model-00001-of-00001.safetensors",
235
+ "visual.vision_model.encoder.layers.13.layer_norm2.bias": "model-00001-of-00001.safetensors",
236
+ "visual.vision_model.encoder.layers.13.mlp.fc1.weight": "model-00001-of-00001.safetensors",
237
+ "visual.vision_model.encoder.layers.13.mlp.fc1.bias": "model-00001-of-00001.safetensors",
238
+ "visual.vision_model.encoder.layers.13.mlp.fc2.weight": "model-00001-of-00001.safetensors",
239
+ "visual.vision_model.encoder.layers.13.mlp.fc2.bias": "model-00001-of-00001.safetensors",
240
+ "visual.vision_model.encoder.layers.14.layer_norm1.weight": "model-00001-of-00001.safetensors",
241
+ "visual.vision_model.encoder.layers.14.layer_norm1.bias": "model-00001-of-00001.safetensors",
242
+ "visual.vision_model.encoder.layers.14.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
243
+ "visual.vision_model.encoder.layers.14.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
244
+ "visual.vision_model.encoder.layers.14.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
245
+ "visual.vision_model.encoder.layers.14.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
246
+ "visual.vision_model.encoder.layers.14.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
247
+ "visual.vision_model.encoder.layers.14.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
248
+ "visual.vision_model.encoder.layers.14.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
249
+ "visual.vision_model.encoder.layers.14.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
250
+ "visual.vision_model.encoder.layers.14.layer_norm2.weight": "model-00001-of-00001.safetensors",
251
+ "visual.vision_model.encoder.layers.14.layer_norm2.bias": "model-00001-of-00001.safetensors",
252
+ "visual.vision_model.encoder.layers.14.mlp.fc1.weight": "model-00001-of-00001.safetensors",
253
+ "visual.vision_model.encoder.layers.14.mlp.fc1.bias": "model-00001-of-00001.safetensors",
254
+ "visual.vision_model.encoder.layers.14.mlp.fc2.weight": "model-00001-of-00001.safetensors",
255
+ "visual.vision_model.encoder.layers.14.mlp.fc2.bias": "model-00001-of-00001.safetensors",
256
+ "visual.vision_model.encoder.layers.15.layer_norm1.weight": "model-00001-of-00001.safetensors",
257
+ "visual.vision_model.encoder.layers.15.layer_norm1.bias": "model-00001-of-00001.safetensors",
258
+ "visual.vision_model.encoder.layers.15.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
259
+ "visual.vision_model.encoder.layers.15.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
260
+ "visual.vision_model.encoder.layers.15.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
261
+ "visual.vision_model.encoder.layers.15.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
262
+ "visual.vision_model.encoder.layers.15.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
263
+ "visual.vision_model.encoder.layers.15.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
264
+ "visual.vision_model.encoder.layers.15.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
265
+ "visual.vision_model.encoder.layers.15.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
266
+ "visual.vision_model.encoder.layers.15.layer_norm2.weight": "model-00001-of-00001.safetensors",
267
+ "visual.vision_model.encoder.layers.15.layer_norm2.bias": "model-00001-of-00001.safetensors",
268
+ "visual.vision_model.encoder.layers.15.mlp.fc1.weight": "model-00001-of-00001.safetensors",
269
+ "visual.vision_model.encoder.layers.15.mlp.fc1.bias": "model-00001-of-00001.safetensors",
270
+ "visual.vision_model.encoder.layers.15.mlp.fc2.weight": "model-00001-of-00001.safetensors",
271
+ "visual.vision_model.encoder.layers.15.mlp.fc2.bias": "model-00001-of-00001.safetensors",
272
+ "visual.vision_model.encoder.layers.16.layer_norm1.weight": "model-00001-of-00001.safetensors",
273
+ "visual.vision_model.encoder.layers.16.layer_norm1.bias": "model-00001-of-00001.safetensors",
274
+ "visual.vision_model.encoder.layers.16.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
275
+ "visual.vision_model.encoder.layers.16.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
276
+ "visual.vision_model.encoder.layers.16.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
277
+ "visual.vision_model.encoder.layers.16.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
278
+ "visual.vision_model.encoder.layers.16.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
279
+ "visual.vision_model.encoder.layers.16.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
280
+ "visual.vision_model.encoder.layers.16.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
281
+ "visual.vision_model.encoder.layers.16.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
282
+ "visual.vision_model.encoder.layers.16.layer_norm2.weight": "model-00001-of-00001.safetensors",
283
+ "visual.vision_model.encoder.layers.16.layer_norm2.bias": "model-00001-of-00001.safetensors",
284
+ "visual.vision_model.encoder.layers.16.mlp.fc1.weight": "model-00001-of-00001.safetensors",
285
+ "visual.vision_model.encoder.layers.16.mlp.fc1.bias": "model-00001-of-00001.safetensors",
286
+ "visual.vision_model.encoder.layers.16.mlp.fc2.weight": "model-00001-of-00001.safetensors",
287
+ "visual.vision_model.encoder.layers.16.mlp.fc2.bias": "model-00001-of-00001.safetensors",
288
+ "visual.vision_model.encoder.layers.17.layer_norm1.weight": "model-00001-of-00001.safetensors",
289
+ "visual.vision_model.encoder.layers.17.layer_norm1.bias": "model-00001-of-00001.safetensors",
290
+ "visual.vision_model.encoder.layers.17.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
291
+ "visual.vision_model.encoder.layers.17.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
292
+ "visual.vision_model.encoder.layers.17.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
293
+ "visual.vision_model.encoder.layers.17.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
294
+ "visual.vision_model.encoder.layers.17.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
295
+ "visual.vision_model.encoder.layers.17.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
296
+ "visual.vision_model.encoder.layers.17.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
297
+ "visual.vision_model.encoder.layers.17.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
298
+ "visual.vision_model.encoder.layers.17.layer_norm2.weight": "model-00001-of-00001.safetensors",
299
+ "visual.vision_model.encoder.layers.17.layer_norm2.bias": "model-00001-of-00001.safetensors",
300
+ "visual.vision_model.encoder.layers.17.mlp.fc1.weight": "model-00001-of-00001.safetensors",
301
+ "visual.vision_model.encoder.layers.17.mlp.fc1.bias": "model-00001-of-00001.safetensors",
302
+ "visual.vision_model.encoder.layers.17.mlp.fc2.weight": "model-00001-of-00001.safetensors",
303
+ "visual.vision_model.encoder.layers.17.mlp.fc2.bias": "model-00001-of-00001.safetensors",
304
+ "visual.vision_model.encoder.layers.18.layer_norm1.weight": "model-00001-of-00001.safetensors",
305
+ "visual.vision_model.encoder.layers.18.layer_norm1.bias": "model-00001-of-00001.safetensors",
306
+ "visual.vision_model.encoder.layers.18.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
307
+ "visual.vision_model.encoder.layers.18.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
308
+ "visual.vision_model.encoder.layers.18.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
309
+ "visual.vision_model.encoder.layers.18.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
310
+ "visual.vision_model.encoder.layers.18.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
311
+ "visual.vision_model.encoder.layers.18.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
312
+ "visual.vision_model.encoder.layers.18.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
313
+ "visual.vision_model.encoder.layers.18.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
314
+ "visual.vision_model.encoder.layers.18.layer_norm2.weight": "model-00001-of-00001.safetensors",
315
+ "visual.vision_model.encoder.layers.18.layer_norm2.bias": "model-00001-of-00001.safetensors",
316
+ "visual.vision_model.encoder.layers.18.mlp.fc1.weight": "model-00001-of-00001.safetensors",
317
+ "visual.vision_model.encoder.layers.18.mlp.fc1.bias": "model-00001-of-00001.safetensors",
318
+ "visual.vision_model.encoder.layers.18.mlp.fc2.weight": "model-00001-of-00001.safetensors",
319
+ "visual.vision_model.encoder.layers.18.mlp.fc2.bias": "model-00001-of-00001.safetensors",
320
+ "visual.vision_model.encoder.layers.19.layer_norm1.weight": "model-00001-of-00001.safetensors",
321
+ "visual.vision_model.encoder.layers.19.layer_norm1.bias": "model-00001-of-00001.safetensors",
322
+ "visual.vision_model.encoder.layers.19.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
323
+ "visual.vision_model.encoder.layers.19.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
324
+ "visual.vision_model.encoder.layers.19.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
325
+ "visual.vision_model.encoder.layers.19.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
326
+ "visual.vision_model.encoder.layers.19.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
327
+ "visual.vision_model.encoder.layers.19.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
328
+ "visual.vision_model.encoder.layers.19.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
329
+ "visual.vision_model.encoder.layers.19.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
330
+ "visual.vision_model.encoder.layers.19.layer_norm2.weight": "model-00001-of-00001.safetensors",
331
+ "visual.vision_model.encoder.layers.19.layer_norm2.bias": "model-00001-of-00001.safetensors",
332
+ "visual.vision_model.encoder.layers.19.mlp.fc1.weight": "model-00001-of-00001.safetensors",
333
+ "visual.vision_model.encoder.layers.19.mlp.fc1.bias": "model-00001-of-00001.safetensors",
334
+ "visual.vision_model.encoder.layers.19.mlp.fc2.weight": "model-00001-of-00001.safetensors",
335
+ "visual.vision_model.encoder.layers.19.mlp.fc2.bias": "model-00001-of-00001.safetensors",
336
+ "visual.vision_model.encoder.layers.20.layer_norm1.weight": "model-00001-of-00001.safetensors",
337
+ "visual.vision_model.encoder.layers.20.layer_norm1.bias": "model-00001-of-00001.safetensors",
338
+ "visual.vision_model.encoder.layers.20.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
339
+ "visual.vision_model.encoder.layers.20.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
340
+ "visual.vision_model.encoder.layers.20.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
341
+ "visual.vision_model.encoder.layers.20.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
342
+ "visual.vision_model.encoder.layers.20.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
343
+ "visual.vision_model.encoder.layers.20.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
344
+ "visual.vision_model.encoder.layers.20.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
345
+ "visual.vision_model.encoder.layers.20.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
346
+ "visual.vision_model.encoder.layers.20.layer_norm2.weight": "model-00001-of-00001.safetensors",
347
+ "visual.vision_model.encoder.layers.20.layer_norm2.bias": "model-00001-of-00001.safetensors",
348
+ "visual.vision_model.encoder.layers.20.mlp.fc1.weight": "model-00001-of-00001.safetensors",
349
+ "visual.vision_model.encoder.layers.20.mlp.fc1.bias": "model-00001-of-00001.safetensors",
350
+ "visual.vision_model.encoder.layers.20.mlp.fc2.weight": "model-00001-of-00001.safetensors",
351
+ "visual.vision_model.encoder.layers.20.mlp.fc2.bias": "model-00001-of-00001.safetensors",
352
+ "visual.vision_model.encoder.layers.21.layer_norm1.weight": "model-00001-of-00001.safetensors",
353
+ "visual.vision_model.encoder.layers.21.layer_norm1.bias": "model-00001-of-00001.safetensors",
354
+ "visual.vision_model.encoder.layers.21.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
355
+ "visual.vision_model.encoder.layers.21.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
356
+ "visual.vision_model.encoder.layers.21.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
357
+ "visual.vision_model.encoder.layers.21.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
358
+ "visual.vision_model.encoder.layers.21.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
359
+ "visual.vision_model.encoder.layers.21.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
360
+ "visual.vision_model.encoder.layers.21.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
361
+ "visual.vision_model.encoder.layers.21.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
362
+ "visual.vision_model.encoder.layers.21.layer_norm2.weight": "model-00001-of-00001.safetensors",
363
+ "visual.vision_model.encoder.layers.21.layer_norm2.bias": "model-00001-of-00001.safetensors",
364
+ "visual.vision_model.encoder.layers.21.mlp.fc1.weight": "model-00001-of-00001.safetensors",
365
+ "visual.vision_model.encoder.layers.21.mlp.fc1.bias": "model-00001-of-00001.safetensors",
366
+ "visual.vision_model.encoder.layers.21.mlp.fc2.weight": "model-00001-of-00001.safetensors",
367
+ "visual.vision_model.encoder.layers.21.mlp.fc2.bias": "model-00001-of-00001.safetensors",
368
+ "visual.vision_model.encoder.layers.22.layer_norm1.weight": "model-00001-of-00001.safetensors",
369
+ "visual.vision_model.encoder.layers.22.layer_norm1.bias": "model-00001-of-00001.safetensors",
370
+ "visual.vision_model.encoder.layers.22.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
371
+ "visual.vision_model.encoder.layers.22.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
372
+ "visual.vision_model.encoder.layers.22.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
373
+ "visual.vision_model.encoder.layers.22.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
374
+ "visual.vision_model.encoder.layers.22.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
375
+ "visual.vision_model.encoder.layers.22.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
376
+ "visual.vision_model.encoder.layers.22.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
377
+ "visual.vision_model.encoder.layers.22.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
378
+ "visual.vision_model.encoder.layers.22.layer_norm2.weight": "model-00001-of-00001.safetensors",
379
+ "visual.vision_model.encoder.layers.22.layer_norm2.bias": "model-00001-of-00001.safetensors",
380
+ "visual.vision_model.encoder.layers.22.mlp.fc1.weight": "model-00001-of-00001.safetensors",
381
+ "visual.vision_model.encoder.layers.22.mlp.fc1.bias": "model-00001-of-00001.safetensors",
382
+ "visual.vision_model.encoder.layers.22.mlp.fc2.weight": "model-00001-of-00001.safetensors",
383
+ "visual.vision_model.encoder.layers.22.mlp.fc2.bias": "model-00001-of-00001.safetensors",
384
+ "visual.vision_model.encoder.layers.23.layer_norm1.weight": "model-00001-of-00001.safetensors",
385
+ "visual.vision_model.encoder.layers.23.layer_norm1.bias": "model-00001-of-00001.safetensors",
386
+ "visual.vision_model.encoder.layers.23.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
387
+ "visual.vision_model.encoder.layers.23.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
388
+ "visual.vision_model.encoder.layers.23.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
389
+ "visual.vision_model.encoder.layers.23.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
390
+ "visual.vision_model.encoder.layers.23.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
391
+ "visual.vision_model.encoder.layers.23.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
392
+ "visual.vision_model.encoder.layers.23.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
393
+ "visual.vision_model.encoder.layers.23.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
394
+ "visual.vision_model.encoder.layers.23.layer_norm2.weight": "model-00001-of-00001.safetensors",
395
+ "visual.vision_model.encoder.layers.23.layer_norm2.bias": "model-00001-of-00001.safetensors",
396
+ "visual.vision_model.encoder.layers.23.mlp.fc1.weight": "model-00001-of-00001.safetensors",
397
+ "visual.vision_model.encoder.layers.23.mlp.fc1.bias": "model-00001-of-00001.safetensors",
398
+ "visual.vision_model.encoder.layers.23.mlp.fc2.weight": "model-00001-of-00001.safetensors",
399
+ "visual.vision_model.encoder.layers.23.mlp.fc2.bias": "model-00001-of-00001.safetensors",
400
+ "visual.vision_model.encoder.layers.24.layer_norm1.weight": "model-00001-of-00001.safetensors",
401
+ "visual.vision_model.encoder.layers.24.layer_norm1.bias": "model-00001-of-00001.safetensors",
402
+ "visual.vision_model.encoder.layers.24.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
403
+ "visual.vision_model.encoder.layers.24.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
404
+ "visual.vision_model.encoder.layers.24.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
405
+ "visual.vision_model.encoder.layers.24.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
406
+ "visual.vision_model.encoder.layers.24.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
407
+ "visual.vision_model.encoder.layers.24.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
408
+ "visual.vision_model.encoder.layers.24.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
409
+ "visual.vision_model.encoder.layers.24.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
410
+ "visual.vision_model.encoder.layers.24.layer_norm2.weight": "model-00001-of-00001.safetensors",
411
+ "visual.vision_model.encoder.layers.24.layer_norm2.bias": "model-00001-of-00001.safetensors",
412
+ "visual.vision_model.encoder.layers.24.mlp.fc1.weight": "model-00001-of-00001.safetensors",
413
+ "visual.vision_model.encoder.layers.24.mlp.fc1.bias": "model-00001-of-00001.safetensors",
414
+ "visual.vision_model.encoder.layers.24.mlp.fc2.weight": "model-00001-of-00001.safetensors",
415
+ "visual.vision_model.encoder.layers.24.mlp.fc2.bias": "model-00001-of-00001.safetensors",
416
+ "visual.vision_model.encoder.layers.25.layer_norm1.weight": "model-00001-of-00001.safetensors",
417
+ "visual.vision_model.encoder.layers.25.layer_norm1.bias": "model-00001-of-00001.safetensors",
418
+ "visual.vision_model.encoder.layers.25.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
419
+ "visual.vision_model.encoder.layers.25.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
420
+ "visual.vision_model.encoder.layers.25.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
421
+ "visual.vision_model.encoder.layers.25.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
422
+ "visual.vision_model.encoder.layers.25.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
423
+ "visual.vision_model.encoder.layers.25.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
424
+ "visual.vision_model.encoder.layers.25.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
425
+ "visual.vision_model.encoder.layers.25.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
426
+ "visual.vision_model.encoder.layers.25.layer_norm2.weight": "model-00001-of-00001.safetensors",
427
+ "visual.vision_model.encoder.layers.25.layer_norm2.bias": "model-00001-of-00001.safetensors",
428
+ "visual.vision_model.encoder.layers.25.mlp.fc1.weight": "model-00001-of-00001.safetensors",
429
+ "visual.vision_model.encoder.layers.25.mlp.fc1.bias": "model-00001-of-00001.safetensors",
430
+ "visual.vision_model.encoder.layers.25.mlp.fc2.weight": "model-00001-of-00001.safetensors",
431
+ "visual.vision_model.encoder.layers.25.mlp.fc2.bias": "model-00001-of-00001.safetensors",
432
+ "visual.vision_model.encoder.layers.26.layer_norm1.weight": "model-00001-of-00001.safetensors",
433
+ "visual.vision_model.encoder.layers.26.layer_norm1.bias": "model-00001-of-00001.safetensors",
434
+ "visual.vision_model.encoder.layers.26.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
435
+ "visual.vision_model.encoder.layers.26.self_attn.k_proj.bias": "model-00001-of-00001.safetensors",
436
+ "visual.vision_model.encoder.layers.26.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
437
+ "visual.vision_model.encoder.layers.26.self_attn.v_proj.bias": "model-00001-of-00001.safetensors",
438
+ "visual.vision_model.encoder.layers.26.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
439
+ "visual.vision_model.encoder.layers.26.self_attn.q_proj.bias": "model-00001-of-00001.safetensors",
440
+ "visual.vision_model.encoder.layers.26.self_attn.out_proj.weight": "model-00001-of-00001.safetensors",
441
+ "visual.vision_model.encoder.layers.26.self_attn.out_proj.bias": "model-00001-of-00001.safetensors",
442
+ "visual.vision_model.encoder.layers.26.layer_norm2.weight": "model-00001-of-00001.safetensors",
443
+ "visual.vision_model.encoder.layers.26.layer_norm2.bias": "model-00001-of-00001.safetensors",
444
+ "visual.vision_model.encoder.layers.26.mlp.fc1.weight": "model-00001-of-00001.safetensors",
445
+ "visual.vision_model.encoder.layers.26.mlp.fc1.bias": "model-00001-of-00001.safetensors",
446
+ "visual.vision_model.encoder.layers.26.mlp.fc2.weight": "model-00001-of-00001.safetensors",
447
+ "visual.vision_model.encoder.layers.26.mlp.fc2.bias": "model-00001-of-00001.safetensors",
448
+ "visual.vision_model.post_layernorm.weight": "model-00001-of-00001.safetensors",
449
+ "visual.vision_model.post_layernorm.bias": "model-00001-of-00001.safetensors",
450
+ "visual.vision_model.head.probe": "model-00001-of-00001.safetensors",
451
+ "visual.vision_model.head.attention.in_proj_weight": "model-00001-of-00001.safetensors",
452
+ "visual.vision_model.head.attention.in_proj_bias": "model-00001-of-00001.safetensors",
453
+ "visual.vision_model.head.attention.out_proj.weight": "model-00001-of-00001.safetensors",
454
+ "visual.vision_model.head.attention.out_proj.bias": "model-00001-of-00001.safetensors",
455
+ "visual.vision_model.head.layernorm.weight": "model-00001-of-00001.safetensors",
456
+ "visual.vision_model.head.layernorm.bias": "model-00001-of-00001.safetensors",
457
+ "visual.vision_model.head.mlp.fc1.weight": "model-00001-of-00001.safetensors",
458
+ "visual.vision_model.head.mlp.fc1.bias": "model-00001-of-00001.safetensors",
459
+ "visual.vision_model.head.mlp.fc2.weight": "model-00001-of-00001.safetensors",
460
+ "visual.vision_model.head.mlp.fc2.bias": "model-00001-of-00001.safetensors",
461
+ "model.embed_tokens.weight": "model-00001-of-00001.safetensors",
462
+ "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
463
+ "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
464
+ "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
465
+ "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
466
+ "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
467
+ "model.layers.0.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
468
+ "model.layers.0.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
469
+ "model.layers.0.input_layernorm.weight": "model-00001-of-00001.safetensors",
470
+ "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
471
+ "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
472
+ "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
473
+ "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
474
+ "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
475
+ "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
476
+ "model.layers.1.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
477
+ "model.layers.1.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
478
+ "model.layers.1.input_layernorm.weight": "model-00001-of-00001.safetensors",
479
+ "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
480
+ "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
481
+ "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
482
+ "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
483
+ "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
484
+ "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
485
+ "model.layers.2.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
486
+ "model.layers.2.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
487
+ "model.layers.2.input_layernorm.weight": "model-00001-of-00001.safetensors",
488
+ "model.layers.2.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
489
+ "model.layers.3.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
490
+ "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
491
+ "model.layers.3.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
492
+ "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
493
+ "model.layers.3.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
494
+ "model.layers.3.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
495
+ "model.layers.3.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
496
+ "model.layers.3.input_layernorm.weight": "model-00001-of-00001.safetensors",
497
+ "model.layers.3.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
498
+ "model.layers.4.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
499
+ "model.layers.4.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
500
+ "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
501
+ "model.layers.4.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
502
+ "model.layers.4.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
503
+ "model.layers.4.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
504
+ "model.layers.4.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
505
+ "model.layers.4.input_layernorm.weight": "model-00001-of-00001.safetensors",
506
+ "model.layers.4.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
507
+ "model.layers.5.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
508
+ "model.layers.5.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
509
+ "model.layers.5.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
510
+ "model.layers.5.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
511
+ "model.layers.5.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
512
+ "model.layers.5.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
513
+ "model.layers.5.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
514
+ "model.layers.5.input_layernorm.weight": "model-00001-of-00001.safetensors",
515
+ "model.layers.5.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
516
+ "model.layers.6.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
517
+ "model.layers.6.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
518
+ "model.layers.6.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
519
+ "model.layers.6.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
520
+ "model.layers.6.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
521
+ "model.layers.6.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
522
+ "model.layers.6.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
523
+ "model.layers.6.input_layernorm.weight": "model-00001-of-00001.safetensors",
524
+ "model.layers.6.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
525
+ "model.layers.7.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
526
+ "model.layers.7.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
527
+ "model.layers.7.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
528
+ "model.layers.7.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
529
+ "model.layers.7.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
530
+ "model.layers.7.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
531
+ "model.layers.7.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
532
+ "model.layers.7.input_layernorm.weight": "model-00001-of-00001.safetensors",
533
+ "model.layers.7.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
534
+ "model.layers.8.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
535
+ "model.layers.8.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
536
+ "model.layers.8.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
537
+ "model.layers.8.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
538
+ "model.layers.8.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
539
+ "model.layers.8.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
540
+ "model.layers.8.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
541
+ "model.layers.8.input_layernorm.weight": "model-00001-of-00001.safetensors",
542
+ "model.layers.8.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
543
+ "model.layers.9.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
544
+ "model.layers.9.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
545
+ "model.layers.9.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
546
+ "model.layers.9.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
547
+ "model.layers.9.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
548
+ "model.layers.9.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
549
+ "model.layers.9.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
550
+ "model.layers.9.input_layernorm.weight": "model-00001-of-00001.safetensors",
551
+ "model.layers.9.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
552
+ "model.layers.10.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
553
+ "model.layers.10.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
554
+ "model.layers.10.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
555
+ "model.layers.10.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
556
+ "model.layers.10.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
557
+ "model.layers.10.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
558
+ "model.layers.10.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
559
+ "model.layers.10.input_layernorm.weight": "model-00001-of-00001.safetensors",
560
+ "model.layers.10.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
561
+ "model.layers.11.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
562
+ "model.layers.11.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
563
+ "model.layers.11.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
564
+ "model.layers.11.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
565
+ "model.layers.11.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
566
+ "model.layers.11.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
567
+ "model.layers.11.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
568
+ "model.layers.11.input_layernorm.weight": "model-00001-of-00001.safetensors",
569
+ "model.layers.11.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
570
+ "model.layers.12.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
571
+ "model.layers.12.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
572
+ "model.layers.12.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
573
+ "model.layers.12.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
574
+ "model.layers.12.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
575
+ "model.layers.12.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
576
+ "model.layers.12.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
577
+ "model.layers.12.input_layernorm.weight": "model-00001-of-00001.safetensors",
578
+ "model.layers.12.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
579
+ "model.layers.13.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
580
+ "model.layers.13.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
581
+ "model.layers.13.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
582
+ "model.layers.13.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
583
+ "model.layers.13.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
584
+ "model.layers.13.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
585
+ "model.layers.13.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
586
+ "model.layers.13.input_layernorm.weight": "model-00001-of-00001.safetensors",
587
+ "model.layers.13.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
588
+ "model.layers.14.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
589
+ "model.layers.14.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
590
+ "model.layers.14.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
591
+ "model.layers.14.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
592
+ "model.layers.14.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
593
+ "model.layers.14.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
594
+ "model.layers.14.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
595
+ "model.layers.14.input_layernorm.weight": "model-00001-of-00001.safetensors",
596
+ "model.layers.14.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
597
+ "model.layers.15.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
598
+ "model.layers.15.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
599
+ "model.layers.15.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
600
+ "model.layers.15.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
601
+ "model.layers.15.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
602
+ "model.layers.15.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
603
+ "model.layers.15.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
604
+ "model.layers.15.input_layernorm.weight": "model-00001-of-00001.safetensors",
605
+ "model.layers.15.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
606
+ "model.layers.16.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
607
+ "model.layers.16.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
608
+ "model.layers.16.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
609
+ "model.layers.16.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
610
+ "model.layers.16.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
611
+ "model.layers.16.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
612
+ "model.layers.16.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
613
+ "model.layers.16.input_layernorm.weight": "model-00001-of-00001.safetensors",
614
+ "model.layers.16.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
615
+ "model.layers.17.self_attn.q_proj.weight": "model-00001-of-00001.safetensors",
616
+ "model.layers.17.self_attn.k_proj.weight": "model-00001-of-00001.safetensors",
617
+ "model.layers.17.self_attn.v_proj.weight": "model-00001-of-00001.safetensors",
618
+ "model.layers.17.self_attn.o_proj.weight": "model-00001-of-00001.safetensors",
619
+ "model.layers.17.mlp.gate_proj.weight": "model-00001-of-00001.safetensors",
620
+ "model.layers.17.mlp.up_proj.weight": "model-00001-of-00001.safetensors",
621
+ "model.layers.17.mlp.down_proj.weight": "model-00001-of-00001.safetensors",
622
+ "model.layers.17.input_layernorm.weight": "model-00001-of-00001.safetensors",
623
+ "model.layers.17.post_attention_layernorm.weight": "model-00001-of-00001.safetensors",
624
+ "model.norm.weight": "model-00001-of-00001.safetensors",
625
+ "lm_head.weight": "model-00001-of-00001.safetensors"
626
+ }
627
+ }
checkpoint-420/optimizer-00001-of-00001.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c47f3e32e7bf9d312ec8cee5d22be648a75a5ab0172892f752f0761f24d2d93
3
+ size 7245150152
checkpoint-420/optimizer.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-420/preprocessor_config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_convert_rgb": true,
3
+ "do_normalize": true,
4
+ "do_rescale": true,
5
+ "do_resize": true,
6
+ "image_mean": [
7
+ 0.5,
8
+ 0.5,
9
+ 0.5
10
+ ],
11
+ "image_processor_type": "SiglipImageProcessor",
12
+ "image_std": [
13
+ 0.5,
14
+ 0.5,
15
+ 0.5
16
+ ],
17
+ "max_pixels": 2822400,
18
+ "merge_size": 2,
19
+ "min_pixels": 147384,
20
+ "patch_size": 14,
21
+ "resample": 3,
22
+ "rescale_factor": 0.00392156862745098,
23
+ "size": {
24
+ "max_pixels": 2822400,
25
+ "min_pixels": 147384
26
+ },
27
+ "temporal_conv_size": 1,
28
+ "temporal_patch_size": 1
29
+ }
checkpoint-420/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:29cde6478b49d2682dfa8c033ed7da4f33ab76ccd1941cfe950453d3923dfec3
3
+ size 32008
checkpoint-420/scheduler.pdparams ADDED
Binary file (51 Bytes). View file