ganlinyang commited on
Commit
f3a90ef
·
verified ·
1 Parent(s): ab74849

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -36,3 +36,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
36
  2B/internvla_2B_ckpt/internvl3_2B/checkpoint-10000/unwrapped_model/tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
  2B/internvla_2B_ckpt/internvl3_2B/checkpoint-34750/unwrapped_model/tokenizer.json filter=lfs diff=lfs merge=lfs -text
38
  2B/internvla_2B_ckpt/internvl3_2B_embodied_0921/checkpoint-34750/unwrapped_model/tokenizer.json filter=lfs diff=lfs merge=lfs -text
 
 
36
  2B/internvla_2B_ckpt/internvl3_2B/checkpoint-10000/unwrapped_model/tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
  2B/internvla_2B_ckpt/internvl3_2B/checkpoint-34750/unwrapped_model/tokenizer.json filter=lfs diff=lfs merge=lfs -text
38
  2B/internvla_2B_ckpt/internvl3_2B_embodied_0921/checkpoint-34750/unwrapped_model/tokenizer.json filter=lfs diff=lfs merge=lfs -text
39
+ 2B/internvla_2B_ckpt/internvl3_2B_embodied_0922/checkpoint-34750/unwrapped_model/tokenizer.json filter=lfs diff=lfs merge=lfs -text
2B/internvla_2B_ckpt/internvl3_2B_embodied_0922/checkpoint-34750/metadata.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"global_step": 34750}
2B/internvla_2B_ckpt/internvl3_2B_embodied_0922/checkpoint-34750/unwrapped_model/added_tokens.json ADDED
@@ -0,0 +1,558 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</box>": 151673,
3
+ "</img>": 151666,
4
+ "</quad>": 151669,
5
+ "</ref>": 151671,
6
+ "</tool_call>": 151658,
7
+ "<IMG_CONTEXT>": 151667,
8
+ "<box>": 151672,
9
+ "<eoi>": 151676,
10
+ "<eot>": 151678,
11
+ "<img>": 151665,
12
+ "<left_arm_eoa>": 151680,
13
+ "<left_arm_eost>": 151684,
14
+ "<left_arm_soa>": 151679,
15
+ "<left_arm_sost>": 151683,
16
+ "<pad>": 151674,
17
+ "<quad>": 151668,
18
+ "<ref>": 151670,
19
+ "<right_arm_eoa>": 151682,
20
+ "<right_arm_eost>": 151686,
21
+ "<right_arm_soa>": 151681,
22
+ "<right_arm_sost>": 151685,
23
+ "<soi>": 151675,
24
+ "<sot>": 151677,
25
+ "<state_100>": 151786,
26
+ "<state_101>": 151787,
27
+ "<state_102>": 151788,
28
+ "<state_103>": 151789,
29
+ "<state_104>": 151790,
30
+ "<state_105>": 151791,
31
+ "<state_106>": 151792,
32
+ "<state_107>": 151793,
33
+ "<state_108>": 151794,
34
+ "<state_109>": 151795,
35
+ "<state_10>": 151696,
36
+ "<state_110>": 151796,
37
+ "<state_111>": 151797,
38
+ "<state_112>": 151798,
39
+ "<state_113>": 151799,
40
+ "<state_114>": 151800,
41
+ "<state_115>": 151801,
42
+ "<state_116>": 151802,
43
+ "<state_117>": 151803,
44
+ "<state_118>": 151804,
45
+ "<state_119>": 151805,
46
+ "<state_11>": 151697,
47
+ "<state_120>": 151806,
48
+ "<state_121>": 151807,
49
+ "<state_122>": 151808,
50
+ "<state_123>": 151809,
51
+ "<state_124>": 151810,
52
+ "<state_125>": 151811,
53
+ "<state_126>": 151812,
54
+ "<state_127>": 151813,
55
+ "<state_128>": 151814,
56
+ "<state_129>": 151815,
57
+ "<state_12>": 151698,
58
+ "<state_130>": 151816,
59
+ "<state_131>": 151817,
60
+ "<state_132>": 151818,
61
+ "<state_133>": 151819,
62
+ "<state_134>": 151820,
63
+ "<state_135>": 151821,
64
+ "<state_136>": 151822,
65
+ "<state_137>": 151823,
66
+ "<state_138>": 151824,
67
+ "<state_139>": 151825,
68
+ "<state_13>": 151699,
69
+ "<state_140>": 151826,
70
+ "<state_141>": 151827,
71
+ "<state_142>": 151828,
72
+ "<state_143>": 151829,
73
+ "<state_144>": 151830,
74
+ "<state_145>": 151831,
75
+ "<state_146>": 151832,
76
+ "<state_147>": 151833,
77
+ "<state_148>": 151834,
78
+ "<state_149>": 151835,
79
+ "<state_14>": 151700,
80
+ "<state_150>": 151836,
81
+ "<state_151>": 151837,
82
+ "<state_152>": 151838,
83
+ "<state_153>": 151839,
84
+ "<state_154>": 151840,
85
+ "<state_155>": 151841,
86
+ "<state_156>": 151842,
87
+ "<state_157>": 151843,
88
+ "<state_158>": 151844,
89
+ "<state_159>": 151845,
90
+ "<state_15>": 151701,
91
+ "<state_160>": 151846,
92
+ "<state_161>": 151847,
93
+ "<state_162>": 151848,
94
+ "<state_163>": 151849,
95
+ "<state_164>": 151850,
96
+ "<state_165>": 151851,
97
+ "<state_166>": 151852,
98
+ "<state_167>": 151853,
99
+ "<state_168>": 151854,
100
+ "<state_169>": 151855,
101
+ "<state_16>": 151702,
102
+ "<state_170>": 151856,
103
+ "<state_171>": 151857,
104
+ "<state_172>": 151858,
105
+ "<state_173>": 151859,
106
+ "<state_174>": 151860,
107
+ "<state_175>": 151861,
108
+ "<state_176>": 151862,
109
+ "<state_177>": 151863,
110
+ "<state_178>": 151864,
111
+ "<state_179>": 151865,
112
+ "<state_17>": 151703,
113
+ "<state_180>": 151866,
114
+ "<state_181>": 151867,
115
+ "<state_182>": 151868,
116
+ "<state_183>": 151869,
117
+ "<state_184>": 151870,
118
+ "<state_185>": 151871,
119
+ "<state_186>": 151872,
120
+ "<state_187>": 151873,
121
+ "<state_188>": 151874,
122
+ "<state_189>": 151875,
123
+ "<state_18>": 151704,
124
+ "<state_190>": 151876,
125
+ "<state_191>": 151877,
126
+ "<state_192>": 151878,
127
+ "<state_193>": 151879,
128
+ "<state_194>": 151880,
129
+ "<state_195>": 151881,
130
+ "<state_196>": 151882,
131
+ "<state_197>": 151883,
132
+ "<state_198>": 151884,
133
+ "<state_199>": 151885,
134
+ "<state_19>": 151705,
135
+ "<state_1>": 151687,
136
+ "<state_200>": 151886,
137
+ "<state_201>": 151887,
138
+ "<state_202>": 151888,
139
+ "<state_203>": 151889,
140
+ "<state_204>": 151890,
141
+ "<state_205>": 151891,
142
+ "<state_206>": 151892,
143
+ "<state_207>": 151893,
144
+ "<state_208>": 151894,
145
+ "<state_209>": 151895,
146
+ "<state_20>": 151706,
147
+ "<state_210>": 151896,
148
+ "<state_211>": 151897,
149
+ "<state_212>": 151898,
150
+ "<state_213>": 151899,
151
+ "<state_214>": 151900,
152
+ "<state_215>": 151901,
153
+ "<state_216>": 151902,
154
+ "<state_217>": 151903,
155
+ "<state_218>": 151904,
156
+ "<state_219>": 151905,
157
+ "<state_21>": 151707,
158
+ "<state_220>": 151906,
159
+ "<state_221>": 151907,
160
+ "<state_222>": 151908,
161
+ "<state_223>": 151909,
162
+ "<state_224>": 151910,
163
+ "<state_225>": 151911,
164
+ "<state_226>": 151912,
165
+ "<state_227>": 151913,
166
+ "<state_228>": 151914,
167
+ "<state_229>": 151915,
168
+ "<state_22>": 151708,
169
+ "<state_230>": 151916,
170
+ "<state_231>": 151917,
171
+ "<state_232>": 151918,
172
+ "<state_233>": 151919,
173
+ "<state_234>": 151920,
174
+ "<state_235>": 151921,
175
+ "<state_236>": 151922,
176
+ "<state_237>": 151923,
177
+ "<state_238>": 151924,
178
+ "<state_239>": 151925,
179
+ "<state_23>": 151709,
180
+ "<state_240>": 151926,
181
+ "<state_241>": 151927,
182
+ "<state_242>": 151928,
183
+ "<state_243>": 151929,
184
+ "<state_244>": 151930,
185
+ "<state_245>": 151931,
186
+ "<state_246>": 151932,
187
+ "<state_247>": 151933,
188
+ "<state_248>": 151934,
189
+ "<state_249>": 151935,
190
+ "<state_24>": 151710,
191
+ "<state_250>": 151936,
192
+ "<state_251>": 151937,
193
+ "<state_252>": 151938,
194
+ "<state_253>": 151939,
195
+ "<state_254>": 151940,
196
+ "<state_255>": 151941,
197
+ "<state_256>": 151942,
198
+ "<state_257>": 151943,
199
+ "<state_258>": 151944,
200
+ "<state_259>": 151945,
201
+ "<state_25>": 151711,
202
+ "<state_260>": 151946,
203
+ "<state_261>": 151947,
204
+ "<state_262>": 151948,
205
+ "<state_263>": 151949,
206
+ "<state_264>": 151950,
207
+ "<state_265>": 151951,
208
+ "<state_266>": 151952,
209
+ "<state_267>": 151953,
210
+ "<state_268>": 151954,
211
+ "<state_269>": 151955,
212
+ "<state_26>": 151712,
213
+ "<state_270>": 151956,
214
+ "<state_271>": 151957,
215
+ "<state_272>": 151958,
216
+ "<state_273>": 151959,
217
+ "<state_274>": 151960,
218
+ "<state_275>": 151961,
219
+ "<state_276>": 151962,
220
+ "<state_277>": 151963,
221
+ "<state_278>": 151964,
222
+ "<state_279>": 151965,
223
+ "<state_27>": 151713,
224
+ "<state_280>": 151966,
225
+ "<state_281>": 151967,
226
+ "<state_282>": 151968,
227
+ "<state_283>": 151969,
228
+ "<state_284>": 151970,
229
+ "<state_285>": 151971,
230
+ "<state_286>": 151972,
231
+ "<state_287>": 151973,
232
+ "<state_288>": 151974,
233
+ "<state_289>": 151975,
234
+ "<state_28>": 151714,
235
+ "<state_290>": 151976,
236
+ "<state_291>": 151977,
237
+ "<state_292>": 151978,
238
+ "<state_293>": 151979,
239
+ "<state_294>": 151980,
240
+ "<state_295>": 151981,
241
+ "<state_296>": 151982,
242
+ "<state_297>": 151983,
243
+ "<state_298>": 151984,
244
+ "<state_299>": 151985,
245
+ "<state_29>": 151715,
246
+ "<state_2>": 151688,
247
+ "<state_300>": 151986,
248
+ "<state_301>": 151987,
249
+ "<state_302>": 151988,
250
+ "<state_303>": 151989,
251
+ "<state_304>": 151990,
252
+ "<state_305>": 151991,
253
+ "<state_306>": 151992,
254
+ "<state_307>": 151993,
255
+ "<state_308>": 151994,
256
+ "<state_309>": 151995,
257
+ "<state_30>": 151716,
258
+ "<state_310>": 151996,
259
+ "<state_311>": 151997,
260
+ "<state_312>": 151998,
261
+ "<state_313>": 151999,
262
+ "<state_314>": 152000,
263
+ "<state_315>": 152001,
264
+ "<state_316>": 152002,
265
+ "<state_317>": 152003,
266
+ "<state_318>": 152004,
267
+ "<state_319>": 152005,
268
+ "<state_31>": 151717,
269
+ "<state_320>": 152006,
270
+ "<state_321>": 152007,
271
+ "<state_322>": 152008,
272
+ "<state_323>": 152009,
273
+ "<state_324>": 152010,
274
+ "<state_325>": 152011,
275
+ "<state_326>": 152012,
276
+ "<state_327>": 152013,
277
+ "<state_328>": 152014,
278
+ "<state_329>": 152015,
279
+ "<state_32>": 151718,
280
+ "<state_330>": 152016,
281
+ "<state_331>": 152017,
282
+ "<state_332>": 152018,
283
+ "<state_333>": 152019,
284
+ "<state_334>": 152020,
285
+ "<state_335>": 152021,
286
+ "<state_336>": 152022,
287
+ "<state_337>": 152023,
288
+ "<state_338>": 152024,
289
+ "<state_339>": 152025,
290
+ "<state_33>": 151719,
291
+ "<state_340>": 152026,
292
+ "<state_341>": 152027,
293
+ "<state_342>": 152028,
294
+ "<state_343>": 152029,
295
+ "<state_344>": 152030,
296
+ "<state_345>": 152031,
297
+ "<state_346>": 152032,
298
+ "<state_347>": 152033,
299
+ "<state_348>": 152034,
300
+ "<state_349>": 152035,
301
+ "<state_34>": 151720,
302
+ "<state_350>": 152036,
303
+ "<state_351>": 152037,
304
+ "<state_352>": 152038,
305
+ "<state_353>": 152039,
306
+ "<state_354>": 152040,
307
+ "<state_355>": 152041,
308
+ "<state_356>": 152042,
309
+ "<state_357>": 152043,
310
+ "<state_358>": 152044,
311
+ "<state_359>": 152045,
312
+ "<state_35>": 151721,
313
+ "<state_360>": 152046,
314
+ "<state_361>": 152047,
315
+ "<state_362>": 152048,
316
+ "<state_363>": 152049,
317
+ "<state_364>": 152050,
318
+ "<state_365>": 152051,
319
+ "<state_366>": 152052,
320
+ "<state_367>": 152053,
321
+ "<state_368>": 152054,
322
+ "<state_369>": 152055,
323
+ "<state_36>": 151722,
324
+ "<state_370>": 152056,
325
+ "<state_371>": 152057,
326
+ "<state_372>": 152058,
327
+ "<state_373>": 152059,
328
+ "<state_374>": 152060,
329
+ "<state_375>": 152061,
330
+ "<state_376>": 152062,
331
+ "<state_377>": 152063,
332
+ "<state_378>": 152064,
333
+ "<state_379>": 152065,
334
+ "<state_37>": 151723,
335
+ "<state_380>": 152066,
336
+ "<state_381>": 152067,
337
+ "<state_382>": 152068,
338
+ "<state_383>": 152069,
339
+ "<state_384>": 152070,
340
+ "<state_385>": 152071,
341
+ "<state_386>": 152072,
342
+ "<state_387>": 152073,
343
+ "<state_388>": 152074,
344
+ "<state_389>": 152075,
345
+ "<state_38>": 151724,
346
+ "<state_390>": 152076,
347
+ "<state_391>": 152077,
348
+ "<state_392>": 152078,
349
+ "<state_393>": 152079,
350
+ "<state_394>": 152080,
351
+ "<state_395>": 152081,
352
+ "<state_396>": 152082,
353
+ "<state_397>": 152083,
354
+ "<state_398>": 152084,
355
+ "<state_399>": 152085,
356
+ "<state_39>": 151725,
357
+ "<state_3>": 151689,
358
+ "<state_400>": 152086,
359
+ "<state_401>": 152087,
360
+ "<state_402>": 152088,
361
+ "<state_403>": 152089,
362
+ "<state_404>": 152090,
363
+ "<state_405>": 152091,
364
+ "<state_406>": 152092,
365
+ "<state_407>": 152093,
366
+ "<state_408>": 152094,
367
+ "<state_409>": 152095,
368
+ "<state_40>": 151726,
369
+ "<state_410>": 152096,
370
+ "<state_411>": 152097,
371
+ "<state_412>": 152098,
372
+ "<state_413>": 152099,
373
+ "<state_414>": 152100,
374
+ "<state_415>": 152101,
375
+ "<state_416>": 152102,
376
+ "<state_417>": 152103,
377
+ "<state_418>": 152104,
378
+ "<state_419>": 152105,
379
+ "<state_41>": 151727,
380
+ "<state_420>": 152106,
381
+ "<state_421>": 152107,
382
+ "<state_422>": 152108,
383
+ "<state_423>": 152109,
384
+ "<state_424>": 152110,
385
+ "<state_425>": 152111,
386
+ "<state_426>": 152112,
387
+ "<state_427>": 152113,
388
+ "<state_428>": 152114,
389
+ "<state_429>": 152115,
390
+ "<state_42>": 151728,
391
+ "<state_430>": 152116,
392
+ "<state_431>": 152117,
393
+ "<state_432>": 152118,
394
+ "<state_433>": 152119,
395
+ "<state_434>": 152120,
396
+ "<state_435>": 152121,
397
+ "<state_436>": 152122,
398
+ "<state_437>": 152123,
399
+ "<state_438>": 152124,
400
+ "<state_439>": 152125,
401
+ "<state_43>": 151729,
402
+ "<state_440>": 152126,
403
+ "<state_441>": 152127,
404
+ "<state_442>": 152128,
405
+ "<state_443>": 152129,
406
+ "<state_444>": 152130,
407
+ "<state_445>": 152131,
408
+ "<state_446>": 152132,
409
+ "<state_447>": 152133,
410
+ "<state_448>": 152134,
411
+ "<state_449>": 152135,
412
+ "<state_44>": 151730,
413
+ "<state_450>": 152136,
414
+ "<state_451>": 152137,
415
+ "<state_452>": 152138,
416
+ "<state_453>": 152139,
417
+ "<state_454>": 152140,
418
+ "<state_455>": 152141,
419
+ "<state_456>": 152142,
420
+ "<state_457>": 152143,
421
+ "<state_458>": 152144,
422
+ "<state_459>": 152145,
423
+ "<state_45>": 151731,
424
+ "<state_460>": 152146,
425
+ "<state_461>": 152147,
426
+ "<state_462>": 152148,
427
+ "<state_463>": 152149,
428
+ "<state_464>": 152150,
429
+ "<state_465>": 152151,
430
+ "<state_466>": 152152,
431
+ "<state_467>": 152153,
432
+ "<state_468>": 152154,
433
+ "<state_469>": 152155,
434
+ "<state_46>": 151732,
435
+ "<state_470>": 152156,
436
+ "<state_471>": 152157,
437
+ "<state_472>": 152158,
438
+ "<state_473>": 152159,
439
+ "<state_474>": 152160,
440
+ "<state_475>": 152161,
441
+ "<state_476>": 152162,
442
+ "<state_477>": 152163,
443
+ "<state_478>": 152164,
444
+ "<state_479>": 152165,
445
+ "<state_47>": 151733,
446
+ "<state_480>": 152166,
447
+ "<state_481>": 152167,
448
+ "<state_482>": 152168,
449
+ "<state_483>": 152169,
450
+ "<state_484>": 152170,
451
+ "<state_485>": 152171,
452
+ "<state_486>": 152172,
453
+ "<state_487>": 152173,
454
+ "<state_488>": 152174,
455
+ "<state_489>": 152175,
456
+ "<state_48>": 151734,
457
+ "<state_490>": 152176,
458
+ "<state_491>": 152177,
459
+ "<state_492>": 152178,
460
+ "<state_493>": 152179,
461
+ "<state_494>": 152180,
462
+ "<state_495>": 152181,
463
+ "<state_496>": 152182,
464
+ "<state_497>": 152183,
465
+ "<state_498>": 152184,
466
+ "<state_499>": 152185,
467
+ "<state_49>": 151735,
468
+ "<state_4>": 151690,
469
+ "<state_500>": 152186,
470
+ "<state_501>": 152187,
471
+ "<state_502>": 152188,
472
+ "<state_503>": 152189,
473
+ "<state_504>": 152190,
474
+ "<state_505>": 152191,
475
+ "<state_506>": 152192,
476
+ "<state_507>": 152193,
477
+ "<state_508>": 152194,
478
+ "<state_509>": 152195,
479
+ "<state_50>": 151736,
480
+ "<state_510>": 152196,
481
+ "<state_511>": 152197,
482
+ "<state_512>": 152198,
483
+ "<state_51>": 151737,
484
+ "<state_52>": 151738,
485
+ "<state_53>": 151739,
486
+ "<state_54>": 151740,
487
+ "<state_55>": 151741,
488
+ "<state_56>": 151742,
489
+ "<state_57>": 151743,
490
+ "<state_58>": 151744,
491
+ "<state_59>": 151745,
492
+ "<state_5>": 151691,
493
+ "<state_60>": 151746,
494
+ "<state_61>": 151747,
495
+ "<state_62>": 151748,
496
+ "<state_63>": 151749,
497
+ "<state_64>": 151750,
498
+ "<state_65>": 151751,
499
+ "<state_66>": 151752,
500
+ "<state_67>": 151753,
501
+ "<state_68>": 151754,
502
+ "<state_69>": 151755,
503
+ "<state_6>": 151692,
504
+ "<state_70>": 151756,
505
+ "<state_71>": 151757,
506
+ "<state_72>": 151758,
507
+ "<state_73>": 151759,
508
+ "<state_74>": 151760,
509
+ "<state_75>": 151761,
510
+ "<state_76>": 151762,
511
+ "<state_77>": 151763,
512
+ "<state_78>": 151764,
513
+ "<state_79>": 151765,
514
+ "<state_7>": 151693,
515
+ "<state_80>": 151766,
516
+ "<state_81>": 151767,
517
+ "<state_82>": 151768,
518
+ "<state_83>": 151769,
519
+ "<state_84>": 151770,
520
+ "<state_85>": 151771,
521
+ "<state_86>": 151772,
522
+ "<state_87>": 151773,
523
+ "<state_88>": 151774,
524
+ "<state_89>": 151775,
525
+ "<state_8>": 151694,
526
+ "<state_90>": 151776,
527
+ "<state_91>": 151777,
528
+ "<state_92>": 151778,
529
+ "<state_93>": 151779,
530
+ "<state_94>": 151780,
531
+ "<state_95>": 151781,
532
+ "<state_96>": 151782,
533
+ "<state_97>": 151783,
534
+ "<state_98>": 151784,
535
+ "<state_99>": 151785,
536
+ "<state_9>": 151695,
537
+ "<tool_call>": 151657,
538
+ "<|box_end|>": 151649,
539
+ "<|box_start|>": 151648,
540
+ "<|endoftext|>": 151643,
541
+ "<|file_sep|>": 151664,
542
+ "<|fim_middle|>": 151660,
543
+ "<|fim_pad|>": 151662,
544
+ "<|fim_prefix|>": 151659,
545
+ "<|fim_suffix|>": 151661,
546
+ "<|im_end|>": 151645,
547
+ "<|im_start|>": 151644,
548
+ "<|image_pad|>": 151655,
549
+ "<|object_ref_end|>": 151647,
550
+ "<|object_ref_start|>": 151646,
551
+ "<|quad_end|>": 151651,
552
+ "<|quad_start|>": 151650,
553
+ "<|repo_name|>": 151663,
554
+ "<|video_pad|>": 151656,
555
+ "<|vision_end|>": 151653,
556
+ "<|vision_pad|>": 151654,
557
+ "<|vision_start|>": 151652
558
+ }
2B/internvla_2B_ckpt/internvl3_2B_embodied_0922/checkpoint-34750/unwrapped_model/chat_template.jinja ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if messages[0]['role'] == 'system' %}{{- '<|im_start|>system
2
+ ' + messages[0]['content'] + '<|im_end|>
3
+ ' }}{%- else %}{{- '<|im_start|>system
4
+ 你是书生·万象,英文名是InternVL,是由上海人工智能实验室、清华大学及多家合作单位联合开发的多模态大语言模型。<|im_end|>
5
+ ' }}{%- endif %}{% for message in messages %}{%- if messages[0]['role'] != 'system' or not loop.first %}{{'<|im_start|>' + message['role'] + '
6
+ '}}{% if message['content'] is string %}{{ message['content'] }}{% else %}{% for content in message['content'] %}{% if content['type'] == 'image' %}{{ '<image>
7
+ ' }}{% elif content['type'] == 'video' %}{{ '<video>
8
+ ' }}{% elif content['type'] == 'text' %}{{ content['text'] }}{% endif %}{% endfor %}{% endif %}{{'<|im_end|>
9
+ '}}{%- endif %}{% endfor %}{% if add_generation_prompt %}{{'<|im_start|>assistant
10
+ ' }}{% endif %}
2B/internvla_2B_ckpt/internvl3_2B_embodied_0922/checkpoint-34750/unwrapped_model/config.json ADDED
@@ -0,0 +1,109 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "InternVLChatModel"
4
+ ],
5
+ "auto_map": {
6
+ "AutoConfig": "configuration_internvl_chat.InternVLChatConfig",
7
+ "AutoModel": "modeling_internvl_chat.InternVLChatModel",
8
+ "AutoModelForCausalLM": "modeling_internvl_chat.InternVLChatModel"
9
+ },
10
+ "downsample_ratio": 0.5,
11
+ "dynamic_image_size": false,
12
+ "force_image_size": 224,
13
+ "hidden_size": 1536,
14
+ "image_fold": null,
15
+ "llm_config": {
16
+ "_name_or_path": "./pretrained/Qwen2.5-32B-Instruct",
17
+ "architectures": [
18
+ "Qwen2ForCausalLM"
19
+ ],
20
+ "attention_dropout": 0.0,
21
+ "attn_implementation": "flash_attention_2",
22
+ "bos_token_id": 151643,
23
+ "eos_token_id": 151643,
24
+ "hidden_act": "silu",
25
+ "hidden_size": 1536,
26
+ "initializer_range": 0.02,
27
+ "intermediate_size": 8960,
28
+ "max_position_embeddings": 32768,
29
+ "max_window_layers": 70,
30
+ "model_type": "qwen2",
31
+ "moe_config": null,
32
+ "num_attention_heads": 12,
33
+ "num_hidden_layers": 28,
34
+ "num_key_value_heads": 2,
35
+ "rms_norm_eps": 1e-06,
36
+ "rope_scaling": {
37
+ "factor": 2.0,
38
+ "rope_type": "dynamic",
39
+ "type": "dynamic"
40
+ },
41
+ "rope_theta": 1000000.0,
42
+ "sliding_window": null,
43
+ "torch_dtype": "bfloat16",
44
+ "use_bfloat16": true,
45
+ "use_cache": false,
46
+ "use_sliding_window": false,
47
+ "vocab_size": 152199
48
+ },
49
+ "max_dynamic_patch": 12,
50
+ "min_dynamic_patch": 1,
51
+ "model_type": "internvl_chat",
52
+ "pad2square": false,
53
+ "ps_version": "v2",
54
+ "select_layer": -2,
55
+ "system_message": null,
56
+ "template": "internvl2_5",
57
+ "tie_word_embeddings": false,
58
+ "torch_dtype": "bfloat16",
59
+ "transformers_version": null,
60
+ "use_backbone_lora": 0,
61
+ "use_llm_lora": 0,
62
+ "use_thumbnail": false,
63
+ "vision_config": {
64
+ "_name_or_path": "OpenGVLab/InternViT-6B-448px-V1-5",
65
+ "architectures": [
66
+ "InternVisionModel"
67
+ ],
68
+ "attention_dropout": 0.0,
69
+ "auto_map": {
70
+ "AutoConfig": "configuration_intern_vit.InternVisionConfig",
71
+ "AutoModel": "modeling_intern_vit.InternVisionModel"
72
+ },
73
+ "capacity_factor": 1.2,
74
+ "drop_path_rate": 0.0,
75
+ "dropout": 0.0,
76
+ "eval_capacity_factor": 1.4,
77
+ "hidden_act": "gelu",
78
+ "hidden_size": 1024,
79
+ "image_size": 224,
80
+ "initializer_factor": 0.1,
81
+ "initializer_range": 1e-10,
82
+ "intermediate_size": 4096,
83
+ "laux_allreduce": "all_nodes",
84
+ "layer_norm_eps": 1e-06,
85
+ "model_type": "intern_vit_6b",
86
+ "moe_coeff_ratio": 0.5,
87
+ "moe_intermediate_size": 768,
88
+ "moe_output_scale": 4.0,
89
+ "noisy_gate_policy": "RSample_before",
90
+ "norm_type": "layer_norm",
91
+ "num_attention_heads": 16,
92
+ "num_channels": 3,
93
+ "num_experts": 8,
94
+ "num_hidden_layers": 24,
95
+ "num_routed_experts": 4,
96
+ "num_shared_experts": 4,
97
+ "patch_size": 14,
98
+ "qk_normalization": false,
99
+ "qkv_bias": true,
100
+ "shared_expert_intermediate_size": 3072,
101
+ "torch_dtype": "bfloat16",
102
+ "use_bfloat16": true,
103
+ "use_flash_attn": true,
104
+ "use_moe": false,
105
+ "use_residual": true,
106
+ "use_rts": false,
107
+ "use_weighted_residual": false
108
+ }
109
+ }
2B/internvla_2B_ckpt/internvl3_2B_embodied_0922/checkpoint-34750/unwrapped_model/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
2B/internvla_2B_ckpt/internvl3_2B_embodied_0922/checkpoint-34750/unwrapped_model/projector.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f18c585d7091d807790959cbb0cd6c172db3acbf4bb425b36000603ac717c99
3
+ size 75659598
2B/internvla_2B_ckpt/internvl3_2B_embodied_0922/checkpoint-34750/unwrapped_model/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ded12a81e7439dd2a14f9a471dea6b1081d9556f2a7bc86e348903c8a0970cc4
3
+ size 4179794046
2B/internvla_2B_ckpt/internvl3_2B_embodied_0922/checkpoint-34750/unwrapped_model/special_tokens_map.json ADDED
@@ -0,0 +1,103 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ {
4
+ "content": "<soi>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false
9
+ },
10
+ {
11
+ "content": "<eoi>",
12
+ "lstrip": false,
13
+ "normalized": false,
14
+ "rstrip": false,
15
+ "single_word": false
16
+ },
17
+ {
18
+ "content": "<sot>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ {
25
+ "content": "<eot>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ },
31
+ {
32
+ "content": "<left_arm_soa>",
33
+ "lstrip": false,
34
+ "normalized": false,
35
+ "rstrip": false,
36
+ "single_word": false
37
+ },
38
+ {
39
+ "content": "<left_arm_eoa>",
40
+ "lstrip": false,
41
+ "normalized": false,
42
+ "rstrip": false,
43
+ "single_word": false
44
+ },
45
+ {
46
+ "content": "<right_arm_soa>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false
51
+ },
52
+ {
53
+ "content": "<right_arm_eoa>",
54
+ "lstrip": false,
55
+ "normalized": false,
56
+ "rstrip": false,
57
+ "single_word": false
58
+ },
59
+ {
60
+ "content": "<left_arm_sost>",
61
+ "lstrip": false,
62
+ "normalized": false,
63
+ "rstrip": false,
64
+ "single_word": false
65
+ },
66
+ {
67
+ "content": "<left_arm_eost>",
68
+ "lstrip": false,
69
+ "normalized": false,
70
+ "rstrip": false,
71
+ "single_word": false
72
+ },
73
+ {
74
+ "content": "<right_arm_sost>",
75
+ "lstrip": false,
76
+ "normalized": false,
77
+ "rstrip": false,
78
+ "single_word": false
79
+ },
80
+ {
81
+ "content": "<right_arm_eost>",
82
+ "lstrip": false,
83
+ "normalized": false,
84
+ "rstrip": false,
85
+ "single_word": false
86
+ }
87
+ ],
88
+ "bos_token": {
89
+ "content": "<s>",
90
+ "lstrip": false,
91
+ "normalized": false,
92
+ "rstrip": false,
93
+ "single_word": false
94
+ },
95
+ "eos_token": {
96
+ "content": "</s>",
97
+ "lstrip": false,
98
+ "normalized": false,
99
+ "rstrip": false,
100
+ "single_word": false
101
+ },
102
+ "pad_token": "<pad>"
103
+ }
2B/internvla_2B_ckpt/internvl3_2B_embodied_0922/checkpoint-34750/unwrapped_model/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2192afdcb300f6aedd8fba31c8e57a894f79180bc0e0b1ee0ad5d33b1267b0fb
3
+ size 11517337
2B/internvla_2B_ckpt/internvl3_2B_embodied_0922/checkpoint-34750/unwrapped_model/tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff
 
2B/internvla_2B_ckpt/internvl3_2B_embodied_0922/checkpoint-34750/unwrapped_model/vocab.json ADDED
The diff for this file is too large to render. See raw diff