Upload folder using huggingface_hub
Browse files- __pycache__/llm.cpython-310.pyc +0 -0
- infer_video.py +68 -24
- internvl3_2b_axmodel/qwen2_p128_l0_together.axmodel +2 -2
- internvl3_2b_axmodel/qwen2_p128_l10_together.axmodel +2 -2
- internvl3_2b_axmodel/qwen2_p128_l11_together.axmodel +2 -2
- internvl3_2b_axmodel/qwen2_p128_l12_together.axmodel +2 -2
- internvl3_2b_axmodel/qwen2_p128_l13_together.axmodel +2 -2
- internvl3_2b_axmodel/qwen2_p128_l14_together.axmodel +2 -2
- internvl3_2b_axmodel/qwen2_p128_l15_together.axmodel +2 -2
- internvl3_2b_axmodel/qwen2_p128_l16_together.axmodel +2 -2
- internvl3_2b_axmodel/qwen2_p128_l17_together.axmodel +2 -2
- internvl3_2b_axmodel/qwen2_p128_l18_together.axmodel +2 -2
- internvl3_2b_axmodel/qwen2_p128_l19_together.axmodel +2 -2
- internvl3_2b_axmodel/qwen2_p128_l1_together.axmodel +2 -2
- internvl3_2b_axmodel/qwen2_p128_l20_together.axmodel +2 -2
- internvl3_2b_axmodel/qwen2_p128_l21_together.axmodel +2 -2
- internvl3_2b_axmodel/qwen2_p128_l22_together.axmodel +2 -2
- internvl3_2b_axmodel/qwen2_p128_l23_together.axmodel +2 -2
- internvl3_2b_axmodel/qwen2_p128_l24_together.axmodel +2 -2
- internvl3_2b_axmodel/qwen2_p128_l25_together.axmodel +2 -2
- internvl3_2b_axmodel/qwen2_p128_l26_together.axmodel +2 -2
- internvl3_2b_axmodel/qwen2_p128_l27_together.axmodel +2 -2
- internvl3_2b_axmodel/qwen2_p128_l2_together.axmodel +2 -2
- internvl3_2b_axmodel/qwen2_p128_l3_together.axmodel +2 -2
- internvl3_2b_axmodel/qwen2_p128_l4_together.axmodel +2 -2
- internvl3_2b_axmodel/qwen2_p128_l5_together.axmodel +2 -2
- internvl3_2b_axmodel/qwen2_p128_l6_together.axmodel +2 -2
- internvl3_2b_axmodel/qwen2_p128_l7_together.axmodel +2 -2
- internvl3_2b_axmodel/qwen2_p128_l8_together.axmodel +2 -2
- internvl3_2b_axmodel/qwen2_p128_l9_together.axmodel +2 -2
- internvl3_2b_axmodel/qwen2_post.axmodel +1 -1
__pycache__/llm.cpython-310.pyc
ADDED
|
Binary file (18.8 kB). View file
|
|
|
infer_video.py
CHANGED
|
@@ -11,7 +11,8 @@ from tqdm import tqdm
|
|
| 11 |
import torchvision.transforms as T
|
| 12 |
from PIL import Image
|
| 13 |
import argparse
|
| 14 |
-
|
|
|
|
| 15 |
|
| 16 |
|
| 17 |
"""
|
|
@@ -70,8 +71,8 @@ def find_closest_aspect_ratio(aspect_ratio, target_ratios, width, height, image_
|
|
| 70 |
best_ratio = ratio
|
| 71 |
return best_ratio
|
| 72 |
|
| 73 |
-
def dynamic_preprocess(image, min_num=1, max_num=12, image_size=448, use_thumbnail=False):
|
| 74 |
-
orig_width,
|
| 75 |
aspect_ratio = orig_width / orig_height
|
| 76 |
|
| 77 |
# calculate the existing image aspect ratio
|
|
@@ -90,7 +91,8 @@ def dynamic_preprocess(image, min_num=1, max_num=12, image_size=448, use_thumbna
|
|
| 90 |
blocks = target_aspect_ratio[0] * target_aspect_ratio[1]
|
| 91 |
|
| 92 |
# resize the image
|
| 93 |
-
resized_img = image.resize((target_width, target_height))
|
|
|
|
| 94 |
processed_images = []
|
| 95 |
for i in range(blocks):
|
| 96 |
box = (
|
|
@@ -100,11 +102,13 @@ def dynamic_preprocess(image, min_num=1, max_num=12, image_size=448, use_thumbna
|
|
| 100 |
((i // (target_width // image_size)) + 1) * image_size
|
| 101 |
)
|
| 102 |
# split the image
|
| 103 |
-
split_img = resized_img.crop(box)
|
|
|
|
| 104 |
processed_images.append(split_img)
|
| 105 |
assert len(processed_images) == blocks
|
| 106 |
if use_thumbnail and len(processed_images) != 1:
|
| 107 |
-
thumbnail_img = image.resize((image_size, image_size))
|
|
|
|
| 108 |
processed_images.append(thumbnail_img)
|
| 109 |
return processed_images
|
| 110 |
|
|
@@ -165,23 +169,63 @@ def get_index(bound, fps, max_frame, first_idx=0, num_segments=32):
|
|
| 165 |
return frame_indices
|
| 166 |
|
| 167 |
|
| 168 |
-
def load_video(video_path, bound=None, input_size=448, max_num=1, num_segments=32):
|
| 169 |
-
|
| 170 |
-
|
| 171 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 172 |
|
| 173 |
-
pixel_values_list, num_patches_list = [], []
|
| 174 |
-
transform = build_transform(input_size=input_size)
|
| 175 |
frame_indices = get_index(bound, fps, max_frame, first_idx=0, num_segments=num_segments)
|
|
|
|
|
|
|
| 176 |
for frame_index in frame_indices:
|
| 177 |
-
|
| 178 |
-
|
| 179 |
-
|
| 180 |
-
|
| 181 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 182 |
pixel_values_list.append(pixel_values)
|
| 183 |
-
|
| 184 |
-
return pixel_values, num_patches_list
|
| 185 |
|
| 186 |
|
| 187 |
if __name__ == "__main__":
|
|
@@ -200,7 +244,6 @@ if __name__ == "__main__":
|
|
| 200 |
help="Your question that you want to ask the model.")
|
| 201 |
args = parser.parse_args()
|
| 202 |
|
| 203 |
-
|
| 204 |
hf_model_path = args.hf_model
|
| 205 |
axmodel_path = args.axmodel_path
|
| 206 |
vit_axmodel_path = args.vit_model
|
|
@@ -209,10 +252,10 @@ if __name__ == "__main__":
|
|
| 209 |
config = AutoConfig.from_pretrained(hf_model_path, trust_remote_code=True)
|
| 210 |
tokenizer = AutoTokenizer.from_pretrained(hf_model_path, trust_remote_code=True, use_fast=False)
|
| 211 |
# set the max number of tiles in `max_num`
|
| 212 |
-
|
| 213 |
# Frame1: <image>\nFrame2: <image>\n...\nFrame8: <image>\n{question}
|
| 214 |
|
| 215 |
-
pixel_values_list = [e[None, ...] for e in pixel_values]
|
| 216 |
if pixel_values_list is not None:
|
| 217 |
print(f"输入帧数: {len(pixel_values_list)}")
|
| 218 |
print("preprocess image done!")
|
|
@@ -221,7 +264,7 @@ if __name__ == "__main__":
|
|
| 221 |
vit_session = InferenceSession(vit_axmodel_path)
|
| 222 |
vit_output_list = []
|
| 223 |
for idx, pixel_values in enumerate(pixel_values_list):
|
| 224 |
-
vit_output = vit_session.run(None, {"image": pixel_values
|
| 225 |
vit_output_list.append(vit_output.copy()) # 避免 vit 输出结果使用同一块内存
|
| 226 |
|
| 227 |
print(f"vit_output.shape is {vit_output_list[0].shape}, vit feature extract done!")
|
|
@@ -288,7 +331,7 @@ if __name__ == "__main__":
|
|
| 288 |
]
|
| 289 |
print(f"slice_indexs is {slice_indexs}")
|
| 290 |
prefill_len = prefill_slice_len * slice_indexs[-1] if slice_indexs[-1] != 0 else prefill_slice_len # 这里的 128 就是 prefill_slice_len
|
| 291 |
-
|
| 292 |
if prefill_len > 0:
|
| 293 |
for slice_index in slice_indexs:
|
| 294 |
indices = np.array(
|
|
@@ -385,6 +428,7 @@ if __name__ == "__main__":
|
|
| 385 |
next_token = token_ids[start_indice]
|
| 386 |
indices = np.array([start_indice], np.uint32).reshape((1, 1))
|
| 387 |
data = embeds[next_token, :].reshape((1, 1, cfg.hidden_size)).astype(bfloat16)
|
|
|
|
| 388 |
for i in range(cfg.num_hidden_layers):
|
| 389 |
input_feed = {
|
| 390 |
"K_cache": k_caches[i],
|
|
|
|
| 11 |
import torchvision.transforms as T
|
| 12 |
from PIL import Image
|
| 13 |
import argparse
|
| 14 |
+
import cv2
|
| 15 |
+
# from decord import VideoReader, cpu
|
| 16 |
|
| 17 |
|
| 18 |
"""
|
|
|
|
| 71 |
best_ratio = ratio
|
| 72 |
return best_ratio
|
| 73 |
|
| 74 |
+
def dynamic_preprocess(image:np.array, min_num=1, max_num=12, image_size=448, use_thumbnail=False):
|
| 75 |
+
orig_height, orig_width, = image.shape[:2]
|
| 76 |
aspect_ratio = orig_width / orig_height
|
| 77 |
|
| 78 |
# calculate the existing image aspect ratio
|
|
|
|
| 91 |
blocks = target_aspect_ratio[0] * target_aspect_ratio[1]
|
| 92 |
|
| 93 |
# resize the image
|
| 94 |
+
# resized_img = image.resize((target_width, target_height))
|
| 95 |
+
resized_img = cv2.resize(image, (target_width, target_height))
|
| 96 |
processed_images = []
|
| 97 |
for i in range(blocks):
|
| 98 |
box = (
|
|
|
|
| 102 |
((i // (target_width // image_size)) + 1) * image_size
|
| 103 |
)
|
| 104 |
# split the image
|
| 105 |
+
# split_img = resized_img.crop(box)
|
| 106 |
+
split_img = resized_img[box[1]:box[3], box[0]:box[2]]
|
| 107 |
processed_images.append(split_img)
|
| 108 |
assert len(processed_images) == blocks
|
| 109 |
if use_thumbnail and len(processed_images) != 1:
|
| 110 |
+
# thumbnail_img = image.resize((image_size, image_size))
|
| 111 |
+
thumbnail_img = cv2.resize(image, (image_size, image_size))
|
| 112 |
processed_images.append(thumbnail_img)
|
| 113 |
return processed_images
|
| 114 |
|
|
|
|
| 169 |
return frame_indices
|
| 170 |
|
| 171 |
|
| 172 |
+
# def load_video(video_path, bound=None, input_size=448, max_num=1, num_segments=32):
|
| 173 |
+
# vr = VideoReader(video_path, ctx=cpu(0), num_threads=1)
|
| 174 |
+
# max_frame = len(vr) - 1
|
| 175 |
+
# fps = float(vr.get_avg_fps())
|
| 176 |
+
|
| 177 |
+
# pixel_values_list, num_patches_list = [], []
|
| 178 |
+
# transform = build_transform(input_size=input_size)
|
| 179 |
+
# frame_indices = get_index(bound, fps, max_frame, first_idx=0, num_segments=num_segments)
|
| 180 |
+
# for frame_index in frame_indices:
|
| 181 |
+
# img = Image.fromarray(vr[frame_index].asnumpy()).convert('RGB')
|
| 182 |
+
# img = dynamic_preprocess(img, image_size=input_size, use_thumbnail=True, max_num=max_num)
|
| 183 |
+
# pixel_values = [transform(tile) for tile in img]
|
| 184 |
+
# pixel_values = torch.stack(pixel_values)
|
| 185 |
+
# num_patches_list.append(pixel_values.shape[0])
|
| 186 |
+
# pixel_values_list.append(pixel_values)
|
| 187 |
+
# pixel_values = torch.cat(pixel_values_list)
|
| 188 |
+
# return pixel_values, num_patches_list
|
| 189 |
+
|
| 190 |
+
def pre_process(image, input_size=448, max_num=12):
|
| 191 |
+
images = dynamic_preprocess(image, image_size=input_size, use_thumbnail=True, max_num=max_num)
|
| 192 |
+
pixel_values = [img_preprocess(image, input_size) for image in images]
|
| 193 |
+
pixel_values = np.concatenate(pixel_values, axis=0)
|
| 194 |
+
return pixel_values
|
| 195 |
+
|
| 196 |
+
def img_preprocess(img, input_size):
|
| 197 |
+
IMAGENET_MEAN = np.array([0.485, 0.456, 0.406], dtype=np.float32)
|
| 198 |
+
IMAGENET_STD = np.array((0.229, 0.224, 0.225), dtype=np.float32)
|
| 199 |
+
img = cv2.cvtColor(img, cv2.COLOR_BGR2RGB)
|
| 200 |
+
img = cv2.resize(img, (input_size, input_size))
|
| 201 |
+
img = img.astype(np.float32) / 255.0
|
| 202 |
+
img = (img - IMAGENET_MEAN) / IMAGENET_STD
|
| 203 |
+
img = img.transpose(2, 0, 1).reshape(1, 3, input_size, input_size)
|
| 204 |
+
return img
|
| 205 |
+
def load_video_opencv(video_path, bound=None, num_segments=32):
|
| 206 |
+
cap = cv2.VideoCapture(video_path)
|
| 207 |
+
if not cap.isOpened():
|
| 208 |
+
raise IOError(f"Cannot open video: {video_path}")
|
| 209 |
+
|
| 210 |
+
fps = cap.get(cv2.CAP_PROP_FPS)
|
| 211 |
+
max_frame = int(cap.get(cv2.CAP_PROP_FRAME_COUNT)) - 1
|
| 212 |
|
|
|
|
|
|
|
| 213 |
frame_indices = get_index(bound, fps, max_frame, first_idx=0, num_segments=num_segments)
|
| 214 |
+
|
| 215 |
+
images_list = []
|
| 216 |
for frame_index in frame_indices:
|
| 217 |
+
cap.set(cv2.CAP_PROP_POS_FRAMES, frame_index)
|
| 218 |
+
ret, frame = cap.read()
|
| 219 |
+
if not ret:
|
| 220 |
+
print(f"⚠ Failed to read frame {frame_index}")
|
| 221 |
+
continue
|
| 222 |
+
images_list.append(frame)
|
| 223 |
+
pixel_values_list = []
|
| 224 |
+
cap.release()
|
| 225 |
+
for img in images_list:
|
| 226 |
+
pixel_values = pre_process(img, input_size=448, max_num=1)
|
| 227 |
pixel_values_list.append(pixel_values)
|
| 228 |
+
return pixel_values_list
|
|
|
|
| 229 |
|
| 230 |
|
| 231 |
if __name__ == "__main__":
|
|
|
|
| 244 |
help="Your question that you want to ask the model.")
|
| 245 |
args = parser.parse_args()
|
| 246 |
|
|
|
|
| 247 |
hf_model_path = args.hf_model
|
| 248 |
axmodel_path = args.axmodel_path
|
| 249 |
vit_axmodel_path = args.vit_model
|
|
|
|
| 252 |
config = AutoConfig.from_pretrained(hf_model_path, trust_remote_code=True)
|
| 253 |
tokenizer = AutoTokenizer.from_pretrained(hf_model_path, trust_remote_code=True, use_fast=False)
|
| 254 |
# set the max number of tiles in `max_num`
|
| 255 |
+
pixel_values_list = load_video_opencv(video_path, num_segments=8)
|
| 256 |
# Frame1: <image>\nFrame2: <image>\n...\nFrame8: <image>\n{question}
|
| 257 |
|
| 258 |
+
# pixel_values_list = [e[None, ...] for e in pixel_values]
|
| 259 |
if pixel_values_list is not None:
|
| 260 |
print(f"输入帧数: {len(pixel_values_list)}")
|
| 261 |
print("preprocess image done!")
|
|
|
|
| 264 |
vit_session = InferenceSession(vit_axmodel_path)
|
| 265 |
vit_output_list = []
|
| 266 |
for idx, pixel_values in enumerate(pixel_values_list):
|
| 267 |
+
vit_output = vit_session.run(None, {"image": pixel_values})[0]
|
| 268 |
vit_output_list.append(vit_output.copy()) # 避免 vit 输出结果使用同一块内存
|
| 269 |
|
| 270 |
print(f"vit_output.shape is {vit_output_list[0].shape}, vit feature extract done!")
|
|
|
|
| 331 |
]
|
| 332 |
print(f"slice_indexs is {slice_indexs}")
|
| 333 |
prefill_len = prefill_slice_len * slice_indexs[-1] if slice_indexs[-1] != 0 else prefill_slice_len # 这里的 128 就是 prefill_slice_len
|
| 334 |
+
# prefill_len = token_len if token_len < lastN else lastN
|
| 335 |
if prefill_len > 0:
|
| 336 |
for slice_index in slice_indexs:
|
| 337 |
indices = np.array(
|
|
|
|
| 428 |
next_token = token_ids[start_indice]
|
| 429 |
indices = np.array([start_indice], np.uint32).reshape((1, 1))
|
| 430 |
data = embeds[next_token, :].reshape((1, 1, cfg.hidden_size)).astype(bfloat16)
|
| 431 |
+
#print(data.shape)
|
| 432 |
for i in range(cfg.num_hidden_layers):
|
| 433 |
input_feed = {
|
| 434 |
"K_cache": k_caches[i],
|
internvl3_2b_axmodel/qwen2_p128_l0_together.axmodel
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b1dd071450c814b4ff63d1c8625ea462a12e6c1f0e2d9450ab79cf93cdf70ae7
|
| 3 |
+
size 81360616
|
internvl3_2b_axmodel/qwen2_p128_l10_together.axmodel
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2fff56a436e3b92ff19e8558a4ef0505acf7166914edf359262f39200f24cecd
|
| 3 |
+
size 81360616
|
internvl3_2b_axmodel/qwen2_p128_l11_together.axmodel
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:146176cd6941ca5256c8982bcb3099f84c99c599474704589275df6998400549
|
| 3 |
+
size 81360616
|
internvl3_2b_axmodel/qwen2_p128_l12_together.axmodel
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:137d2f2df726ebe1c7bc3a1c7ae7b3b5d2bc3a715537a95aa309b3412c4d4e71
|
| 3 |
+
size 81360616
|
internvl3_2b_axmodel/qwen2_p128_l13_together.axmodel
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2592a19121c884a18e6b05a3b030b26e95f353121d32adffefa7faa5c486eca2
|
| 3 |
+
size 81360616
|
internvl3_2b_axmodel/qwen2_p128_l14_together.axmodel
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e721a2515cc165434190010e03d14688100a9e9be933eff15f8dbe54efabc947
|
| 3 |
+
size 81360616
|
internvl3_2b_axmodel/qwen2_p128_l15_together.axmodel
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:24a626bbd133f62b3f3ce23c4aa4cd5277a57c9783129b31ea2c372494f2a9ed
|
| 3 |
+
size 81360616
|
internvl3_2b_axmodel/qwen2_p128_l16_together.axmodel
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:593e7816e379c5f8e07b16095aab81a10c7be5ee8314186e25d430b30e4da38b
|
| 3 |
+
size 81360616
|
internvl3_2b_axmodel/qwen2_p128_l17_together.axmodel
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a859330bdb75de08a56aabacf312aa97ac4abc3862103d3bde312b7eaee33123
|
| 3 |
+
size 81360616
|
internvl3_2b_axmodel/qwen2_p128_l18_together.axmodel
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7d5b40d734848f644249a560a82655429cdbfcbed99cfe97d280aa41e27a39fb
|
| 3 |
+
size 81360616
|
internvl3_2b_axmodel/qwen2_p128_l19_together.axmodel
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c185d30a7c2a824f06756bad0e70eb9f50f73bca0f4817a69e1a393928a80944
|
| 3 |
+
size 81360616
|
internvl3_2b_axmodel/qwen2_p128_l1_together.axmodel
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f99d6fa520114d30900e7659e28d1c5dccd257faad27d439734e4be755fa3300
|
| 3 |
+
size 81360616
|
internvl3_2b_axmodel/qwen2_p128_l20_together.axmodel
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c763458a03e27b42b99aa785997191ece692c2e557b29481ca2b580cc6b8fab0
|
| 3 |
+
size 81360616
|
internvl3_2b_axmodel/qwen2_p128_l21_together.axmodel
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:de52167d4fd5bc023884f76c3858acfb0062df830eb7e28f83871ae2c687026a
|
| 3 |
+
size 81360616
|
internvl3_2b_axmodel/qwen2_p128_l22_together.axmodel
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4643590e0d72b5d039eeed0ec0ce6b855b002b5e43f9a8a6cbe6a6c2cd57159d
|
| 3 |
+
size 81360616
|
internvl3_2b_axmodel/qwen2_p128_l23_together.axmodel
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3bcaf68c2cebbe8366868371dc0603bca4f9850bff5b897b3b9a712d391b775f
|
| 3 |
+
size 81360616
|
internvl3_2b_axmodel/qwen2_p128_l24_together.axmodel
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0098dac4e276ef6490a883f6348fce60a7f190ce97a93c7831f411725cff04ab
|
| 3 |
+
size 81360616
|
internvl3_2b_axmodel/qwen2_p128_l25_together.axmodel
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ff65751a675eb14684357f327b11cccb331b3306edd0a2391f6d88a56cbe625c
|
| 3 |
+
size 81360616
|
internvl3_2b_axmodel/qwen2_p128_l26_together.axmodel
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:15075f5c69f967e5adf64c795436967075feabc5b31ffd50f07dbaaff85f2262
|
| 3 |
+
size 81360616
|
internvl3_2b_axmodel/qwen2_p128_l27_together.axmodel
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:943ebf93b6ef43f6884f1cfff8236e5f869bf0dae32fb1f34c4e7fab919bf734
|
| 3 |
+
size 81360616
|
internvl3_2b_axmodel/qwen2_p128_l2_together.axmodel
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:df93cf776113bed20d6779dd895e2e61d13c9928aa073e0a595e65c2584bc7df
|
| 3 |
+
size 81364200
|
internvl3_2b_axmodel/qwen2_p128_l3_together.axmodel
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:58beca6260114f56c9aa5a6696725161588fa3d2c7ab6324d00715daaa6afbe1
|
| 3 |
+
size 81360616
|
internvl3_2b_axmodel/qwen2_p128_l4_together.axmodel
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dc0da75d869839eeba66a3dd5863add115c2a97de83703ae807918feaec1ffa0
|
| 3 |
+
size 81360616
|
internvl3_2b_axmodel/qwen2_p128_l5_together.axmodel
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d9c4a078e5288d4287607c0c9cccb7f1259a89d38de8b60118505dbd535af4d2
|
| 3 |
+
size 81360616
|
internvl3_2b_axmodel/qwen2_p128_l6_together.axmodel
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3e74bdc2623b3eb8876c68c982d6e3d48ae804b79412dd7304994d65ae650a89
|
| 3 |
+
size 81360616
|
internvl3_2b_axmodel/qwen2_p128_l7_together.axmodel
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:342fed4e97d800cc9824379bac2792c4a51f824a2b2236c50dfd476689e1b553
|
| 3 |
+
size 81360616
|
internvl3_2b_axmodel/qwen2_p128_l8_together.axmodel
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ec4266b62c2040b7a0c689e3cc0bfec1b6634875b0b8681bd9f7e2ce3975e654
|
| 3 |
+
size 81360616
|
internvl3_2b_axmodel/qwen2_p128_l9_together.axmodel
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e6ee2e4269d546948822160db37943cc5f6d2d444823c502465c8ed445f06f81
|
| 3 |
+
size 81360616
|
internvl3_2b_axmodel/qwen2_post.axmodel
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 254396183
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6339b7f4ec86a8f2f1a496431f5027eaf30434b2026b3376ff34391d3c8d73b9
|
| 3 |
size 254396183
|