id
int64
0
190k
prompt
stringlengths
21
13.4M
docstring
stringlengths
1
12k
179,576
import argparse import asyncio import dataclasses from enum import Enum, auto import json import logging import time from typing import List, Union import threading from fastapi import FastAPI, Request from fastapi.responses import StreamingResponse import numpy as np import requests import uvicorn from videollava.constants import CONTROLLER_HEART_BEAT_EXPIRATION from videollava.utils import build_logger, server_error_msg async def get_worker_address(request: Request): data = await request.json() addr = controller.get_worker_address(data["model"]) return {"address": addr}
null
179,577
import argparse import asyncio import dataclasses from enum import Enum, auto import json import logging import time from typing import List, Union import threading from fastapi import FastAPI, Request from fastapi.responses import StreamingResponse import numpy as np import requests import uvicorn from videollava.constants import CONTROLLER_HEART_BEAT_EXPIRATION from videollava.utils import build_logger, server_error_msg async def receive_heart_beat(request: Request): data = await request.json() exist = controller.receive_heart_beat( data["worker_name"], data["queue_length"]) return {"exist": exist}
null
179,578
import argparse import asyncio import dataclasses from enum import Enum, auto import json import logging import time from typing import List, Union import threading from fastapi import FastAPI, Request from fastapi.responses import StreamingResponse import numpy as np import requests import uvicorn from videollava.constants import CONTROLLER_HEART_BEAT_EXPIRATION from videollava.utils import build_logger, server_error_msg async def worker_api_generate_stream(request: Request): params = await request.json() generator = controller.worker_api_generate_stream(params) return StreamingResponse(generator)
null
179,579
import argparse import asyncio import dataclasses from enum import Enum, auto import json import logging import time from typing import List, Union import threading from fastapi import FastAPI, Request from fastapi.responses import StreamingResponse import numpy as np import requests import uvicorn from videollava.constants import CONTROLLER_HEART_BEAT_EXPIRATION from videollava.utils import build_logger, server_error_msg async def worker_api_get_status(request: Request): return controller.worker_api_get_status()
null
179,580
import shutil import subprocess import torch import gradio as gr from fastapi import FastAPI import os from PIL import Image import tempfile from decord import VideoReader, cpu from transformers import TextStreamer from videollava.constants import DEFAULT_IMAGE_TOKEN from videollava.conversation import conv_templates, SeparatorStyle, Conversation from videollava.serve.gradio_utils import Chat, tos_markdown, learn_more_markdown, title_markdown, block_css def save_image_to_local(image): def save_video_to_local(video_path): conv_mode = "llava_v1" device = 'cuda' dtype = torch.float16 handler = Chat(model_path, conv_mode=conv_mode, load_8bit=load_8bit, load_4bit=load_8bit, device=device, cache_dir=cache_dir) if not os.path.exists("temp"): os.makedirs("temp") with gr.Blocks(title='Video-LLaVA🚀', theme=gr.themes.Default(), css=block_css) as demo: gr.Markdown(title_markdown) state = gr.State() state_ = gr.State() first_run = gr.State() images_tensor = gr.State() with gr.Row(): with gr.Column(scale=3): image1 = gr.Image(label="Input Image", type="filepath") video = gr.Video(label="Input Video") cur_dir = os.path.dirname(os.path.abspath(__file__)) gr.Examples( examples=[ [ f"{cur_dir}/examples/extreme_ironing.jpg", "What is unusual about this image?", ], [ f"{cur_dir}/examples/waterview.jpg", "What are the things I should be cautious about when I visit here?", ], [ f"{cur_dir}/examples/desert.jpg", "If there are factual errors in the questions, point it out; if not, proceed answering the question. What’s happening in the desert?", ], ], inputs=[image1, textbox], ) with gr.Column(scale=7): chatbot = gr.Chatbot(label="Video-LLaVA", bubble_full_width=True).style(height=750) with gr.Row(): with gr.Column(scale=8): textbox.render() with gr.Column(scale=1, min_width=50): submit_btn = gr.Button( value="Send", variant="primary", interactive=True ) with gr.Row(elem_id="buttons") as button_row: upvote_btn = gr.Button(value="👍 Upvote", interactive=True) downvote_btn = gr.Button(value="👎 Downvote", interactive=True) flag_btn = gr.Button(value="⚠️ Flag", interactive=True) # stop_btn = gr.Button(value="⏹️ Stop Generation", interactive=False) regenerate_btn = gr.Button(value="🔄 Regenerate", interactive=True) clear_btn = gr.Button(value="🗑️ Clear history", interactive=True) with gr.Row(): gr.Examples( examples=[ [ f"{cur_dir}/examples/sample_img_22.png", f"{cur_dir}/examples/sample_demo_22.mp4", "Are the instruments in the pictures used in the video?", ], [ f"{cur_dir}/examples/sample_img_13.png", f"{cur_dir}/examples/sample_demo_13.mp4", "Does the flag in the image appear in the video?", ], [ f"{cur_dir}/examples/sample_img_8.png", f"{cur_dir}/examples/sample_demo_8.mp4", "Are the image and the video depicting the same place?", ], ], inputs=[image1, video, textbox], ) gr.Examples( examples=[ [ f"{cur_dir}/examples/sample_demo_1.mp4", "Why is this video funny?", ], [ f"{cur_dir}/examples/sample_demo_3.mp4", "Can you identify any safety hazards in this video?" ], [ f"{cur_dir}/examples/sample_demo_9.mp4", "Describe the video.", ], [ f"{cur_dir}/examples/sample_demo_22.mp4", "Describe the activity in the video.", ], ], inputs=[video, textbox], ) gr.Markdown(tos_markdown) gr.Markdown(learn_more_markdown) submit_btn.click(generate, [image1, video, textbox, first_run, state, state_, images_tensor], [state, state_, chatbot, first_run, textbox, images_tensor, image1, video]) regenerate_btn.click(regenerate, [state, state_], [state, state_, chatbot, first_run]).then( generate, [image1, video, textbox, first_run, state, state_, images_tensor], [state, state_, chatbot, first_run, textbox, images_tensor, image1, video]) clear_btn.click(clear_history, [state, state_], [image1, video, textbox, first_run, state, state_, chatbot, images_tensor]) DEFAULT_IMAGE_TOKEN = "<image>" class Conversation: def get_prompt(self): def append_message(self, role, message): def get_images(self, return_pil=False): def expand2square(pil_img, background_color=(122, 116, 104)): def to_gradio_chatbot(self): def copy(self): def dict(self): conv_templates = { "default": conv_vicuna_v0, "v0": conv_vicuna_v0, "v1": conv_vicuna_v1, "vicuna_v1": conv_vicuna_v1, "llama_2": conv_llama_2, "plain": conv_llava_plain, "v0_plain": conv_llava_plain, "llava_v0": conv_llava_v0, "v0_mmtag": conv_llava_v0_mmtag, "llava_v1": conv_llava_v1, "v1_mmtag": conv_llava_v1_mmtag, "llava_llama_2": conv_llava_llama_2, "mpt": conv_mpt, } def generate(image1, video, textbox_in, first_run, state, state_, images_tensor): flag = 1 if not textbox_in: if len(state_.messages) > 0: textbox_in = state_.messages[-1][1] state_.messages.pop(-1) flag = 0 else: return "Please enter instruction" image1 = image1 if image1 else "none" video = video if video else "none" # assert not (os.path.exists(image1) and os.path.exists(video)) if type(state) is not Conversation: state = conv_templates[conv_mode].copy() state_ = conv_templates[conv_mode].copy() images_tensor = [] first_run = False if len(state.messages) > 0 else True text_en_in = textbox_in.replace("picture", "image") # images_tensor = [[], []] image_processor = handler.image_processor if os.path.exists(image1) and not os.path.exists(video): tensor = image_processor.preprocess(image1, return_tensors='pt')['pixel_values'][0] # print(tensor.shape) tensor = tensor.to(handler.model.device, dtype=dtype) images_tensor.append(tensor) video_processor = handler.video_processor if not os.path.exists(image1) and os.path.exists(video): tensor = video_processor(video, return_tensors='pt')['pixel_values'][0] # print(tensor.shape) tensor = tensor.to(handler.model.device, dtype=dtype) images_tensor.append(tensor) if os.path.exists(image1) and os.path.exists(video): tensor = video_processor(video, return_tensors='pt')['pixel_values'][0] # print(tensor.shape) tensor = tensor.to(handler.model.device, dtype=dtype) images_tensor.append(tensor) tensor = image_processor.preprocess(image1, return_tensors='pt')['pixel_values'][0] # print(tensor.shape) tensor = tensor.to(handler.model.device, dtype=dtype) images_tensor.append(tensor) if os.path.exists(image1) and not os.path.exists(video): text_en_in = DEFAULT_IMAGE_TOKEN + '\n' + text_en_in if not os.path.exists(image1) and os.path.exists(video): text_en_in = ''.join([DEFAULT_IMAGE_TOKEN] * handler.model.get_video_tower().config.num_frames) + '\n' + text_en_in if os.path.exists(image1) and os.path.exists(video): text_en_in = ''.join([DEFAULT_IMAGE_TOKEN] * handler.model.get_video_tower().config.num_frames) + '\n' + text_en_in + '\n' + DEFAULT_IMAGE_TOKEN # print(text_en_in) text_en_out, state_ = handler.generate(images_tensor, text_en_in, first_run=first_run, state=state_) state_.messages[-1] = (state_.roles[1], text_en_out) text_en_out = text_en_out.split('#')[0] textbox_out = text_en_out show_images = "" if os.path.exists(image1): filename = save_image_to_local(image1) show_images += f'<img src="./file={filename}" style="display: inline-block;width: 250px;max-height: 400px;">' if os.path.exists(video): filename = save_video_to_local(video) show_images += f'<video controls playsinline width="500" style="display: inline-block;" src="./file={filename}"></video>' if flag: state.append_message(state.roles[0], textbox_in + "\n" + show_images) state.append_message(state.roles[1], textbox_out) return (state, state_, state.to_gradio_chatbot(), False, gr.update(value=None, interactive=True), images_tensor, gr.update(value=image1 if os.path.exists(image1) else None, interactive=True), gr.update(value=video if os.path.exists(video) else None, interactive=True))
null
179,581
import shutil import subprocess import torch import gradio as gr from fastapi import FastAPI import os from PIL import Image import tempfile from decord import VideoReader, cpu from transformers import TextStreamer from videollava.constants import DEFAULT_IMAGE_TOKEN from videollava.conversation import conv_templates, SeparatorStyle, Conversation from videollava.serve.gradio_utils import Chat, tos_markdown, learn_more_markdown, title_markdown, block_css def regenerate(state, state_): state.messages.pop(-1) state_.messages.pop(-1) if len(state.messages) > 0: return state, state_, state.to_gradio_chatbot(), False return (state, state_, state.to_gradio_chatbot(), True)
null
179,582
import shutil import subprocess import torch import gradio as gr from fastapi import FastAPI import os from PIL import Image import tempfile from decord import VideoReader, cpu from transformers import TextStreamer from videollava.constants import DEFAULT_IMAGE_TOKEN from videollava.conversation import conv_templates, SeparatorStyle, Conversation from videollava.serve.gradio_utils import Chat, tos_markdown, learn_more_markdown, title_markdown, block_css conv_mode = "llava_v1" with gr.Blocks(title='Video-LLaVA🚀', theme=gr.themes.Default(), css=block_css) as demo: gr.Markdown(title_markdown) state = gr.State() state_ = gr.State() first_run = gr.State() images_tensor = gr.State() with gr.Row(): with gr.Column(scale=3): image1 = gr.Image(label="Input Image", type="filepath") video = gr.Video(label="Input Video") cur_dir = os.path.dirname(os.path.abspath(__file__)) gr.Examples( examples=[ [ f"{cur_dir}/examples/extreme_ironing.jpg", "What is unusual about this image?", ], [ f"{cur_dir}/examples/waterview.jpg", "What are the things I should be cautious about when I visit here?", ], [ f"{cur_dir}/examples/desert.jpg", "If there are factual errors in the questions, point it out; if not, proceed answering the question. What’s happening in the desert?", ], ], inputs=[image1, textbox], ) with gr.Column(scale=7): chatbot = gr.Chatbot(label="Video-LLaVA", bubble_full_width=True).style(height=750) with gr.Row(): with gr.Column(scale=8): textbox.render() with gr.Column(scale=1, min_width=50): submit_btn = gr.Button( value="Send", variant="primary", interactive=True ) with gr.Row(elem_id="buttons") as button_row: upvote_btn = gr.Button(value="👍 Upvote", interactive=True) downvote_btn = gr.Button(value="👎 Downvote", interactive=True) flag_btn = gr.Button(value="⚠️ Flag", interactive=True) # stop_btn = gr.Button(value="⏹️ Stop Generation", interactive=False) regenerate_btn = gr.Button(value="🔄 Regenerate", interactive=True) clear_btn = gr.Button(value="🗑️ Clear history", interactive=True) with gr.Row(): gr.Examples( examples=[ [ f"{cur_dir}/examples/sample_img_22.png", f"{cur_dir}/examples/sample_demo_22.mp4", "Are the instruments in the pictures used in the video?", ], [ f"{cur_dir}/examples/sample_img_13.png", f"{cur_dir}/examples/sample_demo_13.mp4", "Does the flag in the image appear in the video?", ], [ f"{cur_dir}/examples/sample_img_8.png", f"{cur_dir}/examples/sample_demo_8.mp4", "Are the image and the video depicting the same place?", ], ], inputs=[image1, video, textbox], ) gr.Examples( examples=[ [ f"{cur_dir}/examples/sample_demo_1.mp4", "Why is this video funny?", ], [ f"{cur_dir}/examples/sample_demo_3.mp4", "Can you identify any safety hazards in this video?" ], [ f"{cur_dir}/examples/sample_demo_9.mp4", "Describe the video.", ], [ f"{cur_dir}/examples/sample_demo_22.mp4", "Describe the activity in the video.", ], ], inputs=[video, textbox], ) gr.Markdown(tos_markdown) gr.Markdown(learn_more_markdown) submit_btn.click(generate, [image1, video, textbox, first_run, state, state_, images_tensor], [state, state_, chatbot, first_run, textbox, images_tensor, image1, video]) regenerate_btn.click(regenerate, [state, state_], [state, state_, chatbot, first_run]).then( generate, [image1, video, textbox, first_run, state, state_, images_tensor], [state, state_, chatbot, first_run, textbox, images_tensor, image1, video]) clear_btn.click(clear_history, [state, state_], [image1, video, textbox, first_run, state, state_, chatbot, images_tensor]) conv_templates = { "default": conv_vicuna_v0, "v0": conv_vicuna_v0, "v1": conv_vicuna_v1, "vicuna_v1": conv_vicuna_v1, "llama_2": conv_llama_2, "plain": conv_llava_plain, "v0_plain": conv_llava_plain, "llava_v0": conv_llava_v0, "v0_mmtag": conv_llava_v0_mmtag, "llava_v1": conv_llava_v1, "v1_mmtag": conv_llava_v1_mmtag, "llava_llama_2": conv_llava_llama_2, "mpt": conv_mpt, } def clear_history(state, state_): state = conv_templates[conv_mode].copy() state_ = conv_templates[conv_mode].copy() return (gr.update(value=None, interactive=True), gr.update(value=None, interactive=True), \ gr.update(value=None, interactive=True), \ True, state, state_, state.to_gradio_chatbot(), [])
null
179,583
import os import copy import random from dataclasses import dataclass, field import json import logging import pathlib from typing import Dict, Optional, Sequence, List import torch import transformers from videollava.constants import IGNORE_INDEX, IMAGE_TOKEN_INDEX, DEFAULT_IMAGE_TOKEN, DEFAULT_IM_START_TOKEN, \ DEFAULT_IM_END_TOKEN, DEFAULT_VIDEO_TOKEN, DEFAULT_VID_START_TOKEN, DEFAULT_VID_END_TOKEN, MAX_IMAGE_LENGTH, \ MAX_VIDEO_LENGTH from torch.utils.data import Dataset from videollava.train.llava_trainer import LLaVATrainer from videollava import conversation as conversation_lib from videollava.model import * from videollava.mm_utils import tokenizer_image_token from PIL import Image from videollava.utils import order_pick_k class DataArguments: lazy_preprocess: bool = False is_multimodal: bool = False image_aspect_ratio: str = 'square' # =================================================================== data_path: Optional[List[str]] = field(default=None, metadata={"help": "Path to the training data."}) image_folder: Optional[str] = field(default=None) video_folder: Optional[str] = field(default=None) num_frames: int = 8 # =================================================================== DEFAULT_IMAGE_TOKEN = "<image>" DEFAULT_IM_START_TOKEN = "<im_start>" DEFAULT_IM_END_TOKEN = "<im_end>" DEFAULT_VIDEO_TOKEN = "<video>" DEFAULT_VID_START_TOKEN = "<vid_start>" DEFAULT_VID_END_TOKEN = "<vid_end>" MAX_IMAGE_LENGTH = 16 MAX_VIDEO_LENGTH = 1 def preprocess_multimodal( sources: Sequence[str], data_args: DataArguments ) -> Dict: is_multimodal = data_args.is_multimodal if not is_multimodal: return sources for source in sources: for sentence in source: # ====================================================================================================== if sentence['value'].startswith(DEFAULT_IMAGE_TOKEN) or sentence['value'].startswith(DEFAULT_VIDEO_TOKEN): # run with multi-im, multi-vid, multi-im & multi-vid # <video><video><image><image>\nxxxxxxxxxxxxx # must <video> first # <image>\nxxxxxxxxxxxxx -> <image>\nxxxxxxxxxxxxx # <video>\nxxxxxxxxxxxxx -> <video>\nxxxxxxxxxxxxx if "mmtag" in conversation_lib.default_conversation.version: sentence['value'] = sentence['value'].replace(DEFAULT_IMAGE_TOKEN, '<Image>' + DEFAULT_IMAGE_TOKEN + '</Image>') IMAGE_TOKEN_NUM = sentence['value'].count(DEFAULT_IMAGE_TOKEN) if IMAGE_TOKEN_NUM > MAX_IMAGE_LENGTH: sentence['value'] = sentence['value'].replace(DEFAULT_IMAGE_TOKEN * IMAGE_TOKEN_NUM, DEFAULT_IMAGE_TOKEN * MAX_IMAGE_LENGTH).strip() VIDEO_TOKEN_NUM = sentence['value'].count(DEFAULT_VIDEO_TOKEN) if VIDEO_TOKEN_NUM > MAX_VIDEO_LENGTH: raise ValueError(f"{sentence['value']}") sentence['value'] = sentence['value'].replace(DEFAULT_VIDEO_TOKEN * VIDEO_TOKEN_NUM, DEFAULT_VIDEO_TOKEN * MAX_VIDEO_LENGTH).strip() # a <video> is treated as `num_frames * <image>` replace_token, vid_replace_token = DEFAULT_IMAGE_TOKEN, DEFAULT_IMAGE_TOKEN * data_args.num_frames if data_args.mm_use_im_start_end: replace_token = DEFAULT_IM_START_TOKEN + replace_token + DEFAULT_IM_END_TOKEN vid_replace_token = DEFAULT_VID_START_TOKEN + vid_replace_token + DEFAULT_VID_END_TOKEN # <video><video><image><image>\nxxxxxxxxxxxxx -> `num_frames*<image>``num_frames*<image>`<image><image>\nxxxxxxxxxxxxx # <video>\nxxxxxxxxxxxxx -> `num_frames*<image>`\nxxxxxxxxxxxxx # print('before replace_token:', [sentence['value']]) sentence["value"] = sentence["value"].replace(DEFAULT_IMAGE_TOKEN, replace_token) sentence['value'] = sentence['value'].replace(DEFAULT_VIDEO_TOKEN, vid_replace_token) # print('after replace_token:', [sentence['value']]) # ====================================================================================================== return sources
null
179,584
import os import copy import random from dataclasses import dataclass, field import json import logging import pathlib from typing import Dict, Optional, Sequence, List import torch import transformers from videollava.constants import IGNORE_INDEX, IMAGE_TOKEN_INDEX, DEFAULT_IMAGE_TOKEN, DEFAULT_IM_START_TOKEN, \ DEFAULT_IM_END_TOKEN, DEFAULT_VIDEO_TOKEN, DEFAULT_VID_START_TOKEN, DEFAULT_VID_END_TOKEN, MAX_IMAGE_LENGTH, \ MAX_VIDEO_LENGTH from torch.utils.data import Dataset from videollava.train.llava_trainer import LLaVATrainer from videollava import conversation as conversation_lib from videollava.model import * from videollava.mm_utils import tokenizer_image_token from PIL import Image from videollava.utils import order_pick_k def expand2square(pil_img, background_color): width, height = pil_img.size if width == height: return pil_img elif width > height: result = Image.new(pil_img.mode, (width, width), background_color) result.paste(pil_img, (0, (width - height) // 2)) return result else: result = Image.new(pil_img.mode, (height, height), background_color) result.paste(pil_img, ((height - width) // 2, 0)) return result
null
179,585
import os import copy import random from dataclasses import dataclass, field import json import logging import pathlib from typing import Dict, Optional, Sequence, List import torch import transformers from videollava.constants import IGNORE_INDEX, IMAGE_TOKEN_INDEX, DEFAULT_IMAGE_TOKEN, DEFAULT_IM_START_TOKEN, \ DEFAULT_IM_END_TOKEN, DEFAULT_VIDEO_TOKEN, DEFAULT_VID_START_TOKEN, DEFAULT_VID_END_TOKEN, MAX_IMAGE_LENGTH, \ MAX_VIDEO_LENGTH from torch.utils.data import Dataset from videollava.train.llava_trainer import LLaVATrainer from videollava import conversation as conversation_lib from videollava.model import * from videollava.mm_utils import tokenizer_image_token from PIL import Image from videollava.utils import order_pick_k local_rank = None def rank0_print(*args): if local_rank == 0: print(*args) class ModelArguments: model_name_or_path: Optional[str] = field(default="facebook/opt-125m") version: Optional[str] = field(default="v0") freeze_backbone: bool = field(default=False) tune_mm_mlp_adapter: bool = field(default=False) vision_tower: Optional[str] = field(default=None) mm_vision_select_layer: Optional[int] = field(default=-1) # default to the last layer pretrain_mm_mlp_adapter: Optional[str] = field(default=None) mm_projector_type: Optional[str] = field(default='linear') mm_use_im_start_end: bool = field(default=False) mm_use_im_patch_token: bool = field(default=True) mm_vision_select_feature: Optional[str] = field(default="patch") # =================================================================== image_tower: Optional[str] = field(default=None) video_tower: Optional[str] = field(default=None) # =================================================================== class DataArguments: lazy_preprocess: bool = False is_multimodal: bool = False image_aspect_ratio: str = 'square' # =================================================================== data_path: Optional[List[str]] = field(default=None, metadata={"help": "Path to the training data."}) image_folder: Optional[str] = field(default=None) video_folder: Optional[str] = field(default=None) num_frames: int = 8 # =================================================================== class TrainingArguments(transformers.TrainingArguments): cache_dir: Optional[str] = field(default=None) optim: str = field(default="adamw_torch") remove_unused_columns: bool = field(default=False) freeze_mm_mlp_adapter: bool = field(default=False) mpt_attn_impl: Optional[str] = field(default="triton") model_max_length: int = field( default=512, metadata={ "help": "Maximum sequence length. Sequences will be right padded (and possibly truncated)." }, ) double_quant: bool = field( default=True, metadata={"help": "Compress the quantization statistics through double quantization."} ) quant_type: str = field( default="nf4", metadata={"help": "Quantization data type to use. Should be one of `fp4` or `nf4`."} ) bits: int = field( default=16, metadata={"help": "How many bits to use."} ) lora_enable: bool = False lora_r: int = 64 lora_alpha: int = 16 lora_dropout: float = 0.05 lora_weight_path: str = "" lora_bias: str = "none" mm_projector_lr: Optional[float] = None group_by_modality_length: bool = field(default=False) # ================================================ tokenizer_model_max_length: Optional[int] = None # ================================================ def get_peft_state_maybe_zero_3(named_params, bias): if bias == "none": to_return = {k: t for k, t in named_params if "lora_" in k} elif bias == "all": to_return = {k: t for k, t in named_params if "lora_" in k or "bias" in k} elif bias == "lora_only": to_return = {} maybe_lora_bias = {} lora_bias_names = set() for k, t in named_params: if "lora_" in k: to_return[k] = t bias_name = k.split("lora_")[0] + "bias" lora_bias_names.add(bias_name) elif "bias" in k: maybe_lora_bias[k] = t for k, t in maybe_lora_bias: if bias_name in lora_bias_names: to_return[bias_name] = t else: raise NotImplementedError to_return = {k: maybe_zero_3(v, ignore_status=True) for k, v in to_return.items()} return to_return def get_peft_state_non_lora_maybe_zero_3(named_params, require_grad_only=True): to_return = {k: t for k, t in named_params if "lora_" not in k} if require_grad_only: to_return = {k: t for k, t in to_return.items() if t.requires_grad} to_return = {k: maybe_zero_3(v, ignore_status=True).cpu() for k, v in to_return.items()} return to_return def find_all_linear_names(model): cls = torch.nn.Linear lora_module_names = set() multimodal_keywords = ['mm_projector', 'vision_tower', 'vision_resampler'] for name, module in model.named_modules(): if any(mm_keyword in name for mm_keyword in multimodal_keywords): continue if isinstance(module, cls): names = name.split('.') lora_module_names.add(names[0] if len(names) == 1 else names[-1]) if 'lm_head' in lora_module_names: # needed for 16-bit lora_module_names.remove('lm_head') return list(lora_module_names) def safe_save_model_for_hf_trainer(trainer: transformers.Trainer, output_dir: str): """Collects the state dict and dump to disk.""" if getattr(trainer.args, "tune_mm_mlp_adapter", False): # Only save Adapter keys_to_match = ['mm_projector'] if getattr(trainer.args, "use_im_start_end", False): keys_to_match.extend(['embed_tokens', 'embed_in']) weight_to_save = get_mm_adapter_state_maybe_zero_3(trainer.model.named_parameters(), keys_to_match) trainer.model.config.save_pretrained(output_dir) current_folder = output_dir.split('/')[-1] parent_folder = os.path.dirname(output_dir) if trainer.args.local_rank == 0 or trainer.args.local_rank == -1: if current_folder.startswith('checkpoint-'): mm_projector_folder = os.path.join(parent_folder, "mm_projector") os.makedirs(mm_projector_folder, exist_ok=True) torch.save(weight_to_save, os.path.join(mm_projector_folder, f'{current_folder}.bin')) else: torch.save(weight_to_save, os.path.join(output_dir, f'mm_projector.bin')) return if trainer.deepspeed: torch.cuda.synchronize() trainer.save_model(output_dir) return state_dict = trainer.model.state_dict() if trainer.args.should_save: cpu_state_dict = { key: value.cpu() for key, value in state_dict.items() } del state_dict trainer._save(output_dir, state_dict=cpu_state_dict) # noqa def smart_tokenizer_and_embedding_resize( special_tokens_dict: Dict, tokenizer: transformers.PreTrainedTokenizer, model: transformers.PreTrainedModel, ): """Resize tokenizer and embedding. Note: This is the unoptimized version that may make your embedding size not be divisible by 64. """ num_new_tokens = tokenizer.add_special_tokens(special_tokens_dict) model.resize_token_embeddings(len(tokenizer)) if num_new_tokens > 0: input_embeddings = model.get_input_embeddings().weight.data output_embeddings = model.get_output_embeddings().weight.data input_embeddings_avg = input_embeddings[:-num_new_tokens].mean( dim=0, keepdim=True) output_embeddings_avg = output_embeddings[:-num_new_tokens].mean( dim=0, keepdim=True) input_embeddings[-num_new_tokens:] = input_embeddings_avg output_embeddings[-num_new_tokens:] = output_embeddings_avg def make_supervised_data_module(tokenizer: transformers.PreTrainedTokenizer, data_args) -> Dict: """Make dataset and collator for supervised fine-tuning.""" train_dataset = LazySupervisedDataset(tokenizer=tokenizer, data_path=data_args.data_path, data_args=data_args) data_collator = DataCollatorForSupervisedDataset(tokenizer=tokenizer) return dict(train_dataset=train_dataset, eval_dataset=None, data_collator=data_collator) class LLaVATrainer(Trainer): def _get_train_sampler(self) -> Optional[torch.utils.data.Sampler]: if self.train_dataset is None or not has_length(self.train_dataset): return None if self.args.group_by_modality_length: lengths = self.train_dataset.modality_lengths return LengthGroupedSampler( self.args.train_batch_size, world_size=self.args.world_size * self.args.gradient_accumulation_steps, lengths=lengths, group_by_modality=True, ) else: return super()._get_train_sampler() def create_optimizer(self): """ Setup the optimizer. We provide a reasonable default that works well. If you want to use something else, you can pass a tuple in the Trainer's init through `optimizers`, or subclass and override this method in a subclass. """ if is_sagemaker_mp_enabled(): return super().create_optimizer() if self.sharded_ddp == ShardedDDPOption.SIMPLE: return super().create_optimizer() opt_model = self.model if self.optimizer is None: decay_parameters = get_parameter_names(opt_model, ALL_LAYERNORM_LAYERS) decay_parameters = [name for name in decay_parameters if "bias" not in name] if self.args.mm_projector_lr is not None: projector_parameters = [name for name, _ in opt_model.named_parameters() if "mm_projector" in name] optimizer_grouped_parameters = [ { "params": [ p for n, p in opt_model.named_parameters() if (n in decay_parameters and n not in projector_parameters and p.requires_grad) ], "weight_decay": self.args.weight_decay, }, { "params": [ p for n, p in opt_model.named_parameters() if (n not in decay_parameters and n not in projector_parameters and p.requires_grad) ], "weight_decay": 0.0, }, { "params": [ p for n, p in opt_model.named_parameters() if (n in decay_parameters and n in projector_parameters and p.requires_grad) ], "weight_decay": self.args.weight_decay, "lr": self.args.mm_projector_lr, }, { "params": [ p for n, p in opt_model.named_parameters() if (n not in decay_parameters and n in projector_parameters and p.requires_grad) ], "weight_decay": 0.0, "lr": self.args.mm_projector_lr, }, ] else: optimizer_grouped_parameters = [ { "params": [ p for n, p in opt_model.named_parameters() if (n in decay_parameters and p.requires_grad) ], "weight_decay": self.args.weight_decay, }, { "params": [ p for n, p in opt_model.named_parameters() if (n not in decay_parameters and p.requires_grad) ], "weight_decay": 0.0, }, ] optimizer_cls, optimizer_kwargs = Trainer.get_optimizer_cls_and_kwargs(self.args) if self.sharded_ddp == ShardedDDPOption.SIMPLE: self.optimizer = OSS( params=optimizer_grouped_parameters, optim=optimizer_cls, **optimizer_kwargs, ) else: self.optimizer = optimizer_cls(optimizer_grouped_parameters, **optimizer_kwargs) if optimizer_cls.__name__ == "Adam8bit": import bitsandbytes manager = bitsandbytes.optim.GlobalOptimManager.get_instance() skipped = 0 for module in opt_model.modules(): if isinstance(module, nn.Embedding): skipped += sum({p.data_ptr(): p.numel() for p in module.parameters()}.values()) logger.info(f"skipped {module}: {skipped/2**20}M params") manager.register_module_override(module, "weight", {"optim_bits": 32}) logger.debug(f"bitsandbytes: will optimize {module} in fp32") logger.info(f"skipped: {skipped/2**20}M params") return self.optimizer def _save_checkpoint(self, model, trial, metrics=None): if getattr(self.args, 'tune_mm_mlp_adapter', False): from transformers.trainer_utils import PREFIX_CHECKPOINT_DIR checkpoint_folder = f"{PREFIX_CHECKPOINT_DIR}-{self.state.global_step}" run_dir = self._get_output_dir(trial=trial) output_dir = os.path.join(run_dir, checkpoint_folder) # Only save Adapter keys_to_match = ['mm_projector', 'vision_resampler'] if getattr(self.args, "use_im_start_end", False): keys_to_match.extend(['embed_tokens', 'embed_in']) weight_to_save = get_mm_adapter_state_maybe_zero_3(self.model.named_parameters(), keys_to_match) if self.args.local_rank == 0 or self.args.local_rank == -1: self.model.config.save_pretrained(output_dir) torch.save(weight_to_save, os.path.join(output_dir, f'mm_projector.bin')) else: super(LLaVATrainer, self)._save_checkpoint(model, trial, metrics) def _save(self, output_dir: Optional[str] = None, state_dict=None): if getattr(self.args, 'tune_mm_mlp_adapter', False): pass else: super(LLaVATrainer, self)._save(output_dir, state_dict) def train(): global local_rank parser = transformers.HfArgumentParser( (ModelArguments, DataArguments, TrainingArguments)) model_args, data_args, training_args = parser.parse_args_into_dataclasses() local_rank = training_args.local_rank compute_dtype = (torch.float16 if training_args.fp16 else (torch.bfloat16 if training_args.bf16 else torch.float32)) bnb_model_from_pretrained_args = {} if training_args.bits in [4, 8]: from transformers import BitsAndBytesConfig bnb_model_from_pretrained_args.update(dict( device_map={"": training_args.device}, load_in_4bit=training_args.bits == 4, load_in_8bit=training_args.bits == 8, quantization_config=BitsAndBytesConfig( load_in_4bit=training_args.bits == 4, load_in_8bit=training_args.bits == 8, llm_int8_skip_modules=["mm_projector"], llm_int8_threshold=6.0, llm_int8_has_fp16_weight=False, bnb_4bit_compute_dtype=compute_dtype, bnb_4bit_use_double_quant=training_args.double_quant, bnb_4bit_quant_type=training_args.quant_type # {'fp4', 'nf4'} ) )) # ========================================================================== if model_args.image_tower is not None or model_args.video_tower is not None: # ========================================================================== if 'mpt' in model_args.model_name_or_path: config = transformers.AutoConfig.from_pretrained(model_args.model_name_or_path, trust_remote_code=True) config.attn_config['attn_impl'] = training_args.mpt_attn_impl model = LlavaMPTForCausalLM.from_pretrained( model_args.model_name_or_path, config=config, cache_dir=training_args.cache_dir, **bnb_model_from_pretrained_args ) else: model = LlavaLlamaForCausalLM.from_pretrained( model_args.model_name_or_path, cache_dir=training_args.cache_dir, **bnb_model_from_pretrained_args ) else: model = transformers.LlamaForCausalLM.from_pretrained( model_args.model_name_or_path, cache_dir=training_args.cache_dir, **bnb_model_from_pretrained_args ) model.config.use_cache = False if model_args.freeze_backbone: model.model.requires_grad_(False) if training_args.bits in [4, 8]: from peft import prepare_model_for_kbit_training model.config.torch_dtype=(torch.float32 if training_args.fp16 else (torch.bfloat16 if training_args.bf16 else torch.float32)) model = prepare_model_for_kbit_training(model, use_gradient_checkpointing=training_args.gradient_checkpointing) if training_args.gradient_checkpointing: if hasattr(model, "enable_input_require_grads"): model.enable_input_require_grads() else: def make_inputs_require_grad(module, input, output): output.requires_grad_(True) model.get_input_embeddings().register_forward_hook(make_inputs_require_grad) if training_args.lora_enable: from peft import LoraConfig, get_peft_model lora_config = LoraConfig( r=training_args.lora_r, lora_alpha=training_args.lora_alpha, target_modules=find_all_linear_names(model), lora_dropout=training_args.lora_dropout, bias=training_args.lora_bias, task_type="CAUSAL_LM", ) if training_args.bits == 16: if training_args.bf16: model.to(torch.bfloat16) if training_args.fp16: model.to(torch.float16) rank0_print("Adding LoRA adapters...") model = get_peft_model(model, lora_config) if 'mpt' in model_args.model_name_or_path: tokenizer = transformers.AutoTokenizer.from_pretrained( model_args.model_name_or_path, cache_dir=training_args.cache_dir, model_max_length=training_args.model_max_length, padding_side="right" ) else: tokenizer = transformers.AutoTokenizer.from_pretrained( model_args.model_name_or_path, cache_dir=training_args.cache_dir, model_max_length=training_args.model_max_length, padding_side="right", use_fast=False, ) if model_args.version == "v0": if tokenizer.pad_token is None: smart_tokenizer_and_embedding_resize( special_tokens_dict=dict(pad_token="[PAD]"), tokenizer=tokenizer, model=model, ) elif model_args.version == "v0.5": tokenizer.pad_token = tokenizer.unk_token else: tokenizer.pad_token = tokenizer.unk_token if model_args.version in conversation_lib.conv_templates: conversation_lib.default_conversation = conversation_lib.conv_templates[model_args.version] else: conversation_lib.default_conversation = conversation_lib.conv_templates["vicuna_v1"] # ============================================================================================================= if model_args.image_tower is not None or model_args.video_tower is not None: # print(model_args) model.get_model().initialize_vision_modules( model_args=model_args, fsdp=training_args.fsdp ) if model_args.image_tower is not None: image_tower = model.get_image_tower() image_tower.to(dtype=torch.bfloat16 if training_args.bf16 else torch.float16, device=training_args.device) data_args.image_processor = image_tower.image_processor data_args.is_multimodal = True if model_args.video_tower is not None: video_tower = model.get_video_tower() video_tower.to(dtype=torch.bfloat16 if training_args.bf16 else torch.float16, device=training_args.device) data_args.video_processor = video_tower.video_processor data_args.is_multimodal = True data_args.num_frames = video_tower.config.num_frames # ============================================================================================================= model.config.image_aspect_ratio = data_args.image_aspect_ratio model.config.tokenizer_padding_side = tokenizer.padding_side # ============================================================================================================= tokenizer_model_max_length = training_args.tokenizer_model_max_length model.config.tokenizer_model_max_length = tokenizer.model_max_length if tokenizer_model_max_length is None else tokenizer_model_max_length # ============================================================================================================= model.config.tune_mm_mlp_adapter = training_args.tune_mm_mlp_adapter = model_args.tune_mm_mlp_adapter if model_args.tune_mm_mlp_adapter: model.requires_grad_(False) for p in model.get_model().mm_projector.parameters(): p.requires_grad = True model.config.freeze_mm_mlp_adapter = training_args.freeze_mm_mlp_adapter if training_args.freeze_mm_mlp_adapter: for p in model.get_model().mm_projector.parameters(): p.requires_grad = False if training_args.bits in [4, 8]: model.get_model().mm_projector.to(dtype=compute_dtype, device=training_args.device) model.config.mm_use_im_start_end = data_args.mm_use_im_start_end = model_args.mm_use_im_start_end model.config.mm_projector_lr = training_args.mm_projector_lr training_args.use_im_start_end = model_args.mm_use_im_start_end model.config.mm_use_im_patch_token = model_args.mm_use_im_patch_token model.initialize_vision_tokenizer(model_args, tokenizer=tokenizer) if training_args.bits in [4, 8]: from peft.tuners.lora import LoraLayer for name, module in model.named_modules(): if isinstance(module, LoraLayer): if training_args.bf16: module = module.to(torch.bfloat16) if 'norm' in name: module = module.to(torch.float32) if 'lm_head' in name or 'embed_tokens' in name: if hasattr(module, 'weight'): if training_args.bf16 and module.weight.dtype == torch.float32: module = module.to(torch.bfloat16) data_module = make_supervised_data_module(tokenizer=tokenizer, data_args=data_args) trainer = LLaVATrainer(model=model, tokenizer=tokenizer, args=training_args, **data_module) if list(pathlib.Path(training_args.output_dir).glob("checkpoint-*")): trainer.train(resume_from_checkpoint=True) else: trainer.train() trainer.save_state() model.config.use_cache = True if training_args.lora_enable: state_dict = get_peft_state_maybe_zero_3( model.named_parameters(), training_args.lora_bias ) non_lora_state_dict = get_peft_state_non_lora_maybe_zero_3( model.named_parameters() ) if training_args.local_rank == 0 or training_args.local_rank == -1: model.config.save_pretrained(training_args.output_dir) model.save_pretrained(training_args.output_dir, state_dict=state_dict) torch.save(non_lora_state_dict, os.path.join(training_args.output_dir, 'non_lora_trainables.bin')) else: safe_save_model_for_hf_trainer(trainer=trainer, output_dir=training_args.output_dir)
null
179,589
import os from .clip_encoder import CLIPVisionTower from .languagebind import LanguageBindImageTower, LanguageBindVideoTower class CLIPVisionTower(nn.Module): def __init__(self, vision_tower, args, delay_load=False): super().__init__() self.is_loaded = False self.vision_tower_name = vision_tower self.select_layer = args.mm_vision_select_layer self.select_feature = getattr(args, 'mm_vision_select_feature', 'patch') if not delay_load: self.load_model() else: self.cfg_only = CLIPVisionConfig.from_pretrained(self.vision_tower_name) def load_model(self): self.image_processor = CLIPImageProcessor.from_pretrained(self.vision_tower_name) self.vision_tower = CLIPVisionModel.from_pretrained(self.vision_tower_name) self.vision_tower.requires_grad_(False) self.is_loaded = True def feature_select(self, image_forward_outs): image_features = image_forward_outs.hidden_states[self.select_layer] if self.select_feature == 'patch': image_features = image_features[:, 1:] elif self.select_feature == 'cls_patch': image_features = image_features else: raise ValueError(f'Unexpected select feature: {self.select_feature}') return image_features def forward(self, images): if type(images) is list: image_features = [] for image in images: image_forward_out = self.vision_tower(image.to(device=self.device, dtype=self.dtype).unsqueeze(0), output_hidden_states=True) image_feature = self.feature_select(image_forward_out).to(image.dtype) image_features.append(image_feature) else: image_forward_outs = self.vision_tower(images.to(device=self.device, dtype=self.dtype), output_hidden_states=True) image_features = self.feature_select(image_forward_outs).to(images.dtype) return image_features def dummy_feature(self): return torch.zeros(1, self.hidden_size, device=self.device, dtype=self.dtype) def dtype(self): return self.vision_tower.dtype def device(self): return self.vision_tower.device def config(self): if self.is_loaded: return self.vision_tower.config else: return self.cfg_only def hidden_size(self): return self.config.hidden_size def num_patches(self): return (self.config.image_size // self.config.patch_size) ** 2 class LanguageBindImageTower(nn.Module): def __init__(self, image_tower, args, delay_load=False, cache_dir='./cache_dir'): super().__init__() self.is_loaded = False self.image_tower_name = image_tower self.select_layer = args.mm_vision_select_layer self.select_feature = getattr(args, 'mm_vision_select_feature', 'patch') self.cache_dir = cache_dir if not delay_load: self.load_model() else: self.cfg_only = LanguageBindImageConfig.from_pretrained(self.image_tower_name, cache_dir=self.cache_dir) ############################################################ def load_model(self): model = LanguageBindImage.from_pretrained(self.image_tower_name, cache_dir=self.cache_dir) self.image_tower = model.vision_model self.image_tower.requires_grad_(False) self.image_processor = LanguageBindImageProcessor(model.config) self.is_loaded = True def feature_select(self, image_forward_outs): image_features = image_forward_outs.hidden_states[self.select_layer] if self.select_feature == 'patch': image_features = image_features[:, 1:] elif self.select_feature == 'cls_patch': image_features = image_features else: raise ValueError(f'Unexpected select feature: {self.select_feature}') return image_features def forward(self, images): if type(images) is list: image_features = [] for image in images: image_forward_out = self.image_tower(image.to(device=self.device, dtype=self.dtype).unsqueeze(0), output_hidden_states=True) image_feature = self.feature_select(image_forward_out).to(image.dtype) image_features.append(image_feature) else: # print('images', images.shape) image_forward_outs = self.image_tower(images.to(device=self.device, dtype=self.dtype), output_hidden_states=True) # print('image_forward_outs', len(image_forward_outs), image_forward_outs[0].shape) image_features = self.feature_select(image_forward_outs).to(images.dtype) # print('image_features', image_features.shape) return image_features def dummy_feature(self): return torch.zeros(1, self.hidden_size, device=self.device, dtype=self.dtype) def dtype(self): return self.image_tower.embeddings.class_embedding.dtype ############# def device(self): return self.image_tower.embeddings.class_embedding.device ############## def config(self): if self.is_loaded: return self.image_tower.config else: return self.cfg_only def hidden_size(self): return self.config.hidden_size def num_patches(self): return (self.config.image_size // self.config.patch_size) ** 2 def build_image_tower(image_tower_cfg, **kwargs): image_tower = getattr(image_tower_cfg, 'mm_image_tower', getattr(image_tower_cfg, 'image_tower', None)) is_absolute_path_exists = os.path.exists(image_tower) if is_absolute_path_exists or image_tower.startswith("openai") or image_tower.startswith("laion"): return CLIPVisionTower(image_tower, args=image_tower_cfg, **kwargs) if image_tower.endswith('LanguageBind_Image'): return LanguageBindImageTower(image_tower, args=image_tower_cfg, cache_dir='./cache_dir', **kwargs) raise ValueError(f'Unknown image tower: {image_tower}')
null
179,590
import os from .clip_encoder import CLIPVisionTower from .languagebind import LanguageBindImageTower, LanguageBindVideoTower class LanguageBindVideoTower(nn.Module): def __init__(self, video_tower, args, delay_load=False, cache_dir='./cache_dir'): super().__init__() self.is_loaded = False self.video_tower_name = video_tower self.select_layer = args.mm_vision_select_layer self.select_feature = getattr(args, 'mm_vision_select_feature', 'patch') self.cache_dir = cache_dir if not delay_load: self.load_model() else: self.cfg_only = LanguageBindVideoConfig.from_pretrained(self.video_tower_name, cache_dir=self.cache_dir) ############################################################ def load_model(self): model = LanguageBindVideo.from_pretrained(self.video_tower_name, cache_dir=self.cache_dir) self.video_processor = LanguageBindVideoProcessor(model.config) # model = LanguageBindImage.from_pretrained('LanguageBind/LanguageBind_Image', cache_dir=self.cache_dir) self.video_tower = model.vision_model self.video_tower.requires_grad_(False) self.is_loaded = True def feature_select(self, video_forward_outs): video_features = video_forward_outs.hidden_states[self.select_layer] # b t n c return video_features # return all # b, t, n, c = video_features.shape # if self.select_feature == 'patch': # video_features = video_features[:, :, 1:] # else: # raise ValueError(f'Unexpected select feature: {self.select_feature}') # return video_features def forward(self, videos): if type(videos) is list: video_features = [] for video in videos: video_forward_out = self.video_tower(video.to(device=self.device, dtype=self.dtype).unsqueeze(0), output_hidden_states=True) video_feature = self.feature_select(video_forward_out).to(video.dtype) video_features.append(video_feature) else: video_forward_outs = self.video_tower(videos.to(device=self.device, dtype=self.dtype), output_hidden_states=True) video_features = self.feature_select(video_forward_outs).to(videos.dtype) return video_features def dummy_feature(self): return torch.zeros(1, self.hidden_size, device=self.device, dtype=self.dtype) def dtype(self): return self.video_tower.embeddings.class_embedding.dtype ############# # return torch.randn(1).cuda().dtype def device(self): return self.video_tower.embeddings.class_embedding.device ############## # return torch.randn(1).cuda().device def config(self): if self.is_loaded: return self.video_tower.config else: return self.cfg_only def hidden_size(self): return self.config.hidden_size def num_patches(self): return (self.config.image_size // self.config.patch_size) ** 2 def build_video_tower(video_tower_cfg, **kwargs): video_tower = getattr(video_tower_cfg, 'mm_video_tower', getattr(video_tower_cfg, 'video_tower', None)) if video_tower.endswith('LanguageBind_Video_merge'): return LanguageBindVideoTower(video_tower, args=video_tower_cfg, cache_dir='./cache_dir', **kwargs) raise ValueError(f'Unknown video tower: {video_tower}')
null
179,591
import cv2 import torch from PIL import Image from torch import nn from torchvision import transforms from transformers import ProcessorMixin, BatchEncoding from transformers.image_processing_utils import BatchFeature def make_list_of_images(x): if not isinstance(x, list): return [x] return x
null
179,592
import cv2 import torch from PIL import Image from torch import nn from torchvision import transforms from transformers import ProcessorMixin, BatchEncoding from transformers.image_processing_utils import BatchFeature OPENAI_DATASET_MEAN = (0.48145466, 0.4578275, 0.40821073) OPENAI_DATASET_STD = (0.26862954, 0.26130258, 0.27577711) class DepthNorm(nn.Module): def __init__( self, max_depth=0, min_depth=0.01, ): super().__init__() self.max_depth = max_depth self.min_depth = min_depth self.scale = 1000.0 # nyuv2 abs.depth def forward(self, image): # image = np.array(image) depth_img = image / self.scale # (H, W) in meters depth_img = depth_img.clip(min=self.min_depth) if self.max_depth != 0: depth_img = depth_img.clip(max=self.max_depth) depth_img /= self.max_depth # 0-1 else: depth_img /= depth_img.max() depth_img = torch.from_numpy(depth_img).unsqueeze(0).repeat(3, 1, 1) # assume image return depth_img.to(torch.get_default_dtype()) def get_depth_transform(config): config = config.vision_config transform = transforms.Compose( [ DepthNorm(max_depth=config.max_depth), transforms.Resize(224, interpolation=transforms.InterpolationMode.BICUBIC), transforms.CenterCrop(224), transforms.Normalize(OPENAI_DATASET_MEAN, OPENAI_DATASET_STD), # assume image # transforms.Normalize((0.5, ), (0.5, )) # 0-1 to norm distribution # transforms.Normalize((0.0418, ), (0.0295, )) # sun rgb-d imagebind # transforms.Normalize((0.02, ), (0.00295, )) # nyuv2 ] ) return transform
null
179,593
import cv2 import torch from PIL import Image from torch import nn from torchvision import transforms from transformers import ProcessorMixin, BatchEncoding from transformers.image_processing_utils import BatchFeature def opencv_loader(path): return cv2.imread(path, cv2.IMREAD_UNCHANGED).astype('float32') def load_and_transform_depth(depth_path, transform): depth = opencv_loader(depth_path) depth_outputs = transform(depth) return depth_outputs
null
179,594
import math from typing import Optional, Tuple, Union import torch from einops import rearrange from peft import LoraConfig, get_peft_model from torch import nn from torch.nn import functional as F from transformers import PreTrainedModel, add_start_docstrings from transformers.modeling_outputs import BaseModelOutput, BaseModelOutputWithPooling from transformers.models.clip.modeling_clip import CLIPMLP, CLIPAttention, CLIPTextEmbeddings, CLIPVisionEmbeddings, \ CLIPVisionModelWithProjection, CLIPTextModelWithProjection, _expand_mask, CLIPOutput, clip_loss from transformers.utils import add_start_docstrings_to_model_forward, replace_return_docstrings from .configuration_depth import LanguageBindDepthConfig, CLIPVisionConfig, CLIPTextConfig The provided code snippet includes necessary dependencies for implementing the `_make_causal_mask` function. Write a Python function `def _make_causal_mask( input_ids_shape: torch.Size, dtype: torch.dtype, device: torch.device, past_key_values_length: int = 0 )` to solve the following problem: Make causal mask used for bi-directional self-attention. Here is the function: def _make_causal_mask( input_ids_shape: torch.Size, dtype: torch.dtype, device: torch.device, past_key_values_length: int = 0 ): """ Make causal mask used for bi-directional self-attention. """ bsz, tgt_len = input_ids_shape mask = torch.full((tgt_len, tgt_len), torch.finfo(dtype).min, device=device) mask_cond = torch.arange(mask.size(-1), device=device) mask.masked_fill_(mask_cond < (mask_cond + 1).view(mask.size(-1), 1), 0) mask = mask.to(dtype) if past_key_values_length > 0: mask = torch.cat([torch.zeros(tgt_len, past_key_values_length, dtype=dtype, device=device), mask], dim=-1) return mask[None, None, :, :].expand(bsz, 1, tgt_len, tgt_len + past_key_values_length)
Make causal mask used for bi-directional self-attention.
179,595
import math from typing import Optional, Tuple, Union import torch from einops import rearrange from peft import LoraConfig, get_peft_model from torch import nn from torch.nn import functional as F from transformers import PreTrainedModel, add_start_docstrings from transformers.modeling_outputs import BaseModelOutput, BaseModelOutputWithPooling from transformers.models.clip.modeling_clip import CLIPMLP, CLIPAttention, CLIPTextEmbeddings, CLIPVisionEmbeddings, \ CLIPVisionModelWithProjection, CLIPTextModelWithProjection, _expand_mask, CLIPOutput, clip_loss from transformers.utils import add_start_docstrings_to_model_forward, replace_return_docstrings from .configuration_audio import LanguageBindAudioConfig, CLIPVisionConfig, CLIPTextConfig The provided code snippet includes necessary dependencies for implementing the `_make_causal_mask` function. Write a Python function `def _make_causal_mask( input_ids_shape: torch.Size, dtype: torch.dtype, device: torch.device, past_key_values_length: int = 0 )` to solve the following problem: Make causal mask used for bi-directional self-attention. Here is the function: def _make_causal_mask( input_ids_shape: torch.Size, dtype: torch.dtype, device: torch.device, past_key_values_length: int = 0 ): """ Make causal mask used for bi-directional self-attention. """ bsz, tgt_len = input_ids_shape mask = torch.full((tgt_len, tgt_len), torch.finfo(dtype).min, device=device) mask_cond = torch.arange(mask.size(-1), device=device) mask.masked_fill_(mask_cond < (mask_cond + 1).view(mask.size(-1), 1), 0) mask = mask.to(dtype) if past_key_values_length > 0: mask = torch.cat([torch.zeros(tgt_len, past_key_values_length, dtype=dtype, device=device), mask], dim=-1) return mask[None, None, :, :].expand(bsz, 1, tgt_len, tgt_len + past_key_values_length)
Make causal mask used for bi-directional self-attention.
179,596
import cv2 import numpy as np import torch from torchvision import transforms from transformers import ProcessorMixin, BatchEncoding from transformers.image_processing_utils import BatchFeature from torch.nn import functional as F def make_list_of_images(x): if not isinstance(x, list): return [x] return x
null
179,597
import cv2 import numpy as np import torch from torchvision import transforms from transformers import ProcessorMixin, BatchEncoding from transformers.image_processing_utils import BatchFeature from torch.nn import functional as F def int16_to_float32_torch(x): return (x / 32767.0).type(torch.float32)
null
179,598
import cv2 import numpy as np import torch from torchvision import transforms from transformers import ProcessorMixin, BatchEncoding from transformers.image_processing_utils import BatchFeature from torch.nn import functional as F def float32_to_int16_torch(x): x = torch.clamp(x, min=-1., max=1.) return (x * 32767.).type(torch.int16)
null
179,599
import cv2 import numpy as np import torch from torchvision import transforms from transformers import ProcessorMixin, BatchEncoding from transformers.image_processing_utils import BatchFeature from torch.nn import functional as F class AudioTransform: def __init__(self, config): self.sample_rate = config.audio_sample_rate self.num_mel_bins = config.num_mel_bins self.target_length = config.target_length self.audio_mean = config.audio_mean self.audio_std = config.audio_std # mean=-4.2677393 # std=4.5689974 self.norm = transforms.Normalize(mean=self.audio_mean, std=self.audio_std) def __call__(self, audio_data_and_origin_sr): audio_data, origin_sr = audio_data_and_origin_sr if self.sample_rate != origin_sr: # print(audio_data.shape, origin_sr) audio_data = torchaudio.functional.resample(audio_data, orig_freq=origin_sr, new_freq=self.sample_rate) waveform_melspec = self.waveform2melspec(audio_data[0]) return self.norm(waveform_melspec) def waveform2melspec(self, audio_data): max_len = self.target_length * self.sample_rate // 100 if audio_data.shape[-1] > max_len: mel = self.get_mel(audio_data) # split to three parts chunk_frames = self.target_length total_frames = mel.shape[0] ranges = np.array_split(list(range(0, total_frames - chunk_frames + 1)), 3) # print('total_frames-chunk_frames:', total_frames-chunk_frames, # 'len(audio_data):', len(audio_data), # 'chunk_frames:', chunk_frames, # 'total_frames:', total_frames) if len(ranges[1]) == 0: # if the audio is too short, we just use the first chunk ranges[1] = [0] if len(ranges[2]) == 0: # if the audio is too short, we just use the first chunk ranges[2] = [0] # randomly choose index for each part # idx_front = np.random.choice(ranges[0]) # idx_middle = np.random.choice(ranges[1]) # idx_back = np.random.choice(ranges[2]) idx_front = ranges[0][0] # fixed idx_middle = ranges[1][0] idx_back = ranges[2][0] # select mel mel_chunk_front = mel[idx_front:idx_front + chunk_frames, :] mel_chunk_middle = mel[idx_middle:idx_middle + chunk_frames, :] mel_chunk_back = mel[idx_back:idx_back + chunk_frames, :] # stack mel_fusion = torch.stack([mel_chunk_front, mel_chunk_middle, mel_chunk_back], dim=0) elif audio_data.shape[-1] < max_len: # padding if too short n_repeat = int(max_len / len(audio_data)) audio_data = audio_data.repeat(n_repeat) audio_data = F.pad( audio_data, (0, max_len - len(audio_data)), mode="constant", value=0, ) mel = self.get_mel(audio_data) mel_fusion = torch.stack([mel, mel, mel], dim=0) else: # if equal mel = self.get_mel(audio_data) mel_fusion = torch.stack([mel, mel, mel], dim=0) # twice check p = self.target_length - mel_fusion.shape[1] # if abs(p) / self.target_length > 0.2: # logging.warning( # "Large gap between audio n_frames(%d) and " # "target_length (%d). Is the audio_target_length " # "setting correct?", # mel_fusion.shape[1], # self.target_length, # ) # cut and pad if p > 0: m = torch.nn.ZeroPad2d((0, 0, 0, p)) mel_fusion = m(mel_fusion) elif p < 0: mel_fusion = mel_fusion[:, 0: self.target_length, :] mel_fusion = mel_fusion.transpose(1, 2) # [3, target_length, mel_bins] -> [3, mel_bins, target_length] return mel_fusion def get_mel(self, audio_data): # mel shape: (n_mels, T) audio_data -= audio_data.mean() mel = torchaudio.compliance.kaldi.fbank( audio_data.unsqueeze(0), htk_compat=True, sample_frequency=self.sample_rate, use_energy=False, window_type="hanning", num_mel_bins=self.num_mel_bins, dither=0.0, frame_length=25, frame_shift=DEFAULT_AUDIO_FRAME_SHIFT_MS, ) return mel # (T, n_mels) def get_audio_transform(config): config = config.vision_config return AudioTransform(config)
null
179,600
import cv2 import numpy as np import torch from torchvision import transforms from transformers import ProcessorMixin, BatchEncoding from transformers.image_processing_utils import BatchFeature from torch.nn import functional as F def torchaudio_loader(path): return torchaudio.load(path) def load_and_transform_audio( audio_path, transform, ): waveform_and_sr = torchaudio_loader(audio_path) audio_outputs = transform(waveform_and_sr) return audio_outputs
null
179,601
import torch from PIL import Image from torchvision import transforms from transformers import ProcessorMixin, BatchEncoding from transformers.image_processing_utils import BatchFeature def make_list_of_images(x): if not isinstance(x, list): return [x] return x
null
179,602
import torch from PIL import Image from torchvision import transforms from transformers import ProcessorMixin, BatchEncoding from transformers.image_processing_utils import BatchFeature OPENAI_DATASET_MEAN = (0.48145466, 0.4578275, 0.40821073) OPENAI_DATASET_STD = (0.26862954, 0.26130258, 0.27577711) def get_thermal_transform(config): config = config.vision_config transform = transforms.Compose( [ transforms.ToTensor(), transforms.Resize(224, interpolation=transforms.InterpolationMode.BICUBIC), transforms.CenterCrop(224), transforms.Normalize(OPENAI_DATASET_MEAN, OPENAI_DATASET_STD) # assume image ] ) return transform
null
179,603
import torch from PIL import Image from torchvision import transforms from transformers import ProcessorMixin, BatchEncoding from transformers.image_processing_utils import BatchFeature def load_and_transform_thermal(thermal_path, transform): thermal = Image.open(thermal_path) thermal_outputs = transform(thermal) return thermal_outputs
null
179,604
import math from typing import Optional, Tuple, Union import torch from einops import rearrange from peft import LoraConfig, get_peft_model from torch import nn from torch.nn import functional as F from transformers import PreTrainedModel, add_start_docstrings from transformers.modeling_outputs import BaseModelOutput, BaseModelOutputWithPooling from transformers.models.clip.modeling_clip import CLIPMLP, CLIPAttention, CLIPTextEmbeddings, CLIPVisionEmbeddings, \ CLIPVisionModelWithProjection, CLIPTextModelWithProjection, _expand_mask, CLIPOutput, clip_loss from transformers.utils import add_start_docstrings_to_model_forward, replace_return_docstrings from .configuration_thermal import LanguageBindThermalConfig, CLIPVisionConfig, CLIPTextConfig The provided code snippet includes necessary dependencies for implementing the `_make_causal_mask` function. Write a Python function `def _make_causal_mask( input_ids_shape: torch.Size, dtype: torch.dtype, device: torch.device, past_key_values_length: int = 0 )` to solve the following problem: Make causal mask used for bi-directional self-attention. Here is the function: def _make_causal_mask( input_ids_shape: torch.Size, dtype: torch.dtype, device: torch.device, past_key_values_length: int = 0 ): """ Make causal mask used for bi-directional self-attention. """ bsz, tgt_len = input_ids_shape mask = torch.full((tgt_len, tgt_len), torch.finfo(dtype).min, device=device) mask_cond = torch.arange(mask.size(-1), device=device) mask.masked_fill_(mask_cond < (mask_cond + 1).view(mask.size(-1), 1), 0) mask = mask.to(dtype) if past_key_values_length > 0: mask = torch.cat([torch.zeros(tgt_len, past_key_values_length, dtype=dtype, device=device), mask], dim=-1) return mask[None, None, :, :].expand(bsz, 1, tgt_len, tgt_len + past_key_values_length)
Make causal mask used for bi-directional self-attention.
179,606
import torch from PIL import Image from torchvision import transforms from transformers import ProcessorMixin, BatchEncoding from transformers.image_processing_utils import BatchFeature OPENAI_DATASET_MEAN = (0.48145466, 0.4578275, 0.40821073) OPENAI_DATASET_STD = (0.26862954, 0.26130258, 0.27577711) def get_image_transform(config): config = config.vision_config transform = transforms.Compose( [ transforms.ToTensor(), transforms.Resize(224, interpolation=transforms.InterpolationMode.BICUBIC), transforms.CenterCrop(224), transforms.Normalize(OPENAI_DATASET_MEAN, OPENAI_DATASET_STD) # assume image ] ) return transform
null
179,607
import torch from PIL import Image from torchvision import transforms from transformers import ProcessorMixin, BatchEncoding from transformers.image_processing_utils import BatchFeature def load_and_transform_image(image_path, transform): image = Image.open(image_path).convert('RGB') if isinstance(image_path, str) else image_path image_outputs = transform(image) return image_outputs
null
179,608
import math from typing import Optional, Tuple, Union import torch from einops import rearrange from peft import LoraConfig, get_peft_model from torch import nn from torch.nn import functional as F from transformers import PreTrainedModel, add_start_docstrings from transformers.modeling_outputs import BaseModelOutput, BaseModelOutputWithPooling from transformers.models.clip.modeling_clip import CLIPMLP, CLIPAttention, CLIPTextEmbeddings, CLIPVisionEmbeddings, \ CLIPVisionModelWithProjection, CLIPTextModelWithProjection, _expand_mask, CLIPOutput, clip_loss from transformers.utils import add_start_docstrings_to_model_forward, replace_return_docstrings from .configuration_image import LanguageBindImageConfig, CLIPVisionConfig, CLIPTextConfig The provided code snippet includes necessary dependencies for implementing the `_make_causal_mask` function. Write a Python function `def _make_causal_mask( input_ids_shape: torch.Size, dtype: torch.dtype, device: torch.device, past_key_values_length: int = 0 )` to solve the following problem: Make causal mask used for bi-directional self-attention. Here is the function: def _make_causal_mask( input_ids_shape: torch.Size, dtype: torch.dtype, device: torch.device, past_key_values_length: int = 0 ): """ Make causal mask used for bi-directional self-attention. """ bsz, tgt_len = input_ids_shape mask = torch.full((tgt_len, tgt_len), torch.finfo(dtype).min, device=device) mask_cond = torch.arange(mask.size(-1), device=device) mask.masked_fill_(mask_cond < (mask_cond + 1).view(mask.size(-1), 1), 0) mask = mask.to(dtype) if past_key_values_length > 0: mask = torch.cat([torch.zeros(tgt_len, past_key_values_length, dtype=dtype, device=device), mask], dim=-1) return mask[None, None, :, :].expand(bsz, 1, tgt_len, tgt_len + past_key_values_length)
Make causal mask used for bi-directional self-attention.
179,609
import math from typing import Optional, Tuple, Union import torch from einops import rearrange from peft import LoraConfig, get_peft_model from torch import nn from torch.nn import functional as F from transformers import PreTrainedModel, add_start_docstrings from transformers.modeling_outputs import BaseModelOutput, BaseModelOutputWithPooling from transformers.models.clip.modeling_clip import CLIPMLP, CLIPAttention, CLIPTextEmbeddings, CLIPVisionEmbeddings, \ CLIPVisionModelWithProjection, CLIPTextModelWithProjection, _expand_mask, CLIPOutput, clip_loss from transformers.utils import add_start_docstrings_to_model_forward, replace_return_docstrings from .configuration_video import LanguageBindVideoConfig, CLIPVisionConfig, CLIPTextConfig The provided code snippet includes necessary dependencies for implementing the `_make_causal_mask` function. Write a Python function `def _make_causal_mask( input_ids_shape: torch.Size, dtype: torch.dtype, device: torch.device, past_key_values_length: int = 0 )` to solve the following problem: Make causal mask used for bi-directional self-attention. Here is the function: def _make_causal_mask( input_ids_shape: torch.Size, dtype: torch.dtype, device: torch.device, past_key_values_length: int = 0 ): """ Make causal mask used for bi-directional self-attention. """ bsz, tgt_len = input_ids_shape mask = torch.full((tgt_len, tgt_len), torch.finfo(dtype).min, device=device) mask_cond = torch.arange(mask.size(-1), device=device) mask.masked_fill_(mask_cond < (mask_cond + 1).view(mask.size(-1), 1), 0) mask = mask.to(dtype) if past_key_values_length > 0: mask = torch.cat([torch.zeros(tgt_len, past_key_values_length, dtype=dtype, device=device), mask], dim=-1) return mask[None, None, :, :].expand(bsz, 1, tgt_len, tgt_len + past_key_values_length)
Make causal mask used for bi-directional self-attention.
179,610
import torch import cv2 import decord import numpy as np from PIL import Image from decord import VideoReader, cpu from torchvision import transforms from transformers import ProcessorMixin, BatchEncoding from transformers.image_processing_utils import BatchFeature from pytorchvideo.data.encoded_video import EncodedVideo from torchvision.transforms import Compose, Lambda, ToTensor from torchvision.transforms._transforms_video import NormalizeVideo, RandomCropVideo, RandomHorizontalFlipVideo, CenterCropVideo from pytorchvideo.transforms import ApplyTransformToKey, ShortSideScale, UniformTemporalSubsample def make_list_of_images(x): if not isinstance(x, list): return [x] return x
null
179,611
import torch import cv2 import decord import numpy as np from PIL import Image from decord import VideoReader, cpu from torchvision import transforms from transformers import ProcessorMixin, BatchEncoding from transformers.image_processing_utils import BatchFeature from pytorchvideo.data.encoded_video import EncodedVideo from torchvision.transforms import Compose, Lambda, ToTensor from torchvision.transforms._transforms_video import NormalizeVideo, RandomCropVideo, RandomHorizontalFlipVideo, CenterCropVideo from pytorchvideo.transforms import ApplyTransformToKey, ShortSideScale, UniformTemporalSubsample OPENAI_DATASET_MEAN = (0.48145466, 0.4578275, 0.40821073) OPENAI_DATASET_STD = (0.26862954, 0.26130258, 0.27577711) def get_video_transform(config): config = config.vision_config if config.video_decode_backend == 'pytorchvideo': transform = ApplyTransformToKey( key="video", transform=Compose( [ UniformTemporalSubsample(config.num_frames), Lambda(lambda x: x / 255.0), NormalizeVideo(mean=OPENAI_DATASET_MEAN, std=OPENAI_DATASET_STD), ShortSideScale(size=224), CenterCropVideo(224), RandomHorizontalFlipVideo(p=0.5), ] ), ) elif config.video_decode_backend == 'decord': transform = Compose( [ # UniformTemporalSubsample(num_frames), Lambda(lambda x: x / 255.0), NormalizeVideo(mean=OPENAI_DATASET_MEAN, std=OPENAI_DATASET_STD), ShortSideScale(size=224), CenterCropVideo(224), RandomHorizontalFlipVideo(p=0.5), ] ) elif config.video_decode_backend == 'opencv': transform = Compose( [ # UniformTemporalSubsample(num_frames), Lambda(lambda x: x / 255.0), NormalizeVideo(mean=OPENAI_DATASET_MEAN, std=OPENAI_DATASET_STD), ShortSideScale(size=224), CenterCropVideo(224), RandomHorizontalFlipVideo(p=0.5), ] ) else: raise NameError('video_decode_backend should specify in (pytorchvideo, decord, opencv)') return transform
null
179,612
import torch import cv2 import decord import numpy as np from PIL import Image from decord import VideoReader, cpu from torchvision import transforms from transformers import ProcessorMixin, BatchEncoding from transformers.image_processing_utils import BatchFeature from pytorchvideo.data.encoded_video import EncodedVideo from torchvision.transforms import Compose, Lambda, ToTensor from torchvision.transforms._transforms_video import NormalizeVideo, RandomCropVideo, RandomHorizontalFlipVideo, CenterCropVideo from pytorchvideo.transforms import ApplyTransformToKey, ShortSideScale, UniformTemporalSubsample decord.bridge.set_bridge('torch') def load_and_transform_video( video_path, transform, video_decode_backend='opencv', clip_start_sec=0.0, clip_end_sec=None, num_frames=8, ): if video_decode_backend == 'pytorchvideo': # decord pyav video = EncodedVideo.from_path(video_path, decoder="decord", decode_audio=False) duration = video.duration start_sec = clip_start_sec # secs end_sec = clip_end_sec if clip_end_sec is not None else duration # secs video_data = video.get_clip(start_sec=start_sec, end_sec=end_sec) video_outputs = transform(video_data) elif video_decode_backend == 'decord': decord.bridge.set_bridge('torch') decord_vr = VideoReader(video_path, ctx=cpu(0)) duration = len(decord_vr) frame_id_list = np.linspace(0, duration-1, num_frames, dtype=int) video_data = decord_vr.get_batch(frame_id_list) video_data = video_data.permute(3, 0, 1, 2) # (T, H, W, C) -> (C, T, H, W) video_outputs = transform(video_data) elif video_decode_backend == 'opencv': cv2_vr = cv2.VideoCapture(video_path) duration = int(cv2_vr.get(cv2.CAP_PROP_FRAME_COUNT)) frame_id_list = np.linspace(0, duration-1, num_frames, dtype=int) video_data = [] for frame_idx in frame_id_list: cv2_vr.set(1, frame_idx) _, frame = cv2_vr.read() frame = cv2.cvtColor(frame, cv2.COLOR_BGR2RGB) video_data.append(torch.from_numpy(frame).permute(2, 0, 1)) cv2_vr.release() video_data = torch.stack(video_data, dim=1) video_outputs = transform(video_data) else: raise NameError('video_decode_backend should specify in (pytorchvideo, decord, opencv)') return video_outputs
null
179,613
import argparse import torch from tqdm import tqdm from transformers import AutoTokenizer, AutoModelForCausalLM from videollava import LlavaLlamaForCausalLM def apply_delta(base_model_path, target_model_path, delta_path): print("Loading base model") base = AutoModelForCausalLM.from_pretrained( base_model_path, torch_dtype=torch.float16, low_cpu_mem_usage=True) print("Loading delta") delta = LlavaLlamaForCausalLM.from_pretrained(delta_path, torch_dtype=torch.float16, low_cpu_mem_usage=True) delta_tokenizer = AutoTokenizer.from_pretrained(delta_path) print("Applying delta") for name, param in tqdm(delta.state_dict().items(), desc="Applying delta"): if name not in base.state_dict(): assert name in ['model.mm_projector.weight', 'model.mm_projector.bias'], f'{name} not in base model' continue if param.data.shape == base.state_dict()[name].shape: param.data += base.state_dict()[name] else: assert name in ['model.embed_tokens.weight', 'lm_head.weight'], \ f'{name} dimension mismatch: {param.data.shape} vs {base.state_dict()[name].shape}' bparam = base.state_dict()[name] param.data[:bparam.shape[0], :bparam.shape[1]] += bparam print("Saving target model") delta.save_pretrained(target_model_path) delta_tokenizer.save_pretrained(target_model_path)
null
179,622
import math import torch import triton_pre_mlir as triton import triton_pre_mlir.language as tl def _fwd_kernel(Q, K, V, Bias, Out, Lse, TMP, softmax_scale, stride_qb, stride_qh, stride_qm, stride_kb, stride_kh, stride_kn, stride_vb, stride_vh, stride_vn, stride_bb, stride_bh, stride_bm, stride_ob, stride_oh, stride_om, nheads, seqlen_q, seqlen_k, seqlen_q_rounded, headdim, CACHE_KEY_SEQLEN_Q, CACHE_KEY_SEQLEN_K, BIAS_TYPE: tl.constexpr, IS_CAUSAL: tl.constexpr, BLOCK_HEADDIM: tl.constexpr, EVEN_M: tl.constexpr, EVEN_N: tl.constexpr, EVEN_HEADDIM: tl.constexpr, BLOCK_M: tl.constexpr, BLOCK_N: tl.constexpr): def _flash_attn_forward(q, k, v, bias=None, causal=False, softmax_scale=None): (batch, seqlen_q, nheads, d) = q.shape (_, seqlen_k, _, _) = k.shape assert k.shape == (batch, seqlen_k, nheads, d) assert v.shape == (batch, seqlen_k, nheads, d) assert d <= 128, 'FlashAttention only support head dimensions up to 128' assert q.dtype == k.dtype == v.dtype, 'All tensors must have the same type' assert q.dtype in [torch.float16, torch.bfloat16], 'Only support fp16 and bf16' assert q.is_cuda and k.is_cuda and v.is_cuda softmax_scale = softmax_scale or 1.0 / math.sqrt(d) has_bias = bias is not None bias_type = 'none' if has_bias: assert bias.dtype in [q.dtype, torch.float] assert bias.is_cuda assert bias.dim() == 4 if bias.stride(-1) != 1: bias = bias.contiguous() if bias.shape[2:] == (1, seqlen_k): bias_type = 'vector' elif bias.shape[2:] == (seqlen_q, seqlen_k): bias_type = 'matrix' else: raise RuntimeError('Last 2 dimensions of bias must be (1, seqlen_k) or (seqlen_q, seqlen_k)') bias = bias.expand(batch, nheads, seqlen_q, seqlen_k) bias_strides = (bias.stride(0), bias.stride(1), bias.stride(2)) if has_bias else (0, 0, 0) seqlen_q_rounded = math.ceil(seqlen_q / 128) * 128 lse = torch.empty((batch, nheads, seqlen_q_rounded), device=q.device, dtype=torch.float32) tmp = torch.empty((batch, nheads, seqlen_q_rounded), device=q.device, dtype=torch.float32) o = torch.empty_like(q) BLOCK_HEADDIM = max(triton.next_power_of_2(d), 16) BLOCK = 128 num_warps = 4 if d <= 64 else 8 grid = lambda META: (triton.cdiv(seqlen_q, META['BLOCK_M']), batch * nheads) _fwd_kernel[grid](q, k, v, bias, o, lse, tmp, softmax_scale, q.stride(0), q.stride(2), q.stride(1), k.stride(0), k.stride(2), k.stride(1), v.stride(0), v.stride(2), v.stride(1), *bias_strides, o.stride(0), o.stride(2), o.stride(1), nheads, seqlen_q, seqlen_k, seqlen_q_rounded, d, seqlen_q // 32, seqlen_k // 32, bias_type, causal, BLOCK_HEADDIM, BLOCK_M=BLOCK, BLOCK_N=BLOCK, num_warps=num_warps, num_stages=1) return (o, lse, softmax_scale)
null
179,629
import math import warnings from collections.abc import Sequence from functools import partial from typing import Optional, Tuple, Union import torch from torch import nn from .norm import NORM_CLASS_REGISTRY def generic_param_init_fn_(module: nn.Module, init_fn_, n_layers: int, d_model: Optional[int]=None, init_div_is_residual: Union[int, float, str, bool]=True, emb_init_std: Optional[float]=None, emb_init_uniform_lim: Optional[Union[Tuple[float, float], float]]=None, verbose: int=0, **kwargs): def kaiming_normal_param_init_fn_(module: nn.Module, n_layers: int, d_model: Optional[int]=None, init_div_is_residual: Union[int, float, str, bool]=True, emb_init_std: Optional[float]=None, emb_init_uniform_lim: Optional[Union[Tuple[float, float], float]]=None, init_gain: float=0, fan_mode: str='fan_in', init_nonlinearity: str='leaky_relu', verbose: int=0, **kwargs): del kwargs if verbose > 1: warnings.warn(f'Using nn.init.kaiming_normal_ init fn with parameters: ' + f'a={init_gain}, mode={fan_mode}, nonlinearity={init_nonlinearity}') kaiming_normal_ = partial(torch.nn.init.kaiming_normal_, a=init_gain, mode=fan_mode, nonlinearity=init_nonlinearity) generic_param_init_fn_(module=module, init_fn_=kaiming_normal_, d_model=d_model, n_layers=n_layers, init_div_is_residual=init_div_is_residual, emb_init_std=emb_init_std, emb_init_uniform_lim=emb_init_uniform_lim, verbose=verbose)
null
179,631
import math import warnings from collections.abc import Sequence from functools import partial from typing import Optional, Tuple, Union import torch from torch import nn from .norm import NORM_CLASS_REGISTRY def generic_param_init_fn_(module: nn.Module, init_fn_, n_layers: int, d_model: Optional[int]=None, init_div_is_residual: Union[int, float, str, bool]=True, emb_init_std: Optional[float]=None, emb_init_uniform_lim: Optional[Union[Tuple[float, float], float]]=None, verbose: int=0, **kwargs): def xavier_normal_param_init_fn_(module: nn.Module, n_layers: int, d_model: Optional[int]=None, init_div_is_residual: Union[int, float, str, bool]=True, emb_init_std: Optional[float]=None, emb_init_uniform_lim: Optional[Union[Tuple[float, float], float]]=None, init_gain: float=0, verbose: int=0, **kwargs): xavier_normal_ = partial(torch.nn.init.xavier_normal_, gain=init_gain) if verbose > 1: warnings.warn(f'Using torch.nn.init.xavier_normal_ init fn with parameters: ' + f'gain={init_gain}') generic_param_init_fn_(module=module, init_fn_=xavier_normal_, d_model=d_model, n_layers=n_layers, init_div_is_residual=init_div_is_residual, emb_init_std=emb_init_std, emb_init_uniform_lim=emb_init_uniform_lim, verbose=verbose)
null
179,635
import argparse import torch from tqdm import tqdm from transformers import AutoTokenizer, AutoModelForCausalLM from videollava.model.utils import auto_upgrade def auto_upgrade(config): cfg = AutoConfig.from_pretrained(config) if 'llava' in config and 'llava' not in cfg.model_type: assert cfg.model_type == 'llama' print("You are using newer LLaVA code base, while the checkpoint of v0 is from older code base.") print("You must upgrade the checkpoint to the new code base (this can be done automatically).") confirm = input("Please confirm that you want to upgrade the checkpoint. [Y/N]") if confirm.lower() in ["y", "yes"]: print("Upgrading checkpoint...") assert len(cfg.architectures) == 1 setattr(cfg.__class__, "model_type", "llava") cfg.architectures[0] = 'LlavaLlamaForCausalLM' cfg.save_pretrained(config) print("Checkpoint upgraded.") else: print("Checkpoint upgrade aborted.") exit(1) def make_delta(base_model_path, target_model_path, delta_path, hub_repo_id): print("Loading base model") base = AutoModelForCausalLM.from_pretrained( base_model_path, torch_dtype=torch.float16, low_cpu_mem_usage=True) print("Loading target model") auto_upgrade(target_model_path) target = AutoModelForCausalLM.from_pretrained(target_model_path, torch_dtype=torch.float16, low_cpu_mem_usage=True) print("Calculating delta") for name, param in tqdm(target.state_dict().items(), desc="Calculating delta"): if name not in base.state_dict(): assert name in ['model.mm_projector.weight', 'model.mm_projector.bias'], f'{name} not in base model' continue if param.data.shape == base.state_dict()[name].shape: param.data -= base.state_dict()[name] else: assert name in ['model.embed_tokens.weight', 'lm_head.weight'], f'{name} dimension mismatch: {param.data.shape} vs {base.state_dict()[name].shape}' bparam = base.state_dict()[name] param.data[:bparam.shape[0], :bparam.shape[1]] -= bparam print("Saving delta") if hub_repo_id: kwargs = {"push_to_hub": True, "repo_id": hub_repo_id} else: kwargs = {} target.save_pretrained(delta_path, **kwargs) target_tokenizer = AutoTokenizer.from_pretrained(target_model_path) target_tokenizer.save_pretrained(delta_path, **kwargs)
null
179,636
import argparse import torch from transformers import AutoTokenizer, AutoModelForCausalLM from videollava.model import * from videollava.model.utils import auto_upgrade def auto_upgrade(config): cfg = AutoConfig.from_pretrained(config) if 'llava' in config and 'llava' not in cfg.model_type: assert cfg.model_type == 'llama' print("You are using newer LLaVA code base, while the checkpoint of v0 is from older code base.") print("You must upgrade the checkpoint to the new code base (this can be done automatically).") confirm = input("Please confirm that you want to upgrade the checkpoint. [Y/N]") if confirm.lower() in ["y", "yes"]: print("Upgrading checkpoint...") assert len(cfg.architectures) == 1 setattr(cfg.__class__, "model_type", "llava") cfg.architectures[0] = 'LlavaLlamaForCausalLM' cfg.save_pretrained(config) print("Checkpoint upgraded.") else: print("Checkpoint upgrade aborted.") exit(1) def consolidate_ckpt(src_path, dst_path): print("Loading model") auto_upgrade(src_path) src_model = AutoModelForCausalLM.from_pretrained(src_path, torch_dtype=torch.float16, low_cpu_mem_usage=True) src_tokenizer = AutoTokenizer.from_pretrained(src_path, use_fast=False) src_model.save_pretrained(dst_path) src_tokenizer.save_pretrained(dst_path)
null
179,638
import argparse import torch from tqdm import tqdm import json from videollava.constants import IMAGE_TOKEN_INDEX, DEFAULT_IMAGE_TOKEN, DEFAULT_IM_START_TOKEN, DEFAULT_IM_END_TOKEN from videollava.conversation import conv_templates, SeparatorStyle from videollava.model.builder import load_pretrained_model from videollava.utils import disable_torch_init from videollava.mm_utils import tokenizer_image_token, get_model_name_from_path, KeywordsStoppingCriteria from PIL import Image import requests from PIL import Image from io import BytesIO def load_image(image_file): if image_file.startswith('http') or image_file.startswith('https'): response = requests.get(image_file) image = Image.open(BytesIO(response.content)).convert('RGB') else: image = Image.open(image_file).convert('RGB') return image IMAGE_TOKEN_INDEX = -200 DEFAULT_IMAGE_TOKEN = "<image>" DEFAULT_IM_START_TOKEN = "<im_start>" DEFAULT_IM_END_TOKEN = "<im_end>" class SeparatorStyle(Enum): """Different separator style.""" SINGLE = auto() TWO = auto() MPT = auto() PLAIN = auto() LLAMA_2 = auto() conv_templates = { "default": conv_vicuna_v0, "v0": conv_vicuna_v0, "v1": conv_vicuna_v1, "vicuna_v1": conv_vicuna_v1, "llama_2": conv_llama_2, "plain": conv_llava_plain, "v0_plain": conv_llava_plain, "llava_v0": conv_llava_v0, "v0_mmtag": conv_llava_v0_mmtag, "llava_v1": conv_llava_v1, "v1_mmtag": conv_llava_v1_mmtag, "llava_llama_2": conv_llava_llama_2, "mpt": conv_mpt, } def load_pretrained_model(model_path, model_base, model_name, load_8bit=False, load_4bit=False, device_map="auto", device="cuda", **kwargs): kwargs = {"device_map": device_map, **kwargs} if device != "cuda": kwargs['device_map'] = {"": device} if load_8bit: kwargs['load_in_8bit'] = True elif load_4bit: kwargs['load_in_4bit'] = True kwargs['quantization_config'] = BitsAndBytesConfig( load_in_4bit=True, bnb_4bit_compute_dtype=torch.float16, bnb_4bit_use_double_quant=True, bnb_4bit_quant_type='nf4' ) else: kwargs['torch_dtype'] = torch.float16 if 'llava' in model_name.lower(): # Load LLaVA model if 'lora' in model_name.lower() and model_base is None: warnings.warn('There is `lora` in model name but no `model_base` is provided. If you are loading a LoRA model, please provide the `model_base` argument. Detailed instruction: https://github.com/haotian-liu/LLaVA#launch-a-model-worker-lora-weights-unmerged.') if 'lora' in model_name.lower() and model_base is not None: lora_cfg_pretrained = AutoConfig.from_pretrained(model_path) tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=False) print('Loading LLaVA from base model...') model = LlavaLlamaForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, config=lora_cfg_pretrained, **kwargs) token_num, tokem_dim = model.lm_head.out_features, model.lm_head.in_features if model.lm_head.weight.shape[0] != token_num: model.lm_head.weight = torch.nn.Parameter(torch.empty(token_num, tokem_dim, device=model.device, dtype=model.dtype)) model.model.embed_tokens.weight = torch.nn.Parameter(torch.empty(token_num, tokem_dim, device=model.device, dtype=model.dtype)) print('Loading additional LLaVA weights...') if os.path.exists(os.path.join(model_path, 'non_lora_trainables.bin')): non_lora_trainables = torch.load(os.path.join(model_path, 'non_lora_trainables.bin'), map_location='cpu') else: # this is probably from HF Hub from huggingface_hub import hf_hub_download def load_from_hf(repo_id, filename, subfolder=None): cache_file = hf_hub_download( repo_id=repo_id, filename=filename, subfolder=subfolder) return torch.load(cache_file, map_location='cpu') non_lora_trainables = load_from_hf(model_path, 'non_lora_trainables.bin') non_lora_trainables = {(k[11:] if k.startswith('base_model.') else k): v for k, v in non_lora_trainables.items()} if any(k.startswith('model.model.') for k in non_lora_trainables): non_lora_trainables = {(k[6:] if k.startswith('model.') else k): v for k, v in non_lora_trainables.items()} model.load_state_dict(non_lora_trainables, strict=False) from peft import PeftModel print('Loading LoRA weights...') model = PeftModel.from_pretrained(model, model_path) print('Merging LoRA weights...') model = model.merge_and_unload() print('Model is loaded...') elif model_base is not None: # this may be mm projector only print('Loading LLaVA from base model...') if 'mpt' in model_name.lower(): if not os.path.isfile(os.path.join(model_path, 'configuration_mpt.py')): shutil.copyfile(os.path.join(model_base, 'configuration_mpt.py'), os.path.join(model_path, 'configuration_mpt.py')) tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=True) cfg_pretrained = AutoConfig.from_pretrained(model_path, trust_remote_code=True) model = LlavaMPTForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, config=cfg_pretrained, **kwargs) else: tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=False) cfg_pretrained = AutoConfig.from_pretrained(model_path) model = LlavaLlamaForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, config=cfg_pretrained, **kwargs) mm_projector_weights = torch.load(os.path.join(model_path, 'mm_projector.bin'), map_location='cpu') mm_projector_weights = {k: v.to(torch.float16) for k, v in mm_projector_weights.items()} model.load_state_dict(mm_projector_weights, strict=False) else: if 'mpt' in model_name.lower(): tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=True) model = LlavaMPTForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs) else: tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=False) model = LlavaLlamaForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs) else: # Load language model if model_base is not None: # PEFT model from peft import PeftModel tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=False) model = AutoModelForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, **kwargs) print(f"Loading LoRA weights from {model_path}") model = PeftModel.from_pretrained(model, model_path) print(f"Merging weights") model = model.merge_and_unload() print('Convert to FP16...') model.to(torch.float16) else: use_fast = False if 'mpt' in model_name.lower(): tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=True) model = AutoModelForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, trust_remote_code=True, **kwargs) else: tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=False) model = AutoModelForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs) # ========================================================================================================== processor = {'image': None, 'video': None} if 'llava' in model_name.lower(): mm_use_im_start_end = getattr(model.config, "mm_use_im_start_end", False) mm_use_im_patch_token = getattr(model.config, "mm_use_im_patch_token", True) if mm_use_im_patch_token: tokenizer.add_tokens([DEFAULT_IMAGE_PATCH_TOKEN], special_tokens=True) tokenizer.add_tokens([DEFAULT_VIDEO_PATCH_TOKEN], special_tokens=True) if mm_use_im_start_end: tokenizer.add_tokens([DEFAULT_IM_START_TOKEN, DEFAULT_IM_END_TOKEN], special_tokens=True) tokenizer.add_tokens([DEFAULT_VID_START_TOKEN, DEFAULT_VID_END_TOKEN], special_tokens=True) model.resize_token_embeddings(len(tokenizer)) if model.config.mm_image_tower is not None: image_tower = model.get_image_tower() if not image_tower.is_loaded: image_tower.load_model() image_tower.to(device=device, dtype=torch.float16) image_processor = image_tower.image_processor processor['image'] = image_processor if model.config.mm_video_tower is not None: video_tower = model.get_video_tower() if not video_tower.is_loaded: video_tower.load_model() video_tower.to(device=device, dtype=torch.float16) video_processor = video_tower.video_processor processor['video'] = video_processor # ========================================================================================================== if hasattr(model.config, "max_sequence_length"): context_len = model.config.max_sequence_length else: context_len = 2048 return tokenizer, model, processor, context_len def disable_torch_init(): """ Disable the redundant torch default initialization to accelerate model creation. """ import torch setattr(torch.nn.Linear, "reset_parameters", lambda self: None) setattr(torch.nn.LayerNorm, "reset_parameters", lambda self: None) def tokenizer_image_token(prompt, tokenizer, image_token_index=IMAGE_TOKEN_INDEX, return_tensors=None): prompt_chunks = [tokenizer(chunk).input_ids for chunk in prompt.split('<image>')] def insert_separator(X, sep): return [ele for sublist in zip(X, [sep]*len(X)) for ele in sublist][:-1] input_ids = [] offset = 0 if len(prompt_chunks) > 0 and len(prompt_chunks[0]) > 0 and prompt_chunks[0][0] == tokenizer.bos_token_id: offset = 1 input_ids.append(prompt_chunks[0][0]) for x in insert_separator(prompt_chunks, [image_token_index] * (offset + 1)): input_ids.extend(x[offset:]) if return_tensors is not None: if return_tensors == 'pt': return torch.tensor(input_ids, dtype=torch.long) raise ValueError(f'Unsupported tensor type: {return_tensors}') return input_ids def get_model_name_from_path(model_path): model_path = model_path.strip("/") model_paths = model_path.split("/") if model_paths[-1].startswith('checkpoint-'): return model_paths[-2] + "_" + model_paths[-1] else: return model_paths[-1] class KeywordsStoppingCriteria(StoppingCriteria): def __init__(self, keywords, tokenizer, input_ids): self.keywords = keywords self.keyword_ids = [] self.max_keyword_len = 0 for keyword in keywords: cur_keyword_ids = tokenizer(keyword).input_ids if len(cur_keyword_ids) > 1 and cur_keyword_ids[0] == tokenizer.bos_token_id: cur_keyword_ids = cur_keyword_ids[1:] if len(cur_keyword_ids) > self.max_keyword_len: self.max_keyword_len = len(cur_keyword_ids) self.keyword_ids.append(torch.tensor(cur_keyword_ids)) self.tokenizer = tokenizer self.start_len = input_ids.shape[1] def call_for_batch(self, output_ids: torch.LongTensor, scores: torch.FloatTensor, **kwargs) -> bool: offset = min(output_ids.shape[1] - self.start_len, self.max_keyword_len) self.keyword_ids = [keyword_id.to(output_ids.device) for keyword_id in self.keyword_ids] for keyword_id in self.keyword_ids: if (output_ids[0, -keyword_id.shape[0]:] == keyword_id).all(): return True outputs = self.tokenizer.batch_decode(output_ids[:, -offset:], skip_special_tokens=True)[0] for keyword in self.keywords: if keyword in outputs: return True return False def __call__(self, output_ids: torch.LongTensor, scores: torch.FloatTensor, **kwargs) -> bool: outputs = [] for i in range(output_ids.shape[0]): outputs.append(self.call_for_batch(output_ids[i].unsqueeze(0), scores)) return all(outputs) def eval_model(args): # Model disable_torch_init() model_name = get_model_name_from_path(args.model_path) tokenizer, model, image_processor, context_len = load_pretrained_model(args.model_path, args.model_base, model_name, True) with open(args.questions_file) as f: llvqa_data = json.load(f) for i, llddata in enumerate(tqdm(llvqa_data)): filename = llddata["img_path"] if args.lang == "en": message = llddata["question"] + "\nChoose between one of the options as follows:\n" elif args.lang == "zh": message = llddata["question"] + "\在下列选项中选择一个:\n" else: raise NotImplementedError("Q-Bench does not support languages other than English (en) and Chinese (zh) yet. Contact us (https://github.com/VQAssessment/Q-Bench/) to convert Q-Bench into more languages.") for choice, ans in zip(["A.", "B.", "C.", "D."], llddata["candidates"]): message += f"{choice} {ans}\n" qs = message if model.config.mm_use_im_start_end: qs = DEFAULT_IM_START_TOKEN + DEFAULT_IMAGE_TOKEN + DEFAULT_IM_END_TOKEN + '\n' + qs else: qs = DEFAULT_IMAGE_TOKEN + '\n' + qs if 'llama-2' in model_name.lower(): conv_mode = "llava_llama_2" elif "v1" in model_name.lower(): conv_mode = "llava_v1" elif "mpt" in model_name.lower(): conv_mode = "mpt" else: conv_mode = "llava_v0" if args.conv_mode is not None and conv_mode != args.conv_mode: print('[WARNING] the auto inferred conversation mode is {}, while `--conv-mode` is {}, using {}'.format(conv_mode, args.conv_mode, args.conv_mode)) else: args.conv_mode = conv_mode conv = conv_templates[args.conv_mode].copy() conv.append_message(conv.roles[0], qs) conv.append_message(conv.roles[1], None) prompt = conv.get_prompt() image = load_image(args.image_folder + filename) image_tensor = image_processor.preprocess(image, return_tensors='pt')['pixel_values'].half().cuda() input_ids = tokenizer_image_token(prompt, tokenizer, IMAGE_TOKEN_INDEX, return_tensors='pt').unsqueeze(0).cuda() stop_str = conv.sep if conv.sep_style != SeparatorStyle.TWO else conv.sep2 keywords = [stop_str] stopping_criteria = KeywordsStoppingCriteria(keywords, tokenizer, input_ids) with torch.inference_mode(): output_ids = model.generate( input_ids, images=image_tensor, num_beams=1, do_sample=False, temperature=0, max_new_tokens=1024, use_cache=True, stopping_criteria=[stopping_criteria]) input_token_len = input_ids.shape[1] n_diff_input_output = (input_ids != output_ids[:, :input_token_len]).sum().item() if n_diff_input_output > 0: print(f'[Warning] {n_diff_input_output} output_ids are not the same as the input_ids') outputs = tokenizer.batch_decode(output_ids[:, input_token_len:], skip_special_tokens=True)[0] outputs = outputs.strip() if outputs.endswith(stop_str): outputs = outputs[:-len(stop_str)] outputs = outputs.strip() llddata["response"] = outputs with open(args.answers_file, "a") as wf: json.dump(llddata, wf)
null
179,642
import os import argparse import json import re from videollava.eval.m4c_evaluator import TextVQAAccuracyEvaluator def get_args(): parser = argparse.ArgumentParser() parser.add_argument('--annotation-file', type=str) parser.add_argument('--result-file', type=str) parser.add_argument('--result-dir', type=str) return parser.parse_args()
null
179,643
import os import argparse import json import re from videollava.eval.m4c_evaluator import TextVQAAccuracyEvaluator def prompt_processor(prompt): if prompt.startswith('OCR tokens: '): pattern = r"Question: (.*?) Short answer:" match = re.search(pattern, prompt, re.DOTALL) question = match.group(1) elif 'Reference OCR token: ' in prompt and len(prompt.split('\n')) == 3: if prompt.startswith('Reference OCR token:'): question = prompt.split('\n')[1] else: question = prompt.split('\n')[0] elif len(prompt.split('\n')) == 2: question = prompt.split('\n')[0] else: assert False return question.lower() class TextVQAAccuracyEvaluator: def __init__(self): self.answer_processor = EvalAIAnswerProcessor() def _compute_answer_scores(self, raw_answers): """ compute the accuracy (soft score) of human answers """ answers = [self.answer_processor(a) for a in raw_answers] assert len(answers) == 10 gt_answers = list(enumerate(answers)) unique_answers = set(answers) unique_answer_scores = {} for unique_answer in unique_answers: accs = [] for gt_answer in gt_answers: other_answers = [item for item in gt_answers if item != gt_answer] matching_answers = [ item for item in other_answers if item[1] == unique_answer ] acc = min(1, float(len(matching_answers)) / 3) accs.append(acc) unique_answer_scores[unique_answer] = sum(accs) / len(accs) return unique_answer_scores def eval_pred_list(self, pred_list): pred_scores = [] for entry in tqdm(pred_list): pred_answer = self.answer_processor(entry["pred_answer"]) unique_answer_scores = self._compute_answer_scores(entry["gt_answers"]) score = unique_answer_scores.get(pred_answer, 0.0) pred_scores.append(score) accuracy = sum(pred_scores) / len(pred_scores) return accuracy def eval_single(annotation_file, result_file): experiment_name = os.path.splitext(os.path.basename(result_file))[0] print(experiment_name) annotations = json.load(open(annotation_file))['data'] annotations = {(annotation['image_id'], annotation['question'].lower()): annotation for annotation in annotations} results = [json.loads(line) for line in open(result_file)] pred_list = [] for result in results: annotation = annotations[(result['question_id'], prompt_processor(result['prompt']))] pred_list.append({ "pred_answer": result['text'], "gt_answers": annotation['answers'], }) evaluator = TextVQAAccuracyEvaluator() print('Samples: {}\nAccuracy: {:.2f}%\n'.format(len(pred_list), 100. * evaluator.eval_pred_list(pred_list)))
null
179,650
import argparse from transformers import AutoTokenizer, AutoModelForCausalLM, StoppingCriteria import torch import os import json from tqdm import tqdm import shortuuid from videollava.conversation import default_conversation from videollava.utils import disable_torch_init class KeywordsStoppingCriteria(StoppingCriteria): def __init__(self, keywords, tokenizer, input_ids): self.keywords = keywords self.tokenizer = tokenizer self.start_len = None self.input_ids = input_ids def __call__(self, output_ids: torch.LongTensor, scores: torch.FloatTensor, **kwargs) -> bool: if self.start_len is None: self.start_len = self.input_ids.shape[1] else: outputs = self.tokenizer.batch_decode(output_ids[:, self.start_len:], skip_special_tokens=True)[0] for keyword in self.keywords: if keyword in outputs: return True return False default_conversation = conv_vicuna_v1 def disable_torch_init(): """ Disable the redundant torch default initialization to accelerate model creation. """ import torch setattr(torch.nn.Linear, "reset_parameters", lambda self: None) setattr(torch.nn.LayerNorm, "reset_parameters", lambda self: None) def eval_model(model_name, questions_file, answers_file): # Model disable_torch_init() model_name = os.path.expanduser(model_name) tokenizer = AutoTokenizer.from_pretrained(model_name, use_fast=False) model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.float16).cuda() ques_file = open(os.path.expanduser(questions_file), "r") ans_file = open(os.path.expanduser(answers_file), "w") for i, line in enumerate(tqdm(ques_file)): idx = json.loads(line)["question_id"] qs = json.loads(line)["text"] cat = json.loads(line)["category"] conv = default_conversation.copy() conv.append_message(conv.roles[0], qs) prompt = conv.get_prompt() inputs = tokenizer([prompt]) input_ids = torch.as_tensor(inputs.input_ids).cuda() stopping_criteria = KeywordsStoppingCriteria([conv.sep], tokenizer, input_ids) output_ids = model.generate( input_ids, do_sample=True, use_cache=True, temperature=0.7, max_new_tokens=1024, stopping_criteria=[stopping_criteria]) outputs = tokenizer.batch_decode(output_ids, skip_special_tokens=True)[0] try: index = outputs.index(conv.sep, len(prompt)) except ValueError: outputs += conv.sep index = outputs.index(conv.sep, len(prompt)) outputs = outputs[len(prompt) + len(conv.roles[1]) + 2:index].strip() ans_id = shortuuid.uuid() ans_file.write(json.dumps({"question_id": idx, "text": outputs, "answer_id": ans_id, "model_id": model_name, "metadata": {}}) + "\n") ans_file.flush() ans_file.close()
null
179,651
import argparse import torch import os import json from tqdm import tqdm import shortuuid from videollava.constants import IMAGE_TOKEN_INDEX, DEFAULT_IMAGE_TOKEN, DEFAULT_IM_START_TOKEN, DEFAULT_IM_END_TOKEN from videollava.conversation import conv_templates, SeparatorStyle from videollava.model.builder import load_pretrained_model from videollava.utils import disable_torch_init from videollava.mm_utils import tokenizer_image_token, get_model_name_from_path, KeywordsStoppingCriteria from PIL import Image import math def get_chunk(lst, n, k): chunks = split_list(lst, n) return chunks[k] IMAGE_TOKEN_INDEX = -200 DEFAULT_IMAGE_TOKEN = "<image>" DEFAULT_IM_START_TOKEN = "<im_start>" DEFAULT_IM_END_TOKEN = "<im_end>" class SeparatorStyle(Enum): """Different separator style.""" SINGLE = auto() TWO = auto() MPT = auto() PLAIN = auto() LLAMA_2 = auto() conv_templates = { "default": conv_vicuna_v0, "v0": conv_vicuna_v0, "v1": conv_vicuna_v1, "vicuna_v1": conv_vicuna_v1, "llama_2": conv_llama_2, "plain": conv_llava_plain, "v0_plain": conv_llava_plain, "llava_v0": conv_llava_v0, "v0_mmtag": conv_llava_v0_mmtag, "llava_v1": conv_llava_v1, "v1_mmtag": conv_llava_v1_mmtag, "llava_llama_2": conv_llava_llama_2, "mpt": conv_mpt, } def load_pretrained_model(model_path, model_base, model_name, load_8bit=False, load_4bit=False, device_map="auto", device="cuda", **kwargs): kwargs = {"device_map": device_map, **kwargs} if device != "cuda": kwargs['device_map'] = {"": device} if load_8bit: kwargs['load_in_8bit'] = True elif load_4bit: kwargs['load_in_4bit'] = True kwargs['quantization_config'] = BitsAndBytesConfig( load_in_4bit=True, bnb_4bit_compute_dtype=torch.float16, bnb_4bit_use_double_quant=True, bnb_4bit_quant_type='nf4' ) else: kwargs['torch_dtype'] = torch.float16 if 'llava' in model_name.lower(): # Load LLaVA model if 'lora' in model_name.lower() and model_base is None: warnings.warn('There is `lora` in model name but no `model_base` is provided. If you are loading a LoRA model, please provide the `model_base` argument. Detailed instruction: https://github.com/haotian-liu/LLaVA#launch-a-model-worker-lora-weights-unmerged.') if 'lora' in model_name.lower() and model_base is not None: lora_cfg_pretrained = AutoConfig.from_pretrained(model_path) tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=False) print('Loading LLaVA from base model...') model = LlavaLlamaForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, config=lora_cfg_pretrained, **kwargs) token_num, tokem_dim = model.lm_head.out_features, model.lm_head.in_features if model.lm_head.weight.shape[0] != token_num: model.lm_head.weight = torch.nn.Parameter(torch.empty(token_num, tokem_dim, device=model.device, dtype=model.dtype)) model.model.embed_tokens.weight = torch.nn.Parameter(torch.empty(token_num, tokem_dim, device=model.device, dtype=model.dtype)) print('Loading additional LLaVA weights...') if os.path.exists(os.path.join(model_path, 'non_lora_trainables.bin')): non_lora_trainables = torch.load(os.path.join(model_path, 'non_lora_trainables.bin'), map_location='cpu') else: # this is probably from HF Hub from huggingface_hub import hf_hub_download def load_from_hf(repo_id, filename, subfolder=None): cache_file = hf_hub_download( repo_id=repo_id, filename=filename, subfolder=subfolder) return torch.load(cache_file, map_location='cpu') non_lora_trainables = load_from_hf(model_path, 'non_lora_trainables.bin') non_lora_trainables = {(k[11:] if k.startswith('base_model.') else k): v for k, v in non_lora_trainables.items()} if any(k.startswith('model.model.') for k in non_lora_trainables): non_lora_trainables = {(k[6:] if k.startswith('model.') else k): v for k, v in non_lora_trainables.items()} model.load_state_dict(non_lora_trainables, strict=False) from peft import PeftModel print('Loading LoRA weights...') model = PeftModel.from_pretrained(model, model_path) print('Merging LoRA weights...') model = model.merge_and_unload() print('Model is loaded...') elif model_base is not None: # this may be mm projector only print('Loading LLaVA from base model...') if 'mpt' in model_name.lower(): if not os.path.isfile(os.path.join(model_path, 'configuration_mpt.py')): shutil.copyfile(os.path.join(model_base, 'configuration_mpt.py'), os.path.join(model_path, 'configuration_mpt.py')) tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=True) cfg_pretrained = AutoConfig.from_pretrained(model_path, trust_remote_code=True) model = LlavaMPTForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, config=cfg_pretrained, **kwargs) else: tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=False) cfg_pretrained = AutoConfig.from_pretrained(model_path) model = LlavaLlamaForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, config=cfg_pretrained, **kwargs) mm_projector_weights = torch.load(os.path.join(model_path, 'mm_projector.bin'), map_location='cpu') mm_projector_weights = {k: v.to(torch.float16) for k, v in mm_projector_weights.items()} model.load_state_dict(mm_projector_weights, strict=False) else: if 'mpt' in model_name.lower(): tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=True) model = LlavaMPTForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs) else: tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=False) model = LlavaLlamaForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs) else: # Load language model if model_base is not None: # PEFT model from peft import PeftModel tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=False) model = AutoModelForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, **kwargs) print(f"Loading LoRA weights from {model_path}") model = PeftModel.from_pretrained(model, model_path) print(f"Merging weights") model = model.merge_and_unload() print('Convert to FP16...') model.to(torch.float16) else: use_fast = False if 'mpt' in model_name.lower(): tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=True) model = AutoModelForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, trust_remote_code=True, **kwargs) else: tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=False) model = AutoModelForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs) # ========================================================================================================== processor = {'image': None, 'video': None} if 'llava' in model_name.lower(): mm_use_im_start_end = getattr(model.config, "mm_use_im_start_end", False) mm_use_im_patch_token = getattr(model.config, "mm_use_im_patch_token", True) if mm_use_im_patch_token: tokenizer.add_tokens([DEFAULT_IMAGE_PATCH_TOKEN], special_tokens=True) tokenizer.add_tokens([DEFAULT_VIDEO_PATCH_TOKEN], special_tokens=True) if mm_use_im_start_end: tokenizer.add_tokens([DEFAULT_IM_START_TOKEN, DEFAULT_IM_END_TOKEN], special_tokens=True) tokenizer.add_tokens([DEFAULT_VID_START_TOKEN, DEFAULT_VID_END_TOKEN], special_tokens=True) model.resize_token_embeddings(len(tokenizer)) if model.config.mm_image_tower is not None: image_tower = model.get_image_tower() if not image_tower.is_loaded: image_tower.load_model() image_tower.to(device=device, dtype=torch.float16) image_processor = image_tower.image_processor processor['image'] = image_processor if model.config.mm_video_tower is not None: video_tower = model.get_video_tower() if not video_tower.is_loaded: video_tower.load_model() video_tower.to(device=device, dtype=torch.float16) video_processor = video_tower.video_processor processor['video'] = video_processor # ========================================================================================================== if hasattr(model.config, "max_sequence_length"): context_len = model.config.max_sequence_length else: context_len = 2048 return tokenizer, model, processor, context_len def disable_torch_init(): """ Disable the redundant torch default initialization to accelerate model creation. """ import torch setattr(torch.nn.Linear, "reset_parameters", lambda self: None) setattr(torch.nn.LayerNorm, "reset_parameters", lambda self: None) def tokenizer_image_token(prompt, tokenizer, image_token_index=IMAGE_TOKEN_INDEX, return_tensors=None): prompt_chunks = [tokenizer(chunk).input_ids for chunk in prompt.split('<image>')] def insert_separator(X, sep): return [ele for sublist in zip(X, [sep]*len(X)) for ele in sublist][:-1] input_ids = [] offset = 0 if len(prompt_chunks) > 0 and len(prompt_chunks[0]) > 0 and prompt_chunks[0][0] == tokenizer.bos_token_id: offset = 1 input_ids.append(prompt_chunks[0][0]) for x in insert_separator(prompt_chunks, [image_token_index] * (offset + 1)): input_ids.extend(x[offset:]) if return_tensors is not None: if return_tensors == 'pt': return torch.tensor(input_ids, dtype=torch.long) raise ValueError(f'Unsupported tensor type: {return_tensors}') return input_ids def get_model_name_from_path(model_path): model_path = model_path.strip("/") model_paths = model_path.split("/") if model_paths[-1].startswith('checkpoint-'): return model_paths[-2] + "_" + model_paths[-1] else: return model_paths[-1] class KeywordsStoppingCriteria(StoppingCriteria): def __init__(self, keywords, tokenizer, input_ids): self.keywords = keywords self.keyword_ids = [] self.max_keyword_len = 0 for keyword in keywords: cur_keyword_ids = tokenizer(keyword).input_ids if len(cur_keyword_ids) > 1 and cur_keyword_ids[0] == tokenizer.bos_token_id: cur_keyword_ids = cur_keyword_ids[1:] if len(cur_keyword_ids) > self.max_keyword_len: self.max_keyword_len = len(cur_keyword_ids) self.keyword_ids.append(torch.tensor(cur_keyword_ids)) self.tokenizer = tokenizer self.start_len = input_ids.shape[1] def call_for_batch(self, output_ids: torch.LongTensor, scores: torch.FloatTensor, **kwargs) -> bool: offset = min(output_ids.shape[1] - self.start_len, self.max_keyword_len) self.keyword_ids = [keyword_id.to(output_ids.device) for keyword_id in self.keyword_ids] for keyword_id in self.keyword_ids: if (output_ids[0, -keyword_id.shape[0]:] == keyword_id).all(): return True outputs = self.tokenizer.batch_decode(output_ids[:, -offset:], skip_special_tokens=True)[0] for keyword in self.keywords: if keyword in outputs: return True return False def __call__(self, output_ids: torch.LongTensor, scores: torch.FloatTensor, **kwargs) -> bool: outputs = [] for i in range(output_ids.shape[0]): outputs.append(self.call_for_batch(output_ids[i].unsqueeze(0), scores)) return all(outputs) def eval_model(args): # Model disable_torch_init() model_path = os.path.expanduser(args.model_path) model_name = get_model_name_from_path(model_path) tokenizer, model, processor, context_len = load_pretrained_model(model_path, args.model_base, model_name) image_processor = processor['image'] questions = [json.loads(q) for q in open(os.path.expanduser(args.question_file), "r")] questions = get_chunk(questions, args.num_chunks, args.chunk_idx) answers_file = os.path.expanduser(args.answers_file) os.makedirs(os.path.dirname(answers_file), exist_ok=True) ans_file = open(answers_file, "w") for line in tqdm(questions): idx = line["question_id"] image_file = line["image"] qs = line["text"] cur_prompt = qs if model.config.mm_use_im_start_end: qs = DEFAULT_IM_START_TOKEN + DEFAULT_IMAGE_TOKEN + DEFAULT_IM_END_TOKEN + '\n' + qs else: qs = DEFAULT_IMAGE_TOKEN + '\n' + qs conv = conv_templates[args.conv_mode].copy() conv.append_message(conv.roles[0], qs) conv.append_message(conv.roles[1], None) prompt = conv.get_prompt() input_ids = tokenizer_image_token(prompt, tokenizer, IMAGE_TOKEN_INDEX, return_tensors='pt').unsqueeze(0).cuda() image = Image.open(os.path.join(args.image_folder, image_file)) image_tensor = image_processor.preprocess(image, return_tensors='pt')['pixel_values'][0] stop_str = conv.sep if conv.sep_style != SeparatorStyle.TWO else conv.sep2 keywords = [stop_str] stopping_criteria = KeywordsStoppingCriteria(keywords, tokenizer, input_ids) with torch.inference_mode(): output_ids = model.generate( input_ids, images=image_tensor.unsqueeze(0).half().cuda(), do_sample=True if args.temperature > 0 else False, temperature=args.temperature, top_p=args.top_p, num_beams=args.num_beams, # no_repeat_ngram_size=3, max_new_tokens=1024, use_cache=True) input_token_len = input_ids.shape[1] n_diff_input_output = (input_ids != output_ids[:, :input_token_len]).sum().item() if n_diff_input_output > 0: print(f'[Warning] {n_diff_input_output} output_ids are not the same as the input_ids') outputs = tokenizer.batch_decode(output_ids[:, input_token_len:], skip_special_tokens=True)[0] outputs = outputs.strip() if outputs.endswith(stop_str): outputs = outputs[:-len(stop_str)] outputs = outputs.strip() ans_id = shortuuid.uuid() ans_file.write(json.dumps({"question_id": idx, "prompt": cur_prompt, "text": outputs, "answer_id": ans_id, "model_id": model_name, "metadata": {}}) + "\n") ans_file.flush() ans_file.close()
null
179,652
import argparse import torch import os import json import pandas as pd from tqdm import tqdm import shortuuid from videollava.constants import IMAGE_TOKEN_INDEX, DEFAULT_IMAGE_TOKEN, DEFAULT_IM_START_TOKEN, DEFAULT_IM_END_TOKEN from videollava.conversation import conv_templates, SeparatorStyle from videollava.model.builder import load_pretrained_model from videollava.utils import disable_torch_init from videollava.mm_utils import tokenizer_image_token, process_images, load_image_from_base64, get_model_name_from_path from PIL import Image import math all_options = ['A', 'B', 'C', 'D'] def get_chunk(lst, n, k): chunks = split_list(lst, n) return chunks[k] def is_none(value): if value is None: return True if type(value) is float and math.isnan(value): return True if type(value) is str and value.lower() == 'nan': return True if type(value) is str and value.lower() == 'none': return True return False def get_options(row, options): parsed_options = [] for option in options: option_value = row[option] if is_none(option_value): break parsed_options.append(option_value) return parsed_options IMAGE_TOKEN_INDEX = -200 DEFAULT_IMAGE_TOKEN = "<image>" DEFAULT_IM_START_TOKEN = "<im_start>" DEFAULT_IM_END_TOKEN = "<im_end>" class SeparatorStyle(Enum): """Different separator style.""" SINGLE = auto() TWO = auto() MPT = auto() PLAIN = auto() LLAMA_2 = auto() conv_templates = { "default": conv_vicuna_v0, "v0": conv_vicuna_v0, "v1": conv_vicuna_v1, "vicuna_v1": conv_vicuna_v1, "llama_2": conv_llama_2, "plain": conv_llava_plain, "v0_plain": conv_llava_plain, "llava_v0": conv_llava_v0, "v0_mmtag": conv_llava_v0_mmtag, "llava_v1": conv_llava_v1, "v1_mmtag": conv_llava_v1_mmtag, "llava_llama_2": conv_llava_llama_2, "mpt": conv_mpt, } def load_pretrained_model(model_path, model_base, model_name, load_8bit=False, load_4bit=False, device_map="auto", device="cuda", **kwargs): kwargs = {"device_map": device_map, **kwargs} if device != "cuda": kwargs['device_map'] = {"": device} if load_8bit: kwargs['load_in_8bit'] = True elif load_4bit: kwargs['load_in_4bit'] = True kwargs['quantization_config'] = BitsAndBytesConfig( load_in_4bit=True, bnb_4bit_compute_dtype=torch.float16, bnb_4bit_use_double_quant=True, bnb_4bit_quant_type='nf4' ) else: kwargs['torch_dtype'] = torch.float16 if 'llava' in model_name.lower(): # Load LLaVA model if 'lora' in model_name.lower() and model_base is None: warnings.warn('There is `lora` in model name but no `model_base` is provided. If you are loading a LoRA model, please provide the `model_base` argument. Detailed instruction: https://github.com/haotian-liu/LLaVA#launch-a-model-worker-lora-weights-unmerged.') if 'lora' in model_name.lower() and model_base is not None: lora_cfg_pretrained = AutoConfig.from_pretrained(model_path) tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=False) print('Loading LLaVA from base model...') model = LlavaLlamaForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, config=lora_cfg_pretrained, **kwargs) token_num, tokem_dim = model.lm_head.out_features, model.lm_head.in_features if model.lm_head.weight.shape[0] != token_num: model.lm_head.weight = torch.nn.Parameter(torch.empty(token_num, tokem_dim, device=model.device, dtype=model.dtype)) model.model.embed_tokens.weight = torch.nn.Parameter(torch.empty(token_num, tokem_dim, device=model.device, dtype=model.dtype)) print('Loading additional LLaVA weights...') if os.path.exists(os.path.join(model_path, 'non_lora_trainables.bin')): non_lora_trainables = torch.load(os.path.join(model_path, 'non_lora_trainables.bin'), map_location='cpu') else: # this is probably from HF Hub from huggingface_hub import hf_hub_download def load_from_hf(repo_id, filename, subfolder=None): cache_file = hf_hub_download( repo_id=repo_id, filename=filename, subfolder=subfolder) return torch.load(cache_file, map_location='cpu') non_lora_trainables = load_from_hf(model_path, 'non_lora_trainables.bin') non_lora_trainables = {(k[11:] if k.startswith('base_model.') else k): v for k, v in non_lora_trainables.items()} if any(k.startswith('model.model.') for k in non_lora_trainables): non_lora_trainables = {(k[6:] if k.startswith('model.') else k): v for k, v in non_lora_trainables.items()} model.load_state_dict(non_lora_trainables, strict=False) from peft import PeftModel print('Loading LoRA weights...') model = PeftModel.from_pretrained(model, model_path) print('Merging LoRA weights...') model = model.merge_and_unload() print('Model is loaded...') elif model_base is not None: # this may be mm projector only print('Loading LLaVA from base model...') if 'mpt' in model_name.lower(): if not os.path.isfile(os.path.join(model_path, 'configuration_mpt.py')): shutil.copyfile(os.path.join(model_base, 'configuration_mpt.py'), os.path.join(model_path, 'configuration_mpt.py')) tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=True) cfg_pretrained = AutoConfig.from_pretrained(model_path, trust_remote_code=True) model = LlavaMPTForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, config=cfg_pretrained, **kwargs) else: tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=False) cfg_pretrained = AutoConfig.from_pretrained(model_path) model = LlavaLlamaForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, config=cfg_pretrained, **kwargs) mm_projector_weights = torch.load(os.path.join(model_path, 'mm_projector.bin'), map_location='cpu') mm_projector_weights = {k: v.to(torch.float16) for k, v in mm_projector_weights.items()} model.load_state_dict(mm_projector_weights, strict=False) else: if 'mpt' in model_name.lower(): tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=True) model = LlavaMPTForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs) else: tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=False) model = LlavaLlamaForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs) else: # Load language model if model_base is not None: # PEFT model from peft import PeftModel tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=False) model = AutoModelForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, **kwargs) print(f"Loading LoRA weights from {model_path}") model = PeftModel.from_pretrained(model, model_path) print(f"Merging weights") model = model.merge_and_unload() print('Convert to FP16...') model.to(torch.float16) else: use_fast = False if 'mpt' in model_name.lower(): tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=True) model = AutoModelForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, trust_remote_code=True, **kwargs) else: tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=False) model = AutoModelForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs) # ========================================================================================================== processor = {'image': None, 'video': None} if 'llava' in model_name.lower(): mm_use_im_start_end = getattr(model.config, "mm_use_im_start_end", False) mm_use_im_patch_token = getattr(model.config, "mm_use_im_patch_token", True) if mm_use_im_patch_token: tokenizer.add_tokens([DEFAULT_IMAGE_PATCH_TOKEN], special_tokens=True) tokenizer.add_tokens([DEFAULT_VIDEO_PATCH_TOKEN], special_tokens=True) if mm_use_im_start_end: tokenizer.add_tokens([DEFAULT_IM_START_TOKEN, DEFAULT_IM_END_TOKEN], special_tokens=True) tokenizer.add_tokens([DEFAULT_VID_START_TOKEN, DEFAULT_VID_END_TOKEN], special_tokens=True) model.resize_token_embeddings(len(tokenizer)) if model.config.mm_image_tower is not None: image_tower = model.get_image_tower() if not image_tower.is_loaded: image_tower.load_model() image_tower.to(device=device, dtype=torch.float16) image_processor = image_tower.image_processor processor['image'] = image_processor if model.config.mm_video_tower is not None: video_tower = model.get_video_tower() if not video_tower.is_loaded: video_tower.load_model() video_tower.to(device=device, dtype=torch.float16) video_processor = video_tower.video_processor processor['video'] = video_processor # ========================================================================================================== if hasattr(model.config, "max_sequence_length"): context_len = model.config.max_sequence_length else: context_len = 2048 return tokenizer, model, processor, context_len def disable_torch_init(): """ Disable the redundant torch default initialization to accelerate model creation. """ import torch setattr(torch.nn.Linear, "reset_parameters", lambda self: None) setattr(torch.nn.LayerNorm, "reset_parameters", lambda self: None) def load_image_from_base64(image): return Image.open(BytesIO(base64.b64decode(image))) def process_images(images, image_processor, model_cfg): image_aspect_ratio = getattr(model_cfg, "image_aspect_ratio", None) new_images = [] if image_aspect_ratio == 'pad': for image in images: image = expand2square(image, tuple(int(x*255) for x in image_processor.image_mean)) image = image_processor.preprocess(image, return_tensors='pt')['pixel_values'][0] new_images.append(image) else: return image_processor(images, return_tensors='pt')['pixel_values'] if all(x.shape == new_images[0].shape for x in new_images): new_images = torch.stack(new_images, dim=0) return new_images def tokenizer_image_token(prompt, tokenizer, image_token_index=IMAGE_TOKEN_INDEX, return_tensors=None): prompt_chunks = [tokenizer(chunk).input_ids for chunk in prompt.split('<image>')] def insert_separator(X, sep): return [ele for sublist in zip(X, [sep]*len(X)) for ele in sublist][:-1] input_ids = [] offset = 0 if len(prompt_chunks) > 0 and len(prompt_chunks[0]) > 0 and prompt_chunks[0][0] == tokenizer.bos_token_id: offset = 1 input_ids.append(prompt_chunks[0][0]) for x in insert_separator(prompt_chunks, [image_token_index] * (offset + 1)): input_ids.extend(x[offset:]) if return_tensors is not None: if return_tensors == 'pt': return torch.tensor(input_ids, dtype=torch.long) raise ValueError(f'Unsupported tensor type: {return_tensors}') return input_ids def get_model_name_from_path(model_path): model_path = model_path.strip("/") model_paths = model_path.split("/") if model_paths[-1].startswith('checkpoint-'): return model_paths[-2] + "_" + model_paths[-1] else: return model_paths[-1] def eval_model(args): # Model disable_torch_init() model_path = os.path.expanduser(args.model_path) model_name = get_model_name_from_path(model_path) tokenizer, model, processor, context_len = load_pretrained_model(model_path, args.model_base, model_name) image_processor = processor['image'] questions = pd.read_table(os.path.expanduser(args.question_file)) questions = get_chunk(questions, args.num_chunks, args.chunk_idx) answers_file = os.path.expanduser(args.answers_file) os.makedirs(os.path.dirname(answers_file), exist_ok=True) ans_file = open(answers_file, "w") if 'plain' in model_name and 'finetune' not in model_name.lower() and 'mmtag' not in args.conv_mode: args.conv_mode = args.conv_mode + '_mmtag' print(f'It seems that this is a plain model, but it is not using a mmtag prompt, auto switching to {args.conv_mode}.') for index, row in tqdm(questions.iterrows(), total=len(questions)): options = get_options(row, all_options) cur_option_char = all_options[:len(options)] if args.all_rounds: num_rounds = len(options) else: num_rounds = 1 for round_idx in range(num_rounds): idx = row['index'] question = row['question'] hint = row['hint'] image = load_image_from_base64(row['image']) if not is_none(hint): question = hint + '\n' + question for option_char, option in zip(all_options[:len(options)], options): question = question + '\n' + option_char + '. ' + option qs = cur_prompt = question if model.config.mm_use_im_start_end: qs = DEFAULT_IM_START_TOKEN + DEFAULT_IMAGE_TOKEN + DEFAULT_IM_END_TOKEN + '\n' + qs else: qs = DEFAULT_IMAGE_TOKEN + '\n' + qs if args.single_pred_prompt: if args.lang == 'cn': qs = qs + '\n' + "请直接回答选项字母。" else: qs = qs + '\n' + "Answer with the option's letter from the given choices directly." conv = conv_templates[args.conv_mode].copy() conv.append_message(conv.roles[0], qs) conv.append_message(conv.roles[1], None) prompt = conv.get_prompt() input_ids = tokenizer_image_token(prompt, tokenizer, IMAGE_TOKEN_INDEX, return_tensors='pt').unsqueeze(0).cuda() image_tensor = process_images([image], image_processor, model.config)[0] # image_tensor = image_processor.preprocess(image, return_tensors='pt')['pixel_values'][0] stop_str = conv.sep if conv.sep_style != SeparatorStyle.TWO else conv.sep2 with torch.inference_mode(): output_ids = model.generate( input_ids, images=image_tensor.unsqueeze(0).half().cuda(), do_sample=True if args.temperature > 0 else False, temperature=args.temperature, top_p=args.top_p, num_beams=args.num_beams, # no_repeat_ngram_size=3, max_new_tokens=1024, use_cache=True) input_token_len = input_ids.shape[1] n_diff_input_output = (input_ids != output_ids[:, :input_token_len]).sum().item() if n_diff_input_output > 0: print(f'[Warning] {n_diff_input_output} output_ids are not the same as the input_ids') outputs = tokenizer.batch_decode(output_ids[:, input_token_len:], skip_special_tokens=True)[0] outputs = outputs.strip() if outputs.endswith(stop_str): outputs = outputs[:-len(stop_str)] outputs = outputs.strip() ans_id = shortuuid.uuid() ans_file.write(json.dumps({"question_id": idx, "round_id": round_idx, "prompt": cur_prompt, "text": outputs, "options": options, "option_char": cur_option_char, "answer_id": ans_id, "model_id": model_name, "metadata": {}}) + "\n") ans_file.flush() # rotate options options = options[1:] + options[:1] cur_option_char = cur_option_char[1:] + cur_option_char[:1] ans_file.close()
null
179,653
import argparse import openai import json import os from tqdm import tqdm import pandas as pd import numpy as np from collections import Counter import time len_data = 0 num_run = 1 for k, v in grade_results.items(): if sub_set is not None and k not in sub_set: continue for i in range(num_run): score = v['score'][i] caps = set(data[k]['capability']) for c in caps: cap_socres[c][i] += score cap_socres['total'][i] += score index = cap_set_list.index(caps) cap_socres2[cap_set_names[index]][i] += score cap_socres2['total'][i] += score for k, v in cap_socres.items(): cap_socres[k] = np.array(v) / counter[k] * 100 for k, v in cap_socres.items(): cap_socres[k] = round(v.mean(), decimal_places) for k, v in cap_socres2.items(): cap_socres2[k] = round(np.mean(np.array(v) / counter2[k] * 100), decimal_places) def need_more_runs(): need_more_runs = False if len(grade_results) > 0: for k, v in grade_results.items(): if len(v['score']) < num_run: need_more_runs = True break return need_more_runs or len(grade_results) < len_data
null
179,654
from collections import defaultdict from tqdm import tqdm import argparse import os.path import glob import json import math def loadFile(name): # load standard json file if os.path.isfile(name): with open(name) as file: data = json.load(file) # load file chunks if too big elif os.path.isdir(name.split(".")[0]): data = {} chunks = glob.glob('{dir}/{dir}_*.{ext}'.format(dir=name.split(".")[0], ext=name.split(".")[1])) for chunk in chunks: with open(chunk) as file: data.update(json.load(file)) else: raise Exception("Can't find {}".format(name)) return data
null
179,655
from collections import defaultdict from tqdm import tqdm import argparse import os.path import glob import json import math def wavg(l, w): if sum(w) == 0: return None return float(sum(l[i] * w[i] for i in range(len(l)))) / sum(w)
null
179,656
from collections import defaultdict from tqdm import tqdm import argparse import os.path import glob import json import math def getWordsNum(question): return len(question["question"].split())
null
179,657
from collections import defaultdict from tqdm import tqdm import argparse import os.path import glob import json import math def getStepsNum(question): return len([c for c in question["semantic"] if not (any([o in "{}: {}".format(c["operation"], c["argument"]) for o in ["exist", "query: name", "choose name"]]))])
null
179,658
from collections import defaultdict from tqdm import tqdm import argparse import os.path import glob import json import math def toSlice(strSlice): sliceLims = (int(n) for n in strSlice.split(':')) return apply(slice, sliceLims)
null
179,659
from collections import defaultdict from tqdm import tqdm import argparse import os.path import glob import json import math def intsFromSlice(strSlice): slice_obj = get_slice_obj(slicearg) return (range(slice_obj.start or 0, slice_obj.stop or -1, slice_obj.step or 1))
null
179,660
from collections import defaultdict from tqdm import tqdm import argparse import os.path import glob import json import math def belongs(element, group, question): # normalization () if "Common" in question["types"]["detailed"]: group = ["color", "material", "shape"] return element in group
null
179,661
from collections import defaultdict from tqdm import tqdm import argparse import os.path import glob import json import math predictions = loadFile(args.predictions.format(tier=args.tier)) predictions = {p["questionId"]: p["prediction"] for p in predictions} def toScore(b): return float(1 if b else 0) def avg(l): if len(l) == 0: return 0 return float(sum(l)) / len(l) scores = { "accuracy": [], # list of accuracies per question (1 if correct else 0). Will be averaged ultimately. "binary": [], # list of accuracies per a binary question (1 if correct else 0). Will be averaged ultimately. "open": [], # list of accuracies per an open question (1 if correct else 0). Will be averaged ultimately. "validity": [], # list of validity per question (1 if valid else 0). "plausibility": [], # list of plausibility per question (1 if plausible else 0). "consistency": [], # list of consistency scores for entailed questions. "accuracyPerStructuralType": defaultdict(list), # list of question accuracies for each structural type (e.g. compare, logic questions). "accuracyPerSemanticType": defaultdict(list), # list of question accuracies for each semantic type (e.g. questions about an object, an attribute, a relation). "accuracyPerLength": defaultdict(list), # list of question accuracies per question's word number. "accuracyPerSteps": defaultdict(list), # list of question accuracies per question's reasoning length (steps number). "grounding": [] # list of grounding scores for each question. } scores["distribution"] = chiSquare(dist["gold"], dist["predicted"]) / 100 def updateConsistency(questionId, question, questions): inferredQuestions = [eid for eid in question["entailed"] if eid != questionId] if correct and len(inferredQuestions) > 0: cosnsitencyScores = [] for eid in inferredQuestions: gold = questions[eid]["answer"] predicted = predictions[eid] score = toScore(predicted == gold) cosnsitencyScores.append(score) scores["consistency"].append(avg(cosnsitencyScores))
null
179,662
from collections import defaultdict from tqdm import tqdm import argparse import os.path import glob import json import math args = parser.parse_args() if not args.consistency: print("Please consider using --consistency to compute consistency scores for entailed questions.") print("If you do so, please provide answers to all questions in val_all_questions.json.\n") if not args.grounding: print("Please consider using --grounding to compute attention scores.") print("If you do so, please provide attention maps through --attentions.\n") if args.grounding: with open(args.attentions.format(tier=args.tier)) as attentionsFile: attentions = json.load(attentionsFile) attentions = {a["questionId"]: a["attention"] for a in attentions} scores = { "accuracy": [], # list of accuracies per question (1 if correct else 0). Will be averaged ultimately. "binary": [], # list of accuracies per a binary question (1 if correct else 0). Will be averaged ultimately. "open": [], # list of accuracies per an open question (1 if correct else 0). Will be averaged ultimately. "validity": [], # list of validity per question (1 if valid else 0). "plausibility": [], # list of plausibility per question (1 if plausible else 0). "consistency": [], # list of consistency scores for entailed questions. "accuracyPerStructuralType": defaultdict(list), # list of question accuracies for each structural type (e.g. compare, logic questions). "accuracyPerSemanticType": defaultdict(list), # list of question accuracies for each semantic type (e.g. questions about an object, an attribute, a relation). "accuracyPerLength": defaultdict(list), # list of question accuracies per question's word number. "accuracyPerSteps": defaultdict(list), # list of question accuracies per question's reasoning length (steps number). "grounding": [] # list of grounding scores for each question. } def intersectionRate(c1, c2): return float(intersectionSize(c1, c2)) / size(c1) def getCell(i, j): edge = float(1) / args.mapSize return (edge * i, edge * j, edge * (i + 1), edge * (j + 1)) def getRegion(sceneGraph, objectId): obj = sceneGraph["objects"][objectId] x0 = float(obj["x"]) / sceneGraph["width"] y0 = float(obj["y"]) / sceneGraph["height"] x1 = float(obj["x"] + obj["w"]) / sceneGraph["width"] y1 = float(obj["y"] + obj["h"]) / sceneGraph["height"] return (x0, y0, x1, y1) scores["distribution"] = chiSquare(dist["gold"], dist["predicted"]) / 100 def computeGroundingScore(question, sceneGraph, attentionMap): ## prepare gold regions regions = [] # add question regions regions += [getRegion(sceneGraph, pointer) for pointer in question["annotations"]["question"].values()] # add answer regions regions += [getRegion(sceneGraph, pointer) for pointer in question["annotations"]["fullAnswer"].values()] # add all the image if the question refers to the whole scene if any(("scene" in c) for c in question["semantic"]): regions.append((0, 0, 1, 1)) # prepare attention map if args.objectFeatures: cells = [((x0, y0, x1, y1), attention) for x0, y0, x1, y1, attention in cells] else: cells = [(getCell(i, j), attentionMap[i][j]) for i in range(args.mapSize) for j in range(args.mapSize)] # compare attention map to gold regions scores = [] for region in regions: for cell, attention in cells: scores.append(attention * intersectionRate(cell, region)) return sum(scores)
null
179,663
from collections import defaultdict from tqdm import tqdm import argparse import os.path import glob import json import math def chiSquare(goldDist, predictedDist): sumScore, sumOverall = 0, 0 for group in goldDist: score, overall = 0, 0 for ans in goldDist[group]: e = goldDist[group][ans] o = predictedDist[group].get(ans, 0) score += ((float(o - e) ** 2) / e) overall += goldDist[group][ans] sumScore += score * overall sumOverall += overall avgScore = float(sumScore) / sumOverall return avgScore
null
179,664
import os import argparse import json from tqdm import tqdm from videollava.eval.video.run_inference_video_qa import get_model_output from videollava.mm_utils import get_model_name_from_path from videollava.model.builder import load_pretrained_model The provided code snippet includes necessary dependencies for implementing the `parse_args` function. Write a Python function `def parse_args()` to solve the following problem: Parse command-line arguments. Here is the function: def parse_args(): """ Parse command-line arguments. """ parser = argparse.ArgumentParser() # Define the command-line arguments parser.add_argument('--model_path', help='', required=True) parser.add_argument('--cache_dir', help='', required=True) parser.add_argument('--video_dir', help='Directory containing video files.', required=True) parser.add_argument('--gt_file', help='Path to the ground truth file.', required=True) parser.add_argument('--output_dir', help='Directory to save the model results JSON.', required=True) parser.add_argument('--output_name', help='Name of the file for storing results JSON.', required=True) # parser.add_argument("--model-name", type=str, required=True) parser.add_argument("--device", type=str, required=False, default='cuda:0') parser.add_argument('--model_base', help='', default=None, type=str, required=False) parser.add_argument("--model_max_length", type=int, required=False, default=2048) # parser.add_argument("--conv-mode", type=str, required=False, default='video-chatgpt_v1') # parser.add_argument("--projection_path", type=str, required=True) return parser.parse_args()
Parse command-line arguments.
179,665
import os import argparse import json from tqdm import tqdm from videollava.eval.video.run_inference_video_qa import get_model_output from videollava.mm_utils import get_model_name_from_path from videollava.model.builder import load_pretrained_model def get_model_output(model, video_processor, tokenizer, video, qs, args): if model.config.mm_use_im_start_end: qs = DEFAULT_VID_START_TOKEN + ''.join([DEFAULT_IMAGE_TOKEN]*8) + DEFAULT_VID_END_TOKEN + '\n' + qs else: qs = ''.join([DEFAULT_IMAGE_TOKEN]*8) + '\n' + qs conv_mode = "llava_v1" args.conv_mode = conv_mode conv = conv_templates[args.conv_mode].copy() conv.append_message(conv.roles[0], qs) conv.append_message(conv.roles[1], None) prompt = conv.get_prompt() video_tensor = video_processor.preprocess(video, return_tensors='pt')['pixel_values'][0].half().to(args.device) # print(video_tensor.shape) input_ids = tokenizer_image_token(prompt, tokenizer, IMAGE_TOKEN_INDEX, return_tensors='pt').unsqueeze(0).to(args.device) stop_str = conv.sep if conv.sep_style != SeparatorStyle.TWO else conv.sep2 keywords = [stop_str] stopping_criteria = KeywordsStoppingCriteria(keywords, tokenizer, input_ids) with torch.inference_mode(): output_ids = model.generate( input_ids, images=[video_tensor], do_sample=True, temperature=0.0, max_new_tokens=1024, use_cache=True, stopping_criteria=[stopping_criteria]) input_token_len = input_ids.shape[1] n_diff_input_output = (input_ids != output_ids[:, :input_token_len]).sum().item() if n_diff_input_output > 0: print(f'[Warning] {n_diff_input_output} output_ids are not the same as the input_ids') outputs = tokenizer.batch_decode(output_ids[:, input_token_len:], skip_special_tokens=True)[0] outputs = outputs.strip() if outputs.endswith(stop_str): outputs = outputs[:-len(stop_str)] outputs = outputs.strip() print(outputs) return outputs def get_model_name_from_path(model_path): model_path = model_path.strip("/") model_paths = model_path.split("/") if model_paths[-1].startswith('checkpoint-'): return model_paths[-2] + "_" + model_paths[-1] else: return model_paths[-1] def load_pretrained_model(model_path, model_base, model_name, load_8bit=False, load_4bit=False, device_map="auto", device="cuda", **kwargs): kwargs = {"device_map": device_map, **kwargs} if device != "cuda": kwargs['device_map'] = {"": device} if load_8bit: kwargs['load_in_8bit'] = True elif load_4bit: kwargs['load_in_4bit'] = True kwargs['quantization_config'] = BitsAndBytesConfig( load_in_4bit=True, bnb_4bit_compute_dtype=torch.float16, bnb_4bit_use_double_quant=True, bnb_4bit_quant_type='nf4' ) else: kwargs['torch_dtype'] = torch.float16 if 'llava' in model_name.lower(): # Load LLaVA model if 'lora' in model_name.lower() and model_base is None: warnings.warn('There is `lora` in model name but no `model_base` is provided. If you are loading a LoRA model, please provide the `model_base` argument. Detailed instruction: https://github.com/haotian-liu/LLaVA#launch-a-model-worker-lora-weights-unmerged.') if 'lora' in model_name.lower() and model_base is not None: lora_cfg_pretrained = AutoConfig.from_pretrained(model_path) tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=False) print('Loading LLaVA from base model...') model = LlavaLlamaForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, config=lora_cfg_pretrained, **kwargs) token_num, tokem_dim = model.lm_head.out_features, model.lm_head.in_features if model.lm_head.weight.shape[0] != token_num: model.lm_head.weight = torch.nn.Parameter(torch.empty(token_num, tokem_dim, device=model.device, dtype=model.dtype)) model.model.embed_tokens.weight = torch.nn.Parameter(torch.empty(token_num, tokem_dim, device=model.device, dtype=model.dtype)) print('Loading additional LLaVA weights...') if os.path.exists(os.path.join(model_path, 'non_lora_trainables.bin')): non_lora_trainables = torch.load(os.path.join(model_path, 'non_lora_trainables.bin'), map_location='cpu') else: # this is probably from HF Hub from huggingface_hub import hf_hub_download def load_from_hf(repo_id, filename, subfolder=None): cache_file = hf_hub_download( repo_id=repo_id, filename=filename, subfolder=subfolder) return torch.load(cache_file, map_location='cpu') non_lora_trainables = load_from_hf(model_path, 'non_lora_trainables.bin') non_lora_trainables = {(k[11:] if k.startswith('base_model.') else k): v for k, v in non_lora_trainables.items()} if any(k.startswith('model.model.') for k in non_lora_trainables): non_lora_trainables = {(k[6:] if k.startswith('model.') else k): v for k, v in non_lora_trainables.items()} model.load_state_dict(non_lora_trainables, strict=False) from peft import PeftModel print('Loading LoRA weights...') model = PeftModel.from_pretrained(model, model_path) print('Merging LoRA weights...') model = model.merge_and_unload() print('Model is loaded...') elif model_base is not None: # this may be mm projector only print('Loading LLaVA from base model...') if 'mpt' in model_name.lower(): if not os.path.isfile(os.path.join(model_path, 'configuration_mpt.py')): shutil.copyfile(os.path.join(model_base, 'configuration_mpt.py'), os.path.join(model_path, 'configuration_mpt.py')) tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=True) cfg_pretrained = AutoConfig.from_pretrained(model_path, trust_remote_code=True) model = LlavaMPTForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, config=cfg_pretrained, **kwargs) else: tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=False) cfg_pretrained = AutoConfig.from_pretrained(model_path) model = LlavaLlamaForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, config=cfg_pretrained, **kwargs) mm_projector_weights = torch.load(os.path.join(model_path, 'mm_projector.bin'), map_location='cpu') mm_projector_weights = {k: v.to(torch.float16) for k, v in mm_projector_weights.items()} model.load_state_dict(mm_projector_weights, strict=False) else: if 'mpt' in model_name.lower(): tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=True) model = LlavaMPTForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs) else: tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=False) model = LlavaLlamaForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs) else: # Load language model if model_base is not None: # PEFT model from peft import PeftModel tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=False) model = AutoModelForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, **kwargs) print(f"Loading LoRA weights from {model_path}") model = PeftModel.from_pretrained(model, model_path) print(f"Merging weights") model = model.merge_and_unload() print('Convert to FP16...') model.to(torch.float16) else: use_fast = False if 'mpt' in model_name.lower(): tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=True) model = AutoModelForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, trust_remote_code=True, **kwargs) else: tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=False) model = AutoModelForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs) # ========================================================================================================== processor = {'image': None, 'video': None} if 'llava' in model_name.lower(): mm_use_im_start_end = getattr(model.config, "mm_use_im_start_end", False) mm_use_im_patch_token = getattr(model.config, "mm_use_im_patch_token", True) if mm_use_im_patch_token: tokenizer.add_tokens([DEFAULT_IMAGE_PATCH_TOKEN], special_tokens=True) tokenizer.add_tokens([DEFAULT_VIDEO_PATCH_TOKEN], special_tokens=True) if mm_use_im_start_end: tokenizer.add_tokens([DEFAULT_IM_START_TOKEN, DEFAULT_IM_END_TOKEN], special_tokens=True) tokenizer.add_tokens([DEFAULT_VID_START_TOKEN, DEFAULT_VID_END_TOKEN], special_tokens=True) model.resize_token_embeddings(len(tokenizer)) if model.config.mm_image_tower is not None: image_tower = model.get_image_tower() if not image_tower.is_loaded: image_tower.load_model() image_tower.to(device=device, dtype=torch.float16) image_processor = image_tower.image_processor processor['image'] = image_processor if model.config.mm_video_tower is not None: video_tower = model.get_video_tower() if not video_tower.is_loaded: video_tower.load_model() video_tower.to(device=device, dtype=torch.float16) video_processor = video_tower.video_processor processor['video'] = video_processor # ========================================================================================================== if hasattr(model.config, "max_sequence_length"): context_len = model.config.max_sequence_length else: context_len = 2048 return tokenizer, model, processor, context_len The provided code snippet includes necessary dependencies for implementing the `run_inference` function. Write a Python function `def run_inference(args)` to solve the following problem: Run inference on a set of video files using the provided model. Args: args: Command-line arguments. Here is the function: def run_inference(args): """ Run inference on a set of video files using the provided model. Args: args: Command-line arguments. """# Initialize the model model_name = get_model_name_from_path(args.model_path) tokenizer, model, processor, context_len = load_pretrained_model(args.model_path, args.model_base, model_name) model = model.to(args.device) # Load the ground truth file with open(args.gt_file) as file: gt_contents = json.load(file) # Create the output directory if it doesn't exist if not os.path.exists(args.output_dir): os.makedirs(args.output_dir) output_list = [] # List to store the output results # conv_mode = args.conv_mode video_formats = ['.mp4', '.avi', '.mov', '.mkv'] # Iterate over each sample in the ground truth file for sample in tqdm(gt_contents): video_name = sample['video_name'] sample_set = sample question = sample['Q'] try: # Load the video file for fmt in video_formats: # Added this line temp_path = os.path.join(args.video_dir, f"{video_name}{fmt}") if os.path.exists(temp_path): video_path = temp_path output = get_model_output(model, processor['video'], tokenizer, video_path, question, args) sample_set['pred'] = output output_list.append(sample_set) break except Exception as e: print(f"Error processing video file '{video_name}': {e}") # Save the output list to a JSON file with open(os.path.join(args.output_dir, f"{args.output_name}.json"), 'w') as file: json.dump(output_list, file)
Run inference on a set of video files using the provided model. Args: args: Command-line arguments.
179,666
import openai import os import argparse import json import ast from multiprocessing.pool import Pool def parse_args(): parser = argparse.ArgumentParser(description="question-answer-generation-using-gpt-3") parser.add_argument("--pred_path", required=True, help="The path to file containing prediction.") parser.add_argument("--output_dir", required=True, help="The path to save annotation json files.") parser.add_argument("--output_json", required=True, help="The path to save annotation final combined json file.") parser.add_argument("--api_key", required=True, help="OpenAI API key.") parser.add_argument("--api_base", default="", type=str, help="OpenAI API base.") parser.add_argument("--num_tasks", required=True, type=int, help="Number of splits.") args = parser.parse_args() return args
null
179,667
import openai import os import argparse import json import ast from multiprocessing.pool import Pool The provided code snippet includes necessary dependencies for implementing the `annotate` function. Write a Python function `def annotate(prediction_set, caption_files, output_dir, args)` to solve the following problem: Evaluates question and answer pairs using GPT-3 and returns a score for temporal understanding. Here is the function: def annotate(prediction_set, caption_files, output_dir, args): """ Evaluates question and answer pairs using GPT-3 and returns a score for temporal understanding. """ # Set the OpenAI API key. openai.api_key = args.api_key if args.api_base is not None: openai.api_base = args.api_base for file in caption_files: key = file[:-5] # Strip file extension qa_set = prediction_set[key] question = qa_set['q'] answer = qa_set['a'] pred = qa_set['pred'] try: # Compute the temporal understanding score completion = openai.ChatCompletion.create( model="gpt-3.5-turbo", messages=[ { "role": "system", "content": "You are an intelligent chatbot designed for evaluating the temporal understanding of generative outputs for video-based question-answer pairs. " "Your task is to compare the predicted answer with the correct answer and determine if they correctly reflect the temporal sequence of events in the video content. Here's how you can accomplish the task:" "------" "##INSTRUCTIONS: " "- Focus on the temporal consistency between the predicted answer and the correct answer. The predicted answer should correctly reflect the sequence of events or details as they are presented in the video content.\n" "- Consider synonyms or paraphrases as valid matches, but only if the temporal order is maintained.\n" "- Evaluate the temporal accuracy of the prediction compared to the answer." }, { "role": "user", "content": "Please evaluate the following video-based question-answer pair:\n\n" f"Question: {question}\n" f"Correct Answer: {answer}\n" f"Predicted Answer: {pred}\n\n" "Provide your evaluation only as a temporal accuracy score where the temporal accuracy score is an integer value between 0 and 5, with 5 indicating the highest level of temporal consistency. " "Please generate the response in the form of a Python dictionary string with keys 'score', where its value is the temporal accuracy score in INTEGER, not STRING." "DO NOT PROVIDE ANY OTHER OUTPUT TEXT OR EXPLANATION. Only provide the Python dictionary string. " "For example, your response should look like this: {''score': 4.8}." } ] ) # Convert response to a Python dictionary. response_message = completion["choices"][0]["message"]["content"] response_dict = ast.literal_eval(response_message) result_qa_pair = [response_dict, qa_set] # Save the question-answer pairs to a json file. with open(f"{output_dir}/{key}.json", "w") as f: json.dump(result_qa_pair, f) except Exception as e: print(f"Error processing file '{key}': {e}")
Evaluates question and answer pairs using GPT-3 and returns a score for temporal understanding.
179,669
import openai import os import argparse import json import ast from multiprocessing.pool import Pool The provided code snippet includes necessary dependencies for implementing the `annotate` function. Write a Python function `def annotate(prediction_set, caption_files, output_dir, args)` to solve the following problem: Evaluates question and answer pairs using GPT-3 Returns a score for correctness. Here is the function: def annotate(prediction_set, caption_files, output_dir, args): """ Evaluates question and answer pairs using GPT-3 Returns a score for correctness. """ # Set the OpenAI API key. openai.api_key = args.api_key if args.api_base is not None: openai.api_base = args.api_base for file in caption_files: key = file[:-5] # Strip file extension qa_set = prediction_set[key] question = qa_set['q'] answer = qa_set['a'] pred = qa_set['pred'] try: # Compute the correctness score completion = openai.ChatCompletion.create( model="gpt-3.5-turbo", messages=[ { "role": "system", "content": "You are an intelligent chatbot designed for evaluating the factual accuracy of generative outputs for video-based question-answer pairs. " "Your task is to compare the predicted answer with the correct answer and determine if they are factually consistent. Here's how you can accomplish the task:" "------" "##INSTRUCTIONS: " "- Focus on the factual consistency between the predicted answer and the correct answer. The predicted answer should not contain any misinterpretations or misinformation.\n" "- The predicted answer must be factually accurate and align with the video content.\n" "- Consider synonyms or paraphrases as valid matches.\n" "- Evaluate the factual accuracy of the prediction compared to the answer." }, { "role": "user", "content": "Please evaluate the following video-based question-answer pair:\n\n" f"Question: {question}\n" f"Correct Answer: {answer}\n" f"Predicted Answer: {pred}\n\n" "Provide your evaluation only as a factual accuracy score where the factual accuracy score is an integer value between 0 and 5, with 5 indicating the highest level of factual consistency. " "Please generate the response in the form of a Python dictionary string with keys 'score', where its value is the factual accuracy score in INTEGER, not STRING." "DO NOT PROVIDE ANY OTHER OUTPUT TEXT OR EXPLANATION. Only provide the Python dictionary string. " "For example, your response should look like this: {''score': 4.8}." } ] ) # Convert response to a Python dictionary. response_message = completion["choices"][0]["message"]["content"] response_dict = ast.literal_eval(response_message) result_qa_pair = [response_dict, qa_set] # Save the question-answer pairs to a json file. with open(f"{output_dir}/{key}.json", "w") as f: json.dump(result_qa_pair, f) except Exception as e: print(f"Error processing file '{key}': {e}")
Evaluates question and answer pairs using GPT-3 Returns a score for correctness.
179,671
import os import argparse import json from tqdm import tqdm from videollava.eval.video.run_inference_video_qa import get_model_output from videollava.mm_utils import get_model_name_from_path from videollava.model.builder import load_pretrained_model def get_model_output(model, video_processor, tokenizer, video, qs, args): if model.config.mm_use_im_start_end: qs = DEFAULT_VID_START_TOKEN + ''.join([DEFAULT_IMAGE_TOKEN]*8) + DEFAULT_VID_END_TOKEN + '\n' + qs else: qs = ''.join([DEFAULT_IMAGE_TOKEN]*8) + '\n' + qs conv_mode = "llava_v1" args.conv_mode = conv_mode conv = conv_templates[args.conv_mode].copy() conv.append_message(conv.roles[0], qs) conv.append_message(conv.roles[1], None) prompt = conv.get_prompt() video_tensor = video_processor.preprocess(video, return_tensors='pt')['pixel_values'][0].half().to(args.device) # print(video_tensor.shape) input_ids = tokenizer_image_token(prompt, tokenizer, IMAGE_TOKEN_INDEX, return_tensors='pt').unsqueeze(0).to(args.device) stop_str = conv.sep if conv.sep_style != SeparatorStyle.TWO else conv.sep2 keywords = [stop_str] stopping_criteria = KeywordsStoppingCriteria(keywords, tokenizer, input_ids) with torch.inference_mode(): output_ids = model.generate( input_ids, images=[video_tensor], do_sample=True, temperature=0.0, max_new_tokens=1024, use_cache=True, stopping_criteria=[stopping_criteria]) input_token_len = input_ids.shape[1] n_diff_input_output = (input_ids != output_ids[:, :input_token_len]).sum().item() if n_diff_input_output > 0: print(f'[Warning] {n_diff_input_output} output_ids are not the same as the input_ids') outputs = tokenizer.batch_decode(output_ids[:, input_token_len:], skip_special_tokens=True)[0] outputs = outputs.strip() if outputs.endswith(stop_str): outputs = outputs[:-len(stop_str)] outputs = outputs.strip() print(outputs) return outputs def get_model_name_from_path(model_path): model_path = model_path.strip("/") model_paths = model_path.split("/") if model_paths[-1].startswith('checkpoint-'): return model_paths[-2] + "_" + model_paths[-1] else: return model_paths[-1] def load_pretrained_model(model_path, model_base, model_name, load_8bit=False, load_4bit=False, device_map="auto", device="cuda", **kwargs): kwargs = {"device_map": device_map, **kwargs} if device != "cuda": kwargs['device_map'] = {"": device} if load_8bit: kwargs['load_in_8bit'] = True elif load_4bit: kwargs['load_in_4bit'] = True kwargs['quantization_config'] = BitsAndBytesConfig( load_in_4bit=True, bnb_4bit_compute_dtype=torch.float16, bnb_4bit_use_double_quant=True, bnb_4bit_quant_type='nf4' ) else: kwargs['torch_dtype'] = torch.float16 if 'llava' in model_name.lower(): # Load LLaVA model if 'lora' in model_name.lower() and model_base is None: warnings.warn('There is `lora` in model name but no `model_base` is provided. If you are loading a LoRA model, please provide the `model_base` argument. Detailed instruction: https://github.com/haotian-liu/LLaVA#launch-a-model-worker-lora-weights-unmerged.') if 'lora' in model_name.lower() and model_base is not None: lora_cfg_pretrained = AutoConfig.from_pretrained(model_path) tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=False) print('Loading LLaVA from base model...') model = LlavaLlamaForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, config=lora_cfg_pretrained, **kwargs) token_num, tokem_dim = model.lm_head.out_features, model.lm_head.in_features if model.lm_head.weight.shape[0] != token_num: model.lm_head.weight = torch.nn.Parameter(torch.empty(token_num, tokem_dim, device=model.device, dtype=model.dtype)) model.model.embed_tokens.weight = torch.nn.Parameter(torch.empty(token_num, tokem_dim, device=model.device, dtype=model.dtype)) print('Loading additional LLaVA weights...') if os.path.exists(os.path.join(model_path, 'non_lora_trainables.bin')): non_lora_trainables = torch.load(os.path.join(model_path, 'non_lora_trainables.bin'), map_location='cpu') else: # this is probably from HF Hub from huggingface_hub import hf_hub_download def load_from_hf(repo_id, filename, subfolder=None): cache_file = hf_hub_download( repo_id=repo_id, filename=filename, subfolder=subfolder) return torch.load(cache_file, map_location='cpu') non_lora_trainables = load_from_hf(model_path, 'non_lora_trainables.bin') non_lora_trainables = {(k[11:] if k.startswith('base_model.') else k): v for k, v in non_lora_trainables.items()} if any(k.startswith('model.model.') for k in non_lora_trainables): non_lora_trainables = {(k[6:] if k.startswith('model.') else k): v for k, v in non_lora_trainables.items()} model.load_state_dict(non_lora_trainables, strict=False) from peft import PeftModel print('Loading LoRA weights...') model = PeftModel.from_pretrained(model, model_path) print('Merging LoRA weights...') model = model.merge_and_unload() print('Model is loaded...') elif model_base is not None: # this may be mm projector only print('Loading LLaVA from base model...') if 'mpt' in model_name.lower(): if not os.path.isfile(os.path.join(model_path, 'configuration_mpt.py')): shutil.copyfile(os.path.join(model_base, 'configuration_mpt.py'), os.path.join(model_path, 'configuration_mpt.py')) tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=True) cfg_pretrained = AutoConfig.from_pretrained(model_path, trust_remote_code=True) model = LlavaMPTForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, config=cfg_pretrained, **kwargs) else: tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=False) cfg_pretrained = AutoConfig.from_pretrained(model_path) model = LlavaLlamaForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, config=cfg_pretrained, **kwargs) mm_projector_weights = torch.load(os.path.join(model_path, 'mm_projector.bin'), map_location='cpu') mm_projector_weights = {k: v.to(torch.float16) for k, v in mm_projector_weights.items()} model.load_state_dict(mm_projector_weights, strict=False) else: if 'mpt' in model_name.lower(): tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=True) model = LlavaMPTForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs) else: tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=False) model = LlavaLlamaForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs) else: # Load language model if model_base is not None: # PEFT model from peft import PeftModel tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=False) model = AutoModelForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, **kwargs) print(f"Loading LoRA weights from {model_path}") model = PeftModel.from_pretrained(model, model_path) print(f"Merging weights") model = model.merge_and_unload() print('Convert to FP16...') model.to(torch.float16) else: use_fast = False if 'mpt' in model_name.lower(): tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=True) model = AutoModelForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, trust_remote_code=True, **kwargs) else: tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=False) model = AutoModelForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs) # ========================================================================================================== processor = {'image': None, 'video': None} if 'llava' in model_name.lower(): mm_use_im_start_end = getattr(model.config, "mm_use_im_start_end", False) mm_use_im_patch_token = getattr(model.config, "mm_use_im_patch_token", True) if mm_use_im_patch_token: tokenizer.add_tokens([DEFAULT_IMAGE_PATCH_TOKEN], special_tokens=True) tokenizer.add_tokens([DEFAULT_VIDEO_PATCH_TOKEN], special_tokens=True) if mm_use_im_start_end: tokenizer.add_tokens([DEFAULT_IM_START_TOKEN, DEFAULT_IM_END_TOKEN], special_tokens=True) tokenizer.add_tokens([DEFAULT_VID_START_TOKEN, DEFAULT_VID_END_TOKEN], special_tokens=True) model.resize_token_embeddings(len(tokenizer)) if model.config.mm_image_tower is not None: image_tower = model.get_image_tower() if not image_tower.is_loaded: image_tower.load_model() image_tower.to(device=device, dtype=torch.float16) image_processor = image_tower.image_processor processor['image'] = image_processor if model.config.mm_video_tower is not None: video_tower = model.get_video_tower() if not video_tower.is_loaded: video_tower.load_model() video_tower.to(device=device, dtype=torch.float16) video_processor = video_tower.video_processor processor['video'] = video_processor # ========================================================================================================== if hasattr(model.config, "max_sequence_length"): context_len = model.config.max_sequence_length else: context_len = 2048 return tokenizer, model, processor, context_len The provided code snippet includes necessary dependencies for implementing the `run_inference` function. Write a Python function `def run_inference(args)` to solve the following problem: Run inference on a set of video files using the provided model. Args: args: Command-line arguments. Here is the function: def run_inference(args): """ Run inference on a set of video files using the provided model. Args: args: Command-line arguments. """ # Initialize the model model_name = get_model_name_from_path(args.model_path) tokenizer, model, processor, context_len = load_pretrained_model(args.model_path, args.model_base, model_name) model = model.to(args.device) # Load the ground truth file with open(args.gt_file) as file: gt_contents = json.load(file) # Create the output directory if it doesn't exist if not os.path.exists(args.output_dir): os.makedirs(args.output_dir) output_list = [] # List to store the output results # conv_mode = args.conv_mode video_formats = ['.mp4', '.avi', '.mov', '.mkv'] # Iterate over each sample in the ground truth file for sample in tqdm(gt_contents): video_name = sample['video_name'] sample_set = sample question_1 = sample['Q1'] question_2 = sample['Q2'] try: # Load the video file for fmt in video_formats: # Added this line temp_path = os.path.join(args.video_dir, f"{video_name}{fmt}") if os.path.exists(temp_path): video_path = temp_path # Run inference on the video for the first question and add the output to the list output_1 = get_model_output(model, processor['video'], tokenizer, video_path, question_1, args) sample_set['pred1'] = output_1 # Run inference on the video for the second question and add the output to the list output_2 = get_model_output(model, processor['video'], tokenizer, video_path, question_2, args) sample_set['pred2'] = output_2 output_list.append(sample_set) break except Exception as e: print(f"Error processing video file '{video_name}': {e}") # Save the output list to a JSON file with open(os.path.join(args.output_dir, f"{args.output_name}.json"), 'w') as file: json.dump(output_list, file)
Run inference on a set of video files using the provided model. Args: args: Command-line arguments.
179,673
import openai import os import argparse import json import ast from multiprocessing.pool import Pool The provided code snippet includes necessary dependencies for implementing the `annotate` function. Write a Python function `def annotate(prediction_set, caption_files, output_dir, args)` to solve the following problem: Evaluates question and answer pairs using GPT-3 and returns a score for consistency. Here is the function: def annotate(prediction_set, caption_files, output_dir, args): """ Evaluates question and answer pairs using GPT-3 and returns a score for consistency. """ # Set the OpenAI API key. openai.api_key = args.api_key if args.api_base is not None: openai.api_base = args.api_base for file in caption_files: key = file[:-5] # Strip file extension qa_set = prediction_set[key] question1 = qa_set['q1'] question2 = qa_set['q2'] answer = qa_set['a'] pred1 = qa_set['pred1'] pred2 = qa_set['pred2'] try: # Compute the consistency score completion = openai.ChatCompletion.create( model="gpt-3.5-turbo", messages=[ { "role": "system", "content": "You are an intelligent chatbot designed for evaluating the consistency of generative outputs for similar video-based question-answer pairs. " "You will be given two very similar questions, a common answer common to both the questions and predicted answers for the two questions ." "Your task is to compare the predicted answers for two very similar question, with a common correct answer and determine if they are consistent. Here's how you can accomplish the task:" "------" "##INSTRUCTIONS: " "- Focus on the consistency between the two predicted answers and the correct answer. Both predicted answers should correspond to the correct answer and to each other, and should not contain any contradictions or significant differences in the conveyed information.\n" "- Both predicted answers must be consistent with each other and the correct answer, in terms of the information they provide about the video content.\n" "- Consider synonyms or paraphrases as valid matches, but only if they maintain the consistency in the conveyed information.\n" "- Evaluate the consistency of the two predicted answers compared to the correct answer." }, { "role": "user", "content": "Please evaluate the following video-based question-answer pair:\n\n" f"Question 1: {question1}\n" f"Question 2: {question2}\n" f"Correct Answer: {answer}\n" f"Predicted Answer to Question 1: {pred1}\n" f"Predicted Answer to Question 2: {pred2}\n\n" "Provide your evaluation only as a consistency score where the consistency score is an integer value between 0 and 5, with 5 indicating the highest level of consistency. " "Please generate the response in the form of a Python dictionary string with keys 'score', where its value is the consistency score in INTEGER, not STRING." "DO NOT PROVIDE ANY OTHER OUTPUT TEXT OR EXPLANATION. Only provide the Python dictionary string. " "For example, your response should look like this: {''score': 4.8}." } ] ) # Convert response to a Python dictionary. response_message = completion["choices"][0]["message"]["content"] response_dict = ast.literal_eval(response_message) result_qa_pair = [response_dict, qa_set] # Save the question-answer pairs to a json file. with open(f"{output_dir}/{key}.json", "w") as f: json.dump(result_qa_pair, f) except Exception as e: print(f"Error processing file '{key}': {e}")
Evaluates question and answer pairs using GPT-3 and returns a score for consistency.
179,675
import openai import os import argparse import json import ast from multiprocessing.pool import Pool The provided code snippet includes necessary dependencies for implementing the `annotate` function. Write a Python function `def annotate(prediction_set, caption_files, output_dir, args)` to solve the following problem: Evaluates question and answer pairs using GPT-3 and returns a score for detailed orientation. Here is the function: def annotate(prediction_set, caption_files, output_dir, args): """ Evaluates question and answer pairs using GPT-3 and returns a score for detailed orientation. """ # Set the OpenAI API key. openai.api_key = args.api_key if args.api_base is not None: openai.api_base = args.api_base for file in caption_files: key = file[:-5] # Strip file extension qa_set = prediction_set[key] question = qa_set['q'] answer = qa_set['a'] pred = qa_set['pred'] try: # Compute the detailed-orientation score completion = openai.ChatCompletion.create( model="gpt-3.5-turbo", messages=[ { "role": "system", "content": "You are an intelligent chatbot designed for evaluating the detail orientation of generative outputs for video-based question-answer pairs. " "Your task is to compare the predicted answer with the correct answer and determine its level of detail, considering both completeness and specificity. Here's how you can accomplish the task:" "------" "##INSTRUCTIONS: " "- Check if the predicted answer covers all major points from the video. The response should not leave out any key aspects.\n" "- Evaluate whether the predicted answer includes specific details rather than just generic points. It should provide comprehensive information that is tied to specific elements of the video.\n" "- Consider synonyms or paraphrases as valid matches.\n" "- Provide a single evaluation score that reflects the level of detail orientation of the prediction, considering both completeness and specificity." }, { "role": "user", "content": "Please evaluate the following video-based question-answer pair:\n\n" f"Question: {question}\n" f"Correct Answer: {answer}\n" f"Predicted Answer: {pred}\n\n" "Provide your evaluation only as a detail orientation score where the detail orientation score is an integer value between 0 and 5, with 5 indicating the highest level of detail orientation. " "Please generate the response in the form of a Python dictionary string with keys 'score', where its value is the detail orientation score in INTEGER, not STRING." "DO NOT PROVIDE ANY OTHER OUTPUT TEXT OR EXPLANATION. Only provide the Python dictionary string. " "For example, your response should look like this: {''score': 4.8}." } ] ) # Convert response to a Python dictionary. response_message = completion["choices"][0]["message"]["content"] response_dict = ast.literal_eval(response_message) result_qa_pair = [response_dict, qa_set] # Save the question-answer pairs to a json file. with open(f"{output_dir}/{key}.json", "w") as f: json.dump(result_qa_pair, f) except Exception as e: print(f"Error processing file '{key}': {e}")
Evaluates question and answer pairs using GPT-3 and returns a score for detailed orientation.
179,677
import openai import os import argparse import json import ast from multiprocessing.pool import Pool The provided code snippet includes necessary dependencies for implementing the `annotate` function. Write a Python function `def annotate(prediction_set, caption_files, output_dir, args)` to solve the following problem: Evaluates question and answer pairs using GPT-3 and returns a score for contextual understanding. Here is the function: def annotate(prediction_set, caption_files, output_dir, args): """ Evaluates question and answer pairs using GPT-3 and returns a score for contextual understanding. """ # Set the OpenAI API key. openai.api_key = args.api_key if args.api_base is not None: openai.api_base = args.api_base for file in caption_files: key = file[:-5] # Strip file extension qa_set = prediction_set[key] question = qa_set['q'] answer = qa_set['a'] pred = qa_set['pred'] try: # Compute the contextual understanding score completion = openai.ChatCompletion.create( model="gpt-3.5-turbo", messages=[ { "role": "system", "content": "You are an intelligent chatbot designed for evaluating the contextual understanding of generative outputs for video-based question-answer pairs. " "Your task is to compare the predicted answer with the correct answer and determine if the generated response aligns with the overall context of the video content. Here's how you can accomplish the task:" "------" "##INSTRUCTIONS: " "- Evaluate whether the predicted answer aligns with the overall context of the video content. It should not provide information that is out of context or misaligned.\n" "- The predicted answer must capture the main themes and sentiments of the video.\n" "- Consider synonyms or paraphrases as valid matches.\n" "- Provide your evaluation of the contextual understanding of the prediction compared to the answer." }, { "role": "user", "content": "Please evaluate the following video-based question-answer pair:\n\n" f"Question: {question}\n" f"Correct Answer: {answer}\n" f"Predicted Answer: {pred}\n\n" "Provide your evaluation only as a contextual understanding score where the contextual understanding score is an integer value between 0 and 5, with 5 indicating the highest level of contextual understanding. " "Please generate the response in the form of a Python dictionary string with keys 'score', where its value is contextual understanding score in INTEGER, not STRING." "DO NOT PROVIDE ANY OTHER OUTPUT TEXT OR EXPLANATION. Only provide the Python dictionary string. " "For example, your response should look like this: {''score': 4.8}." } ] ) # Convert response to a Python dictionary. response_message = completion["choices"][0]["message"]["content"] response_dict = ast.literal_eval(response_message) result_qa_pair = [response_dict, qa_set] # Save the question-answer pairs to a json file. with open(f"{output_dir}/{key}.json", "w") as f: json.dump(result_qa_pair, f) except Exception as e: print(f"Error processing file '{key}': {e}")
Evaluates question and answer pairs using GPT-3 and returns a score for contextual understanding.
179,678
import math import os import argparse import json import torch import transformers from tqdm import tqdm from videollava.conversation import conv_templates, SeparatorStyle from videollava.constants import DEFAULT_IM_START_TOKEN, DEFAULT_IMAGE_TOKEN, DEFAULT_IM_END_TOKEN, IMAGE_TOKEN_INDEX, DEFAULT_VID_START_TOKEN, DEFAULT_VID_END_TOKEN from videollava.mm_utils import get_model_name_from_path, tokenizer_image_token, KeywordsStoppingCriteria from videollava.model.builder import load_pretrained_model from videollava.model.language_model.llava_llama import LlavaLlamaForCausalLM from videollava.train.train import smart_tokenizer_and_embedding_resize The provided code snippet includes necessary dependencies for implementing the `parse_args` function. Write a Python function `def parse_args()` to solve the following problem: Parse command-line arguments. Here is the function: def parse_args(): """ Parse command-line arguments. """ parser = argparse.ArgumentParser() # Define the command-line arguments parser.add_argument('--model_path', help='', required=True) parser.add_argument('--cache_dir', help='', required=True) parser.add_argument('--video_dir', help='Directory containing video files.', required=True) parser.add_argument('--gt_file_question', help='Path to the ground truth file containing question.', required=True) parser.add_argument('--gt_file_answers', help='Path to the ground truth file containing answers.', required=True) parser.add_argument('--output_dir', help='Directory to save the model results JSON.', required=True) parser.add_argument('--output_name', help='Name of the file for storing results JSON.', required=True) parser.add_argument("--num_chunks", type=int, default=1) parser.add_argument("--chunk_idx", type=int, default=0) parser.add_argument("--device", type=str, required=False, default='cuda:0') parser.add_argument('--model_base', help='', default=None, type=str, required=False) parser.add_argument("--model_max_length", type=int, required=False, default=2048) return parser.parse_args()
Parse command-line arguments.
179,679
import math import os import argparse import json import torch import transformers from tqdm import tqdm from videollava.conversation import conv_templates, SeparatorStyle from videollava.constants import DEFAULT_IM_START_TOKEN, DEFAULT_IMAGE_TOKEN, DEFAULT_IM_END_TOKEN, IMAGE_TOKEN_INDEX, DEFAULT_VID_START_TOKEN, DEFAULT_VID_END_TOKEN from videollava.mm_utils import get_model_name_from_path, tokenizer_image_token, KeywordsStoppingCriteria from videollava.model.builder import load_pretrained_model from videollava.model.language_model.llava_llama import LlavaLlamaForCausalLM from videollava.train.train import smart_tokenizer_and_embedding_resize def get_chunk(lst, n, k): chunks = split_list(lst, n) return chunks[k] def get_model_output(model, video_processor, tokenizer, video, qs, args): if model.config.mm_use_im_start_end: qs = DEFAULT_VID_START_TOKEN + ''.join([DEFAULT_IMAGE_TOKEN]*8) + DEFAULT_VID_END_TOKEN + '\n' + qs else: qs = ''.join([DEFAULT_IMAGE_TOKEN]*8) + '\n' + qs conv_mode = "llava_v1" args.conv_mode = conv_mode conv = conv_templates[args.conv_mode].copy() conv.append_message(conv.roles[0], qs) conv.append_message(conv.roles[1], None) prompt = conv.get_prompt() video_tensor = video_processor.preprocess(video, return_tensors='pt')['pixel_values'][0].half().to(args.device) # print(video_tensor.shape) input_ids = tokenizer_image_token(prompt, tokenizer, IMAGE_TOKEN_INDEX, return_tensors='pt').unsqueeze(0).to(args.device) stop_str = conv.sep if conv.sep_style != SeparatorStyle.TWO else conv.sep2 keywords = [stop_str] stopping_criteria = KeywordsStoppingCriteria(keywords, tokenizer, input_ids) with torch.inference_mode(): output_ids = model.generate( input_ids, images=[video_tensor], do_sample=True, temperature=0.0, max_new_tokens=1024, use_cache=True, stopping_criteria=[stopping_criteria]) input_token_len = input_ids.shape[1] n_diff_input_output = (input_ids != output_ids[:, :input_token_len]).sum().item() if n_diff_input_output > 0: print(f'[Warning] {n_diff_input_output} output_ids are not the same as the input_ids') outputs = tokenizer.batch_decode(output_ids[:, input_token_len:], skip_special_tokens=True)[0] outputs = outputs.strip() if outputs.endswith(stop_str): outputs = outputs[:-len(stop_str)] outputs = outputs.strip() print(outputs) return outputs def get_model_name_from_path(model_path): model_path = model_path.strip("/") model_paths = model_path.split("/") if model_paths[-1].startswith('checkpoint-'): return model_paths[-2] + "_" + model_paths[-1] else: return model_paths[-1] def load_pretrained_model(model_path, model_base, model_name, load_8bit=False, load_4bit=False, device_map="auto", device="cuda", **kwargs): kwargs = {"device_map": device_map, **kwargs} if device != "cuda": kwargs['device_map'] = {"": device} if load_8bit: kwargs['load_in_8bit'] = True elif load_4bit: kwargs['load_in_4bit'] = True kwargs['quantization_config'] = BitsAndBytesConfig( load_in_4bit=True, bnb_4bit_compute_dtype=torch.float16, bnb_4bit_use_double_quant=True, bnb_4bit_quant_type='nf4' ) else: kwargs['torch_dtype'] = torch.float16 if 'llava' in model_name.lower(): # Load LLaVA model if 'lora' in model_name.lower() and model_base is None: warnings.warn('There is `lora` in model name but no `model_base` is provided. If you are loading a LoRA model, please provide the `model_base` argument. Detailed instruction: https://github.com/haotian-liu/LLaVA#launch-a-model-worker-lora-weights-unmerged.') if 'lora' in model_name.lower() and model_base is not None: lora_cfg_pretrained = AutoConfig.from_pretrained(model_path) tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=False) print('Loading LLaVA from base model...') model = LlavaLlamaForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, config=lora_cfg_pretrained, **kwargs) token_num, tokem_dim = model.lm_head.out_features, model.lm_head.in_features if model.lm_head.weight.shape[0] != token_num: model.lm_head.weight = torch.nn.Parameter(torch.empty(token_num, tokem_dim, device=model.device, dtype=model.dtype)) model.model.embed_tokens.weight = torch.nn.Parameter(torch.empty(token_num, tokem_dim, device=model.device, dtype=model.dtype)) print('Loading additional LLaVA weights...') if os.path.exists(os.path.join(model_path, 'non_lora_trainables.bin')): non_lora_trainables = torch.load(os.path.join(model_path, 'non_lora_trainables.bin'), map_location='cpu') else: # this is probably from HF Hub from huggingface_hub import hf_hub_download def load_from_hf(repo_id, filename, subfolder=None): cache_file = hf_hub_download( repo_id=repo_id, filename=filename, subfolder=subfolder) return torch.load(cache_file, map_location='cpu') non_lora_trainables = load_from_hf(model_path, 'non_lora_trainables.bin') non_lora_trainables = {(k[11:] if k.startswith('base_model.') else k): v for k, v in non_lora_trainables.items()} if any(k.startswith('model.model.') for k in non_lora_trainables): non_lora_trainables = {(k[6:] if k.startswith('model.') else k): v for k, v in non_lora_trainables.items()} model.load_state_dict(non_lora_trainables, strict=False) from peft import PeftModel print('Loading LoRA weights...') model = PeftModel.from_pretrained(model, model_path) print('Merging LoRA weights...') model = model.merge_and_unload() print('Model is loaded...') elif model_base is not None: # this may be mm projector only print('Loading LLaVA from base model...') if 'mpt' in model_name.lower(): if not os.path.isfile(os.path.join(model_path, 'configuration_mpt.py')): shutil.copyfile(os.path.join(model_base, 'configuration_mpt.py'), os.path.join(model_path, 'configuration_mpt.py')) tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=True) cfg_pretrained = AutoConfig.from_pretrained(model_path, trust_remote_code=True) model = LlavaMPTForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, config=cfg_pretrained, **kwargs) else: tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=False) cfg_pretrained = AutoConfig.from_pretrained(model_path) model = LlavaLlamaForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, config=cfg_pretrained, **kwargs) mm_projector_weights = torch.load(os.path.join(model_path, 'mm_projector.bin'), map_location='cpu') mm_projector_weights = {k: v.to(torch.float16) for k, v in mm_projector_weights.items()} model.load_state_dict(mm_projector_weights, strict=False) else: if 'mpt' in model_name.lower(): tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=True) model = LlavaMPTForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs) else: tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=False) model = LlavaLlamaForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs) else: # Load language model if model_base is not None: # PEFT model from peft import PeftModel tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=False) model = AutoModelForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, **kwargs) print(f"Loading LoRA weights from {model_path}") model = PeftModel.from_pretrained(model, model_path) print(f"Merging weights") model = model.merge_and_unload() print('Convert to FP16...') model.to(torch.float16) else: use_fast = False if 'mpt' in model_name.lower(): tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=True) model = AutoModelForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, trust_remote_code=True, **kwargs) else: tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=False) model = AutoModelForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs) # ========================================================================================================== processor = {'image': None, 'video': None} if 'llava' in model_name.lower(): mm_use_im_start_end = getattr(model.config, "mm_use_im_start_end", False) mm_use_im_patch_token = getattr(model.config, "mm_use_im_patch_token", True) if mm_use_im_patch_token: tokenizer.add_tokens([DEFAULT_IMAGE_PATCH_TOKEN], special_tokens=True) tokenizer.add_tokens([DEFAULT_VIDEO_PATCH_TOKEN], special_tokens=True) if mm_use_im_start_end: tokenizer.add_tokens([DEFAULT_IM_START_TOKEN, DEFAULT_IM_END_TOKEN], special_tokens=True) tokenizer.add_tokens([DEFAULT_VID_START_TOKEN, DEFAULT_VID_END_TOKEN], special_tokens=True) model.resize_token_embeddings(len(tokenizer)) if model.config.mm_image_tower is not None: image_tower = model.get_image_tower() if not image_tower.is_loaded: image_tower.load_model() image_tower.to(device=device, dtype=torch.float16) image_processor = image_tower.image_processor processor['image'] = image_processor if model.config.mm_video_tower is not None: video_tower = model.get_video_tower() if not video_tower.is_loaded: video_tower.load_model() video_tower.to(device=device, dtype=torch.float16) video_processor = video_tower.video_processor processor['video'] = video_processor # ========================================================================================================== if hasattr(model.config, "max_sequence_length"): context_len = model.config.max_sequence_length else: context_len = 2048 return tokenizer, model, processor, context_len The provided code snippet includes necessary dependencies for implementing the `run_inference` function. Write a Python function `def run_inference(args)` to solve the following problem: Run inference on ActivityNet QA DataSet using the Video-ChatGPT model. Args: args: Command-line arguments. Here is the function: def run_inference(args): """ Run inference on ActivityNet QA DataSet using the Video-ChatGPT model. Args: args: Command-line arguments. """ # Initialize the model model_name = get_model_name_from_path(args.model_path) tokenizer, model, processor, context_len = load_pretrained_model(args.model_path, args.model_base, model_name) model = model.to(args.device) # Load both ground truth file containing questions and answers # with open(args.gt_file_question) as file: # gt_questions = json.load(file) # with open(args.gt_file_answers) as file: # gt_answers = json.load(file) gt_questions = json.load(open(args.gt_file_question, "r")) gt_questions = get_chunk(gt_questions, args.num_chunks, args.chunk_idx) gt_answers = json.load(open(args.gt_file_answers, "r")) # gt_answers = get_chunk(gt_answers, args.num_chunks, args.chunk_idx) answers_file = os.path.join(args.output_dir, f"{args.output_name}.json") os.makedirs(args.output_dir, exist_ok=True) ans_file = open(answers_file, "w") # Create the output directory if it doesn't exist if not os.path.exists(args.output_dir): os.makedirs(args.output_dir) output_list = [] # List to store the output results video_formats = ['.mp4', '.avi', '.mov', '.mkv'] # Iterate over each sample in the ground truth file index = 0 for sample in tqdm(gt_questions): video_name = sample['video_name'] question = sample['question'] id = sample['question_id'] answer = gt_answers[index]['answer'] index += 1 sample_set = {'id': id, 'question': question, 'answer': answer} # Load the video file for fmt in tqdm(video_formats): # Added this line temp_path = os.path.join(args.video_dir, f"{video_name}{fmt}") if os.path.exists(temp_path): video_path = temp_path # try: # Run inference on the video and add the output to the list output = get_model_output(model, processor['video'], tokenizer, video_path, question, args) sample_set['pred'] = output output_list.append(sample_set) # except Exception as e: # print(f"Error processing video file '{video_name}': {e}") ans_file.write(json.dumps(sample_set) + "\n") break ans_file.close() # Save the output list to a JSON file # with open(os.path.join(args.output_dir, f"{args.output_name}.json"), 'w') as file: # json.dump(output_list, file)
Run inference on ActivityNet QA DataSet using the Video-ChatGPT model. Args: args: Command-line arguments.
179,681
import math import os import argparse import json import torch import transformers from tqdm import tqdm from videollava.conversation import conv_templates, SeparatorStyle from videollava.constants import DEFAULT_IM_START_TOKEN, DEFAULT_IMAGE_TOKEN, DEFAULT_IM_END_TOKEN, IMAGE_TOKEN_INDEX, DEFAULT_VID_START_TOKEN, DEFAULT_VID_END_TOKEN from videollava.mm_utils import get_model_name_from_path, tokenizer_image_token, KeywordsStoppingCriteria from videollava.model.builder import load_pretrained_model from videollava.model.language_model.llava_llama import LlavaLlamaForCausalLM from videollava.train.train import smart_tokenizer_and_embedding_resize def get_chunk(lst, n, k): chunks = split_list(lst, n) return chunks[k] def get_model_output(model, video_processor, tokenizer, video, qs, args): if model.config.mm_use_im_start_end: qs = DEFAULT_VID_START_TOKEN + ''.join([DEFAULT_IMAGE_TOKEN]*8) + DEFAULT_VID_END_TOKEN + '\n' + qs else: qs = ''.join([DEFAULT_IMAGE_TOKEN]*8) + '\n' + qs conv_mode = "llava_v1" args.conv_mode = conv_mode conv = conv_templates[args.conv_mode].copy() conv.append_message(conv.roles[0], qs) conv.append_message(conv.roles[1], None) prompt = conv.get_prompt() video_tensor = video_processor.preprocess(video, return_tensors='pt')['pixel_values'][0].half().to(args.device) # print(video_tensor.shape) input_ids = tokenizer_image_token(prompt, tokenizer, IMAGE_TOKEN_INDEX, return_tensors='pt').unsqueeze(0).to(args.device) stop_str = conv.sep if conv.sep_style != SeparatorStyle.TWO else conv.sep2 keywords = [stop_str] stopping_criteria = KeywordsStoppingCriteria(keywords, tokenizer, input_ids) with torch.inference_mode(): output_ids = model.generate( input_ids, images=[video_tensor], do_sample=True, temperature=0.0, max_new_tokens=1024, use_cache=True, stopping_criteria=[stopping_criteria]) input_token_len = input_ids.shape[1] n_diff_input_output = (input_ids != output_ids[:, :input_token_len]).sum().item() if n_diff_input_output > 0: print(f'[Warning] {n_diff_input_output} output_ids are not the same as the input_ids') outputs = tokenizer.batch_decode(output_ids[:, input_token_len:], skip_special_tokens=True)[0] outputs = outputs.strip() if outputs.endswith(stop_str): outputs = outputs[:-len(stop_str)] outputs = outputs.strip() print(outputs) return outputs def get_model_name_from_path(model_path): model_path = model_path.strip("/") model_paths = model_path.split("/") if model_paths[-1].startswith('checkpoint-'): return model_paths[-2] + "_" + model_paths[-1] else: return model_paths[-1] def load_pretrained_model(model_path, model_base, model_name, load_8bit=False, load_4bit=False, device_map="auto", device="cuda", **kwargs): kwargs = {"device_map": device_map, **kwargs} if device != "cuda": kwargs['device_map'] = {"": device} if load_8bit: kwargs['load_in_8bit'] = True elif load_4bit: kwargs['load_in_4bit'] = True kwargs['quantization_config'] = BitsAndBytesConfig( load_in_4bit=True, bnb_4bit_compute_dtype=torch.float16, bnb_4bit_use_double_quant=True, bnb_4bit_quant_type='nf4' ) else: kwargs['torch_dtype'] = torch.float16 if 'llava' in model_name.lower(): # Load LLaVA model if 'lora' in model_name.lower() and model_base is None: warnings.warn('There is `lora` in model name but no `model_base` is provided. If you are loading a LoRA model, please provide the `model_base` argument. Detailed instruction: https://github.com/haotian-liu/LLaVA#launch-a-model-worker-lora-weights-unmerged.') if 'lora' in model_name.lower() and model_base is not None: lora_cfg_pretrained = AutoConfig.from_pretrained(model_path) tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=False) print('Loading LLaVA from base model...') model = LlavaLlamaForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, config=lora_cfg_pretrained, **kwargs) token_num, tokem_dim = model.lm_head.out_features, model.lm_head.in_features if model.lm_head.weight.shape[0] != token_num: model.lm_head.weight = torch.nn.Parameter(torch.empty(token_num, tokem_dim, device=model.device, dtype=model.dtype)) model.model.embed_tokens.weight = torch.nn.Parameter(torch.empty(token_num, tokem_dim, device=model.device, dtype=model.dtype)) print('Loading additional LLaVA weights...') if os.path.exists(os.path.join(model_path, 'non_lora_trainables.bin')): non_lora_trainables = torch.load(os.path.join(model_path, 'non_lora_trainables.bin'), map_location='cpu') else: # this is probably from HF Hub from huggingface_hub import hf_hub_download def load_from_hf(repo_id, filename, subfolder=None): cache_file = hf_hub_download( repo_id=repo_id, filename=filename, subfolder=subfolder) return torch.load(cache_file, map_location='cpu') non_lora_trainables = load_from_hf(model_path, 'non_lora_trainables.bin') non_lora_trainables = {(k[11:] if k.startswith('base_model.') else k): v for k, v in non_lora_trainables.items()} if any(k.startswith('model.model.') for k in non_lora_trainables): non_lora_trainables = {(k[6:] if k.startswith('model.') else k): v for k, v in non_lora_trainables.items()} model.load_state_dict(non_lora_trainables, strict=False) from peft import PeftModel print('Loading LoRA weights...') model = PeftModel.from_pretrained(model, model_path) print('Merging LoRA weights...') model = model.merge_and_unload() print('Model is loaded...') elif model_base is not None: # this may be mm projector only print('Loading LLaVA from base model...') if 'mpt' in model_name.lower(): if not os.path.isfile(os.path.join(model_path, 'configuration_mpt.py')): shutil.copyfile(os.path.join(model_base, 'configuration_mpt.py'), os.path.join(model_path, 'configuration_mpt.py')) tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=True) cfg_pretrained = AutoConfig.from_pretrained(model_path, trust_remote_code=True) model = LlavaMPTForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, config=cfg_pretrained, **kwargs) else: tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=False) cfg_pretrained = AutoConfig.from_pretrained(model_path) model = LlavaLlamaForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, config=cfg_pretrained, **kwargs) mm_projector_weights = torch.load(os.path.join(model_path, 'mm_projector.bin'), map_location='cpu') mm_projector_weights = {k: v.to(torch.float16) for k, v in mm_projector_weights.items()} model.load_state_dict(mm_projector_weights, strict=False) else: if 'mpt' in model_name.lower(): tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=True) model = LlavaMPTForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs) else: tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=False) model = LlavaLlamaForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs) else: # Load language model if model_base is not None: # PEFT model from peft import PeftModel tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=False) model = AutoModelForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, **kwargs) print(f"Loading LoRA weights from {model_path}") model = PeftModel.from_pretrained(model, model_path) print(f"Merging weights") model = model.merge_and_unload() print('Convert to FP16...') model.to(torch.float16) else: use_fast = False if 'mpt' in model_name.lower(): tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=True) model = AutoModelForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, trust_remote_code=True, **kwargs) else: tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=False) model = AutoModelForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs) # ========================================================================================================== processor = {'image': None, 'video': None} if 'llava' in model_name.lower(): mm_use_im_start_end = getattr(model.config, "mm_use_im_start_end", False) mm_use_im_patch_token = getattr(model.config, "mm_use_im_patch_token", True) if mm_use_im_patch_token: tokenizer.add_tokens([DEFAULT_IMAGE_PATCH_TOKEN], special_tokens=True) tokenizer.add_tokens([DEFAULT_VIDEO_PATCH_TOKEN], special_tokens=True) if mm_use_im_start_end: tokenizer.add_tokens([DEFAULT_IM_START_TOKEN, DEFAULT_IM_END_TOKEN], special_tokens=True) tokenizer.add_tokens([DEFAULT_VID_START_TOKEN, DEFAULT_VID_END_TOKEN], special_tokens=True) model.resize_token_embeddings(len(tokenizer)) if model.config.mm_image_tower is not None: image_tower = model.get_image_tower() if not image_tower.is_loaded: image_tower.load_model() image_tower.to(device=device, dtype=torch.float16) image_processor = image_tower.image_processor processor['image'] = image_processor if model.config.mm_video_tower is not None: video_tower = model.get_video_tower() if not video_tower.is_loaded: video_tower.load_model() video_tower.to(device=device, dtype=torch.float16) video_processor = video_tower.video_processor processor['video'] = video_processor # ========================================================================================================== if hasattr(model.config, "max_sequence_length"): context_len = model.config.max_sequence_length else: context_len = 2048 return tokenizer, model, processor, context_len The provided code snippet includes necessary dependencies for implementing the `run_inference` function. Write a Python function `def run_inference(args)` to solve the following problem: Run inference on ActivityNet QA DataSet using the Video-ChatGPT model. Args: args: Command-line arguments. Here is the function: def run_inference(args): """ Run inference on ActivityNet QA DataSet using the Video-ChatGPT model. Args: args: Command-line arguments. """ # Initialize the model model_name = get_model_name_from_path(args.model_path) tokenizer, model, processor, context_len = load_pretrained_model(args.model_path, args.model_base, model_name) model = model.to(args.device) # Load both ground truth file containing questions and answers # with open(args.gt_file_question) as file: # gt_questions = json.load(file) # with open(args.gt_file_answers) as file: # gt_answers = json.load(file) gt_questions = json.load(open(args.gt_file_question, "r")) gt_questions = get_chunk(gt_questions, args.num_chunks, args.chunk_idx) gt_answers = json.load(open(args.gt_file_answers, "r")) # gt_answers = get_chunk(gt_answers, args.num_chunks, args.chunk_idx) answers_file = os.path.join(args.output_dir, f"{args.output_name}.json") os.makedirs(args.output_dir, exist_ok=True) ans_file = open(answers_file, "w") # Create the output directory if it doesn't exist if not os.path.exists(args.output_dir): os.makedirs(args.output_dir) output_list = [] # List to store the output results video_formats = ['.mp4', '.avi', '.mov', '.mkv'] # Iterate over each sample in the ground truth file index = 0 for sample in tqdm(gt_questions): video_name = sample['video_name'] question = sample['question'] id = sample['question_id'] answer = gt_answers[index]['answer'] index += 1 sample_set = {'id': id, 'question': question, 'answer': answer} # Load the video file for fmt in tqdm(video_formats): # Added this line temp_path = os.path.join(args.video_dir, f"v_{video_name}{fmt}") if os.path.exists(temp_path): video_path = temp_path # try: # Run inference on the video and add the output to the list output = get_model_output(model, processor['video'], tokenizer, video_path, question, args) sample_set['pred'] = output output_list.append(sample_set) # except Exception as e: # print(f"Error processing video file '{video_name}': {e}") ans_file.write(json.dumps(sample_set) + "\n") break ans_file.close() # Save the output list to a JSON file # with open(os.path.join(args.output_dir, f"{args.output_name}.json"), 'w') as file: # json.dump(output_list, file)
Run inference on ActivityNet QA DataSet using the Video-ChatGPT model. Args: args: Command-line arguments.
179,682
import openai import os import argparse import json import ast from multiprocessing.pool import Pool from tqdm import tqdm def parse_args(): parser = argparse.ArgumentParser(description="question-answer-generation-using-gpt-3") parser.add_argument("--pred_path", default=r'', help="The path to file containing prediction.") parser.add_argument("--output_dir", default=r'', help="The path to save annotation json files.") parser.add_argument("--output_json", default=r'', help="The path to save annotation final combined json file.") parser.add_argument("--api_key", default="", help="OpenAI API key.") parser.add_argument("--api_base", default="", type=str, help="OpenAI API base.") parser.add_argument("--num_tasks", default=1, type=int, help="Number of splits.") args = parser.parse_args() return args
null
179,683
import openai import os import argparse import json import ast from multiprocessing.pool import Pool from tqdm import tqdm The provided code snippet includes necessary dependencies for implementing the `annotate` function. Write a Python function `def annotate(prediction_set, caption_files, output_dir, args)` to solve the following problem: Evaluates question and answer pairs using GPT-3 Returns a score for correctness. Here is the function: def annotate(prediction_set, caption_files, output_dir, args): """ Evaluates question and answer pairs using GPT-3 Returns a score for correctness. """ # Set the OpenAI API key. openai.api_key = args.api_key if args.api_base is not None: openai.api_base = args.api_base for file in caption_files: key = file[:-5] # Strip file extension qa_set = prediction_set[key] question = qa_set['q'] answer = qa_set['a'] pred = qa_set['pred'] try: # Compute the correctness score completion = openai.ChatCompletion.create( model="gpt-3.5-turbo", messages=[ { "role": "system", "content": "You are an intelligent chatbot designed for evaluating the correctness of generative outputs for question-answer pairs. " "Your task is to compare the predicted answer with the correct answer and determine if they match meaningfully. Here's how you can accomplish the task:" "------" "##INSTRUCTIONS: " "- Focus on the meaningful match between the predicted answer and the correct answer.\n" "- Consider synonyms or paraphrases as valid matches.\n" "- Evaluate the correctness of the prediction compared to the answer." }, { "role": "user", "content": "Please evaluate the following video-based question-answer pair:\n\n" f"Question: {question}\n" f"Correct Answer: {answer}\n" f"Predicted Answer: {pred}\n\n" "Provide your evaluation only as a yes/no and score where the score is an integer value between 0 and 5, with 5 indicating the highest meaningful match. " "Please generate the response in the form of a Python dictionary string with keys 'pred' and 'score', where value of 'pred' is a string of 'yes' or 'no' and value of 'score' is in INTEGER, not STRING." "DO NOT PROVIDE ANY OTHER OUTPUT TEXT OR EXPLANATION. Only provide the Python dictionary string. " "For example, your response should look like this: {'pred': 'yes', 'score': 4.8}." } ] ) # Convert response to a Python dictionary. response_message = completion["choices"][0]["message"]["content"] response_dict = ast.literal_eval(response_message) result_qa_pair = [response_dict, qa_set] # Save the question-answer pairs to a json file. with open(f"{output_dir}/{key}.json", "w") as f: json.dump(result_qa_pair, f) except Exception as e: print(f"Error processing file '{key}': {e}")
Evaluates question and answer pairs using GPT-3 Returns a score for correctness.
179,687
import argparse import torch from videollava.constants import ( IMAGE_TOKEN_INDEX, DEFAULT_IMAGE_TOKEN, DEFAULT_IM_START_TOKEN, DEFAULT_IM_END_TOKEN, IMAGE_PLACEHOLDER, ) from videollava.conversation import conv_templates, SeparatorStyle from videollava.model.builder import load_pretrained_model from videollava.utils import disable_torch_init from videollava.mm_utils import ( process_images, tokenizer_image_token, get_model_name_from_path, KeywordsStoppingCriteria, ) from PIL import Image import requests from PIL import Image from io import BytesIO import re def image_parser(args): out = args.image_file.split(args.sep) return out def load_images(image_files): out = [] for image_file in image_files: image = load_image(image_file) out.append(image) return out IMAGE_TOKEN_INDEX = -200 DEFAULT_IMAGE_TOKEN = "<image>" DEFAULT_IM_START_TOKEN = "<im_start>" DEFAULT_IM_END_TOKEN = "<im_end>" IMAGE_PLACEHOLDER = "<image-placeholder>" class SeparatorStyle(Enum): """Different separator style.""" SINGLE = auto() TWO = auto() MPT = auto() PLAIN = auto() LLAMA_2 = auto() conv_templates = { "default": conv_vicuna_v0, "v0": conv_vicuna_v0, "v1": conv_vicuna_v1, "vicuna_v1": conv_vicuna_v1, "llama_2": conv_llama_2, "plain": conv_llava_plain, "v0_plain": conv_llava_plain, "llava_v0": conv_llava_v0, "v0_mmtag": conv_llava_v0_mmtag, "llava_v1": conv_llava_v1, "v1_mmtag": conv_llava_v1_mmtag, "llava_llama_2": conv_llava_llama_2, "mpt": conv_mpt, } def load_pretrained_model(model_path, model_base, model_name, load_8bit=False, load_4bit=False, device_map="auto", device="cuda", **kwargs): kwargs = {"device_map": device_map, **kwargs} if device != "cuda": kwargs['device_map'] = {"": device} if load_8bit: kwargs['load_in_8bit'] = True elif load_4bit: kwargs['load_in_4bit'] = True kwargs['quantization_config'] = BitsAndBytesConfig( load_in_4bit=True, bnb_4bit_compute_dtype=torch.float16, bnb_4bit_use_double_quant=True, bnb_4bit_quant_type='nf4' ) else: kwargs['torch_dtype'] = torch.float16 if 'llava' in model_name.lower(): # Load LLaVA model if 'lora' in model_name.lower() and model_base is None: warnings.warn('There is `lora` in model name but no `model_base` is provided. If you are loading a LoRA model, please provide the `model_base` argument. Detailed instruction: https://github.com/haotian-liu/LLaVA#launch-a-model-worker-lora-weights-unmerged.') if 'lora' in model_name.lower() and model_base is not None: lora_cfg_pretrained = AutoConfig.from_pretrained(model_path) tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=False) print('Loading LLaVA from base model...') model = LlavaLlamaForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, config=lora_cfg_pretrained, **kwargs) token_num, tokem_dim = model.lm_head.out_features, model.lm_head.in_features if model.lm_head.weight.shape[0] != token_num: model.lm_head.weight = torch.nn.Parameter(torch.empty(token_num, tokem_dim, device=model.device, dtype=model.dtype)) model.model.embed_tokens.weight = torch.nn.Parameter(torch.empty(token_num, tokem_dim, device=model.device, dtype=model.dtype)) print('Loading additional LLaVA weights...') if os.path.exists(os.path.join(model_path, 'non_lora_trainables.bin')): non_lora_trainables = torch.load(os.path.join(model_path, 'non_lora_trainables.bin'), map_location='cpu') else: # this is probably from HF Hub from huggingface_hub import hf_hub_download def load_from_hf(repo_id, filename, subfolder=None): cache_file = hf_hub_download( repo_id=repo_id, filename=filename, subfolder=subfolder) return torch.load(cache_file, map_location='cpu') non_lora_trainables = load_from_hf(model_path, 'non_lora_trainables.bin') non_lora_trainables = {(k[11:] if k.startswith('base_model.') else k): v for k, v in non_lora_trainables.items()} if any(k.startswith('model.model.') for k in non_lora_trainables): non_lora_trainables = {(k[6:] if k.startswith('model.') else k): v for k, v in non_lora_trainables.items()} model.load_state_dict(non_lora_trainables, strict=False) from peft import PeftModel print('Loading LoRA weights...') model = PeftModel.from_pretrained(model, model_path) print('Merging LoRA weights...') model = model.merge_and_unload() print('Model is loaded...') elif model_base is not None: # this may be mm projector only print('Loading LLaVA from base model...') if 'mpt' in model_name.lower(): if not os.path.isfile(os.path.join(model_path, 'configuration_mpt.py')): shutil.copyfile(os.path.join(model_base, 'configuration_mpt.py'), os.path.join(model_path, 'configuration_mpt.py')) tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=True) cfg_pretrained = AutoConfig.from_pretrained(model_path, trust_remote_code=True) model = LlavaMPTForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, config=cfg_pretrained, **kwargs) else: tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=False) cfg_pretrained = AutoConfig.from_pretrained(model_path) model = LlavaLlamaForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, config=cfg_pretrained, **kwargs) mm_projector_weights = torch.load(os.path.join(model_path, 'mm_projector.bin'), map_location='cpu') mm_projector_weights = {k: v.to(torch.float16) for k, v in mm_projector_weights.items()} model.load_state_dict(mm_projector_weights, strict=False) else: if 'mpt' in model_name.lower(): tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=True) model = LlavaMPTForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs) else: tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=False) model = LlavaLlamaForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs) else: # Load language model if model_base is not None: # PEFT model from peft import PeftModel tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=False) model = AutoModelForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, **kwargs) print(f"Loading LoRA weights from {model_path}") model = PeftModel.from_pretrained(model, model_path) print(f"Merging weights") model = model.merge_and_unload() print('Convert to FP16...') model.to(torch.float16) else: use_fast = False if 'mpt' in model_name.lower(): tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=True) model = AutoModelForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, trust_remote_code=True, **kwargs) else: tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=False) model = AutoModelForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs) # ========================================================================================================== processor = {'image': None, 'video': None} if 'llava' in model_name.lower(): mm_use_im_start_end = getattr(model.config, "mm_use_im_start_end", False) mm_use_im_patch_token = getattr(model.config, "mm_use_im_patch_token", True) if mm_use_im_patch_token: tokenizer.add_tokens([DEFAULT_IMAGE_PATCH_TOKEN], special_tokens=True) tokenizer.add_tokens([DEFAULT_VIDEO_PATCH_TOKEN], special_tokens=True) if mm_use_im_start_end: tokenizer.add_tokens([DEFAULT_IM_START_TOKEN, DEFAULT_IM_END_TOKEN], special_tokens=True) tokenizer.add_tokens([DEFAULT_VID_START_TOKEN, DEFAULT_VID_END_TOKEN], special_tokens=True) model.resize_token_embeddings(len(tokenizer)) if model.config.mm_image_tower is not None: image_tower = model.get_image_tower() if not image_tower.is_loaded: image_tower.load_model() image_tower.to(device=device, dtype=torch.float16) image_processor = image_tower.image_processor processor['image'] = image_processor if model.config.mm_video_tower is not None: video_tower = model.get_video_tower() if not video_tower.is_loaded: video_tower.load_model() video_tower.to(device=device, dtype=torch.float16) video_processor = video_tower.video_processor processor['video'] = video_processor # ========================================================================================================== if hasattr(model.config, "max_sequence_length"): context_len = model.config.max_sequence_length else: context_len = 2048 return tokenizer, model, processor, context_len def disable_torch_init(): """ Disable the redundant torch default initialization to accelerate model creation. """ import torch setattr(torch.nn.Linear, "reset_parameters", lambda self: None) setattr(torch.nn.LayerNorm, "reset_parameters", lambda self: None) def process_images(images, image_processor, model_cfg): image_aspect_ratio = getattr(model_cfg, "image_aspect_ratio", None) new_images = [] if image_aspect_ratio == 'pad': for image in images: image = expand2square(image, tuple(int(x*255) for x in image_processor.image_mean)) image = image_processor.preprocess(image, return_tensors='pt')['pixel_values'][0] new_images.append(image) else: return image_processor(images, return_tensors='pt')['pixel_values'] if all(x.shape == new_images[0].shape for x in new_images): new_images = torch.stack(new_images, dim=0) return new_images def tokenizer_image_token(prompt, tokenizer, image_token_index=IMAGE_TOKEN_INDEX, return_tensors=None): prompt_chunks = [tokenizer(chunk).input_ids for chunk in prompt.split('<image>')] def insert_separator(X, sep): return [ele for sublist in zip(X, [sep]*len(X)) for ele in sublist][:-1] input_ids = [] offset = 0 if len(prompt_chunks) > 0 and len(prompt_chunks[0]) > 0 and prompt_chunks[0][0] == tokenizer.bos_token_id: offset = 1 input_ids.append(prompt_chunks[0][0]) for x in insert_separator(prompt_chunks, [image_token_index] * (offset + 1)): input_ids.extend(x[offset:]) if return_tensors is not None: if return_tensors == 'pt': return torch.tensor(input_ids, dtype=torch.long) raise ValueError(f'Unsupported tensor type: {return_tensors}') return input_ids def get_model_name_from_path(model_path): model_path = model_path.strip("/") model_paths = model_path.split("/") if model_paths[-1].startswith('checkpoint-'): return model_paths[-2] + "_" + model_paths[-1] else: return model_paths[-1] class KeywordsStoppingCriteria(StoppingCriteria): def __init__(self, keywords, tokenizer, input_ids): self.keywords = keywords self.keyword_ids = [] self.max_keyword_len = 0 for keyword in keywords: cur_keyword_ids = tokenizer(keyword).input_ids if len(cur_keyword_ids) > 1 and cur_keyword_ids[0] == tokenizer.bos_token_id: cur_keyword_ids = cur_keyword_ids[1:] if len(cur_keyword_ids) > self.max_keyword_len: self.max_keyword_len = len(cur_keyword_ids) self.keyword_ids.append(torch.tensor(cur_keyword_ids)) self.tokenizer = tokenizer self.start_len = input_ids.shape[1] def call_for_batch(self, output_ids: torch.LongTensor, scores: torch.FloatTensor, **kwargs) -> bool: offset = min(output_ids.shape[1] - self.start_len, self.max_keyword_len) self.keyword_ids = [keyword_id.to(output_ids.device) for keyword_id in self.keyword_ids] for keyword_id in self.keyword_ids: if (output_ids[0, -keyword_id.shape[0]:] == keyword_id).all(): return True outputs = self.tokenizer.batch_decode(output_ids[:, -offset:], skip_special_tokens=True)[0] for keyword in self.keywords: if keyword in outputs: return True return False def __call__(self, output_ids: torch.LongTensor, scores: torch.FloatTensor, **kwargs) -> bool: outputs = [] for i in range(output_ids.shape[0]): outputs.append(self.call_for_batch(output_ids[i].unsqueeze(0), scores)) return all(outputs) def eval_model(args): # Model disable_torch_init() model_name = get_model_name_from_path(args.model_path) tokenizer, model, image_processor, context_len = load_pretrained_model( args.model_path, args.model_base, model_name ) qs = args.query image_token_se = DEFAULT_IM_START_TOKEN + DEFAULT_IMAGE_TOKEN + DEFAULT_IM_END_TOKEN if IMAGE_PLACEHOLDER in qs: if model.config.mm_use_im_start_end: qs = re.sub(IMAGE_PLACEHOLDER, image_token_se, qs) else: qs = re.sub(IMAGE_PLACEHOLDER, DEFAULT_IMAGE_TOKEN, qs) else: if model.config.mm_use_im_start_end: qs = image_token_se + "\n" + qs else: qs = DEFAULT_IMAGE_TOKEN + "\n" + qs if "llama-2" in model_name.lower(): conv_mode = "llava_llama_2" elif "v1" in model_name.lower(): conv_mode = "llava_v1" elif "mpt" in model_name.lower(): conv_mode = "mpt" else: conv_mode = "llava_v0" if args.conv_mode is not None and conv_mode != args.conv_mode: print( "[WARNING] the auto inferred conversation mode is {}, while `--conv-mode` is {}, using {}".format( conv_mode, args.conv_mode, args.conv_mode ) ) else: args.conv_mode = conv_mode conv = conv_templates[args.conv_mode].copy() conv.append_message(conv.roles[0], qs) conv.append_message(conv.roles[1], None) prompt = conv.get_prompt() image_files = image_parser(args) images = load_images(image_files) images_tensor = process_images( images, image_processor, model.config ).to(model.device, dtype=torch.float16) input_ids = ( tokenizer_image_token(prompt, tokenizer, IMAGE_TOKEN_INDEX, return_tensors="pt") .unsqueeze(0) .cuda() ) stop_str = conv.sep if conv.sep_style != SeparatorStyle.TWO else conv.sep2 keywords = [stop_str] stopping_criteria = KeywordsStoppingCriteria(keywords, tokenizer, input_ids) with torch.inference_mode(): output_ids = model.generate( input_ids, images=images_tensor, do_sample=True if args.temperature > 0 else False, temperature=args.temperature, top_p=args.top_p, num_beams=args.num_beams, max_new_tokens=args.max_new_tokens, use_cache=True, stopping_criteria=[stopping_criteria], ) input_token_len = input_ids.shape[1] n_diff_input_output = (input_ids != output_ids[:, :input_token_len]).sum().item() if n_diff_input_output > 0: print( f"[Warning] {n_diff_input_output} output_ids are not the same as the input_ids" ) outputs = tokenizer.batch_decode( output_ids[:, input_token_len:], skip_special_tokens=True )[0] outputs = outputs.strip() if outputs.endswith(stop_str): outputs = outputs[: -len(stop_str)] outputs = outputs.strip() print(outputs)
null
179,688
import argparse import torch import os import json from tqdm import tqdm import shortuuid from videollava.constants import IMAGE_TOKEN_INDEX, DEFAULT_IMAGE_TOKEN, DEFAULT_IM_START_TOKEN, DEFAULT_IM_END_TOKEN from videollava.conversation import conv_templates, SeparatorStyle from videollava.model.builder import load_pretrained_model from videollava.utils import disable_torch_init from videollava.mm_utils import tokenizer_image_token, get_model_name_from_path, KeywordsStoppingCriteria from PIL import Image import math def get_chunk(lst, n, k): chunks = split_list(lst, n) return chunks[k] IMAGE_TOKEN_INDEX = -200 DEFAULT_IMAGE_TOKEN = "<image>" DEFAULT_IM_START_TOKEN = "<im_start>" DEFAULT_IM_END_TOKEN = "<im_end>" class SeparatorStyle(Enum): """Different separator style.""" SINGLE = auto() TWO = auto() MPT = auto() PLAIN = auto() LLAMA_2 = auto() conv_templates = { "default": conv_vicuna_v0, "v0": conv_vicuna_v0, "v1": conv_vicuna_v1, "vicuna_v1": conv_vicuna_v1, "llama_2": conv_llama_2, "plain": conv_llava_plain, "v0_plain": conv_llava_plain, "llava_v0": conv_llava_v0, "v0_mmtag": conv_llava_v0_mmtag, "llava_v1": conv_llava_v1, "v1_mmtag": conv_llava_v1_mmtag, "llava_llama_2": conv_llava_llama_2, "mpt": conv_mpt, } def load_pretrained_model(model_path, model_base, model_name, load_8bit=False, load_4bit=False, device_map="auto", device="cuda", **kwargs): kwargs = {"device_map": device_map, **kwargs} if device != "cuda": kwargs['device_map'] = {"": device} if load_8bit: kwargs['load_in_8bit'] = True elif load_4bit: kwargs['load_in_4bit'] = True kwargs['quantization_config'] = BitsAndBytesConfig( load_in_4bit=True, bnb_4bit_compute_dtype=torch.float16, bnb_4bit_use_double_quant=True, bnb_4bit_quant_type='nf4' ) else: kwargs['torch_dtype'] = torch.float16 if 'llava' in model_name.lower(): # Load LLaVA model if 'lora' in model_name.lower() and model_base is None: warnings.warn('There is `lora` in model name but no `model_base` is provided. If you are loading a LoRA model, please provide the `model_base` argument. Detailed instruction: https://github.com/haotian-liu/LLaVA#launch-a-model-worker-lora-weights-unmerged.') if 'lora' in model_name.lower() and model_base is not None: lora_cfg_pretrained = AutoConfig.from_pretrained(model_path) tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=False) print('Loading LLaVA from base model...') model = LlavaLlamaForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, config=lora_cfg_pretrained, **kwargs) token_num, tokem_dim = model.lm_head.out_features, model.lm_head.in_features if model.lm_head.weight.shape[0] != token_num: model.lm_head.weight = torch.nn.Parameter(torch.empty(token_num, tokem_dim, device=model.device, dtype=model.dtype)) model.model.embed_tokens.weight = torch.nn.Parameter(torch.empty(token_num, tokem_dim, device=model.device, dtype=model.dtype)) print('Loading additional LLaVA weights...') if os.path.exists(os.path.join(model_path, 'non_lora_trainables.bin')): non_lora_trainables = torch.load(os.path.join(model_path, 'non_lora_trainables.bin'), map_location='cpu') else: # this is probably from HF Hub from huggingface_hub import hf_hub_download def load_from_hf(repo_id, filename, subfolder=None): cache_file = hf_hub_download( repo_id=repo_id, filename=filename, subfolder=subfolder) return torch.load(cache_file, map_location='cpu') non_lora_trainables = load_from_hf(model_path, 'non_lora_trainables.bin') non_lora_trainables = {(k[11:] if k.startswith('base_model.') else k): v for k, v in non_lora_trainables.items()} if any(k.startswith('model.model.') for k in non_lora_trainables): non_lora_trainables = {(k[6:] if k.startswith('model.') else k): v for k, v in non_lora_trainables.items()} model.load_state_dict(non_lora_trainables, strict=False) from peft import PeftModel print('Loading LoRA weights...') model = PeftModel.from_pretrained(model, model_path) print('Merging LoRA weights...') model = model.merge_and_unload() print('Model is loaded...') elif model_base is not None: # this may be mm projector only print('Loading LLaVA from base model...') if 'mpt' in model_name.lower(): if not os.path.isfile(os.path.join(model_path, 'configuration_mpt.py')): shutil.copyfile(os.path.join(model_base, 'configuration_mpt.py'), os.path.join(model_path, 'configuration_mpt.py')) tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=True) cfg_pretrained = AutoConfig.from_pretrained(model_path, trust_remote_code=True) model = LlavaMPTForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, config=cfg_pretrained, **kwargs) else: tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=False) cfg_pretrained = AutoConfig.from_pretrained(model_path) model = LlavaLlamaForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, config=cfg_pretrained, **kwargs) mm_projector_weights = torch.load(os.path.join(model_path, 'mm_projector.bin'), map_location='cpu') mm_projector_weights = {k: v.to(torch.float16) for k, v in mm_projector_weights.items()} model.load_state_dict(mm_projector_weights, strict=False) else: if 'mpt' in model_name.lower(): tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=True) model = LlavaMPTForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs) else: tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=False) model = LlavaLlamaForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs) else: # Load language model if model_base is not None: # PEFT model from peft import PeftModel tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=False) model = AutoModelForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, **kwargs) print(f"Loading LoRA weights from {model_path}") model = PeftModel.from_pretrained(model, model_path) print(f"Merging weights") model = model.merge_and_unload() print('Convert to FP16...') model.to(torch.float16) else: use_fast = False if 'mpt' in model_name.lower(): tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=True) model = AutoModelForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, trust_remote_code=True, **kwargs) else: tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=False) model = AutoModelForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs) # ========================================================================================================== processor = {'image': None, 'video': None} if 'llava' in model_name.lower(): mm_use_im_start_end = getattr(model.config, "mm_use_im_start_end", False) mm_use_im_patch_token = getattr(model.config, "mm_use_im_patch_token", True) if mm_use_im_patch_token: tokenizer.add_tokens([DEFAULT_IMAGE_PATCH_TOKEN], special_tokens=True) tokenizer.add_tokens([DEFAULT_VIDEO_PATCH_TOKEN], special_tokens=True) if mm_use_im_start_end: tokenizer.add_tokens([DEFAULT_IM_START_TOKEN, DEFAULT_IM_END_TOKEN], special_tokens=True) tokenizer.add_tokens([DEFAULT_VID_START_TOKEN, DEFAULT_VID_END_TOKEN], special_tokens=True) model.resize_token_embeddings(len(tokenizer)) if model.config.mm_image_tower is not None: image_tower = model.get_image_tower() if not image_tower.is_loaded: image_tower.load_model() image_tower.to(device=device, dtype=torch.float16) image_processor = image_tower.image_processor processor['image'] = image_processor if model.config.mm_video_tower is not None: video_tower = model.get_video_tower() if not video_tower.is_loaded: video_tower.load_model() video_tower.to(device=device, dtype=torch.float16) video_processor = video_tower.video_processor processor['video'] = video_processor # ========================================================================================================== if hasattr(model.config, "max_sequence_length"): context_len = model.config.max_sequence_length else: context_len = 2048 return tokenizer, model, processor, context_len def disable_torch_init(): """ Disable the redundant torch default initialization to accelerate model creation. """ import torch setattr(torch.nn.Linear, "reset_parameters", lambda self: None) setattr(torch.nn.LayerNorm, "reset_parameters", lambda self: None) def tokenizer_image_token(prompt, tokenizer, image_token_index=IMAGE_TOKEN_INDEX, return_tensors=None): prompt_chunks = [tokenizer(chunk).input_ids for chunk in prompt.split('<image>')] def insert_separator(X, sep): return [ele for sublist in zip(X, [sep]*len(X)) for ele in sublist][:-1] input_ids = [] offset = 0 if len(prompt_chunks) > 0 and len(prompt_chunks[0]) > 0 and prompt_chunks[0][0] == tokenizer.bos_token_id: offset = 1 input_ids.append(prompt_chunks[0][0]) for x in insert_separator(prompt_chunks, [image_token_index] * (offset + 1)): input_ids.extend(x[offset:]) if return_tensors is not None: if return_tensors == 'pt': return torch.tensor(input_ids, dtype=torch.long) raise ValueError(f'Unsupported tensor type: {return_tensors}') return input_ids def get_model_name_from_path(model_path): model_path = model_path.strip("/") model_paths = model_path.split("/") if model_paths[-1].startswith('checkpoint-'): return model_paths[-2] + "_" + model_paths[-1] else: return model_paths[-1] class KeywordsStoppingCriteria(StoppingCriteria): def __init__(self, keywords, tokenizer, input_ids): self.keywords = keywords self.keyword_ids = [] self.max_keyword_len = 0 for keyword in keywords: cur_keyword_ids = tokenizer(keyword).input_ids if len(cur_keyword_ids) > 1 and cur_keyword_ids[0] == tokenizer.bos_token_id: cur_keyword_ids = cur_keyword_ids[1:] if len(cur_keyword_ids) > self.max_keyword_len: self.max_keyword_len = len(cur_keyword_ids) self.keyword_ids.append(torch.tensor(cur_keyword_ids)) self.tokenizer = tokenizer self.start_len = input_ids.shape[1] def call_for_batch(self, output_ids: torch.LongTensor, scores: torch.FloatTensor, **kwargs) -> bool: offset = min(output_ids.shape[1] - self.start_len, self.max_keyword_len) self.keyword_ids = [keyword_id.to(output_ids.device) for keyword_id in self.keyword_ids] for keyword_id in self.keyword_ids: if (output_ids[0, -keyword_id.shape[0]:] == keyword_id).all(): return True outputs = self.tokenizer.batch_decode(output_ids[:, -offset:], skip_special_tokens=True)[0] for keyword in self.keywords: if keyword in outputs: return True return False def __call__(self, output_ids: torch.LongTensor, scores: torch.FloatTensor, **kwargs) -> bool: outputs = [] for i in range(output_ids.shape[0]): outputs.append(self.call_for_batch(output_ids[i].unsqueeze(0), scores)) return all(outputs) def eval_model(args): # Model disable_torch_init() model_path = os.path.expanduser(args.model_path) model_name = get_model_name_from_path(model_path) tokenizer, model, processor, context_len = load_pretrained_model(model_path, args.model_base, model_name) if args.return_gating_logit: from videollava.utils import get_gating_logit_by_hook print(model) fea_hooks = get_gating_logit_by_hook(model) all_gating_logits = {} image_processor = processor['image'] questions = json.load(open(os.path.expanduser(args.question_file), "r")) questions = get_chunk(questions, args.num_chunks, args.chunk_idx) answers_file = os.path.expanduser(args.answers_file) os.makedirs(os.path.dirname(answers_file), exist_ok=True) ans_file = open(answers_file, "w") for i, line in enumerate(tqdm(questions)): idx = line["id"] question = line['conversations'][0] qs = question['value'].replace('<image>', '').strip() cur_prompt = qs if 'image' in line: image_file = line["image"] image = Image.open(os.path.join(args.image_folder, image_file)) image_tensor = image_processor.preprocess(image, return_tensors='pt')['pixel_values'][0] images = image_tensor.unsqueeze(0).half().cuda() if getattr(model.config, 'mm_use_im_start_end', False): qs = DEFAULT_IM_START_TOKEN + DEFAULT_IMAGE_TOKEN + DEFAULT_IM_END_TOKEN + '\n' + qs else: qs = DEFAULT_IMAGE_TOKEN + '\n' + qs cur_prompt = '<image>' + '\n' + cur_prompt else: images = None if args.single_pred_prompt: qs = qs + '\n' + "Answer with the option's letter from the given choices directly." cur_prompt = cur_prompt + '\n' + "Answer with the option's letter from the given choices directly." conv = conv_templates[args.conv_mode].copy() conv.append_message(conv.roles[0], qs) conv.append_message(conv.roles[1], None) prompt = conv.get_prompt() input_ids = tokenizer_image_token(prompt, tokenizer, IMAGE_TOKEN_INDEX, return_tensors='pt').unsqueeze(0).cuda() stop_str = conv.sep if conv.sep_style != SeparatorStyle.TWO else conv.sep2 keywords = [stop_str] stopping_criteria = [KeywordsStoppingCriteria(keywords, tokenizer, input_ids)] if conv.version == "v0" else None with torch.inference_mode(): output_ids = model.generate( input_ids, images=images, do_sample=True if args.temperature > 0 else False, temperature=args.temperature, max_new_tokens=1024, use_cache=True if not args.return_gating_logit else False, stopping_criteria=stopping_criteria, ) if args.return_gating_logit: all_gating_logits[idx] = dict(gating_logit=fea_hooks, images=images if images is None else images.detach().cpu(), input_ids=input_ids.detach().cpu()) print(input_ids.shape, images.shape if images is not None else []) print('The number of hooks is:', len(fea_hooks), 'The shape of the first gating logit is:', fea_hooks[0].fea.shape) # print(output_ids) # import ipdb # ipdb.set_trace() input_token_len = input_ids.shape[1] n_diff_input_output = (input_ids != output_ids[:, :input_token_len]).sum().item() if n_diff_input_output > 0: print(f'[Warning] {n_diff_input_output} output_ids are not the same as the input_ids') outputs = tokenizer.batch_decode(output_ids[:, input_token_len:], skip_special_tokens=True)[0] outputs = outputs.strip() if outputs.endswith(stop_str): outputs = outputs[:-len(stop_str)] outputs = outputs.strip() print(outputs) # prompt for answer if args.answer_prompter: outputs_reasoning = outputs input_ids = tokenizer_image_token(prompt + outputs_reasoning + ' ###\nANSWER:', tokenizer, IMAGE_TOKEN_INDEX, return_tensors='pt').unsqueeze(0).cuda() with torch.inference_mode(): output_ids = model.generate( input_ids, images=images, do_sample=True if args.temperature > 0 else False, temperature=args.temperature, max_new_tokens=64, use_cache=True, stopping_criteria=[stopping_criteria]) input_token_len = input_ids.shape[1] n_diff_input_output = (input_ids != output_ids[:, :input_token_len]).sum().item() if n_diff_input_output > 0: print(f'[Warning] {n_diff_input_output} output_ids are not the same as the input_ids') outputs = tokenizer.batch_decode(output_ids[:, input_token_len:], skip_special_tokens=True)[0] outputs = outputs.strip() if outputs.endswith(stop_str): outputs = outputs[:-len(stop_str)] outputs = outputs.strip() outputs = outputs_reasoning + '\n The answer is ' + outputs ans_id = shortuuid.uuid() ans_file.write(json.dumps({"question_id": idx, "prompt": cur_prompt, "text": outputs, "answer_id": ans_id, "model_id": model_name, "metadata": {}}) + "\n") ans_file.flush() ans_file.close() if args.return_gating_logit: torch.save(all_gating_logits, 'vqa_science_all_gating_logits.pt')
null
179,692
import argparse import torch import os import json from tqdm import tqdm import shortuuid from videollava.constants import IMAGE_TOKEN_INDEX, DEFAULT_IMAGE_TOKEN, DEFAULT_IM_START_TOKEN, DEFAULT_IM_END_TOKEN from videollava.conversation import conv_templates, SeparatorStyle from videollava.model.builder import load_pretrained_model from videollava.utils import disable_torch_init from videollava.mm_utils import tokenizer_image_token, process_images, get_model_name_from_path from torch.utils.data import Dataset, DataLoader from PIL import Image import math def get_chunk(lst, n, k): chunks = split_list(lst, n) return chunks[k] def create_data_loader(questions, image_folder, tokenizer, image_processor, model_config, batch_size=1, num_workers=4): assert batch_size == 1, "batch_size must be 1" dataset = CustomDataset(questions, image_folder, tokenizer, image_processor, model_config) data_loader = DataLoader(dataset, batch_size=batch_size, num_workers=num_workers, shuffle=False) return data_loader def load_pretrained_model(model_path, model_base, model_name, load_8bit=False, load_4bit=False, device_map="auto", device="cuda", **kwargs): kwargs = {"device_map": device_map, **kwargs} if device != "cuda": kwargs['device_map'] = {"": device} if load_8bit: kwargs['load_in_8bit'] = True elif load_4bit: kwargs['load_in_4bit'] = True kwargs['quantization_config'] = BitsAndBytesConfig( load_in_4bit=True, bnb_4bit_compute_dtype=torch.float16, bnb_4bit_use_double_quant=True, bnb_4bit_quant_type='nf4' ) else: kwargs['torch_dtype'] = torch.float16 if 'llava' in model_name.lower(): # Load LLaVA model if 'lora' in model_name.lower() and model_base is None: warnings.warn('There is `lora` in model name but no `model_base` is provided. If you are loading a LoRA model, please provide the `model_base` argument. Detailed instruction: https://github.com/haotian-liu/LLaVA#launch-a-model-worker-lora-weights-unmerged.') if 'lora' in model_name.lower() and model_base is not None: lora_cfg_pretrained = AutoConfig.from_pretrained(model_path) tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=False) print('Loading LLaVA from base model...') model = LlavaLlamaForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, config=lora_cfg_pretrained, **kwargs) token_num, tokem_dim = model.lm_head.out_features, model.lm_head.in_features if model.lm_head.weight.shape[0] != token_num: model.lm_head.weight = torch.nn.Parameter(torch.empty(token_num, tokem_dim, device=model.device, dtype=model.dtype)) model.model.embed_tokens.weight = torch.nn.Parameter(torch.empty(token_num, tokem_dim, device=model.device, dtype=model.dtype)) print('Loading additional LLaVA weights...') if os.path.exists(os.path.join(model_path, 'non_lora_trainables.bin')): non_lora_trainables = torch.load(os.path.join(model_path, 'non_lora_trainables.bin'), map_location='cpu') else: # this is probably from HF Hub from huggingface_hub import hf_hub_download def load_from_hf(repo_id, filename, subfolder=None): cache_file = hf_hub_download( repo_id=repo_id, filename=filename, subfolder=subfolder) return torch.load(cache_file, map_location='cpu') non_lora_trainables = load_from_hf(model_path, 'non_lora_trainables.bin') non_lora_trainables = {(k[11:] if k.startswith('base_model.') else k): v for k, v in non_lora_trainables.items()} if any(k.startswith('model.model.') for k in non_lora_trainables): non_lora_trainables = {(k[6:] if k.startswith('model.') else k): v for k, v in non_lora_trainables.items()} model.load_state_dict(non_lora_trainables, strict=False) from peft import PeftModel print('Loading LoRA weights...') model = PeftModel.from_pretrained(model, model_path) print('Merging LoRA weights...') model = model.merge_and_unload() print('Model is loaded...') elif model_base is not None: # this may be mm projector only print('Loading LLaVA from base model...') if 'mpt' in model_name.lower(): if not os.path.isfile(os.path.join(model_path, 'configuration_mpt.py')): shutil.copyfile(os.path.join(model_base, 'configuration_mpt.py'), os.path.join(model_path, 'configuration_mpt.py')) tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=True) cfg_pretrained = AutoConfig.from_pretrained(model_path, trust_remote_code=True) model = LlavaMPTForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, config=cfg_pretrained, **kwargs) else: tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=False) cfg_pretrained = AutoConfig.from_pretrained(model_path) model = LlavaLlamaForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, config=cfg_pretrained, **kwargs) mm_projector_weights = torch.load(os.path.join(model_path, 'mm_projector.bin'), map_location='cpu') mm_projector_weights = {k: v.to(torch.float16) for k, v in mm_projector_weights.items()} model.load_state_dict(mm_projector_weights, strict=False) else: if 'mpt' in model_name.lower(): tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=True) model = LlavaMPTForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs) else: tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=False) model = LlavaLlamaForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs) else: # Load language model if model_base is not None: # PEFT model from peft import PeftModel tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=False) model = AutoModelForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, **kwargs) print(f"Loading LoRA weights from {model_path}") model = PeftModel.from_pretrained(model, model_path) print(f"Merging weights") model = model.merge_and_unload() print('Convert to FP16...') model.to(torch.float16) else: use_fast = False if 'mpt' in model_name.lower(): tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=True) model = AutoModelForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, trust_remote_code=True, **kwargs) else: tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=False) model = AutoModelForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs) # ========================================================================================================== processor = {'image': None, 'video': None} if 'llava' in model_name.lower(): mm_use_im_start_end = getattr(model.config, "mm_use_im_start_end", False) mm_use_im_patch_token = getattr(model.config, "mm_use_im_patch_token", True) if mm_use_im_patch_token: tokenizer.add_tokens([DEFAULT_IMAGE_PATCH_TOKEN], special_tokens=True) tokenizer.add_tokens([DEFAULT_VIDEO_PATCH_TOKEN], special_tokens=True) if mm_use_im_start_end: tokenizer.add_tokens([DEFAULT_IM_START_TOKEN, DEFAULT_IM_END_TOKEN], special_tokens=True) tokenizer.add_tokens([DEFAULT_VID_START_TOKEN, DEFAULT_VID_END_TOKEN], special_tokens=True) model.resize_token_embeddings(len(tokenizer)) if model.config.mm_image_tower is not None: image_tower = model.get_image_tower() if not image_tower.is_loaded: image_tower.load_model() image_tower.to(device=device, dtype=torch.float16) image_processor = image_tower.image_processor processor['image'] = image_processor if model.config.mm_video_tower is not None: video_tower = model.get_video_tower() if not video_tower.is_loaded: video_tower.load_model() video_tower.to(device=device, dtype=torch.float16) video_processor = video_tower.video_processor processor['video'] = video_processor # ========================================================================================================== if hasattr(model.config, "max_sequence_length"): context_len = model.config.max_sequence_length else: context_len = 2048 return tokenizer, model, processor, context_len def disable_torch_init(): """ Disable the redundant torch default initialization to accelerate model creation. """ import torch setattr(torch.nn.Linear, "reset_parameters", lambda self: None) setattr(torch.nn.LayerNorm, "reset_parameters", lambda self: None) def get_model_name_from_path(model_path): model_path = model_path.strip("/") model_paths = model_path.split("/") if model_paths[-1].startswith('checkpoint-'): return model_paths[-2] + "_" + model_paths[-1] else: return model_paths[-1] def eval_model(args): # Model disable_torch_init() model_path = os.path.expanduser(args.model_path) model_name = get_model_name_from_path(model_path) tokenizer, model, processor, context_len = load_pretrained_model(model_path, args.model_base, model_name) if args.return_gating_logit: from videollava.utils import get_gating_logit_by_hook print(model) fea_hooks = get_gating_logit_by_hook(model) all_gating_logits = {} image_processor = processor['image'] questions = [json.loads(q) for q in open(os.path.expanduser(args.question_file), "r")] questions = get_chunk(questions, args.num_chunks, args.chunk_idx) answers_file = os.path.expanduser(args.answers_file) os.makedirs(os.path.dirname(answers_file), exist_ok=True) ans_file = open(answers_file, "w") if 'plain' in model_name and 'finetune' not in model_name.lower() and 'mmtag' not in args.conv_mode: args.conv_mode = args.conv_mode + '_mmtag' print(f'It seems that this is a plain model, but it is not using a mmtag prompt, auto switching to {args.conv_mode}.') data_loader = create_data_loader(questions, args.image_folder, tokenizer, image_processor, model.config) cnt = -1 for (input_ids, image_tensor), line in tqdm(zip(data_loader, questions), total=len(questions)): cnt += 1 # if cnt == 30: # break idx = line["question_id"] cur_prompt = line["text"] input_ids = input_ids.to(device='cuda', non_blocking=True) with torch.inference_mode(): output_ids = model.generate( input_ids, images=image_tensor.to(dtype=torch.float16, device='cuda', non_blocking=True), do_sample=True if args.temperature > 0 else False, temperature=args.temperature, top_p=args.top_p, num_beams=args.num_beams, max_new_tokens=args.max_new_tokens, use_cache=True if not args.return_gating_logit else False) if args.return_gating_logit: # import ipdb # ipdb.set_trace() all_gating_logits[cnt] = dict(gating_logit=[i.fea for i in fea_hooks], images=image_tensor if image_tensor is None else image_tensor.detach().cpu(), input_ids=input_ids.detach().cpu(), output_ids=output_ids.detach().cpu()) print(input_ids.shape, output_ids.shape, fea_hooks[0].fea.shape, image_tensor.shape if image_tensor is not None else []) assert fea_hooks[0].fea.shape[0] + 1 == output_ids.shape[1] + 575 print('The number of hooks is:', len(fea_hooks)) input_token_len = input_ids.shape[1] n_diff_input_output = (input_ids != output_ids[:, :input_token_len]).sum().item() if n_diff_input_output > 0: print(f'[Warning] {n_diff_input_output} output_ids are not the same as the input_ids') outputs = tokenizer.batch_decode(output_ids[:, input_token_len:], skip_special_tokens=True)[0] outputs = outputs.strip() print(outputs) ans_id = shortuuid.uuid() ans_file.write(json.dumps({"question_id": idx, "prompt": cur_prompt, "text": outputs, "answer_id": ans_id, "model_id": model_name, "metadata": {}}) + "\n") # ans_file.flush() ans_file.close() if args.return_gating_logit: torch.save(all_gating_logits, 'text_qa_all_gating_logits.pt')
null
179,698
import argparse import json import os import time import pandas as pd import tensor_parallel as tp import torch from tqdm import tqdm from transformers import LlamaForCausalLM, LlamaTokenizer, AutoTokenizer, AutoModelForCausalLM def load(ckpt_dir, model_type, cache_dir): # n_gpus = torch.cuda.device_count() n_gpus = 1 if model_type == 'llama': # we use tensor parallel for loading llama tokenizer = AutoTokenizer.from_pretrained(ckpt_dir, use_fast=False, padding_side="left", cache_dir=cache_dir) model = LlamaForCausalLM.from_pretrained(ckpt_dir, low_cpu_mem_usage=True, torch_dtype=torch.float16, cache_dir=cache_dir) model = tp.tensor_parallel(model, [i for i in range(n_gpus)]) tokenizer.pad_token_id = 0 if tokenizer.pad_token_id is None else tokenizer.pad_token_id tokenizer.bos_token_id = 1 elif model_type == 'qwen': from videollava.model.language_model.qwen.tokenization_qwen import QWenTokenizer from videollava.model.language_model.qwen.modeling_qwen import QWenLMHeadModel model = QWenLMHeadModel.from_pretrained(ckpt_dir, low_cpu_mem_usage=True, torch_dtype=torch.float16, cache_dir=cache_dir) model = tp.tensor_parallel(model, [i for i in range(n_gpus)]) tokenizer = QWenTokenizer.from_pretrained(ckpt_dir, use_fast=False, padding_side="left", cache_dir=cache_dir) tokenizer.add_special_tokens({'unk_token': '<|extra_0|>', 'bos_token': '<|extra_1|>', 'eos_token': '<|endoftext|>'}) tokenizer.pad_token = tokenizer.unk_token elif model_type == 'llava': from videollava.mm_utils import get_model_name_from_path from videollava.model.builder import load_pretrained_model load_8bit, load_4bit = False, False model_base = None model_name = get_model_name_from_path(ckpt_dir) tokenizer, model, _, _ = load_pretrained_model(ckpt_dir, model_base, model_name, load_8bit, load_4bit, padding_side="left") model.eval() return model, tokenizer def compute_metric(output_filename): with open(output_filename, 'r') as f: run_results = json.load(f) total_acc = 0 total_num = 0 for task in run_results: acc = 0 pred_answers = run_results[task]['pred_answers'] gold_answers = run_results[task]['gold_answers'] for pred, gold in zip(pred_answers, gold_answers): if pred == gold: acc += 1 print("ACC-%s: %.4f" % (task, acc / len(gold_answers))) total_acc += acc total_num += len(gold_answers) print("ACC-all: %.4f" % (total_acc / total_num))
null
179,699
import argparse import json import os import time import pandas as pd import tensor_parallel as tp import torch from tqdm import tqdm from transformers import LlamaForCausalLM, LlamaTokenizer, AutoTokenizer, AutoModelForCausalLM def format_subject(subject): l = subject.split("_") s = "" for entry in l: s += " " + entry return s def format_example(df, idx, include_answer=True): prompt = df.iloc[idx, 0] k = df.shape[1] - 2 for j in range(k): prompt += "\n{}. {}".format(choices[j], df.iloc[idx, j + 1]) prompt += "\nAnswer:" if include_answer: prompt += " {}\n\n".format(df.iloc[idx, k + 1]) return prompt def gen_prompt(train_df, subject, k=-1): prompt = "The following are multiple choice questions (with answers) about {}.\n\n".format(format_subject(subject)) if k == -1: k = train_df.shape[0] for i in range(k): prompt += format_example(train_df, i) return prompt
null
179,700
import argparse import json import os import time import pandas as pd import tensor_parallel as tp import torch from tqdm import tqdm from transformers import LlamaForCausalLM, LlamaTokenizer, AutoTokenizer, AutoModelForCausalLM def prepare_input(tokenizer, prompts): input_tokens = tokenizer.batch_encode_plus(prompts, return_tensors="pt", padding=True) input_tokens = {k: input_tokens[k] for k in input_tokens if k in ["input_ids", "attention_mask"]} for t in input_tokens: if torch.is_tensor(input_tokens[t]): input_tokens[t] = input_tokens[t].to('cuda') return input_tokens def batch_split(prompts, batch_num): batch_prompts = [] mini_batch = [] for prompt in prompts: mini_batch.append(prompt) if len(mini_batch) == batch_num: batch_prompts.append(mini_batch) mini_batch = [] if len(mini_batch) != 0: batch_prompts.append(mini_batch) return batch_prompts def batch_infer(model, tokenizer, prompts): batch_size = 8 answers = [] for batch_input in tqdm(batch_split(prompts, batch_size)): encode_inputs = prepare_input(tokenizer, batch_input) outputs = model.generate(**encode_inputs, max_new_tokens=1, pad_token_id=tokenizer.pad_token_id) answers.extend(tokenizer.batch_decode(outputs, skip_special_tokens=True)) answers = [answer[-1] for answer in answers] return answers
null
179,702
def get_question_text(problem): question = problem['question'] return question def get_context_text(problem, use_caption): txt_context = problem['hint'] img_context = problem['caption'] if use_caption else "" context = " ".join([txt_context, img_context]).strip() if context == "": context = "N/A" return context def get_choice_text(probelm, options): choices = probelm['choices'] choice_list = [] for i, c in enumerate(choices): choice_list.append("({}) {}".format(options[i], c)) choice_txt = " ".join(choice_list) #print(choice_txt) return choice_txt def get_answer(problem, options): return options[problem['answer']] def get_lecture_text(problem): # \\n: GPT-3 can generate the lecture with more tokens. lecture = problem['lecture'].replace("\n", "\\n") return lecture def get_solution_text(problem): # \\n: GPT-3 can generate the solution with more tokens solution = problem['solution'].replace("\n", "\\n") return solution def create_one_example_gpt4(format, question, context, choice, answer, lecture, solution, test_example=True): input_format, output_format = format.split("-") ## Inputs if input_format == "CQM": input = f"Context: {context}\nQuestion: {question}\nOptions: {choice}\n" elif input_format == "QCM": input = f"Question: {question}\nContext: {context}\nOptions: {choice}\n" # upper bound experiment elif input_format == "QCML": input = f"Question: {question}\nContext: {context}\nOptions: {choice}\nBECAUSE: {lecture}\n" elif input_format == "QCME": input = f"Question: {question}\nContext: {context}\nOptions: {choice}\nBECAUSE: {solution}\n" elif input_format == "QCMLE": input = f"Question: {question}\nContext: {context}\nOptions: {choice}\nBECAUSE: {lecture} {solution}\n" elif input_format == "QCLM": input = f"Question: {question}\nContext: {context}\nBECAUSE: {lecture}\nOptions: {choice}\n" elif input_format == "QCEM": input = f"Question: {question}\nContext: {context}\nBECAUSE: {solution}\nOptions: {choice}\n" elif input_format == "QCLEM": input = f"Question: {question}\nContext: {context}\nBECAUSE: {lecture} {solution}\nOptions: {choice}\n" # Outputs if test_example: output = "Answer:" elif output_format == 'A': output = f"Answer: The answer is {answer}." elif output_format == 'AL': output = f"Answer: The answer is {answer}. BECAUSE: {solution}" elif output_format == 'AE': output = f"Answer: The answer is {answer}. BECAUSE: {lecture}" elif output_format == 'ALE': output = f"Answer: The answer is {answer}. BECAUSE: {lecture} {solution}" elif output_format == 'AEL': output = f"Answer: The answer is {answer}. BECAUSE: {solution} {lecture}" elif output_format == 'LA': output = f"Answer: {lecture} The answer is {answer}." elif output_format == 'EA': output = f"Answer: {solution} The answer is {answer}." elif output_format == 'LEA': output = f"Answer: {lecture} {solution} The answer is {answer}." elif output_format == 'ELA': output = f"Answer: {solution} {lecture} The answer is {answer}." input = input.replace(" ", " ").strip() output = output.replace(" ", " ").strip() if output.endswith("BECAUSE:"): output = output.replace("BECAUSE:", "").strip() user_prompt = {"role": "user", "content": f"Can you explain {input}?"} assistant_prompt = {"role": "assistant", "content": f"{output}"} return user_prompt, assistant_prompt def build_prompt_gpt4(problems, shot_qids, test_qid, args): prompt_array = [{"role": "system", "content": "You are a helpful assistant."}] # n-shot training examples for qid in shot_qids: question = get_question_text(problems[qid]) context = get_context_text(problems[qid], args.use_caption) choice = get_choice_text(problems[qid], args.options) answer = get_answer(problems[qid], args.options) lecture = get_lecture_text(problems[qid]) solution = get_solution_text(problems[qid]) user_prompt, assistant_prompt = create_one_example_gpt4(args.prompt_format, question, context, choice, answer, lecture, solution, test_example=False) prompt_array.append(user_prompt) prompt_array.append(assistant_prompt) # test example question = get_question_text(problems[test_qid]) context = get_context_text(problems[test_qid], args.use_caption) choice = get_choice_text(problems[test_qid], args.options) answer = get_answer(problems[test_qid], args.options) lecture = get_lecture_text(problems[test_qid]) solution = get_solution_text(problems[test_qid]) user_prompt, assistant_prompt = create_one_example_gpt4(args.prompt_format, question, context, choice, answer, lecture, solution, test_example=True) prompt_array.append(user_prompt) prompt_array.append(assistant_prompt) return prompt_array
null
179,708
import argparse from llava.model.builder import load_pretrained_model from llava.mm_utils import get_model_name_from_path def merge_lora(args): model_name = get_model_name_from_path(args.model_path) tokenizer, model, image_processor, context_len = load_pretrained_model(args.model_path, args.model_base, model_name, device_map='cpu') model.save_pretrained(args.save_model_path) tokenizer.save_pretrained(args.save_model_path)
null
179,712
import logging import sys from datetime import datetime from logging import getLogger, basicConfig from pathlib import Path from time import sleep from django.conf import settings from mwmbl.indexer import index_batches, historical from mwmbl.indexer.batch_cache import BatchCache from mwmbl.models import OldIndex from mwmbl.tinysearchengine.copy_index import copy_pages logger = getLogger(__name__) def index_batches(batch_data: Collection[HashedBatch], index_path: str): class BatchCache: def __init__(self, repo_path): def get_cached(self, batch_urls: list[str]) -> dict[str, HashedBatch]: def retrieve_batches(self, num_batches): def retrieve_batch(self, url): def store(self, batch, url): def get_path_from_url(self, url) -> Path: def run(data_path: str): logger.info("Started background process") historical.run() index_path = Path(data_path) / settings.INDEX_NAME batch_cache = BatchCache(Path(data_path) / settings.BATCH_DIR_NAME) while True: try: batch_cache.retrieve_batches(num_batches=10000) except Exception: logger.exception("Error retrieving batches") try: index_batches.run(batch_cache, index_path) except Exception: logger.exception("Error indexing batches") sleep(10)
null
179,713
import os import django import uvicorn from django.core.management import call_command from redis import Redis from mwmbl.indexer.update_urls import update_urls_continuously from mwmbl.redis_url_queue import RedisURLQueue def update_urls_continuously(data_path: str, new_item_queue: RedisURLQueue): batch_cache = BatchCache(Path(data_path) / settings.BATCH_DIR_NAME) while True: try: run(batch_cache, new_item_queue) except Exception: logger.exception("Error updating URLs") sleep(10) class RedisURLQueue: def __init__(self, redis: Redis): self.redis = redis def queue_urls(self, found_urls: list[FoundURL]): with DomainLinkDatabase() as link_db: for url in found_urls: domain = urlparse(url.url).netloc url_score = 1/len(url.url) domain_score = link_db.get_domain_score(domain) + url_score max_urls = get_domain_max_urls(domain) self.redis.zadd(DOMAIN_URLS_KEY.format(domain=domain), {url.url: url_score}) self.redis.zremrangebyrank(DOMAIN_URLS_KEY.format(domain=domain), 0, -max_urls) self.redis.zadd(DOMAIN_SCORE_KEY, {domain: domain_score}, gt=True) # Remove the lowest scoring domains while self.redis.zcard(DOMAIN_SCORE_KEY) > MAX_OTHER_DOMAINS: lowest_scoring_domain = self.redis.zpopmin(DOMAIN_SCORE_KEY) self.redis.delete(DOMAIN_URLS_KEY.format(domain=lowest_scoring_domain)) logger.info(f"Queued {len(found_urls)} URLs, number of domains: {self.redis.zcard(DOMAIN_SCORE_KEY)}") def get_batch(self) -> list[str]: top_scoring_domains = set(self.redis.zrange(DOMAIN_SCORE_KEY, 0, 2000, desc=True)) top_other_domains = top_scoring_domains - DOMAINS.keys() domains = (list(CORE_DOMAINS) + random.sample(DOMAINS.keys(), 50) + random.sample(top_other_domains, 100)) logger.info(f"Getting batch from domains {domains}") # Pop the highest scoring URL from each domain urls = [] for domain in domains: domain_urls_scores = self.redis.zpopmax(DOMAIN_URLS_KEY.format(domain=domain)) # Update the domain score if we removed a URL new_domain_scores = self.redis.zrangebyscore( DOMAIN_URLS_KEY.format(domain=domain), "-inf", "+inf", start=0, num=1, withscores=True) if new_domain_scores: new_domain_score = new_domain_scores[0][1] self.redis.zadd(DOMAIN_SCORE_KEY, {domain: new_domain_score}, gt=True) else: self.redis.zrem(DOMAIN_SCORE_KEY, domain) for url, score in domain_urls_scores: urls.append(url) if len(urls) >= BATCH_SIZE: break logger.info(f"Returning URLs: {urls}") return urls def run(): django.setup() from django.conf import settings from mwmbl import background if settings.STATIC_ROOT: call_command("collectstatic", "--clear", "--noinput") call_command("migrate") mwmbl_app = os.environ["MWMBL_APP"] if mwmbl_app == "update_urls": redis: Redis = Redis.from_url(os.environ.get("REDIS_URL", "redis://127.0.0.1:6379"), decode_responses=True) url_queue = RedisURLQueue(redis) update_urls_continuously(settings.DATA_PATH, url_queue) elif mwmbl_app == "update_batches": background.run(settings.DATA_PATH) elif mwmbl_app == "copy_indexes": background.copy_indexes_continuously() elif mwmbl_app == "server": uvicorn.run("mwmbl.asgi:application", host="0.0.0.0", port=5000) else: raise ValueError(f"Unknown MWMBL_APP: {mwmbl_app}")
null
179,714
from pandas import DataFrame, Series from sklearn.base import BaseEstimator, RegressorMixin, TransformerMixin from mwmbl.tinysearchengine.rank import get_features def get_features(terms, title, url, extract, score, is_complete): def get_features_as_series(item: Series): terms = item['query'].lower().split() features = get_features(terms, item['title'], item['url'], item['extract'], item['score'], True) # features_filtered = {k: v for k, v in features.items() if 'match_score' not in k} return Series(features)
null
179,715
import math import re from abc import abstractmethod from logging import getLogger from operator import itemgetter from typing import Optional from urllib.parse import urlparse from mwmbl.format import format_result_with_pattern, get_query_regex from mwmbl.tokenizer import tokenize, get_bigrams from mwmbl.tinysearchengine.completer import Completer from mwmbl.hn_top_domains_filtered import DOMAINS from mwmbl.tinysearchengine.indexer import TinyIndex, Document, DocumentState SCORE_THRESHOLD = 0.0 def score_result(terms: list[str], result: Document, is_complete: bool): features = get_features(terms, result.title, result.url, result.extract, result.score, is_complete) length_penalty = math.e ** (-LENGTH_PENALTY * len(result.url)) match_score = (4 * features['match_score_title'] + features['match_score_extract'] + 2 * features[ 'match_score_domain'] + 2 * features['match_score_domain_tokenized'] + features['match_score_path']) max_match_terms = max(features[f'match_terms_{name}'] for name in ['title', 'extract', 'domain', 'domain_tokenized', 'path']) if max_match_terms <= len(terms) / 2 and result.state is None: return 0.0 if match_score > MATCH_SCORE_THRESHOLD: return match_score * length_penalty * (features['domain_score'] + DOMAIN_SCORE_SMOOTHING) / 10 # best_match_score = max(features[f'match_score_{name}'] for name in ['title', 'extract', 'domain', 'domain_tokenized']) # score = best_match_score * length_penalty * (features['domain_score'] + DOMAIN_SCORE_SMOOTHING) return 0.0 class Document: title: str url: str extract: str score: float term: Optional[str] = None state: Optional[int] = None def __init__(self, title, url, extract, score, term=None, state=None): # Sometimes the title or extract may be None, probably because of user generated content # It's not allowed to be None though, or things will break self.title = title if title is not None else '' self.url = url self.extract = extract if extract is not None else '' self.score = score self.term = term self.state = state def order_results(terms: list[str], results: list[Document], is_complete: bool) -> list[Document]: if len(results) == 0: return [] results_and_scores = [(score_result(terms, result, is_complete), result) for result in results] ordered_results = sorted(results_and_scores, key=itemgetter(0), reverse=True) filtered_results = [result for score, result in ordered_results if score > SCORE_THRESHOLD] return filtered_results
null
179,716
import math import re from abc import abstractmethod from logging import getLogger from operator import itemgetter from typing import Optional from urllib.parse import urlparse from mwmbl.format import format_result_with_pattern, get_query_regex from mwmbl.tokenizer import tokenize, get_bigrams from mwmbl.tinysearchengine.completer import Completer from mwmbl.hn_top_domains_filtered import DOMAINS from mwmbl.tinysearchengine.indexer import TinyIndex, Document, DocumentState def deduplicate(results, seen_titles): deduplicated_results = [] for result in results: if result.title not in seen_titles: deduplicated_results.append(result) seen_titles.add(result.title) return deduplicated_results
null
179,717
import math import re from abc import abstractmethod from logging import getLogger from operator import itemgetter from typing import Optional from urllib.parse import urlparse from mwmbl.format import format_result_with_pattern, get_query_regex from mwmbl.tokenizer import tokenize, get_bigrams from mwmbl.tinysearchengine.completer import Completer from mwmbl.hn_top_domains_filtered import DOMAINS from mwmbl.tinysearchengine.indexer import TinyIndex, Document, DocumentState class DocumentState(IntEnum): def remove_curate_state(state: DocumentState): if state == DocumentState.ORGANIC_APPROVED: return None if state == DocumentState.FROM_USER_APPROVED: return DocumentState.FROM_USER if state == DocumentState.FROM_GOOGLE_APPROVED: return DocumentState.FROM_GOOGLE return state
null
179,718
import json import os from dataclasses import dataclass, asdict, field from enum import IntEnum from io import UnsupportedOperation from logging import getLogger from mmap import mmap, PROT_READ, PROT_WRITE from typing import TypeVar, Generic, Callable, List, Optional import mmh3 from zstandard import ZstdDecompressor, ZstdCompressor, ZstdError T = TypeVar('T') def _trim_items_to_page(compressor: ZstdCompressor, page_size: int, items:list[T]): # Find max number of items that fit on a page return _binary_search_fitting_size(compressor, page_size, items, 0, len(items)) def _pad_to_page_size(data: bytes, page_size: int): page_length = len(data) if page_length > page_size: raise PageError(f"Data is too big ({page_length}) for page size ({page_size})") padding = b'\x00' * (page_size - page_length) page_data = data + padding return page_data def _get_page_data(page_size: int, items: list[T]): compressor = ZstdCompressor() num_fitting, serialised_data = _trim_items_to_page(compressor, page_size, items) compressed_data = compressor.compress(json.dumps(items[:num_fitting]).encode('utf8')) assert len(compressed_data) <= page_size, "The data shouldn't get bigger" return _pad_to_page_size(compressed_data, page_size)
null
179,719
from logging import getLogger from ninja import NinjaAPI from mwmbl.format import format_result from mwmbl.tinysearchengine.rank import HeuristicRanker def format_result(result, query): tokens = tokenize(query) pattern = get_query_regex(tokens, True, False) return format_result_with_pattern(pattern, result) class HeuristicRanker(Ranker): def order_results(self, terms, pages, is_complete): return order_results(terms, pages, is_complete) def create_router(ranker: HeuristicRanker, version: str) -> NinjaAPI: router = NinjaAPI(urls_namespace=f"search-{version}") @router.get("") def search(request, s: str): results = ranker.search(s, []) return [format_result(result, s) for result in results] @router.get("/complete") def complete(request, q: str): return ranker.complete(q) return router
null
179,720
import re from django.template import Library from django.utils.html import conditional_escape from django.utils.safestring import mark_safe from mwmbl.format import get_query_regex, DOCUMENT_SOURCES, get_document_source from mwmbl.tinysearchengine.indexer import DocumentState from mwmbl.tokenizer import tokenize def get_query_regex(terms, is_complete, is_url): def tokenize(input_text): def format_for_query(text: str, query: str, autoescape=True): escape = conditional_escape if autoescape else lambda x: x tokens = tokenize(query) pattern = get_query_regex(tokens, True, False) matches = re.finditer(pattern, text, re.IGNORECASE) formatted = [] start = 0 for match in matches: formatted.append(escape(text[start:match.start()])) formatted.append(f"<strong>{escape(text[match.start():match.end()])}</strong>") start = match.end() formatted.append(escape(text[start:])) return mark_safe("".join(formatted))
null
179,721
import re from django.template import Library from django.utils.html import conditional_escape from django.utils.safestring import mark_safe from mwmbl.format import get_query_regex, DOCUMENT_SOURCES, get_document_source from mwmbl.tinysearchengine.indexer import DocumentState from mwmbl.tokenizer import tokenize def get_document_source(state: DocumentState): return DOCUMENT_SOURCES.get(state, 'mwmbl') class DocumentState(IntEnum): """ The state of the document in the index. A value of None indicates an organic search result. """ DELETED = -1 FROM_USER = 2 FROM_GOOGLE = 3 ORGANIC_APPROVED = 7 FROM_USER_APPROVED = 8 FROM_GOOGLE_APPROVED = 9 def convert_state_to_source(state: DocumentState, autoescape=True): escape = conditional_escape if autoescape else lambda x: x return escape(get_document_source(state))
null
179,722
from itertools import groupby from urllib.parse import urlparse, parse_qs from django.db import migrations def create_curations_from_user_curation(apps, schema_editor): Curation = apps.get_model('mwmbl', 'Curation') UserCuration = apps.get_model('mwmbl', 'UserCuration') # Order curations by timestamp curations = UserCuration.objects.all().order_by('timestamp') # Group by user and url for (user, url), curations in groupby(curations, lambda c: (c.user, c.url)): curations = list(curations) # Get the original results from the first curation original_results = curations[0].results # Get the new results from the last curation new_results = curations[-1].results # Get the number of changes from the number of curations num_changes = len(curations) # Get the timestamp from the first curation timestamp = curations[0].timestamp # Get the query from the url parsed_url_query = parse_qs(urlparse(url).query) query = parsed_url_query.get("q", [""])[0] # Create the new curation curation = Curation( timestamp=timestamp, query=query, original_results=original_results, new_results=new_results, num_changes=num_changes, user=user, ) curation.save()
null
179,723
import gzip from datetime import datetime, timedelta from glob import glob from itertools import islice from logging import getLogger from urllib.parse import urlparse from pydantic import BaseModel from redis import Redis from mwmbl.crawler.batch import HashedBatch from mwmbl.indexer.update_urls import get_datetime_from_timestamp class HashedBatch(Schema): user_id_hash: str timestamp: int items: list[Item] def get_test_batches(): for path in glob("./devdata/batches/**/*.json.gz", recursive=True): print("Processing path", path) with gzip.open(path) as gzip_file: yield HashedBatch.parse_raw(gzip_file.read())
null
179,724
import gzip import hashlib import json import os from datetime import datetime, timezone, date from queue import Queue, Empty from typing import Union from uuid import uuid4 import boto3 import requests from fastapi import HTTPException from ninja import NinjaAPI from redis import Redis from mwmbl.crawler.batch import Batch, NewBatchRequest, HashedBatch from mwmbl.crawler.stats import MwmblStats, StatsManager from mwmbl.database import Database from mwmbl.indexer.batch_cache import BatchCache from mwmbl.indexer.indexdb import IndexDatabase, BatchInfo, BatchStatus from mwmbl.redis_url_queue import RedisURLQueue from mwmbl.settings import ( ENDPOINT_URL, KEY_ID, APPLICATION_KEY, BUCKET_NAME, MAX_BATCH_SIZE, USER_ID_LENGTH, VERSION, PUBLIC_URL_PREFIX, PUBLIC_USER_ID_LENGTH, FILE_NAME_SUFFIX, DATE_REGEX) stats_manager = StatsManager(Redis.from_url(os.environ.get("REDIS_URL", "redis://127.0.0.1:6379"), decode_responses=True)) def upload(data: bytes, name: str): bucket = get_bucket(name) result = bucket.put(Body=data) return result last_batch = None def _get_user_id_hash(batch: Union[Batch, NewBatchRequest]): return hashlib.sha3_256(batch.user_id.encode('utf8')).hexdigest() def check_public_user_id(public_user_id): if len(public_user_id) != PUBLIC_USER_ID_LENGTH: raise HTTPException(400, f"Incorrect public user ID length, should be {PUBLIC_USER_ID_LENGTH}") def get_batch_url(batch_id, date_str, public_user_id): check_date_str(date_str) check_public_user_id(public_user_id) url = f'{PUBLIC_URL_PREFIX}1/{VERSION}/{date_str}/1/{public_user_id}/{batch_id}{FILE_NAME_SUFFIX}' return url def get_batch_ids_for_prefix(prefix): filenames = get_batches_for_prefix(prefix) filename_endings = sorted(filename.rsplit('/', 1)[1] for filename in filenames) results = {'batch_ids': [get_batch_id_from_file_name(name) for name in filename_endings]} return results def check_date_str(date_str): if not DATE_REGEX.match(date_str): raise HTTPException(400, f"Incorrect date format, should be YYYY-MM-DD") def get_subfolders(prefix): client = boto3.client('s3', endpoint_url=ENDPOINT_URL, aws_access_key_id=KEY_ID, aws_secret_access_key=APPLICATION_KEY) items = client.list_objects(Bucket=BUCKET_NAME, Prefix=prefix, Delimiter='/') item_keys = [item['Prefix'][len(prefix):].strip('/') for item in items['CommonPrefixes']] return item_keys class Batch(Schema): user_id: str items: list[Item] class NewBatchRequest(Schema): user_id: str class HashedBatch(Schema): user_id_hash: str timestamp: int items: list[Item] class MwmblStats(BaseModel): urls_crawled_today: int urls_crawled_daily: dict[str, int] urls_crawled_hourly: list[int] users_crawled_daily: dict[str, int] top_users: dict[str, int] top_domains: dict[str, int] class Database: def __init__(self): self.connection = None def __enter__(self): self.connection = connect(DATABASE_URL) self.connection.set_session(autocommit=True) return self def __exit__(self, exc_type, exc_val, exc_tb): self.connection.close() class BatchCache: num_threads = 20 def __init__(self, repo_path): os.makedirs(repo_path, exist_ok=True) self.path = repo_path def get_cached(self, batch_urls: list[str]) -> dict[str, HashedBatch]: batches = {} for url in batch_urls: path = self.get_path_from_url(url) try: data = gzip.GzipFile(path).read() except FileNotFoundError: logger.exception(f"Missing batch file: {path}") continue try: batch = HashedBatch.parse_raw(data) except ValidationError: logger.exception(f"Unable to parse batch, skipping: '{data}'") continue batches[url] = batch return batches def retrieve_batches(self, num_batches): with Database() as db: index_db = IndexDatabase(db.connection) index_db.create_tables() with Database() as db: index_db = IndexDatabase(db.connection) batches = index_db.get_batches_by_status(BatchStatus.REMOTE, num_batches) logger.info(f"Found {len(batches)} remote batches") if len(batches) == 0: return urls = [batch.url for batch in batches] pool = ThreadPool(self.num_threads) results = pool.imap_unordered(self.retrieve_batch, urls) total_processed = 0 for result in results: total_processed += result logger.info(f"Processed batches with {total_processed} items") index_db.update_batch_status(urls, BatchStatus.LOCAL) def retrieve_batch(self, url): data = json.loads(gzip.decompress(retry_requests.get(url).content)) try: batch = HashedBatch.parse_obj(data) except ValidationError: logger.info(f"Failed to validate batch {data}") return 0 if len(batch.items) > 0: self.store(batch, url) return len(batch.items) def store(self, batch, url): path = self.get_path_from_url(url) logger.debug(f"Storing local batch at {path}") os.makedirs(path.parent, exist_ok=True) with open(path, 'wb') as output_file: data = gzip.compress(batch.json().encode('utf8')) output_file.write(data) def get_path_from_url(self, url) -> Path: url_path = urlparse(url).path return Path(self.path) / url_path.lstrip('/') class BatchStatus(Enum): REMOTE = 0 # The batch only exists in long term storage LOCAL = 10 # We have a copy of the batch locally in Postgresql URLS_UPDATED = 20 # We've updated URLs from the batch INDEXED = 30 # The batch has been indexed class BatchInfo: url: str user_id_hash: str status: BatchStatus class IndexDatabase: def __init__(self, connection): self.connection = connection def create_tables(self): batches_sql = """ CREATE TABLE IF NOT EXISTS batches ( url VARCHAR PRIMARY KEY, user_id_hash VARCHAR NOT NULL, status INT NOT NULL ) """ with self.connection.cursor() as cursor: cursor.execute(batches_sql) def record_batches(self, batch_infos: list[BatchInfo]): sql = """ INSERT INTO batches (url, user_id_hash, status) values %s ON CONFLICT (url) DO NOTHING """ data = [(info.url, info.user_id_hash, info.status.value) for info in batch_infos] with self.connection.cursor() as cursor: execute_values(cursor, sql, data) def get_batches_by_status(self, status: BatchStatus, num_batches=1000) -> list[BatchInfo]: sql = """ SELECT * FROM batches WHERE status = %(status)s LIMIT %(num_batches)s """ with self.connection.cursor() as cursor: cursor.execute(sql, {'status': status.value, 'num_batches': num_batches}) results = cursor.fetchall() return [BatchInfo(url, user_id_hash, status) for url, user_id_hash, status in results] def update_batch_status(self, batch_urls: list[str], status: BatchStatus): if not batch_urls: return sql = """ UPDATE batches SET status = %(status)s WHERE url IN %(urls)s """ with self.connection.cursor() as cursor: cursor.execute(sql, {'status': status.value, 'urls': tuple(batch_urls)}) class RedisURLQueue: def __init__(self, redis: Redis): self.redis = redis def queue_urls(self, found_urls: list[FoundURL]): with DomainLinkDatabase() as link_db: for url in found_urls: domain = urlparse(url.url).netloc url_score = 1/len(url.url) domain_score = link_db.get_domain_score(domain) + url_score max_urls = get_domain_max_urls(domain) self.redis.zadd(DOMAIN_URLS_KEY.format(domain=domain), {url.url: url_score}) self.redis.zremrangebyrank(DOMAIN_URLS_KEY.format(domain=domain), 0, -max_urls) self.redis.zadd(DOMAIN_SCORE_KEY, {domain: domain_score}, gt=True) # Remove the lowest scoring domains while self.redis.zcard(DOMAIN_SCORE_KEY) > MAX_OTHER_DOMAINS: lowest_scoring_domain = self.redis.zpopmin(DOMAIN_SCORE_KEY) self.redis.delete(DOMAIN_URLS_KEY.format(domain=lowest_scoring_domain)) logger.info(f"Queued {len(found_urls)} URLs, number of domains: {self.redis.zcard(DOMAIN_SCORE_KEY)}") def get_batch(self) -> list[str]: top_scoring_domains = set(self.redis.zrange(DOMAIN_SCORE_KEY, 0, 2000, desc=True)) top_other_domains = top_scoring_domains - DOMAINS.keys() domains = (list(CORE_DOMAINS) + random.sample(DOMAINS.keys(), 50) + random.sample(top_other_domains, 100)) logger.info(f"Getting batch from domains {domains}") # Pop the highest scoring URL from each domain urls = [] for domain in domains: domain_urls_scores = self.redis.zpopmax(DOMAIN_URLS_KEY.format(domain=domain)) # Update the domain score if we removed a URL new_domain_scores = self.redis.zrangebyscore( DOMAIN_URLS_KEY.format(domain=domain), "-inf", "+inf", start=0, num=1, withscores=True) if new_domain_scores: new_domain_score = new_domain_scores[0][1] self.redis.zadd(DOMAIN_SCORE_KEY, {domain: new_domain_score}, gt=True) else: self.redis.zrem(DOMAIN_SCORE_KEY, domain) for url, score in domain_urls_scores: urls.append(url) if len(urls) >= BATCH_SIZE: break logger.info(f"Returning URLs: {urls}") return urls MAX_BATCH_SIZE = 100 USER_ID_LENGTH = 36 VERSION = 'v1' PUBLIC_URL_PREFIX = f'https://f004.backblazeb2.com/file/{BUCKET_NAME}/' def create_router(batch_cache: BatchCache, queued_batches: RedisURLQueue, version: str) -> NinjaAPI: router = NinjaAPI(urls_namespace=f"crawler-{version}") @router.post('/batches/') def post_batch(request, batch: Batch): if len(batch.items) > MAX_BATCH_SIZE: raise HTTPException(400, f"Batch size too large (maximum {MAX_BATCH_SIZE}), got {len(batch.items)}") if len(batch.user_id) != USER_ID_LENGTH: raise HTTPException(400, f"User ID length is incorrect, should be {USER_ID_LENGTH} characters") if len(batch.items) == 0: return { 'status': 'ok', } user_id_hash = _get_user_id_hash(batch) now = datetime.now(timezone.utc) seconds = (now - datetime(now.year, now.month, now.day, tzinfo=timezone.utc)).seconds # How to pad a string with zeros: https://stackoverflow.com/a/39402910 # Maximum seconds in a day is 60*60*24 = 86400, so 5 digits is enough padded_seconds = str(seconds).zfill(5) # See discussion here: https://stackoverflow.com/a/13484764 uid = str(uuid4())[:8] filename = f'1/{VERSION}/{now.date()}/1/{user_id_hash}/{padded_seconds}__{uid}.json.gz' # Using an approach from https://stackoverflow.com/a/30476450 epoch_time = (now - datetime(1970, 1, 1, tzinfo=timezone.utc)).total_seconds() hashed_batch = HashedBatch(user_id_hash=user_id_hash, timestamp=epoch_time, items=batch.items) stats_manager.record_batch(hashed_batch) data = gzip.compress(hashed_batch.json().encode('utf8')) upload(data, filename) global last_batch last_batch = hashed_batch batch_url = f'{PUBLIC_URL_PREFIX}{filename}' batch_cache.store(hashed_batch, batch_url) # Record the batch as being local so that we don't retrieve it again when the server restarts infos = [BatchInfo(batch_url, user_id_hash, BatchStatus.LOCAL)] with Database() as db: index_db = IndexDatabase(db.connection) index_db.record_batches(infos) return { 'status': 'ok', 'public_user_id': user_id_hash, 'url': batch_url, } @router.post('/batches/new') def request_new_batch(request, batch_request: NewBatchRequest) -> list[str]: user_id_hash = _get_user_id_hash(batch_request) try: urls = queued_batches.get_batch() except Empty: return [] # TODO: track which URLs are currently being crawled return urls @router.get('/batches/{date_str}/users/{public_user_id}') def get_batches_for_date_and_user(request, date_str, public_user_id): check_date_str(date_str) check_public_user_id(public_user_id) prefix = f'1/{VERSION}/{date_str}/1/{public_user_id}/' return get_batch_ids_for_prefix(prefix) @router.get('/batches/{date_str}/users/{public_user_id}/batch/{batch_id}') def get_batch_from_id(request, date_str, public_user_id, batch_id): url = get_batch_url(batch_id, date_str, public_user_id) data = json.loads(gzip.decompress(requests.get(url).content)) return { 'url': url, 'batch': data, } @router.get('/latest-batch') def get_latest_batch(request) -> list[HashedBatch]: return [] if last_batch is None else [last_batch] @router.get('/batches/{date_str}/users') def get_user_id_hashes_for_date(request, date_str: str): check_date_str(date_str) prefix = f'1/{VERSION}/{date_str}/1/' return get_subfolders(prefix) @router.get('/stats') def get_stats(request) -> MwmblStats: return stats_manager.get_stats() @router.get('/') def status(request): return { 'status': 'ok' } return router
null
179,725
from itertools import islice from logging import getLogger from django.conf import settings from pybloomfilter import BloomFilter from mwmbl.hn_top_domains_filtered import DOMAINS def get_bloom_filter(domain_group: str) -> BloomFilter: try: bloom_filter = BloomFilter.open(settings.DOMAIN_LINKS_BLOOM_FILTER_PATH.format(domain_group=domain_group)) except FileNotFoundError: bloom_filter = BloomFilter(settings.NUM_DOMAINS_IN_BLOOM_FILTER, 1e-6, settings.DOMAIN_LINKS_BLOOM_FILTER_PATH.format(domain_group=domain_group), perm=0o666) return bloom_filter
null
179,726
import glob import gzip import json from collections import defaultdict from urllib.parse import urlparse from mwmbl.indexer.paths import CRAWL_GLOB, LINK_COUNT_PATH def get_urls(): def collect_links(urls): LINK_COUNT_PATH = MWMBL_DATA_DIR / 'crawl-counts.json' def run(): url_links = get_urls() collected = collect_links(url_links) link_counts = {url: len(links) for url, links in collected.items()} with open(LINK_COUNT_PATH, 'w') as output_file: json.dump(link_counts, output_file, indent=2)
null
179,727
import csv import gzip from mwmbl.indexer.fsqueue import FSQueue, ZstdJsonSerializer from mwmbl.indexer.paths import DOMAINS_PATH, DOMAINS_QUEUE_NAME, TINYSEARCH_DATA_DIR BATCH_SIZE = 250 def get_domains(): class ZstdJsonSerializer(Serializer): def __init__(self): def serialize(self, item) -> bytes: def deserialize(self, serialized_item: bytes): class FSQueue: def __init__(self, directory: Union[str, Path], name: str, serializer: Serializer): def _get_dir(self, state: FSState): def _get_path(self, state: FSState, name: str): def _move(self, name: str, old_state: FSState, new_state: FSState): def put(self, item: object): def get(self) -> (str, object): def done(self, item_id: str): def error(self, item_id: str): def unlock_all(self): TINYSEARCH_DATA_DIR = DATA_DIR / 'tinysearch' DOMAINS_QUEUE_NAME = 'domains-queue-fs' def queue_domains(): queue = FSQueue(TINYSEARCH_DATA_DIR, DOMAINS_QUEUE_NAME, ZstdJsonSerializer()) queued = 0 batch = [] for rank, domain in get_domains(): batch.append((rank, domain)) queued += 1 if queued % BATCH_SIZE == 0: queue.put(batch) batch = [] print("Queued:", queued)
null
179,728
from multiprocessing import Process from time import sleep from urllib.parse import urlsplit, urlunsplit import bs4 import requests from mwmbl.indexer.fsqueue import FSQueue, ZstdJsonSerializer from mwmbl.indexer.paths import TINYSEARCH_DATA_DIR, DOMAINS_QUEUE_NAME, DOMAINS_TITLES_QUEUE_NAME NUM_PROCESSES = 10 def get_domain_titles(): domains_queue = FSQueue(TINYSEARCH_DATA_DIR, DOMAINS_QUEUE_NAME, ZstdJsonSerializer()) titles_queue = FSQueue(TINYSEARCH_DATA_DIR, DOMAINS_TITLES_QUEUE_NAME, ZstdJsonSerializer()) while True: items_id, items = domains_queue.get() titles = retrieve_titles(items) # print("Item", item) # print("Title", type(title)) # print("Title item", str(title_item)) # print("Dump", pickle.dumps(title_item)) titles_queue.put(titles) domains_queue.done(items_id) print("Done titles", len(titles)) def run(): for i in range(NUM_PROCESSES): process = Process(target=get_domain_titles) process.start() sleep(3)
null
179,729
from typing import Iterable from urllib.parse import unquote from mwmbl.tinysearchengine.indexer import TokenizedDocument from mwmbl.tokenizer import tokenize, get_bigrams DEFAULT_SCORE = 0 def tokenize_document(url, title_cleaned, extract, score): title_tokens = tokenize(title_cleaned) prepared_url = prepare_url_for_tokenizing(unquote(url)) url_tokens = tokenize(prepared_url) extract_tokens = tokenize(extract) # print("Extract tokens", extract_tokens) tokens = get_index_tokens(title_tokens) | get_index_tokens(url_tokens) | get_index_tokens(extract_tokens) # doc = Document(title_cleaned, url, extract, score) # token_scores = {token: score_result([token], doc, True) for token in tokens} # high_scoring_tokens = [k for k, v in token_scores.items() if v > 0.5] # print("High scoring", len(high_scoring_tokens), token_scores, doc) document = TokenizedDocument(tokens=list(tokens), url=url, title=title_cleaned, extract=extract, score=score) return document class TokenizedDocument(Document): tokens: List[str] = field(default_factory=list) def get_pages(nlp, titles_urls_and_extracts, link_counts) -> Iterable[TokenizedDocument]: for i, (title_cleaned, url, extract) in enumerate(titles_urls_and_extracts): score = link_counts.get(url, DEFAULT_SCORE) yield tokenize_document(url, title_cleaned, extract, score) if i % 1000 == 0: print("Processed", i)
null
179,730
from datetime import date, timedelta from mwmbl.crawler.app import get_batches_for_date from mwmbl.database import Database from mwmbl.indexer.indexdb import BatchInfo, BatchStatus, IndexDatabase DAYS = 20 def get_user_id_hash_from_url(url): return url.split('/')[9] def get_batches_for_date(date_str): check_date_str(date_str) prefix = f'1/{VERSION}/{date_str}/1/' cache_filename = prefix + 'batches.json.gz' cache_url = PUBLIC_URL_PREFIX + cache_filename try: cached_batches = json.loads(gzip.decompress(requests.get(cache_url).content)) print(f"Got cached batches for {date_str}") return cached_batches except gzip.BadGzipFile: pass batches = get_batches_for_prefix(prefix) result = {'batch_urls': [f'{PUBLIC_URL_PREFIX}{batch}' for batch in sorted(batches)]} if date_str != str(date.today()): # Don't cache data from today since it may change data = gzip.compress(json.dumps(result).encode('utf8')) upload(data, cache_filename) print(f"Cached batches for {date_str} in {PUBLIC_URL_PREFIX}{cache_filename}") return result class Database: def __init__(self): self.connection = None def __enter__(self): self.connection = connect(DATABASE_URL) self.connection.set_session(autocommit=True) return self def __exit__(self, exc_type, exc_val, exc_tb): self.connection.close() class BatchStatus(Enum): REMOTE = 0 # The batch only exists in long term storage LOCAL = 10 # We have a copy of the batch locally in Postgresql URLS_UPDATED = 20 # We've updated URLs from the batch INDEXED = 30 # The batch has been indexed class BatchInfo: url: str user_id_hash: str status: BatchStatus class IndexDatabase: def __init__(self, connection): self.connection = connection def create_tables(self): batches_sql = """ CREATE TABLE IF NOT EXISTS batches ( url VARCHAR PRIMARY KEY, user_id_hash VARCHAR NOT NULL, status INT NOT NULL ) """ with self.connection.cursor() as cursor: cursor.execute(batches_sql) def record_batches(self, batch_infos: list[BatchInfo]): sql = """ INSERT INTO batches (url, user_id_hash, status) values %s ON CONFLICT (url) DO NOTHING """ data = [(info.url, info.user_id_hash, info.status.value) for info in batch_infos] with self.connection.cursor() as cursor: execute_values(cursor, sql, data) def get_batches_by_status(self, status: BatchStatus, num_batches=1000) -> list[BatchInfo]: sql = """ SELECT * FROM batches WHERE status = %(status)s LIMIT %(num_batches)s """ with self.connection.cursor() as cursor: cursor.execute(sql, {'status': status.value, 'num_batches': num_batches}) results = cursor.fetchall() return [BatchInfo(url, user_id_hash, status) for url, user_id_hash, status in results] def update_batch_status(self, batch_urls: list[str], status: BatchStatus): if not batch_urls: return sql = """ UPDATE batches SET status = %(status)s WHERE url IN %(urls)s """ with self.connection.cursor() as cursor: cursor.execute(sql, {'status': status.value, 'urls': tuple(batch_urls)}) def run(): for day in range(DAYS): date_str = str(date.today() - timedelta(days=day)) with Database() as db: index_db = IndexDatabase(db.connection) index_db.create_tables() batches = get_batches_for_date(date_str) batch_urls = batches['batch_urls'] print("Historical batches for date", date_str, len(batch_urls)) infos = [BatchInfo(url, get_user_id_hash_from_url(url), BatchStatus.REMOTE) for url in batch_urls] index_db.record_batches(infos)
null
179,731
import glob import gzip import json from itertools import islice from typing import Iterator from mwmbl.indexer.fsqueue import FSQueue, GzipJsonBlobSerializer from mwmbl.indexer.paths import CRAWL_GLOB, TINYSEARCH_DATA_DIR def get_deduped_pages(): seen_urls = set() for path in sorted(glob.glob(CRAWL_GLOB), reverse=True): data = json.load(gzip.open(path)) for item in data['items']: url = item['url'] if url in seen_urls: continue seen_urls.add(url) yield item def queue_deduped_items(deduped_pages): output_queue = FSQueue(TINYSEARCH_DATA_DIR, 'mwmbl-search-items', GzipJsonBlobSerializer()) for batch in grouper(BATCH_SIZE, deduped_pages): data = {'items': batch} output_queue.put(data) def run(): deduped_pages = get_deduped_pages() queue_deduped_items(deduped_pages)
null
179,732
from logging import getLogger from random import Random from urllib.parse import urlparse from redis import Redis from mwmbl.crawler.domains import DomainLinkDatabase, TOP_DOMAINS from mwmbl.crawler.urls import FoundURL from mwmbl.hn_top_domains_filtered import DOMAINS from mwmbl.settings import CORE_DOMAINS MAX_URLS_PER_CORE_DOMAIN = 1000 MAX_URLS_PER_TOP_DOMAIN = 100 MAX_URLS_PER_OTHER_DOMAIN = 5 TOP_DOMAINS = set(islice(DOMAINS, 1000)) CORE_DOMAINS = { 'github.com', 'en.wikipedia.org', 'stackoverflow.com', 'docs.google.com', 'programmers.stackexchange.com', 'developer.mozilla.org', 'arxiv.org', 'www.python.org', } def get_domain_max_urls(domain: str): if domain in CORE_DOMAINS: return MAX_URLS_PER_CORE_DOMAIN elif domain in TOP_DOMAINS: return MAX_URLS_PER_TOP_DOMAIN else: return MAX_URLS_PER_OTHER_DOMAIN
null
179,733
from dataclasses import asdict from datetime import datetime from logging import getLogger from typing import Optional from urllib.parse import urlencode import justext import requests from django.conf import settings from django.contrib.auth.decorators import login_required from django.http import HttpResponseBadRequest from django.shortcuts import render from django.views.decorators.http import require_http_methods from justext.core import html_to_dom, ParagraphMaker, classify_paragraphs, revise_paragraph_classification, \ LENGTH_LOW_DEFAULT, STOPWORDS_LOW_DEFAULT, MAX_LINK_DENSITY_DEFAULT, NO_HEADINGS_DEFAULT, LENGTH_HIGH_DEFAULT, \ STOPWORDS_HIGH_DEFAULT, MAX_HEADING_DISTANCE_DEFAULT, DEFAULT_ENCODING, DEFAULT_ENC_ERRORS, preprocessor from requests.exceptions import RequestException from mwmbl.models import Curation from mwmbl.search_setup import ranker, index_path from mwmbl.settings import NUM_EXTRACT_CHARS from mwmbl.tinysearchengine.indexer import Document, DocumentState, TinyIndex from mwmbl.tinysearchengine.rank import fix_document_state from mwmbl.tokenizer import tokenize from mwmbl.utils import add_term_infos def _get_results_and_activity(request): query = request.GET.get("q") if query: # There may be extra results in the request that we need to add in # format is ?enhanced=google&title=title1&url=url1&extract=extract1&title=title2&url=url2&extract=extract2 # source = request.GET.get("enhanced", "unknown") titles = request.GET.getlist(f"title") urls = request.GET.getlist(f"url") extracts = request.GET.getlist(f"extract") term = " ".join(tokenize(query)) # For now, we only support the Google source additional_results = [ Document(title=title, url=url, extract=extract, score=100.0 * 2 ** -i, term=term, state=DocumentState.FROM_GOOGLE) for i, (title, url, extract) in enumerate(zip(titles, urls, extracts)) ] results = ranker.search(query, additional_results=additional_results) activity = None else: results = None activity = Curation.objects.order_by("-timestamp")[:500] return activity, query, results def index(request): activity, query, results = _get_results_and_activity(request) return render(request, "index.html", { "results": results, "query": query, "user": request.user, "activity": activity, "footer_links": settings.FOOTER_LINKS, })
null