| import torch |
| from PIL import Image |
| from transformers import AutoConfig, AutoModel, AutoTokenizer, LlamaTokenizer, LlamaForCausalLM |
| from decord import VideoReader, cpu |
| import decord |
| from accelerate import init_empty_weights, infer_auto_device_map, load_checkpoint_in_model, dispatch_model |
| import torch.nn as nn |
| import einops |
|
|
| from .modelclass import Model |
|
|
| import torchshow as ts |
|
|
| import sys |
| sys.path.insert(0, "/root/videollm-online/baseline/TimeChat") |
|
|
| import argparse |
| import os |
| import random |
| import json |
| import numpy as np |
| import torch |
| import torch.backends.cudnn as cudnn |
| import torchshow as ts |
| from timechat.common.config import Config |
| from timechat.common.dist_utils import get_rank |
| from timechat.common.registry import registry |
| from timechat.conversation.conversation_video import Chat, Conversation, default_conversation,SeparatorStyle, conv_llava_llama_2 |
| import decord |
| import cv2 |
| import time |
| import subprocess |
| from decord import VideoReader |
| from timechat.processors.video_processor import ToTHWC, ToUint8, load_video |
| decord.bridge.set_bridge('torch') |
|
|
| |
| from timechat.datasets.builders import * |
| from timechat.models import * |
| from timechat.processors import * |
| from timechat.runners import * |
| from timechat.tasks import * |
|
|
| import random as rnd |
| from transformers import StoppingCriteria, StoppingCriteriaList |
| from PIL import Image |
| import gradio as gr |
|
|
| import random as rnd |
| import numpy as np |
|
|
| import re |
|
|
| import math |
| def ceil_time_by_fps(time: float, fps: int, min_time: float, max_time: float): |
| return min(max(math.ceil(time * fps) / fps, min_time), max_time) |
|
|
| |
| def load_video(video_path, start_time, end_time, n_frms=32, height=-1, width=-1, sampling="uniform", return_msg = False): |
| decord.bridge.set_bridge("torch") |
| vr = VideoReader(uri=video_path, height=height, width=width) |
| |
| |
| fps = float(vr.get_avg_fps()) |
| if start_time is not None: |
| start_time = ceil_time_by_fps(start_time, fps, min_time=0, max_time=len(vr)/fps) |
| start_frame = int(start_time * fps) |
| if end_time is not None: |
| end_time = ceil_time_by_fps(end_time, fps, min_time=0, max_time=len(vr)/fps) |
| end_frame = int(end_time * fps + 1) |
| vlen = end_frame - start_frame |
| acc_samples = min(n_frms, vlen) |
| n_frms = min(n_frms, vlen) |
|
|
| if sampling == "uniform": |
| indices = np.arange(start_frame, end_frame, vlen / n_frms).astype(int).tolist() |
| elif sampling == "headtail": |
| indices_h = sorted(rnd.sample(range(vlen // 2), n_frms // 2)) |
| indices_t = sorted(rnd.sample(range(vlen // 2, vlen), n_frms // 2)) |
| indices = indices_h + indices_t |
| elif sampling == 'rand': |
| |
| intervals = np.linspace(start=0, stop=vlen, num=acc_samples + 1).astype(int) |
| ranges = [] |
| for idx, interv in enumerate(intervals[:-1]): |
| ranges.append((interv, intervals[idx + 1] - 1)) |
| try: |
| indices = [rnd.choice(range(x[0], x[1])) for x in ranges] |
| except: |
| indices = np.random.permutation(vlen)[:acc_samples] |
| indices.sort() |
| indices = list(indices) |
| else: |
| raise NotImplementedError |
|
|
| |
| temp_frms = vr.get_batch(indices) |
| |
| tensor_frms = torch.from_numpy(temp_frms) if type(temp_frms) is not torch.Tensor else temp_frms |
| frms = tensor_frms.permute(3, 0, 1, 2).float() |
|
|
| if not return_msg: |
| return frms |
|
|
| fps = float(vr.get_avg_fps()) |
| sec = ", ".join([str(round(f / fps, 1)) for f in indices]) |
| |
| msg = f"The video contains {len(indices)} frames sampled at {sec} seconds. " |
| return frms, msg |
|
|
|
|
| class TimeChat(Model): |
| def __init__(self, device, config=None): |
| """ |
| Initialize the model by loading the pretrained TimeChat model and tokenizer. |
| """ |
| self.device = device |
| self.config = config |
| |
| |
| cfg = Config(config) |
| |
| |
| DIR="/home/docker_shared/asus/zhangyl/model/huggingface/hub/models--ShuhuaiRen--TimeChat-7b/snapshots/e12f42c6c9bd114525e99b2d5b1903d86ea3ce43/" |
| MODEL_DIR=f"{DIR}/timechat_7b.pth" |
| model_config = cfg.model_cfg |
| model_config.device_8bit = int(device.split(':')[1]) if ':' in device else 0 |
| model_config.ckpt = MODEL_DIR |
| model_config.vit_model = "/root/videollm-online/baseline/TimeChat/ckpt/eva_vit_g.pth" |
| model_config.q_former_model = "/root/videollm-online/baseline/TimeChat/ckpt/instruct_blip_vicuna7b_trimmed.pth" |
| |
| |
| model_cls = registry.get_model_class(model_config.arch) |
| self.model = model_cls.from_config(model_config).to(device) |
| self.model.eval() |
| |
| |
| vis_processor_cfg = cfg.datasets_cfg.webvid.vis_processor.train |
| self.vis_processor = registry.get_processor_class(vis_processor_cfg.name).from_config(vis_processor_cfg) |
| self.chat = Chat(self.model, self.vis_processor, device=device) |
| |
| |
| self.frame_fps = config.frame_fps |
| self.MAX_NUM_FRAMES = config.max_frames_num |
| self.num_beams = config.num_beams |
| self.temperature = config.temperature |
| self.height = config.height |
| self.width = config.width |
| |
|
|
| def Run(self, file, inp, start_time, end_time): |
| """ |
| Given the file (video file path) and input prompt (inp), run the model and return the response. |
| """ |
| |
| frames, msg = load_video( |
| video_path=file, |
| start_time=start_time, |
| end_time=end_time, |
| n_frms=self.MAX_NUM_FRAMES, |
| height=self.height, |
| width=self.width, |
| sampling ="uniform", return_msg = True |
| ) |
|
|
|
|
| img_list = [] |
| chat_state = conv_llava_llama_2.copy() |
| chat_state.system = "You are able to understand the visual content that the user provides. Follow the instructions carefully and explain your answers in detail." |
| msg = self.chat.upload_video_without_audio( |
| video_path=(frames, msg), |
| conv=chat_state, |
| img_list=img_list, |
| n_frms=self.MAX_NUM_FRAMES, |
| ) |
| text_input = "You are given a video. Please watch the video, identify all relevant moments that help answer the question. For each moments, determine the starting and ending times and provide a answer. The format should be: 'start time - end time, answer'. For example, ' 90 - 102 seconds, spread margarine on two slices of white bread'." |
|
|
| self.chat.ask(text_input + "\n" + inp, chat_state) |
|
|
| num_beams = self.num_beams |
| temperature = self.temperature |
| llm_message = self.chat.answer(conv=chat_state, |
| img_list=img_list, |
| num_beams=num_beams, |
| temperature=temperature, |
| max_new_tokens=300, |
| max_length=2000)[0] |
| response = llm_message.split('\n')[-1] |
| |
| breakpoint() |
| |
| |
| pattern = r"(\d+\.\d+) - (\d+\.\d+)\s*seconds,\s*(.*)" |
| matches = re.findall(pattern, response) |
| answer_pairs = list() |
| for match in matches: |
| moment_start, moment_end, answer = float(match[0]), float(match[1]), match[2] |
| answer_pairs.append(((moment_start + moment_end) / 2, answer)) |
| |
| breakpoint() |
| return answer_pairs |
|
|
| @staticmethod |
| def name(): |
| """ |
| Return the name of the model |
| """ |
| return "TimeChat" |
|
|