|
|
import torch |
|
|
from torch import Tensor |
|
|
from torch.cuda.amp import autocast |
|
|
from transformers import AutoModelForCausalLM, AutoModel, AutoTokenizer |
|
|
from model.build import MODEL_REGISTRY, BaseModel |
|
|
from modules.build import build_module |
|
|
from optim.utils import no_decay_param_group |
|
|
import torch.nn as nn |
|
|
|
|
|
def last_token_pool(last_hidden_states: Tensor, |
|
|
attention_mask: Tensor) -> Tensor: |
|
|
left_padding = (attention_mask[:, -1].sum() == attention_mask.shape[0]) |
|
|
if left_padding: |
|
|
return last_hidden_states[:, -1] |
|
|
else: |
|
|
sequence_lengths = attention_mask.sum(dim=1) - 1 |
|
|
batch_size = last_hidden_states.shape[0] |
|
|
return last_hidden_states[torch.arange(batch_size, device=last_hidden_states.device), sequence_lengths] |
|
|
|
|
|
@MODEL_REGISTRY.register() |
|
|
class OpenVocab(BaseModel): |
|
|
def __init__(self, cfg): |
|
|
super().__init__(cfg) |
|
|
self.cfg = cfg |
|
|
model_root = "fg-clip-base" |
|
|
self.pm_encoder = AutoModelForCausalLM.from_pretrained(model_root, trust_remote_code=True) |
|
|
|
|
|
if cfg.mode in ['warmup', 'pretrain']: |
|
|
self.frozen_model = AutoModelForCausalLM.from_pretrained(model_root, trust_remote_code=True) |
|
|
self.use_scene_cap = self.cfg.data.args.get("use_scene_cap", False) |
|
|
self.set_training_mode() |
|
|
else: |
|
|
self.text_encoder = AutoModel.from_pretrained('jinaai/jina-clip-v2', trust_remote_code=True) |
|
|
self.tokenizer = AutoTokenizer.from_pretrained('jinaai/jina-clip-v2', trust_remote_code=True) |
|
|
self.text_encoder.text_model.output_tokens = True |
|
|
self.set_downstream_mode() |
|
|
|
|
|
self.head_list = self.cfg.model.heads.head_list |
|
|
for head in self.head_list: |
|
|
setattr(self, head, build_module("heads", getattr(self.cfg.model.heads, head))) |
|
|
|
|
|
def set_training_mode(self): |
|
|
for name, param in self.frozen_model.named_parameters(): |
|
|
param.requires_grad = False |
|
|
|
|
|
for name, param in self.pm_encoder.named_parameters(): |
|
|
if "text_model" in name: |
|
|
param.requires_grad = False |
|
|
|
|
|
self.pm_encoder.train() |
|
|
self.frozen_model.eval() |
|
|
|
|
|
def set_downstream_mode(self): |
|
|
"""Set the model to downstream mode.""" |
|
|
for param in self.pm_encoder.parameters(): |
|
|
param.requires_grad = False |
|
|
|
|
|
for name, param in self.text_encoder.named_parameters(): |
|
|
if "vision_model" in name: |
|
|
param.requires_grad = False |
|
|
|
|
|
self.pm_encoder.eval() |
|
|
self.text_encoder.train() |
|
|
|
|
|
def forward(self, data_dict, mode=None): |
|
|
|
|
|
if 'cur_step' not in data_dict: |
|
|
data_dict['cur_step'] = 1 |
|
|
data_dict['total_steps'] = 1 |
|
|
|
|
|
data_dict['logit_scale'] = self.pm_encoder.logit_scale.exp() |
|
|
if mode == "warmup": |
|
|
B, C, H, W = data_dict["images"].shape |
|
|
data_dict["point_map"] = data_dict["point_map"].to(torch.bfloat16, non_blocking=True).permute(0, 3, 1, 2) |
|
|
data_dict["txt_ids"] = data_dict["txt_ids"].view(B, -1) |
|
|
with torch.autocast("cuda", dtype=torch.bfloat16): |
|
|
pm = data_dict["point_map"] |
|
|
_, data_dict["inter_view_pm_embed"] = self.pm_encoder.get_image_features(pm) |
|
|
with torch.no_grad(): |
|
|
data_dict["inter_view_txt_embed"] = self.frozen_model.get_text_features(data_dict["txt_ids"]) |
|
|
_, data_dict["inter_view_rgb_embed"] = self.frozen_model.get_image_features(data_dict["images"]) |
|
|
elif mode == 'pretrain': |
|
|
pm_basic_features = [] |
|
|
data_dict['point_map'] = data_dict['point_map'].to(torch.bfloat16, non_blocking=True).permute(0, 1, 4, 2, 3) |
|
|
for i in range(data_dict['point_map'].shape[0]): |
|
|
with autocast(dtype=torch.bfloat16): |
|
|
pm = data_dict['point_map'][i] |
|
|
_, pm_feat = self.pm_encoder.get_image_features(data_dict['point_map'][i]) |
|
|
pm_basic_features.append(pm_feat) |
|
|
|
|
|
data_dict['inter_view_pm_embed'] = torch.stack(pm_basic_features, dim=0) |
|
|
data_dict['scene_pm_embed'] = data_dict['inter_view_pm_embed'].mean(dim=1) |
|
|
|
|
|
B_txt = data_dict['txt_ids'].shape[0] |
|
|
lang_basic_features = torch.empty((B_txt, 32, 512), dtype=torch.bfloat16, device=data_dict['txt_ids'].device) |
|
|
rgb_basic_features = torch.empty((B_txt, 32, 512), dtype=torch.bfloat16, device=data_dict['txt_ids'].device) |
|
|
with torch.no_grad(): |
|
|
with autocast(dtype=torch.bfloat16): |
|
|
for i in range(B_txt): |
|
|
lang_basic_features[i] = self.frozen_model.get_text_features(data_dict['txt_ids'][i], walk_short_pos=True) |
|
|
rgb_basic_features[i] = self.frozen_model.get_image_features(data_dict['images'][i])[1] |
|
|
|
|
|
if getattr(self, "use_scene_cap", False): |
|
|
data_dict['scene_text_embed'] = self.frozen_model.get_text_features(data_dict['scene_txt_ids'], walk_short_pos=False) |
|
|
|
|
|
data_dict['inter_view_txt_embed'] = lang_basic_features |
|
|
data_dict['inter_view_rgb_embed'] = rgb_basic_features |
|
|
data_dict['scene_rgb_embed'] = rgb_basic_features.mean(dim=1) |
|
|
elif mode == 'qa': |
|
|
B, V, C, H, W = data_dict['point_map'].shape |
|
|
pm = data_dict['point_map'].reshape(B * V, C, H, W) |
|
|
with torch.no_grad(): |
|
|
with autocast(dtype=torch.bfloat16): |
|
|
_, pm_feats = self.pm_encoder.get_image_features(pm) |
|
|
data_dict['inter_view_pm_embed'] = pm_feats.reshape(B, V, -1) |
|
|
|
|
|
|
|
|
tokenized = self.tokenizer.batch_encode_plus( |
|
|
data_dict['sentence'], |
|
|
padding="max_length", |
|
|
return_tensors="pt", |
|
|
max_length=256, |
|
|
).to(data_dict['inter_view_pm_embed'].device) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
data_dict['txt_ids'] = tokenized['input_ids'] |
|
|
with autocast(dtype=torch.bfloat16): |
|
|
data_dict['inter_view_txt_tokens'] = self.text_encoder.text_model(data_dict['txt_ids'])[-1] |
|
|
data_dict['attention_mask'] = tokenized['attention_mask'].ne(1).bool() |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
if hasattr(self, "qa_head") and self.qa_head is not None: |
|
|
answer_scores = self.qa_head( |
|
|
data_dict['inter_view_pm_embed'], |
|
|
data_dict['inter_view_txt_tokens'], |
|
|
data_dict['attention_mask'] |
|
|
) |
|
|
data_dict['answer_scores'] = answer_scores |
|
|
return data_dict |
|
|
|
|
|
def get_vision_params(self, model): |
|
|
return [(n, p) for n, p in model.named_parameters() if p.requires_grad] |
|
|
|
|
|
def get_text_params(self, model): |
|
|
text_params = [ |
|
|
(n, p) for n, p in model.named_parameters() |
|
|
if "text_model" in n |
|
|
] |
|
|
return text_params |
|
|
|
|
|
def get_opt_params(self): |
|
|
def get_lr(cfg, default_lr): |
|
|
return default_lr if cfg.get("lr") is None else cfg.get("lr") |
|
|
|
|
|
optimizer_grouped_parameters = [] |
|
|
if self.cfg.mode == 'warmup': |
|
|
optimizer_grouped_parameters += no_decay_param_group(self.get_vision_params(self.pm_encoder),get_lr(self.cfg.model.vision, self.cfg.solver.lr)) |
|
|
elif self.cfg.mode == 'pretrain': |
|
|
optimizer_grouped_parameters += no_decay_param_group(self.get_vision_params(self.pm_encoder),get_lr(self.cfg.model.vision, self.cfg.solver.lr)) |
|
|
else: |
|
|
optimizer_grouped_parameters += no_decay_param_group(self.get_text_params(self.text_encoder), get_lr(self.cfg.model.vision, self.cfg.solver.lr)) |
|
|
if "qa_head" in self.head_list: |
|
|
optimizer_grouped_parameters += no_decay_param_group( |
|
|
self.qa_head.named_parameters(), get_lr(self.cfg.model.heads.qa_head, self.cfg.solver.lr) |
|
|
) |
|
|
if "ground_head" in self.head_list: |
|
|
optimizer_grouped_parameters += no_decay_param_group( |
|
|
self.ground_head.named_parameters(), get_lr(self.cfg.model.heads.ground_head, self.cfg.solver.lr) |
|
|
) |
|
|
|
|
|
return optimizer_grouped_parameters |