| """ |
| Copyright (c) 2022, salesforce.com, inc. |
| All rights reserved. |
| SPDX-License-Identifier: BSD-3-Clause |
| For full license text, see the LICENSE_Lavis file in the repo root or https://opensource.org/licenses/BSD-3-Clause |
| """ |
|
|
| import logging |
| import torch |
| from omegaconf import OmegaConf |
|
|
| from .registry import registry |
| from .base_model import BaseModel |
| from .base_processor import BaseProcessor |
| from .blip_processors import * |
| from .blip2 import Blip2Base |
| from .clip_vision_encoder import * |
| from .config import * |
| from .eva_vit import * |
| from .mini_gpt4_llama_v2 import MiniGPT4_Video |
|
|
|
|
|
|
| __all__ = [ |
| "load_model", |
| "BaseModel", |
| "Blip2Base", |
| "MiniGPT4_Video", |
| |
| ] |
|
|
|
|
| def load_model(name, model_type, is_eval=False, device="cpu", checkpoint=None): |
| """ |
| Load supported models. |
| |
| To list all available models and types in registry: |
| >>> from minigpt4.models import model_zoo |
| >>> print(model_zoo) |
| |
| Args: |
| name (str): name of the model. |
| model_type (str): type of the model. |
| is_eval (bool): whether the model is in eval mode. Default: False. |
| device (str): device to use. Default: "cpu". |
| checkpoint (str): path or to checkpoint. Default: None. |
| Note that expecting the checkpoint to have the same keys in state_dict as the model. |
| |
| Returns: |
| model (torch.nn.Module): model. |
| """ |
|
|
| model = registry.get_model_class(name).from_pretrained(model_type=model_type) |
|
|
| if checkpoint is not None: |
| model.load_checkpoint(checkpoint) |
|
|
| if is_eval: |
| model.eval() |
|
|
| if device == "cpu": |
| model = model.float() |
|
|
| return model.to(device) |
|
|
|
|
| def load_preprocess(config): |
| """ |
| Load preprocessor configs and construct preprocessors. |
| |
| If no preprocessor is specified, return BaseProcessor, which does not do any preprocessing. |
| |
| Args: |
| config (dict): preprocessor configs. |
| |
| Returns: |
| vis_processors (dict): preprocessors for visual inputs. |
| txt_processors (dict): preprocessors for text inputs. |
| |
| Key is "train" or "eval" for processors used in training and evaluation respectively. |
| """ |
|
|
| def _build_proc_from_cfg(cfg): |
| return ( |
| registry.get_processor_class(cfg.name).from_config(cfg) |
| if cfg is not None |
| else BaseProcessor() |
| ) |
|
|
| vis_processors = dict() |
| txt_processors = dict() |
|
|
| vis_proc_cfg = config.get("vis_processor") |
| txt_proc_cfg = config.get("text_processor") |
|
|
| if vis_proc_cfg is not None: |
| vis_train_cfg = vis_proc_cfg.get("train") |
| vis_eval_cfg = vis_proc_cfg.get("eval") |
| else: |
| vis_train_cfg = None |
| vis_eval_cfg = None |
|
|
| vis_processors["train"] = _build_proc_from_cfg(vis_train_cfg) |
| vis_processors["eval"] = _build_proc_from_cfg(vis_eval_cfg) |
|
|
| if txt_proc_cfg is not None: |
| txt_train_cfg = txt_proc_cfg.get("train") |
| txt_eval_cfg = txt_proc_cfg.get("eval") |
| else: |
| txt_train_cfg = None |
| txt_eval_cfg = None |
|
|
| txt_processors["train"] = _build_proc_from_cfg(txt_train_cfg) |
| txt_processors["eval"] = _build_proc_from_cfg(txt_eval_cfg) |
|
|
| return vis_processors, txt_processors |
|
|
|
|
| def load_model_and_preprocess(name, model_type, is_eval=False, device="cpu"): |
| """ |
| Load model and its related preprocessors. |
| |
| List all available models and types in registry: |
| >>> from minigpt4.models import model_zoo |
| >>> print(model_zoo) |
| |
| Args: |
| name (str): name of the model. |
| model_type (str): type of the model. |
| is_eval (bool): whether the model is in eval mode. Default: False. |
| device (str): device to use. Default: "cpu". |
| |
| Returns: |
| model (torch.nn.Module): model. |
| vis_processors (dict): preprocessors for visual inputs. |
| txt_processors (dict): preprocessors for text inputs. |
| """ |
| model_cls = registry.get_model_class(name) |
|
|
| |
| model = model_cls.from_pretrained(model_type=model_type) |
|
|
| if is_eval: |
| model.eval() |
|
|
| |
| cfg = OmegaConf.load(model_cls.default_config_path(model_type)) |
| if cfg is not None: |
| preprocess_cfg = cfg.preprocess |
|
|
| vis_processors, txt_processors = load_preprocess(preprocess_cfg) |
| else: |
| vis_processors, txt_processors = None, None |
| logging.info( |
| f"""No default preprocess for model {name} ({model_type}). |
| This can happen if the model is not finetuned on downstream datasets, |
| or it is not intended for direct use without finetuning. |
| """ |
| ) |
|
|
| if device == "cpu" or device == torch.device("cpu"): |
| model = model.float() |
|
|
| return model.to(device), vis_processors, txt_processors |
|
|
|
|
| class ModelZoo: |
| """ |
| A utility class to create string representation of available model architectures and types. |
| |
| >>> from minigpt4.models import model_zoo |
| >>> # list all available models |
| >>> print(model_zoo) |
| >>> # show total number of models |
| >>> print(len(model_zoo)) |
| """ |
|
|
| def __init__(self) -> None: |
| self.model_zoo = { |
| k: list(v.PRETRAINED_MODEL_CONFIG_DICT.keys()) |
| for k, v in registry.mapping["model_name_mapping"].items() |
| } |
|
|
| def __str__(self) -> str: |
| return ( |
| "=" * 50 |
| + "\n" |
| + f"{'Architectures':<30} {'Types'}\n" |
| + "=" * 50 |
| + "\n" |
| + "\n".join( |
| [ |
| f"{name:<30} {', '.join(types)}" |
| for name, types in self.model_zoo.items() |
| ] |
| ) |
| ) |
|
|
| def __iter__(self): |
| return iter(self.model_zoo.items()) |
|
|
| def __len__(self): |
| return sum([len(v) for v in self.model_zoo.values()]) |
|
|
|
|
| model_zoo = ModelZoo() |
|
|