|
|
import os |
|
|
from .clip_encoder import CLIPVisionTower |
|
|
from .imagebind import ImageBindWrapper |
|
|
from .open_clip_encoder import OpenCLIPVisionTower |
|
|
from .siglip_encoder import SigLipVisionTower |
|
|
from .clip_encoder import CLIPVisionTower, CLIPVisionTowerS2 |
|
|
|
|
|
from .eva_clip.eva_clip_encoder import EvaClipVisionTower |
|
|
from .dev_eva_clip.eva_vit import EvaViTWrapper |
|
|
|
|
|
from blip3o.model.nextdit_crossattn import NextDiTCrossAttnConfig, NextDiTCrossAttn |
|
|
from blip3o.model.sana_crossattn import SanaCrossAttnConfig, SanaCrossAttn |
|
|
|
|
|
from diffusers.models import AutoencoderKL |
|
|
from diffusers.schedulers import FlowMatchEulerDiscreteScheduler, DPMSolverMultistepScheduler |
|
|
from diffusers import SanaTransformer2DModel |
|
|
import torch |
|
|
|
|
|
def build_vision_tower(vision_tower_cfg, **kwargs): |
|
|
vision_tower = getattr(vision_tower_cfg, 'mm_vision_tower', getattr(vision_tower_cfg, 'vision_tower', None)) |
|
|
is_absolute_path_exists = os.path.exists(vision_tower) |
|
|
use_s2 = getattr(vision_tower_cfg, 's2', False) |
|
|
if "siglip" in vision_tower: |
|
|
return SigLipVisionTower(vision_tower, vision_tower_cfg=vision_tower_cfg, **kwargs) |
|
|
if "eva" in vision_tower: |
|
|
return EvaClipVisionTower(vision_tower, args=vision_tower_cfg, **kwargs) |
|
|
if is_absolute_path_exists or vision_tower.startswith("openai") or vision_tower.startswith("laion") or "ShareGPT4V" in vision_tower: |
|
|
if use_s2: |
|
|
return CLIPVisionTowerS2(vision_tower, args=vision_tower_cfg, **kwargs) |
|
|
else: |
|
|
return CLIPVisionTower(vision_tower, args=vision_tower_cfg, **kwargs) |
|
|
|
|
|
raise ValueError(f'Unknown vision tower: {vision_tower}') |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
def build_gen_vision_tower(vision_tower_cfg, **kwargs): |
|
|
vision_tower = getattr(vision_tower_cfg, 'gen_vision_tower') |
|
|
is_absolute_path_exists = os.path.exists(vision_tower) |
|
|
use_s2 = getattr(vision_tower_cfg, 's2', False) |
|
|
if "siglip" in vision_tower: |
|
|
return SigLipVisionTower(vision_tower, vision_tower_cfg=vision_tower_cfg, **kwargs) |
|
|
if "eva" in vision_tower: |
|
|
return EvaClipVisionTower(vision_tower, args=vision_tower_cfg, **kwargs) |
|
|
if is_absolute_path_exists or vision_tower.startswith("openai") or vision_tower.startswith("laion") or "ShareGPT4V" in vision_tower: |
|
|
if use_s2: |
|
|
return CLIPVisionTowerS2(vision_tower, args=vision_tower_cfg, **kwargs) |
|
|
else: |
|
|
return CLIPVisionTower(vision_tower, args=vision_tower_cfg, **kwargs) |
|
|
|
|
|
raise ValueError(f'Unknown vision tower: {vision_tower}') |
|
|
|
|
|
|
|
|
|
|
|
def build_dit(vision_tower_cfg, **kwargs): |
|
|
vision_tower_cfg.hidden_size = 896 |
|
|
print("="*20, "vision tower config", vision_tower_cfg, "="*20) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
print("="*20, "Building SANA with hidden size", vision_tower_cfg.hidden_size, "="*20) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
dit = SanaTransformer2DModel.from_pretrained("Efficient-Large-Model/Sana_600M_1024px_diffusers",ignore_missing_keys=True, ignore_mismatched_sizes=True,device_map="cpu", subfolder="transformer") |
|
|
noise_scheduler = DPMSolverMultistepScheduler.from_pretrained("Efficient-Large-Model/Sana_600M_1024px_diffusers",subfolder="scheduler") |
|
|
|
|
|
return dit, noise_scheduler |
|
|
|
|
|
|
|
|
|