| import torch |
| import torch.nn as nn |
| import numpy as np |
| import math |
|
|
| |
|
|
| def modulate(x, shift, scale): |
| return x * (1 + scale.unsqueeze(1)) + shift.unsqueeze(1) |
|
|
|
|
| |
| |
|
|
| class DiTAttention(nn.Module): |
| def __init__(self, cfg): |
| super().__init__() |
| self.t_embedder = TimestepEmbedder(cfg.hidden_size) |
| self.trans_conv = nn.ModuleList([ |
| DiTBlock(cfg) for _ in range(cfg.num_layers) |
| ]) |
|
|
| def forward(self, t, x, mask=None): |
| t = self.t_embedder(t) |
| for block in self.trans_conv: |
| x = block(x, t, mask) |
| return x |
| |
|
|
|
|
| class TimestepEmbedder(nn.Module): |
| """ |
| Embeds scalar timesteps into vector representations. |
| """ |
| def __init__(self, hidden_size, frequency_embedding_size=256): |
| super().__init__() |
| self.mlp = nn.Sequential( |
| nn.Linear(frequency_embedding_size, hidden_size, bias=True), |
| nn.SiLU(), |
| nn.Linear(hidden_size, hidden_size, bias=True), |
| ) |
| self.frequency_embedding_size = frequency_embedding_size |
|
|
| @staticmethod |
| def timestep_embedding(t, dim, max_period=10000): |
| """ |
| Create sinusoidal timestep embeddings. |
| :param t: a 1-D Tensor of N indices, one per batch element. |
| These may be fractional. |
| :param dim: the dimension of the output. |
| :param max_period: controls the minimum frequency of the embeddings. |
| :return: an (N, D) Tensor of positional embeddings. |
| """ |
| |
| half = dim // 2 |
| freqs = torch.exp( |
| -math.log(max_period) * torch.arange(start=0, end=half, dtype=torch.float32) / half |
| ).to(device=t.device) |
| args = t[:, None].float() * freqs[None] |
| embedding = torch.cat([torch.cos(args), torch.sin(args)], dim=-1) |
| if dim % 2: |
| embedding = torch.cat([embedding, torch.zeros_like(embedding[:, :1])], dim=-1) |
| return embedding |
|
|
| def forward(self, t): |
| t_freq = self.timestep_embedding(t, self.frequency_embedding_size) |
| t_emb = self.mlp(t_freq) |
| return t_emb |
|
|
|
|
|
|
| class Mlp(nn.Module): |
| """ MLP as used in Vision Transformer, MLP-Mixer and related networks |
| |
| NOTE: When use_conv=True, expects 2D NCHW tensors, otherwise N*C expected. |
| """ |
| def __init__( |
| self, |
| in_features, |
| hidden_features=None, |
| out_features=None, |
| act_layer=nn.GELU, |
| norm_layer=None, |
| bias=True, |
| drop=0., |
| ): |
| super().__init__() |
| out_features = out_features or in_features |
| hidden_features = hidden_features or in_features |
| linear_layer = nn.Linear |
|
|
| self.fc1 = linear_layer(in_features, hidden_features, bias=bias) |
| self.act = act_layer() |
| self.drop1 = nn.Dropout(drop) |
| self.norm = norm_layer(hidden_features) if norm_layer is not None else nn.Identity() |
| self.fc2 = linear_layer(hidden_features, out_features, bias=bias) |
| self.drop2 = nn.Dropout(drop) |
|
|
| def forward(self, x): |
| x = self.fc1(x) |
| x = self.act(x) |
| x = self.drop1(x) |
| x = self.norm(x) |
| x = self.fc2(x) |
| x = self.drop2(x) |
| return x |
| |
| class DiTBlock(nn.Module): |
| """ |
| A DiT block with adaptive layer norm zero (adaLN-Zero) conditioning. |
| """ |
| def __init__(self, cfg): |
| super().__init__() |
| hidden_size = cfg.model.hidden_channels + cfg.model.edge_dim |
| num_heads = cfg.model.trans_num_heads |
| dropout = cfg.model.trans_dropout |
| mlp_ratio= 4.0 |
| |
| self.norm1 = nn.LayerNorm(hidden_size, elementwise_affine=False, eps=1e-6) |
| self.attn = BertSelfAttention(num_attention_heads=num_heads, hidden_size=hidden_size, dropout_prob=dropout) |
| self.norm2 = nn.LayerNorm(hidden_size, elementwise_affine=False, eps=1e-6) |
| mlp_hidden_dim = int(hidden_size * mlp_ratio) |
| approx_gelu = lambda: nn.GELU(approximate="tanh") |
| self.mlp = Mlp(in_features=hidden_size, hidden_features=mlp_hidden_dim, act_layer=approx_gelu, drop=dropout) |
| self.adaLN_modulation = nn.Sequential( |
| nn.SiLU(), |
| nn.Linear(hidden_size, 6 * hidden_size, bias=True) |
| ) |
|
|
| def forward(self, x, c, mask): |
| shift_msa, scale_msa, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.adaLN_modulation(c).chunk(6, dim=-1) |
| x = x + gate_msa * self.attn(modulate(self.norm1(x), shift_msa, scale_msa), mask) |
| x = x + gate_mlp * self.mlp(modulate(self.norm2(x), shift_mlp, scale_mlp)) |
| return x |
|
|
|
|
| class BertSelfAttention(nn.Module): |
| def __init__(self, num_attention_heads, hidden_size, dropout_prob): |
| super().__init__() |
| self.num_attention_heads = num_attention_heads |
| self.attention_head_size = int(hidden_size / num_attention_heads) |
| self.all_head_size = self.num_attention_heads * self.attention_head_size |
|
|
| self.query = nn.Linear(hidden_size, hidden_size) |
| self.key = nn.Linear(hidden_size, hidden_size) |
| self.value = nn.Linear(hidden_size, hidden_size) |
| self.dropout = nn.Dropout(dropout_prob) |
|
|
| def transpose_for_scores(self, x: torch.Tensor) -> torch.Tensor: |
| new_x_shape = x.size()[:-1] + (self.num_attention_heads, self.attention_head_size) |
| x = x.view(new_x_shape) |
| return x.permute(0, 2, 1, 3) |
|
|
| def forward( |
| self, |
| hidden_states: torch.Tensor, |
| attention_mask: None, |
| ): |
| mixed_query_layer = self.query(hidden_states) |
|
|
| key_layer = self.transpose_for_scores(self.key(hidden_states)) |
| value_layer = self.transpose_for_scores(self.value(hidden_states)) |
| query_layer = self.transpose_for_scores(mixed_query_layer) |
| |
| |
| attention_scores = torch.matmul(query_layer, key_layer.transpose(-1, -2)) |
| attention_scores = attention_scores / math.sqrt(self.attention_head_size) |
| |
| if attention_mask is not None: |
| |
| |
| attention_mask = torch.einsum('bi,bj->bij', attention_mask, attention_mask).unsqueeze(1) |
| attention_mask = attention_mask.masked_fill(attention_mask == 0, -10000.0) |
| attention_scores = attention_scores + attention_mask |
|
|
| |
| attention_probs = nn.functional.softmax(attention_scores, dim=-1) |
|
|
| |
| |
| attention_probs = self.dropout(attention_probs) |
| context_layer = torch.matmul(attention_probs, value_layer) |
| context_layer = context_layer.permute(0, 2, 1, 3).contiguous() |
| new_context_layer_shape = context_layer.size()[:-2] + (self.all_head_size,) |
| context_layer = context_layer.view(new_context_layer_shape) |
| |
| return context_layer |