import torch import torch.nn as nn import torch.nn.functional as F class GELU(nn.Module): def __init__(self): super().__init__() def forward(self,x): return 0.5 * x * ( 1+torch.tanh( torch.sqrt(torch.tensor(2/torch.pi))*(x+0.044715*torch.pow(x,3)) ) ) class UstaMLP(nn.Module): def __init__(self,embedding_dim,hidden_dim,device="cpu"): super().__init__() self.gate_proj = nn.Linear(embedding_dim,hidden_dim,device=device) self.up_proj = nn.Linear(embedding_dim,hidden_dim,device=device) self.down_proj = nn.Linear(hidden_dim,embedding_dim,device=device) self.gelu = GELU().to(device) def forward(self,x): gate = self.gate_proj(x) gate = self.gelu(gate) up = self.up_proj(x) fuse = gate*up outputs = self.down_proj(fuse) return outputs