andylizf's picture
Upload folder using huggingface_hub
5fed0fc verified
import torch
import torch.nn.functional as F
def fused_linear_jsd(X: torch.Tensor, W1: torch.Tensor, B1: torch.Tensor, W2: torch.Tensor, B2: torch.Tensor) -> torch.Tensor:
"""
Baseline fused linear Jensen-Shannon Divergence implementation using PyTorch.
Args:
X: Input tensor of shape (M, K) - input features (float16)
W1: Weight tensor of shape (K, N) - first weight matrix (float16)
B1: Bias tensor of shape (N,) - first bias vector (float32)
W2: Weight tensor of shape (K, N) - second weight matrix (float16)
B2: Bias tensor of shape (N,) - second bias vector (float32)
Returns:
Output tensor of shape (M,) - Jensen-Shannon Divergence per sample (float32)
"""
logits1 = (X.float() @ W1.float()) + B1.float()
logits2 = (X.float() @ W2.float()) + B2.float()
P = torch.softmax(logits1, dim=-1)
Q = torch.softmax(logits2, dim=-1)
Mmid = 0.5 * (P + Q)
eps = 1e-12
jsd = 0.5 * (torch.sum(P * (torch.log(P + eps) - torch.log(Mmid + eps)), dim=-1) +
torch.sum(Q * (torch.log(Q + eps) - torch.log(Mmid + eps)), dim=-1))
return jsd