import torch import torch.nn.functional as F def fused_linear_jsd(X: torch.Tensor, W1: torch.Tensor, B1: torch.Tensor, W2: torch.Tensor, B2: torch.Tensor) -> torch.Tensor: """ Baseline fused linear Jensen-Shannon Divergence implementation using PyTorch. Args: X: Input tensor of shape (M, K) - input features (float16) W1: Weight tensor of shape (K, N) - first weight matrix (float16) B1: Bias tensor of shape (N,) - first bias vector (float32) W2: Weight tensor of shape (K, N) - second weight matrix (float16) B2: Bias tensor of shape (N,) - second bias vector (float32) Returns: Output tensor of shape (M,) - Jensen-Shannon Divergence per sample (float32) """ logits1 = (X.float() @ W1.float()) + B1.float() logits2 = (X.float() @ W2.float()) + B2.float() P = torch.softmax(logits1, dim=-1) Q = torch.softmax(logits2, dim=-1) Mmid = 0.5 * (P + Q) eps = 1e-12 jsd = 0.5 * (torch.sum(P * (torch.log(P + eps) - torch.log(Mmid + eps)), dim=-1) + torch.sum(Q * (torch.log(Q + eps) - torch.log(Mmid + eps)), dim=-1)) return jsd