| ''' |
| NBRDF MLP model |
| |
| Input: Cartesian coordinate for positional samples |
| (1: theta_h, 2: theta_d, 3: phi_d, 4: phi_h = 0) -> (hx, hy, hz, dx, dy, dz) |
| Output: MERL reflectance value |
| |
| - input_size 6 |
| - hidden_size 21 |
| - hidden_layer 3 |
| - output_size 3 |
| |
| @author |
| Copyright (c) 2024-2025 Peter HU. |
| |
| @file |
| reference: https://github.com/asztr/Neural-BRDF |
| |
| ''' |
| |
| import sys |
| import path |
| |
| import numpy as np |
| import torch |
| import torch.nn as nn |
| import torch.nn.functional as F |
| import random |
|
|
| |
| device = torch.device( |
| "cuda" if torch.cuda.is_available() |
| else torch.device("mps") if torch.backends.mps.is_available() |
| else "cpu") |
|
|
|
|
| class MLP(nn.Module): |
| '''Pytorch NBRDF MLP model''' |
| def __init__(self, input_size, hidden_size, output_size) -> None: |
| super().__init__() |
| |
| self.fc1 = nn.Linear(input_size, hidden_size, bias=True) |
| self.fc2 = nn.Linear(hidden_size, hidden_size, bias=True) |
| self.fc3 = nn.Linear(hidden_size, output_size, bias=True) |
| |
|
|
| |
| torch.manual_seed(0) |
| random.seed(0) |
| with torch.no_grad(): |
| for func in [self.fc1, self.fc2, self.fc3]: |
| func.bias.zero_() |
| func.weight.uniform_(0.0, 0.02) |
|
|
|
|
| def forward(self, x): |
| out = self.fc1(x) |
| out = F.leaky_relu(out) |
| out = self.fc2(out) |
| out = F.leaky_relu(out) |
| out = self.fc3(out) |
| out = F.relu(torch.exp(out) - 1.0) |
|
|
| return out |
|
|