| | import math |
| |
|
| | import matplotlib.pyplot as plt |
| | import numpy as np |
| | import torch |
| | import torch.nn as nn |
| | import torch.nn.functional as F |
| |
|
| | try: |
| | |
| | from .canopy_plots import createCHM |
| | except ImportError: |
| | |
| | from hhdc.canopy_plots import createCHM |
| |
|
| |
|
| | class LidarForwardImagingModel(nn.Module): |
| | def __init__( |
| | self, |
| | input_res_m=(2.0, 2.0), |
| | output_res_m=(3.0, 6.0), |
| | footprint_diameter_m=10.0, |
| | b=0.1, |
| | eta=0.5, |
| | ref_altitude=500.0, |
| | ref_photon_count=20.0, |
| | ): |
| | """ |
| | Args: |
| | input_res_m (tuple): Physical size of input pixels (dy, dx) in meters. |
| | output_res_m (tuple): Physical size of output pixels (dy, dx) in meters. |
| | footprint_diameter_m (float): The 1/e^2 beam diameter in meters. |
| | b (float): Background noise. |
| | eta (float): Readout noise. |
| | ref_altitude (float): Reference altitude (km). |
| | ref_photon_count (float): Target photon count. |
| | """ |
| | super().__init__() |
| | self.b = b |
| | self.eta = eta |
| | self.ref_altitude = ref_altitude |
| | self.ref_photon_count = ref_photon_count |
| |
|
| | self.input_res_m = input_res_m |
| | self.output_res_m = output_res_m |
| |
|
| | |
| | in_area = input_res_m[0] * input_res_m[1] |
| | out_area = output_res_m[0] * output_res_m[1] |
| | self.area_scale_factor = out_area / in_area |
| |
|
| | |
| | |
| | sigma_m = footprint_diameter_m / 4.0 |
| |
|
| | avg_input_res = (input_res_m[0] + input_res_m[1]) / 2.0 |
| | sigma_px = sigma_m / avg_input_res |
| |
|
| | |
| | |
| | |
| | kernel_size = int(math.ceil(6 * sigma_px)) |
| | if kernel_size % 2 == 0: |
| | kernel_size += 1 |
| |
|
| | self.register_buffer("kernel", self._create_gaussian_kernel(kernel_size, sigma_px)) |
| |
|
| | print(f"Model Initialized: In {input_res_m}m -> Out {output_res_m}m") |
| | print(f"Footprint (1/e^2): {footprint_diameter_m}m (Sigma: {sigma_m:.2f}m / {sigma_px:.2f} px)") |
| |
|
| | def _create_gaussian_kernel(self, size, sigma): |
| | coords = torch.arange(size).float() - (size - 1) / 2 |
| | x_grid, y_grid = torch.meshgrid(coords, coords, indexing='ij') |
| | kernel = torch.exp(-(x_grid**2 + y_grid**2) / (2 * sigma**2)) |
| | kernel = kernel / kernel.sum() |
| | return kernel.view(1, 1, size, size) |
| |
|
| | def forward(self, X_h, altitude=500.0): |
| | if X_h.ndim == 3: |
| | X_h = X_h.unsqueeze(0) |
| |
|
| | batch_size, num_bins, h_in, w_in = X_h.shape |
| |
|
| | |
| | fov_h_m = h_in * self.input_res_m[0] |
| | fov_w_m = w_in * self.input_res_m[1] |
| |
|
| | out_h = int(fov_h_m / self.output_res_m[0]) |
| | out_w = int(fov_w_m / self.output_res_m[1]) |
| | output_size = (out_h, out_w) |
| |
|
| | |
| | energy_per_tube = X_h.sum(dim=1, keepdim=True) |
| | global_mean_energy = energy_per_tube.mean(dim=(2, 3), keepdim=True) |
| | X_norm = X_h / (global_mean_energy + 1e-8) |
| |
|
| | dist_scale = (self.ref_altitude / altitude) ** 2 |
| | target_intensity = (self.ref_photon_count / self.area_scale_factor) * dist_scale |
| | X_scaled = X_norm * target_intensity |
| |
|
| | |
| | |
| | current_kernel = self.kernel.repeat(num_bins, 1, 1, 1) |
| |
|
| | padding = current_kernel.shape[-1] // 2 |
| | X_blurred = F.conv2d(X_scaled, current_kernel, padding=padding, groups=num_bins) |
| |
|
| | |
| | X_binned = F.interpolate(X_blurred, size=output_size, mode='area') |
| | X_integrated = X_binned * self.area_scale_factor |
| |
|
| | |
| | lambda_val = torch.relu(X_integrated) + self.b |
| | X_l = torch.poisson(lambda_val) |
| | gaussian_noise = torch.randn_like(X_l) * self.eta |
| | Y_l = X_l + gaussian_noise |
| |
|
| | if Y_l.shape[0] == 1: |
| | Y_l = Y_l.squeeze(0) |
| |
|
| | return Y_l |
| |
|