|
|
from math import sin, cos, sqrt, atan2, radians, asin |
|
|
import numpy as np |
|
|
import torch |
|
|
import os |
|
|
from torch.distributed import init_process_group, destroy_process_group |
|
|
import torch.nn.functional as F |
|
|
import random |
|
|
def resample_trajectory(x, length=200): |
|
|
""" |
|
|
Resamples a trajectory to a new length. |
|
|
|
|
|
Parameters: |
|
|
x (np.ndarray): original trajectory, shape (N, 2) |
|
|
length (int): length of resampled trajectory |
|
|
|
|
|
Returns: |
|
|
np.ndarray: resampled trajectory, shape (length, 2) |
|
|
""" |
|
|
len_x = len(x) |
|
|
time_steps = np.arange(length) * (len_x - 1) / (length - 1) |
|
|
x = x.T |
|
|
resampled_trajectory = np.zeros((2, length)) |
|
|
for i in range(2): |
|
|
resampled_trajectory[i] = np.interp(time_steps, np.arange(len_x), x[i]) |
|
|
return resampled_trajectory.T |
|
|
|
|
|
|
|
|
def time_warping(x, length=200): |
|
|
""" |
|
|
Resamples a trajectory to a new length. |
|
|
""" |
|
|
len_x = len(x) |
|
|
time_steps = np.arange(length) * (len_x - 1) / (length - 1) |
|
|
x = x.T |
|
|
warped_trajectory = np.zeros((2, length)) |
|
|
for i in range(2): |
|
|
warped_trajectory[i] = np.interp(time_steps, np.arange(len_x), x[i]) |
|
|
return warped_trajectory.T |
|
|
|
|
|
|
|
|
def gather(consts: torch.Tensor, t: torch.Tensor): |
|
|
""" |
|
|
Gather consts for $t$ and reshape to feature map shape |
|
|
:param consts: (N, 1, 1) |
|
|
:param t: (N, H, W) |
|
|
:return: (N, H, W) |
|
|
""" |
|
|
c = consts.gather(-1, t) |
|
|
return c.reshape(-1, 1, 1) |
|
|
|
|
|
|
|
|
def q_xt_x0(x0, t, alpha_bar): |
|
|
|
|
|
mean = gather(alpha_bar, t) ** 0.5 * x0 |
|
|
var = 1 - gather(alpha_bar, t) |
|
|
|
|
|
eps = torch.randn_like(x0).to(x0.device) |
|
|
xt = mean + (var ** 0.5) * eps |
|
|
return xt, eps |
|
|
|
|
|
|
|
|
def compute_alpha(beta, t): |
|
|
beta = torch.cat([torch.zeros(1).to(beta.device), beta], dim=0) |
|
|
a = (1 - beta).cumprod(dim=0).index_select(0, t + 1).view(-1, 1, 1) |
|
|
return a |
|
|
|
|
|
|
|
|
def p_xt(xt, noise, t, next_t, beta, eta=0): |
|
|
at = compute_alpha(beta.cuda(), t.long()) |
|
|
at_next = compute_alpha(beta, next_t.long()) |
|
|
x0_t = (xt - noise * (1 - at).sqrt()) / at.sqrt() |
|
|
c1 = (eta * ((1 - at / at_next) * (1 - at_next) / (1 - at)).sqrt()) |
|
|
c2 = ((1 - at_next) - c1 ** 2).sqrt() |
|
|
eps = torch.randn(xt.shape, device=xt.device) |
|
|
xt_next = at_next.sqrt() * x0_t + c1 * eps + c2 * noise |
|
|
return xt_next |
|
|
|
|
|
|
|
|
def divide_grids(boundary, grids_num): |
|
|
lati_min, lati_max = boundary['lati_min'], boundary['lati_max'] |
|
|
long_min, long_max = boundary['long_min'], boundary['long_max'] |
|
|
|
|
|
lati_interval = (lati_max - lati_min) / grids_num |
|
|
long_interval = (long_max - long_min) / grids_num |
|
|
|
|
|
latgrids = np.arange(lati_min, lati_max, lati_interval) |
|
|
longrids = np.arange(long_min, long_max, long_interval) |
|
|
return latgrids, longrids |
|
|
|
|
|
|
|
|
|
|
|
def distance(lat1, lon1, lat2, lon2): |
|
|
""" |
|
|
Calculate the great circle distance between two points |
|
|
on the earth (specified in decimal degrees) |
|
|
""" |
|
|
|
|
|
lon1, lat1, lon2, lat2 = map(radians, [lon1, lat1, lon2, lat2]) |
|
|
|
|
|
dlon = lon2 - lon1 |
|
|
dlat = lat2 - lat1 |
|
|
a = sin(dlat/2)**2 + cos(lat1) * cos(lat2) * sin(dlon/2)**2 |
|
|
c = 2 * asin(sqrt(a)) |
|
|
r = 6371 |
|
|
return c * r * 1000 |
|
|
|
|
|
def set_seed(seed): |
|
|
random.seed(seed) |
|
|
np.random.seed(seed) |
|
|
torch.manual_seed(seed) |
|
|
if torch.cuda.is_available(): |
|
|
torch.cuda.manual_seed(seed) |
|
|
torch.cuda.manual_seed_all(seed) |
|
|
torch.backends.cudnn.deterministic = True |
|
|
torch.backends.cudnn.benchmark = False |
|
|
|
|
|
def ddp_setup(): |
|
|
init_process_group(backend="nccl") |
|
|
torch.cuda.set_device(int(os.environ['LOCAL_RANK'])) |
|
|
|
|
|
|
|
|
def destroy_process_group(): |
|
|
destroy_process_group() |
|
|
|
|
|
|
|
|
import torch |
|
|
|
|
|
class IterativeKMeans: |
|
|
def __init__(self, num_clusters, device, num_iters=100, tol=1e-4): |
|
|
self.num_clusters = num_clusters |
|
|
self.num_iters = num_iters |
|
|
self.tol = tol |
|
|
self.cluster_centers = None |
|
|
self.labels = None |
|
|
self.device = device |
|
|
|
|
|
def fit(self, X): |
|
|
|
|
|
X = X.clone().detach().to(self.device) |
|
|
num_samples, num_features = X.shape |
|
|
indices = torch.randperm(num_samples)[:self.num_clusters] |
|
|
self.cluster_centers = X[indices].clone().detach() |
|
|
self.labels = torch.argmin(torch.cdist(X, self.cluster_centers), dim=1).cpu().numpy() |
|
|
|
|
|
for _ in range(self.num_iters): |
|
|
distances = torch.cdist(X, self.cluster_centers) |
|
|
labels = torch.argmin(distances, dim=1) |
|
|
|
|
|
new_cluster_centers = torch.stack([X[labels == i].mean(dim=0) if (labels == i).sum() > 0 else self.cluster_centers[i] for i in range(self.num_clusters)]) |
|
|
center_shift = torch.norm(new_cluster_centers - self.cluster_centers, dim=1).sum().item() |
|
|
if center_shift < self.tol: |
|
|
break |
|
|
self.cluster_centers = new_cluster_centers |
|
|
|
|
|
self.labels = labels.cpu().numpy() |
|
|
return self.cluster_centers, self.labels |
|
|
|
|
|
def update(self, new_X, original_X): |
|
|
combined_X = torch.cat([original_X, new_X], dim=0) |
|
|
combined_X = combined_X.clone().detach().to(self.device) |
|
|
|
|
|
for _ in range(self.num_iters): |
|
|
distances = torch.cdist(combined_X, self.cluster_centers) |
|
|
labels = torch.argmin(distances, dim=1) |
|
|
new_cluster_centers = torch.stack([combined_X[labels == i].mean(dim=0) if (labels == i).sum() > 0 else self.cluster_centers[i] for i in range(self.num_clusters)]) |
|
|
center_shift = torch.norm(new_cluster_centers - self.cluster_centers, dim=1).sum().item() |
|
|
if center_shift < self.tol: |
|
|
break |
|
|
self.cluster_centers = new_cluster_centers |
|
|
|
|
|
self.labels = labels.cpu().numpy() |
|
|
return self.cluster_centers, self.labels |
|
|
|
|
|
def predict(self, X): |
|
|
|
|
|
X = X.clone().detach().to(self.device) |
|
|
distances = torch.cdist(X, self.cluster_centers) |
|
|
labels = torch.argmin(distances, dim=1) |
|
|
return labels |
|
|
|
|
|
def to(self, device): |
|
|
self.device = device |
|
|
if self.cluster_centers is not None: |
|
|
self.cluster_centers = self.cluster_centers.to(device) |
|
|
return self |
|
|
|
|
|
|
|
|
def assign_labels(prototypes, features): |
|
|
|
|
|
distances = F.pairwise_distance(features.unsqueeze(1), prototypes.unsqueeze(0)) |
|
|
|
|
|
labels = torch.argmin(distances, dim=-1) |
|
|
|
|
|
return labels |
|
|
|
|
|
|
|
|
def get_positive_negative_pairs(prototypes, samples): |
|
|
positive_pairs = [] |
|
|
negative_pairs = [] |
|
|
for sample in samples: |
|
|
distances = F.pairwise_distance(sample.unsqueeze(0), prototypes) |
|
|
pos_idx = torch.argmin(distances).item() |
|
|
neg_idx = torch.argmax(distances).item() |
|
|
positive_pairs.append(prototypes[pos_idx]) |
|
|
negative_pairs.append(prototypes[neg_idx]) |
|
|
return torch.stack(positive_pairs), torch.stack(negative_pairs) |
|
|
|
|
|
|
|
|
|
|
|
def mask_data_general(x: torch.Tensor): |
|
|
"""Mask the input data""" |
|
|
mask = torch.ones_like(x) |
|
|
mask[:, :, 1:-1] = 0 |
|
|
return x * mask.float() |
|
|
|
|
|
|
|
|
def continuous_mask_data(x: torch.Tensor, mask_ratio: float = 0.5): |
|
|
""" |
|
|
Mask a continuous block of the input data. |
|
|
It keeps the first and last elements unmasked. |
|
|
""" |
|
|
mask = torch.ones_like(x) |
|
|
|
|
|
traj_length = x.shape[2] |
|
|
if traj_length <= 2: |
|
|
return x * mask.float() |
|
|
|
|
|
masked_length = int((traj_length - 2) * mask_ratio) |
|
|
if masked_length == 0: |
|
|
return x * mask.float() |
|
|
|
|
|
|
|
|
|
|
|
mask_start = random.randint(1, traj_length - 2 - masked_length) |
|
|
mask_end = mask_start + masked_length |
|
|
|
|
|
mask[:, :, mask_start:mask_end] = 0 |
|
|
return x * mask.float() |
|
|
|
|
|
|
|
|
def update_npy(file_path, data): |
|
|
if os.path.exists(file_path): |
|
|
existing_data = np.load(file_path, allow_pickle=True).item() |
|
|
existing_data.update(data) |
|
|
else: |
|
|
existing_data = data |
|
|
np.save(file_path, existing_data) |
|
|
|
|
|
|
|
|
def haversine(lat1, lon1, lat2, lon2): |
|
|
|
|
|
lat1, lon1, lat2, lon2 = map(np.radians, [lat1, lon1, lat2, lon2]) |
|
|
|
|
|
|
|
|
dlat = lat2 - lat1 |
|
|
dlon = lon2 - lon1 |
|
|
a = np.sin(dlat / 2.0)**2 + np.cos(lat1) * np.cos(lat2) * np.sin(dlon / 2.0)**2 |
|
|
c = 2 * np.arcsin(np.sqrt(a)) |
|
|
r = 6371 |
|
|
return c * r * 1000 |
|
|
|
|
|
|
|
|
def continuous_time_based_mask(x: torch.Tensor, points_to_mask: int): |
|
|
""" |
|
|
Mask a continuous block of the input data based on a fixed number of points. |
|
|
It keeps the first and last elements unmasked. |
|
|
|
|
|
Args: |
|
|
x (torch.Tensor): Input tensor of shape (batch, features, length). |
|
|
points_to_mask (int): The number of continuous points to mask. |
|
|
""" |
|
|
mask = torch.ones_like(x) |
|
|
|
|
|
traj_length = x.shape[2] |
|
|
|
|
|
if traj_length <= points_to_mask + 2: |
|
|
|
|
|
return x * mask.float() |
|
|
|
|
|
|
|
|
|
|
|
mask_start = random.randint(1, traj_length - 1 - points_to_mask) |
|
|
mask_end = mask_start + points_to_mask |
|
|
|
|
|
mask[:, :, mask_start:mask_end] = 0 |
|
|
return x * mask.float() |
|
|
|
|
|
|
|
|
def mask_multiple_segments(x: torch.Tensor, points_per_segment: list): |
|
|
""" |
|
|
Mask multiple non-overlapping continuous segments in the input data. |
|
|
Keeps the first and last elements unmasked. |
|
|
|
|
|
Args: |
|
|
x (torch.Tensor): Input tensor of shape (batch, features, length). |
|
|
points_per_segment (list of int): List containing the length of each segment to mask. |
|
|
""" |
|
|
mask = torch.ones_like(x) |
|
|
traj_length = x.shape[2] |
|
|
|
|
|
|
|
|
segments = sorted(points_per_segment, reverse=True) |
|
|
total_mask_points = sum(segments) |
|
|
|
|
|
|
|
|
if traj_length < total_mask_points + 2: |
|
|
return x * mask.float() |
|
|
|
|
|
|
|
|
|
|
|
possible_indices = list(range(1, traj_length - 1)) |
|
|
|
|
|
masked_intervals = [] |
|
|
|
|
|
for seg_length in segments: |
|
|
|
|
|
valid_starts = [] |
|
|
for i in possible_indices: |
|
|
|
|
|
|
|
|
if i + seg_length > traj_length - 1: |
|
|
continue |
|
|
|
|
|
is_valid = True |
|
|
for start, end in masked_intervals: |
|
|
|
|
|
if not (i + seg_length <= start or i >= end): |
|
|
is_valid = False |
|
|
break |
|
|
if is_valid: |
|
|
valid_starts.append(i) |
|
|
|
|
|
if not valid_starts: |
|
|
|
|
|
continue |
|
|
|
|
|
|
|
|
start_pos = random.choice(valid_starts) |
|
|
end_pos = start_pos + seg_length |
|
|
mask[:, :, start_pos:end_pos] = 0 |
|
|
|
|
|
|
|
|
masked_intervals.append((start_pos, end_pos)) |
|
|
|
|
|
|
|
|
indices_to_remove = set(range(start_pos, end_pos)) |
|
|
possible_indices = [idx for idx in possible_indices if idx not in indices_to_remove] |
|
|
|
|
|
return x * mask.float() |
|
|
|
|
|
def get_data_paths(data_config, for_train=True): |
|
|
"""Get the file paths for training or testing data for TKY-like structure. |
|
|
Assumes data_config.traj_path1 points to a directory containing train.h5 and test.h5. |
|
|
""" |
|
|
|
|
|
base_path = data_config.traj_path1 |
|
|
if not isinstance(base_path, str): |
|
|
base_path = str(base_path) |
|
|
|
|
|
|
|
|
if hasattr(data_config, 'dataset') and 'temporal' in data_config.dataset: |
|
|
if for_train: |
|
|
file_path = os.path.join(base_path, "train_temporal.h5") |
|
|
else: |
|
|
file_path = os.path.join(base_path, "test_temporal.h5") |
|
|
else: |
|
|
|
|
|
if for_train: |
|
|
file_path = os.path.join(base_path, "train.h5") |
|
|
else: |
|
|
file_path = os.path.join(base_path, "test.h5") |
|
|
|
|
|
|
|
|
return [file_path] |