File size: 13,400 Bytes
0a027c9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
from math import sin, cos, sqrt, atan2, radians, asin
import numpy as np
import torch
import os
from torch.distributed import init_process_group, destroy_process_group
import torch.nn.functional as F
import random
def resample_trajectory(x, length=200):
    """
    Resamples a trajectory to a new length.

    Parameters:
        x (np.ndarray): original trajectory, shape (N, 2)
        length (int): length of resampled trajectory

    Returns:
        np.ndarray: resampled trajectory, shape (length, 2)
    """
    len_x = len(x)
    time_steps = np.arange(length) * (len_x - 1) / (length - 1)
    x = x.T
    resampled_trajectory = np.zeros((2, length))
    for i in range(2):
        resampled_trajectory[i] = np.interp(time_steps, np.arange(len_x), x[i])
    return resampled_trajectory.T


def time_warping(x, length=200):
    """
    Resamples a trajectory to a new length.
    """
    len_x = len(x)
    time_steps = np.arange(length) * (len_x - 1) / (length - 1)
    x = x.T
    warped_trajectory = np.zeros((2, length))
    for i in range(2):
        warped_trajectory[i] = np.interp(time_steps, np.arange(len_x), x[i])
    return warped_trajectory.T


def gather(consts: torch.Tensor, t: torch.Tensor):
    """
    Gather consts for $t$ and reshape to feature map shape
    :param consts: (N, 1, 1)
    :param t: (N, H, W)
    :return: (N, H, W)
    """
    c = consts.gather(-1, t)
    return c.reshape(-1, 1, 1)


def q_xt_x0(x0, t, alpha_bar):
    # get mean and variance of xt given x0
    mean = gather(alpha_bar, t) ** 0.5 * x0
    var = 1 - gather(alpha_bar, t)
    # sample xt from q(xt | x0)
    eps = torch.randn_like(x0).to(x0.device)
    xt = mean + (var ** 0.5) * eps
    return xt, eps  # also return noise


def compute_alpha(beta, t):
    beta = torch.cat([torch.zeros(1).to(beta.device), beta], dim=0)
    a = (1 - beta).cumprod(dim=0).index_select(0, t + 1).view(-1, 1, 1)
    return a


def p_xt(xt, noise, t, next_t, beta, eta=0):
    at = compute_alpha(beta.cuda(), t.long())
    at_next = compute_alpha(beta, next_t.long())
    x0_t = (xt - noise * (1 - at).sqrt()) / at.sqrt()
    c1 = (eta * ((1 - at / at_next) * (1 - at_next) / (1 - at)).sqrt())
    c2 = ((1 - at_next) - c1 ** 2).sqrt()
    eps = torch.randn(xt.shape, device=xt.device)
    xt_next = at_next.sqrt() * x0_t + c1 * eps + c2 * noise
    return xt_next


def divide_grids(boundary, grids_num):
    lati_min, lati_max = boundary['lati_min'], boundary['lati_max']
    long_min, long_max = boundary['long_min'], boundary['long_max']
    # Divide the latitude and longitude into grids_num intervals.
    lati_interval = (lati_max - lati_min) / grids_num
    long_interval = (long_max - long_min) / grids_num
    # Create arrays of latitude and longitude values.
    latgrids = np.arange(lati_min, lati_max, lati_interval)
    longrids = np.arange(long_min, long_max, long_interval)
    return latgrids, longrids


# calculate the distance between two points
def distance(lat1, lon1, lat2, lon2):
    """
    Calculate the great circle distance between two points 
    on the earth (specified in decimal degrees)
    """
    # convert decimal degrees to radians
    lon1, lat1, lon2, lat2 = map(radians, [lon1, lat1, lon2, lat2])
    # haversine formula
    dlon = lon2 - lon1
    dlat = lat2 - lat1
    a = sin(dlat/2)**2 + cos(lat1) * cos(lat2) * sin(dlon/2)**2
    c = 2 * asin(sqrt(a))
    r = 6371  # Radius of earth in kilometers. Use 3956 for miles
    return c * r * 1000

def set_seed(seed):
    random.seed(seed)
    np.random.seed(seed)
    torch.manual_seed(seed)
    if torch.cuda.is_available():
        torch.cuda.manual_seed(seed)
        torch.cuda.manual_seed_all(seed)
        torch.backends.cudnn.deterministic = True
        torch.backends.cudnn.benchmark = False
        
def ddp_setup():
    init_process_group(backend="nccl")
    torch.cuda.set_device(int(os.environ['LOCAL_RANK']))


def destroy_process_group():
    destroy_process_group()


import torch

class IterativeKMeans:
    def __init__(self, num_clusters, device, num_iters=100, tol=1e-4):
        self.num_clusters = num_clusters
        self.num_iters = num_iters
        self.tol = tol
        self.cluster_centers = None
        self.labels = None
        self.device = device

    def fit(self, X):
        # X = torch.tensor(X, dtype=torch.float32).to(self.device)
        X = X.clone().detach().to(self.device)
        num_samples, num_features = X.shape
        indices = torch.randperm(num_samples)[:self.num_clusters]
        self.cluster_centers = X[indices].clone().detach()
        self.labels = torch.argmin(torch.cdist(X, self.cluster_centers), dim=1).cpu().numpy()

        for _ in range(self.num_iters):
            distances = torch.cdist(X, self.cluster_centers)
            labels = torch.argmin(distances, dim=1)
            # new_cluster_centers = torch.stack([X[labels == i].mean(dim=0) for i in range(self.num_clusters)])
            new_cluster_centers = torch.stack([X[labels == i].mean(dim=0) if (labels == i).sum() > 0 else self.cluster_centers[i] for i in range(self.num_clusters)])
            center_shift = torch.norm(new_cluster_centers - self.cluster_centers, dim=1).sum().item()
            if center_shift < self.tol:
                break
            self.cluster_centers = new_cluster_centers

        self.labels = labels.cpu().numpy()
        return self.cluster_centers, self.labels

    def update(self, new_X, original_X):
        combined_X = torch.cat([original_X, new_X], dim=0)
        combined_X = combined_X.clone().detach().to(self.device)

        for _ in range(self.num_iters):
            distances = torch.cdist(combined_X, self.cluster_centers)
            labels = torch.argmin(distances, dim=1)
            new_cluster_centers = torch.stack([combined_X[labels == i].mean(dim=0) if (labels == i).sum() > 0 else self.cluster_centers[i] for i in range(self.num_clusters)])
            center_shift = torch.norm(new_cluster_centers - self.cluster_centers, dim=1).sum().item()
            if center_shift < self.tol:
                break
            self.cluster_centers = new_cluster_centers

        self.labels = labels.cpu().numpy()
        return self.cluster_centers, self.labels

    def predict(self, X):
        # X = torch.tensor(X, dtype=torch.float32).to(self.device)
        X = X.clone().detach().to(self.device)
        distances = torch.cdist(X, self.cluster_centers)
        labels = torch.argmin(distances, dim=1)
        return labels

    def to(self, device):
        self.device = device
        if self.cluster_centers is not None:
            self.cluster_centers = self.cluster_centers.to(device)
        return self


def assign_labels(prototypes, features):
    # Calculate pairwise distances between all features and prototypes
    distances = F.pairwise_distance(features.unsqueeze(1), prototypes.unsqueeze(0))
    # Find the index of the prototype with the minimum distance (on the second dimension)
    labels = torch.argmin(distances, dim=-1)

    return labels


def get_positive_negative_pairs(prototypes, samples):
    positive_pairs = []
    negative_pairs = []
    for sample in samples:
        distances = F.pairwise_distance(sample.unsqueeze(0), prototypes)
        pos_idx = torch.argmin(distances).item()
        neg_idx = torch.argmax(distances).item()
        positive_pairs.append(prototypes[pos_idx])
        negative_pairs.append(prototypes[neg_idx])
    return torch.stack(positive_pairs), torch.stack(negative_pairs)



def mask_data_general(x: torch.Tensor):
    """Mask the input data""" 
    mask = torch.ones_like(x)
    mask[:, :, 1:-1] = 0
    return x * mask.float()


def continuous_mask_data(x: torch.Tensor, mask_ratio: float = 0.5):
    """
    Mask a continuous block of the input data.
    It keeps the first and last elements unmasked.
    """
    mask = torch.ones_like(x)
    
    traj_length = x.shape[2]
    if traj_length <= 2:
        return x * mask.float()

    masked_length = int((traj_length - 2) * mask_ratio)
    if masked_length == 0:
        return x * mask.float()

    # The start of the mask is between the first and the last but one element.
    # The selection ensures that the mask does not run over the second to last element.
    mask_start = random.randint(1, traj_length - 2 - masked_length)
    mask_end = mask_start + masked_length

    mask[:, :, mask_start:mask_end] = 0
    return x * mask.float()


def update_npy(file_path, data):
    if os.path.exists(file_path):
        existing_data = np.load(file_path, allow_pickle=True).item()
        existing_data.update(data)
    else:
        existing_data = data
    np.save(file_path, existing_data)


def haversine(lat1, lon1, lat2, lon2):
    # Convert degrees to radians
    lat1, lon1, lat2, lon2 = map(np.radians, [lat1, lon1, lat2, lon2])

    # Haversine formula
    dlat = lat2 - lat1
    dlon = lon2 - lon1
    a = np.sin(dlat / 2.0)**2 + np.cos(lat1) * np.cos(lat2) * np.sin(dlon / 2.0)**2
    c = 2 * np.arcsin(np.sqrt(a))
    r = 6371  # Radius of Earth in kilometers
    return c * r * 1000  # Return distance in meters


def continuous_time_based_mask(x: torch.Tensor, points_to_mask: int):
    """
    Mask a continuous block of the input data based on a fixed number of points.
    It keeps the first and last elements unmasked.
    
    Args:
        x (torch.Tensor): Input tensor of shape (batch, features, length).
        points_to_mask (int): The number of continuous points to mask.
    """
    mask = torch.ones_like(x)
    
    traj_length = x.shape[2]
    # 确保有足够的点可以遮蔽 (首尾点+遮蔽段)
    if traj_length <= points_to_mask + 2:
        # 如果轨迹太短,无法满足遮蔽要求,则不进行遮蔽
        return x * mask.float()

    # 随机选择遮蔽的起始位置
    # 起始位置必须在第一个点之后,并确保遮蔽段不会超出倒数第二个点
    mask_start = random.randint(1, traj_length - 1 - points_to_mask)
    mask_end = mask_start + points_to_mask

    mask[:, :, mask_start:mask_end] = 0
    return x * mask.float()


def mask_multiple_segments(x: torch.Tensor, points_per_segment: list):
    """
    Mask multiple non-overlapping continuous segments in the input data.
    Keeps the first and last elements unmasked.

    Args:
        x (torch.Tensor): Input tensor of shape (batch, features, length).
        points_per_segment (list of int): List containing the length of each segment to mask.
    """
    mask = torch.ones_like(x)
    traj_length = x.shape[2]
    
    # Sort segments by length, descending, to place larger gaps first
    segments = sorted(points_per_segment, reverse=True)
    total_mask_points = sum(segments)

    # Check if there's enough space for all masks and endpoints
    if traj_length < total_mask_points + 2:
        return x * mask.float()

    # Generate a list of all possible start indices for masking
    # Exclude first and last points: [1, ..., traj_length-2]
    possible_indices = list(range(1, traj_length - 1))
    
    masked_intervals = []

    for seg_length in segments:
        # Find valid start positions for the current segment
        valid_starts = []
        for i in possible_indices:
            # A start is valid if the segment [i, i + seg_length) does not overlap with existing masked intervals
            # and does not go out of bounds
            if i + seg_length > traj_length - 1:
                continue
            
            is_valid = True
            for start, end in masked_intervals:
                # Check for overlap: [i, i+seg_length) vs [start, end)
                if not (i + seg_length <= start or i >= end):
                    is_valid = False
                    break
            if is_valid:
                valid_starts.append(i)
        
        if not valid_starts:
            # Not enough space for this segment, continue to next (smaller) one
            continue

        # Choose a random start position and apply the mask
        start_pos = random.choice(valid_starts)
        end_pos = start_pos + seg_length
        mask[:, :, start_pos:end_pos] = 0
        
        # Record the masked interval and remove these indices from possible choices
        masked_intervals.append((start_pos, end_pos))
        
        # Update possible_indices by removing the masked range
        indices_to_remove = set(range(start_pos, end_pos))
        possible_indices = [idx for idx in possible_indices if idx not in indices_to_remove]

    return x * mask.float()

def get_data_paths(data_config, for_train=True):
    """Get the file paths for training or testing data for TKY-like structure.
    Assumes data_config.traj_path1 points to a directory containing train.h5 and test.h5.
    """

    base_path = data_config.traj_path1
    if not isinstance(base_path, str):
        base_path = str(base_path)

    # Check if we're using temporal split data
    if hasattr(data_config, 'dataset') and 'temporal' in data_config.dataset:
        if for_train:
            file_path = os.path.join(base_path, "train_temporal.h5")
        else:
            file_path = os.path.join(base_path, "test_temporal.h5")
    else:
        # Use the original file names
        if for_train:
            file_path = os.path.join(base_path, "train.h5")
        else:
            file_path = os.path.join(base_path, "test.h5")
    

    return [file_path]