File size: 13,623 Bytes
ba9f51f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
import pandas as pd
import argparse
import os
import numpy as np
from scipy.ndimage import gaussian_filter1d
from sklearn.preprocessing import MinMaxScaler



#这个文件是用来在原始文件的基础上删除一些Trial,进行特征工程,然后再打上标签
def angle_between_vectors(v1, v2):
    """Calculate the angle between two vectors."""
    v1 = v1 / np.linalg.norm(v1)
    v2 = v2 / np.linalg.norm(v2)
    u_minus_v = v1 - v2
    u_plus_v = v1 + v2
    angle = 2 * np.arctan2(np.linalg.norm(u_minus_v), np.linalg.norm(u_plus_v))
    return np.degrees(angle)
def smooth_velocity(velocity, sigma=5):
    return gaussian_filter1d(velocity, sigma=sigma)
def calculate_angular_velocity(forward_vectors, timestamps):
    """计算角速度"""
    angles = np.array(
        [angle_between_vectors(forward_vectors[i], forward_vectors[i + 1]) if i + 1 < len(forward_vectors) else 0 for i
         in range(len(forward_vectors) - 1)])  # 注意这里的变化,排除了最后一个向量的计算
    # 计算时间间隔
    times = np.diff(timestamps)/1000
    # 计算角速度,对于第一个时间点,将角度设置为0
    if len(times) > 0:  # 检查times是否非空,避免除以空数组
        angular_velocities = np.insert(angles / times, 0, 0)  # 在结果数组的开始插入0
    else:
        angular_velocities = np.array([0])  # 如果times为空,说明只有一个时间点,无法计算速度
    return angular_velocities
def calculate_linear_velocity(positions, timestamps):
    """计算线性速度"""
    distances = np.linalg.norm(np.diff(positions, axis=0), axis=1)
    times = np.diff(timestamps)/1000
    linear_velocities = np.insert(distances / times, 0,0)  # Insert 0 at the start as there's no velocity at the first timestamp
    return linear_velocities
def calculate_direction(start_position, current_position):
    """计算方向"""
    direction = current_position - start_position
    norm = np.linalg.norm(direction)
    return direction / norm if norm != 0 else np.zeros_like(direction)
def calculate_acceleration(velocities, timestamps):
    """根据速度计算加速度"""
    accels = [0]  # 第一个时间点的加速度假设为0
    for i in range(1, len(velocities)):
        accel = (velocities[i] - velocities[i-1]) / ((timestamps[i] - timestamps[i-1])/1000)
        accels.append(accel)
    return np.array(accels)

## 特征工程,先加上三个模态的角速度,和旋转角,再加上手的线性速度和移动距离
def process_single_sequence(df):
    """处理groupBy之后的单个序列DataFrame,计算所有模态的角速度、线性速度、方向和旋转轴"""
    # 提取时间戳
    timestamps = df['TimeStamp'].values
    # 定义所有需要计算的模态
    modalities = ['HMD', 'Hand', 'Leye']
    for modality in modalities:
        # 计算角速度和旋转轴
        if all(f'{modality}ForwardV{i}' in df.columns for i in ['X', 'Y', 'Z']):
            forward_vectors = df[[f'{modality}ForwardVX', f'{modality}ForwardVY', f'{modality}ForwardVZ']].values
            initial_forward_vector = forward_vectors[0]
            df[f'{modality}A'] = [(angle_between_vectors(initial_forward_vector, fv)) for fv in forward_vectors]
            angular_velocity = calculate_angular_velocity(forward_vectors, timestamps)
            smoothed_angular_velocity = smooth_velocity(angular_velocity)
            df[f'{modality}AV'] = smoothed_angular_velocity
            df[f'{modality}AAcc'] = calculate_acceleration(smoothed_angular_velocity, timestamps)
            # 计算旋转轴并进行标准化以仅保留方向信息
            if modality == 'Hand':
                initial_forward_vector = forward_vectors[0]
                rotation_axes = np.array([np.cross(initial_forward_vector, fv) for fv in forward_vectors])
                rotation_axes_normalized = np.array(
                    [axis / np.linalg.norm(axis) if np.linalg.norm(axis) != 0 else np.array([0.0, 0.0, 0.0]) for axis in
                     rotation_axes]
                )
                df[f'{modality}RotationAxis_X'] = rotation_axes_normalized[:, 0]
                df[f'{modality}RotationAxis_Y'] = rotation_axes_normalized[:, 1]
                df[f'{modality}RotationAxis_Z'] = rotation_axes_normalized[:, 2]

        # 对于HMD和Hand,还需计算线性速度和方向还有移动距离
        if modality in ['HMD', 'Hand'] and all(f'{modality}Position{i}' in df.columns for i in ['X', 'Y', 'Z']):
            positions = df[[f'{modality}PositionX', f'{modality}PositionY', f'{modality}PositionZ']].values
            initial_position = positions[0]
            linear_velocity = calculate_linear_velocity(positions, timestamps)
            df[f'{modality}L'] = [np.linalg.norm(pos-initial_position) for pos in positions]
            smoothed_velocity = smooth_velocity(linear_velocity)
            df[f'{modality}LV'] = smoothed_velocity
            df[f'{modality}LAcc'] = calculate_acceleration(smoothed_velocity, timestamps)  # 新特性:加速度
            # 计算方向
            if modality == 'Hand':
                start_position = positions[0]
                directions = np.array([calculate_direction(start_position, pos) for pos in positions])
                df[f'{modality}Direction_X'] = directions[:, 0]
                df[f'{modality}Direction_Y'] = directions[:, 1]
                df[f'{modality}Direction_Z'] = directions[:, 2]

    return df


def label_trials_with_motion_metrics(df):
    # Define a helper function for labeling each group
    def label_group(group):
        # For the 'HandLinearDistance' and 'HandAngularDistance', take the last value in the group as it represents the total
        total_linear_distance = group['HandL'].iloc[-1]
        total_angular_distance = group['HandA'].iloc[-1]
        # Assign these totals to new columns for every row in the group
        group['LLabel'] = total_linear_distance
        group['ALabel'] = total_angular_distance
        return group
    # Apply the labeling function to each trial group
    df_labeled = df.groupby(['ParticipantID', 'BlockID', 'TrialID'], group_keys=False).apply(label_group)
    return df_labeled


def split_data_by_theta_grouped(df):
    grouped = df.groupby(['ParticipantID', 'BlockID', 'TrialID'])
    theta_groups = {}
    # 将每个组添加到对应Theta值的列表中
    for _, group in grouped:
        theta_value = group['Theta'].iloc[0]
        if theta_value not in theta_groups:
            theta_groups[theta_value] = []
        theta_groups[theta_value].append(group)
    train_dfs = []
    test_dfs = []
    for theta_value, groups in theta_groups.items():
        # 计算训练集大小
        n_train = int(len(groups) * 0.8)
        # 随机选择训练集序列
        np.random.seed(1)  # 确保可重复性
        train_indices = np.random.choice(len(groups), size=n_train, replace=False)
        train_groups = [groups[i] for i in train_indices]
        # 选择测试集序列
        test_groups = [groups[i] for i in range(len(groups)) if i not in train_indices]
        # 将训练集和测试集序列合并
        train_df = pd.concat(train_groups)
        test_df = pd.concat(test_groups)
        train_dfs.append(train_df)
        test_dfs.append(test_df)
    # 合并所有训练集和测试集的DataFrame
    final_train_df = pd.concat(train_dfs).sort_index()
    final_test_df = pd.concat(test_dfs).sort_index()
    return final_train_df, final_test_df


def split_data_by_theta_grouped(df):
    grouped = df.groupby(['ParticipantID', 'BlockID', 'TrialID'])
    theta_groups = {}
    # 将每个组添加到对应Theta值的列表中
    for _, group in grouped:
        theta_value = group['Theta'].iloc[0]
        if theta_value not in theta_groups:
            theta_groups[theta_value] = []
        theta_groups[theta_value].append(group)
    train_dfs = []
    test_dfs = []
    for theta_value, groups in theta_groups.items():
        # 计算训练集大小
        n_train = int(len(groups) * 0.8)
        # 随机选择训练集序列
        np.random.seed(1)  # 确保可重复性
        train_indices = np.random.choice(len(groups), size=n_train, replace=False)
        train_groups = [groups[i] for i in train_indices]
        # 选择测试集序列
        test_groups = [groups[i] for i in range(len(groups)) if i not in train_indices]
        # 将训练集和测试集序列合并
        train_df = pd.concat(train_groups)
        test_df = pd.concat(test_groups)
        train_dfs.append(train_df)
        test_dfs.append(test_df)
    # 合并所有训练集和测试集的DataFrame
    final_train_df = pd.concat(train_dfs).sort_index()
    final_test_df = pd.concat(test_dfs).sort_index()
    return final_train_df, final_test_df


# 定义一个函数,将特征缩放到指定的范围内
def preprocess_features(train_df, test_df):
    # 定义包含负值的特征列和其他特征列
    negative_value_features = ['HandAAcc', 'HMDAAcc',"LeyeAAcc","HandLAcc","HMDLAcc"]
    other_features = ["HMDA", "HMDAV", "HandA", "HandAV", "LeyeA", "LeyeAV","HMDL", "HMDLV","HandL", "HandLV"]
    # 为包含负值的特征和其他特征创建两个不同的缩放器
    scaler_negatives = MinMaxScaler(feature_range=(-1, 1))
    scaler_others = MinMaxScaler(feature_range=(0, 1))
    # 对训练集应用fit_transform,对测试集应用transform
    train_df[negative_value_features] = scaler_negatives.fit_transform(train_df[negative_value_features])
    test_df[negative_value_features] = scaler_negatives.transform(test_df[negative_value_features])
    train_df[other_features] = scaler_others.fit_transform(train_df[other_features])
    test_df[other_features] = scaler_others.transform(test_df[other_features])
    return train_df, test_df

 #这个function用于修改Evaluation的数据集
def main(Participant_ID):
    # 读入数据
    input_file_path = f'../Data/Study2Evaluation/{Participant_ID}_Trajectory.csv'
    output_train_path = f'../Data/Study2Evaluation/Preprocessed/{Participant_ID}_train_data_preprocessed_evaluation.csv'
    output_test_path = f'../Data/Study2Evaluation/Preprocessed/{Participant_ID}_test_data_preprocessed_evaluation.csv'

    data=pd.read_csv(input_file_path)
    data_cleaned = data.loc[:, ~data.columns.str.contains('^Unnamed')]
    data_no_error = data_cleaned[data_cleaned['isError'] == False]
    final_data = data_no_error[(data_no_error['TrialID'] != 0) & (data_no_error['TrialID'] != 6) & (data_no_error['TrialID'] != 12)
                               & (data_no_error['TrialID'] != 18) & (data_no_error['TrialID'] != 24) & (data_no_error['TrialID'] != 30) & (data_no_error['TrialID'] != 36)]
    # 特征工程
    processed_groups = final_data.groupby(['ParticipantID','BlockID', 'TrialID']).apply(process_single_sequence)
    processed_df = processed_groups.reset_index(drop=True)

    # 为处理完的数据添加标签
    df_with_features_labelled = label_trials_with_motion_metrics(processed_df.copy())
    final_train_df, final_test_df = split_data_by_theta_grouped(df_with_features_labelled)
    # 特征预处理
    final_train_df_preprocessed, final_test_df_preprocessed = preprocess_features(final_train_df.copy(), final_test_df.copy())
    # 保存处理后的数据集到CSV文件
    final_train_df_preprocessed.to_csv(output_train_path, index=False)
    final_test_df_preprocessed.to_csv(output_test_path, index=False)


# def main(Participant_ID):
#     # 读入数据
#     input_file_path = f'../Data/{Participant_ID}_Trajectory.csv'
#     output_train_path = f'../Data/Study1AllUSers/Preprocessed/{Participant_ID}_train_data_preprocessed.csv'
#     output_test_path = f'../Data/Study1AllUSers/Preprocessed/{Participant_ID}_test_data_preprocessed.csv'
#     data = pd. read_csv(input_file_path)
#     participant_id = int(os.path.basename(input_file_path).split('_')[0])
#     data.insert(0, 'ParticipantID', participant_id)
#     data_cleaned = data.loc[:, ~data.columns.str.contains('^Unnamed')]
#     data_no_error = data_cleaned[data_cleaned['isError'] == False]
#     cleaned_data_path = '../Data/cleaned_data_trimmed.xlsx'
#     cleaned_data = pd.read_excel(cleaned_data_path)
#     filtered_data = pd.merge(data_no_error, cleaned_data, on=['ParticipantID', 'BlockID', 'TrialID'], how='inner')
#     final_data = filtered_data[(filtered_data['TrialID'] != 0) & (filtered_data['TrialID'] != 8) & (filtered_data['TrialID'] != 16) & (filtered_data['TrialID'] != 24)]
#     # 特征工程
#     processed_groups = final_data.groupby(['ParticipantID','BlockID', 'TrialID']).apply(process_single_sequence)
#     processed_df = processed_groups.reset_index(drop=True)
#     # 为处理完的数据添加标签
#     df_with_features_labelled = label_trials_with_motion_metrics(processed_df.copy())
#     final_train_df, final_test_df = split_data_by_theta_grouped(df_with_features_labelled)
#     # 特征预处理
#     final_train_df_preprocessed, final_test_df_preprocessed = preprocess_features(final_train_df.copy(), final_test_df.copy())
#     # 保存处理后的数据集到CSV文件
#     final_train_df_preprocessed.to_csv(output_train_path, index=False)
#     final_test_df_preprocessed.to_csv(output_test_path, index=False)

if __name__ == '__main__':
    for i in range(79, 80):
        main(str(i))