diff --git "a/1791.jsonl" "b/1791.jsonl" new file mode 100644--- /dev/null +++ "b/1791.jsonl" @@ -0,0 +1,783 @@ +{"seq_id":"352345357","text":"import argparse\nimport os\nimport cv2\nimport numpy as np\nfrom tqdm import tqdm\n\ndef split_video(video_path, output_dir):\n # Load input video\n cap = cv2.VideoCapture(video_path)\n fps = int(cap.get(cv2.CAP_PROP_FPS))\n frame_width = int(cap.get(cv2.CAP_PROP_FRAME_WIDTH))\n frame_height = int(cap.get(cv2.CAP_PROP_FRAME_HEIGHT))\n length = int(cap.get(cv2.CAP_PROP_FRAME_COUNT))\n\n for i in tqdm(range(length)):\n ret, img = cap.read()\n if img is None:\n continue\n\n out_fn = os.path.join(output_dir, \"{:06d}.jpg\".format(i))\n cv2.imwrite(out_fn, img)\n\n cap.release()\n\nif __name__ == '__main__':\n parser = argparse.ArgumentParser()\n parser.add_argument('-v', '--video_path', help='The video path')\n parser.add_argument('-o', '--output_dir', help='The output dir', default=\"./output\")\n args = parser.parse_args()\n\n if not os.path.exists(args.output_dir):\n os.makedirs(args.output_dir)\n\n split_video(args.video_path, args.output_dir)","sub_path":"preprocess/LabelMeTools/src/virat/split_video.py","file_name":"split_video.py","file_ext":"py","file_size_in_byte":1008,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"14"} +{"seq_id":"635506816","text":"import os\nimport time\n\n\nimport allure\nimport pytest\nfrom selenium import webdriver\nfrom selenium.webdriver.chrome.options import Options\nfrom selenium.webdriver.common.by import By\n\n\n@allure.feature('百度搜索模块测试')\nclass TestBaiduSearch:\n def get_dir(self):\n \"\"\"\n 获取当前项目地址\n :return:\n \"\"\"\n now_dir = os.getcwd()\n while True:\n now_dir = os.path.split(now_dir)\n if now_dir[1] == 'test_baidu_search':\n now_dir = os.path.join(now_dir[0], 'test_baidu_search')\n break\n now_dir = now_dir[0]\n return now_dir\n\n def setup(self):\n \"\"\"前置动作\"\"\"\n # driver_path = os.path.join(self.get_dir(),'plugin/windows/chromedriver.exe') # Windows下使用\n driver_path = os.path.join(self.get_dir(), 'plugin/linux/chromedriver') # linux下使用\n option = Options()\n option.add_argument(\"--headless\") # linux下使用无头浏览器需要添加这个参数\n option.add_argument(\"--no-sandbox\") # 表示不用跟用户运行chrome\n option.add_argument(\"--disable-dev-shm-usage\") #\n self.driver = webdriver.Chrome(executable_path=driver_path,chrome_options=option)\n self.driver.maximize_window()\n self.driver.implicitly_wait(5)\n\n def teardown(self):\n \"\"\"后置动作\"\"\"\n self.driver.quit()\n\n\n @allure.story('百度搜索测试用例')\n @pytest.mark.parametrize(\"name\", [(\"狗\"), (\"猫\"), (\"欧毅\")])\n def test_baidu_search(self, name):\n self.driver.get(\"https://www.baidu.com/\")\n time.sleep(5)\n self.driver.find_element(By.ID, \"kw\").send_keys(f\"{name}\")\n self.driver.find_element(By.ID, \"su\").click()\n time.sleep(5)\n r = self.driver.title\n assert r == f\"{name}_百度搜索\"\n","sub_path":"test_baidu_search.py","file_name":"test_baidu_search.py","file_ext":"py","file_size_in_byte":1846,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"14"} +{"seq_id":"68299983","text":"from __future__ import division\n\nimport sklearn\nimport numpy as np\nimport os\n\nfrom scipy.special import comb\n\n\ndef compute_a_i(contingency):\n # Return a 1xI matrix corresponding to the marginal total of data points in each of the I predicted clusters\n rep = np.sum(contingency, axis=1)\n return np.reshape(rep, (np.ma.size(rep), 1))\n\n\ndef compute_b_j(contingency):\n # Return a Jx1 matrix corresponding to the marginal total of data points in each of the J true clusters\n rep = np.sum(contingency, axis=0)\n return np.reshape(rep, (1, np.ma.size(rep)))\n\n\ndef pair_based_values(contingency):\n # Return the quadruplet (SS, SD, DS, DD) corresponding to the 4 categories of data pairs\n # For conveniency (since all the criteria originally use the TP, FP, FN, TN values), the name of these variables\n # in pair_based_values() and functions that use results of pair_based_values() will be tp, fp, fn, tn\n # (correspondance : tp <-> SS, fp <-> SD, fn <-> DS, tn <-> DD)\n a_i = compute_a_i(contingency)\n b_j = compute_b_j(contingency)\n N = sum(a_i)\n\n tp = np.sum(comb(contingency, 2))\n J_a_i = np.repeat(a_i, np.ma.size(b_j), axis=1)\n fp = np.sum(np.multiply(J_a_i - contingency, contingency))\n fp = 0.5 * fp\n I_b_j = np.repeat(b_j, np.ma.size(a_i), axis=0)\n fn = np.sum(np.multiply(I_b_j - contingency, contingency))\n fn = 0.5 * fn\n\n tn = np.sum(np.multiply(contingency, N - J_a_i - I_b_j + contingency))\n tn = 0.5 * tn\n\n return tp, fp, fn, tn\n\n\ndef compute_contingency(y_true, y_pred):\n # Custom calculation of contingency table, since adopted notation require IxJ contingency matrix and sklearn contingency\n # function produce JxI matrix\n tmp = np.transpose(sklearn.metrics.cluster.contingency_matrix(y_true, y_pred, sparse=False))\n return np.array(tmp, dtype='float64') # Cast to avoid overflow problems\n\ndef pair_sum_test(y_true, y_pred):\n # Simple test to check if the sum of pairs tp,fp,fn,tn from pair_based_values() is equal to\n # total possible pair : N comb 2\n contingency = compute_contingency(y_true, y_pred)\n tp, fp, fn, tn = pair_based_values(contingency)\n N = np.ma.size(y_true)\n pair_sum = tp + fp + fn + tn\n real_pair_number = comb(N, 2)\n print(\"(TP+FP+FN+TN, comb(N,2) = \", pair_sum, real_pair_number)\n\n\ndef normalize(matrix):\n # Input : Matrix of values (integer/float only), without header/footer/etc, and return a normalized matrix (by\n # feature, ie by column), using Z normalization\n instance_nb = np.size(matrix, axis=0)\n\n mean = np.mean(matrix, axis=0)\n mean = np.expand_dims(mean, 0)\n mean = np.repeat(mean, instance_nb, axis=0)\n\n std = np.std(matrix, axis=0)\n std = np.expand_dims(std, 0)\n std = np.repeat(std, instance_nb, axis=0)\n normalized_matrix = (matrix - mean) / std # Z normalization\n normalized_matrix = np.nan_to_num(normalized_matrix)\n return normalized_matrix\n\n\ndef formate():\n # Note : Function that need to be customized for each dataset that needs to be cleaned\n\n common_directory_original = \"C:/Users/Vincent/Desktop/Stage PFE/Phase 3 - Tests sur datasets/Datasets bruts de l'UCI/\"\n clean_data_directory = \"C:/Users/Vincent/Desktop/Stage PFE/Phase 3 - Tests sur datasets/Datasets propres/\"\n\n sub_directory = \"fertility\"\n filename = \"fertility_Diagnosis.txt\"\n\n\n file = common_directory_original + sub_directory + \"/\" + filename\n f = open(file, 'r')\n\n matrix = []\n header = False\n for line in f:\n if header:\n header = False\n else:\n if line.find(\"?\") == -1 and line.find(\"NaN\") == -1 and line.find(\";;\") == -1 :\n # line = line.replace(\"\\n\",\"\")\n # str = line[-13:]\n # print(str)\n # if str == \"1\\t0\\t0\\t0\\t0\\t0\\t0\" :\n # line = line[0:-13] + \"0\"\n # if str == \"0\\t1\\t0\\t0\\t0\\t0\\t0\" :\n # line = line[0:-13] + \"1\"\n # if str == \"0\\t0\\t1\\t0\\t0\\t0\\t0\" :\n # line = line[0:-13] + \"2\"\n # if str == \"0\\t0\\t0\\t1\\t0\\t0\\t0\" :\n # line = line[0:-13] + \"3\"\n # if str == \"0\\t0\\t0\\t0\\t1\\t0\\t0\" :\n # line = line[0:-13] + \"4\"\n # if str == \"0\\t0\\t0\\t0\\t0\\t1\\t0\" :\n # line = line[0:-13] + \"5\"\n # if str == \"0\\t0\\t0\\t0\\t0\\t0\\t1\" :\n # line = line[0:-13] + \"6\"\n\n line = line.replace(\"O\",\"0\")\n line = line.replace(\"N\",\"1\")\n\n matrix = matrix + [[float(num) for num in line.split(',')]]\n\n nump = np.array(matrix)\n f.close()\n features = nump[:,:-1]\n featuresN = normalize(features)\n classes = np.array(nump[:,-1])\n\n if os.path.isfile(common_directory_original + sub_directory + \"/\" + \"features\"):\n warning = input(\"Ecraser ? (y/n)\")\n if warning == \"n\":\n return\n np.savetxt(common_directory_original + sub_directory + \"/\" + \"features\", featuresN, delimiter=',')\n np.savetxt(common_directory_original + sub_directory + \"/\" + \"classes\", classes, delimiter=',', fmt=\"%i\")\n\n np.savetxt(clean_data_directory + sub_directory +\"_features\", featuresN, delimiter=',')\n np.savetxt(clean_data_directory + sub_directory + \"_classes\", classes, delimiter=',', fmt=\"%i\")\n\ndef load_csv(path):\n # Given a file in csv format (seperator : ',') , return a file\n file = open(path,'r')\n matrix = []\n for line in file:\n matrix = matrix + [[float(num) for num in line.split(',')]]\n file.close()\n return matrix\n\ndef save_csv(path,data):\n # Save a matrix into a csv file\n np.savetxt(path, data, delimiter=',',fmt=\"%i\")","sub_path":"utility.py","file_name":"utility.py","file_ext":"py","file_size_in_byte":5736,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"14"} +{"seq_id":"419580879","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Wed Jul 27 16:12:30 2022\n\n@author: stevenweisberg\n\"\"\"\nimport os\nimport glob\nimport pandas as pd\nimport xlrd\nimport numpy as np\n\nscriptDir = os.getcwd()\n\n# Expects that the data is one file up from the analysis script\nos.chdir(os.path.dirname(os.getcwd()))\nmasterDir = os.getcwd()\n\n\noutdir_backup = os.path.join(masterDir,'DSP_RawData','Script_Output_DO_NOT_TOUCH')\n\nfiles = glob.glob(outdir_backup + os.sep + '*.xlsx')\n\ntallies = pd.DataFrame(columns={'participant','success_40','success_60'})\n\nfor file in files:\n df = pd.read_excel(file)\n \n success_40 = df.groupby('Status')['ParticipantNo'].count()['Success']\n\n \n conds = [\n (df['Status'] == 'Success') & (df['Time Elapsed'] > 39.9),\n (df['Status'] == 'Success') & (df['Time Elapsed'] < 39.9),\n (df['Status'] == 'Failure')\n ]\n \n values = ['Failure','Success','Failure']\n \n \n df['Status_60'] = np.select(conds,values)\n \n try:\n success_60 = df.groupby('Status_60')['ParticipantNo'].count()['Success']\n except:\n success_60 = 0\n \n p_tallies = {'participant':df['ParticipantNo'][0],\n 'success_40':success_40,\n 'success_60':success_60}\n \n tallies = tallies.append(p_tallies,ignore_index=True)\n \nprint('h')","sub_path":"dspy/archive/tallySuccess.py","file_name":"tallySuccess.py","file_ext":"py","file_size_in_byte":1323,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"14"} +{"seq_id":"499480150","text":"import tkinter\nimport turtle\n\nsc = tkinter.Tk()\nsc.geometry(\"1000x1000+100+100\")\n\nfr4 = tkinter.Frame(sc, height=500, width=600, bd=4, bg=\"light green\", takefocus=\"\", relief=tkinter.SUNKEN)\n\nfr4.grid(row=2, column=2, sticky=(tkinter.N, tkinter.E, tkinter.W, tkinter.S))\n\n# Canvas\ncanvas = tkinter.Canvas(fr4, width=750, height=750)\ncanvas.pack()\n\n# Turtle\nturtle1 = turtle.RawTurtle(canvas)\nturtle1.color(\"black\")\nturtle1.shape(\"circle\")\n\n\ndef drag_handler(x, y):\n turtle1.ondrag(None) # disable event inside event handler\n turtle1.goto(x, y)\n print(x,y)\n turtle1.ondrag(drag_handler) # reenable event on event handler exit\n\nturtle1.ondrag(drag_handler)\n\nsc.mainloop()","sub_path":"Remove/draw.py","file_name":"draw.py","file_ext":"py","file_size_in_byte":682,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"14"} +{"seq_id":"307304238","text":"import gzip\nimport json\nimport datetime\nimport numpy as np\nimport pandas as pd\nimport networkx as nx\n\nfrom collections import defaultdict\nfrom scipy.spatial.distance import cdist\nfrom scipy.spatial.distance import cosine\nfrom networkx.readwrite import json_graph\n\nfrom individual_mobility_network import entropy as calculate_entropy\nfrom mobility_distance_functions import spherical_distance\nfrom tak_quadtree import lon_lat_to_quadtree_path\n\n\nclass NumpyEncoder(json.JSONEncoder):\n \"\"\" Special json encoder for numpy types \"\"\"\n def default(self, obj):\n if isinstance(obj, (np.int_, np.intc, np.intp, np.int8, np.int16, np.int32, np.int64,\n np.uint8, np.uint16, np.uint32, np.uint64)):\n return int(obj)\n elif isinstance(obj, (np.float_, np.float16, np.float32, np.float64)):\n return float(obj)\n elif isinstance(obj, (np.ndarray,)):\n return obj.tolist()\n return json.JSONEncoder.default(self, obj)\n\n\ndef get_timeday(time):\n morning_from = datetime.datetime.strptime('06', '%H').time() # 6 hours\n afternoon_from = datetime.datetime.strptime('12', '%H').time() # 6 hours\n evening_from = datetime.datetime.strptime('18', '%H').time() # 4 hours\n night_from = datetime.datetime.strptime('22', '%H').time() # 8 hours\n if morning_from <= time < afternoon_from:\n return 0\n elif afternoon_from <= time < evening_from:\n return 1\n elif evening_from <= time < night_from:\n return 2\n else:\n return 3\n\n\ndef get_trjectory_features(trajectories):\n\n km_list = list()\n traveltime_list = list()\n speed_list = list()\n\n traj_per_day = defaultdict(int)\n km_per_day = defaultdict(float)\n traveltime_per_day = defaultdict(float)\n max_speed_per_day = dict()\n min_speed_per_day = dict()\n\n traj_per_timeday = defaultdict(int)\n km_per_timeday = defaultdict(float)\n traveltime_per_timeday = defaultdict(float)\n speed_per_timeday = defaultdict(list)\n\n for traj in trajectories.values():\n km = traj.length()\n duration = traj.duration()\n speed = km / (duration/3600)\n day = traj.start_time().date()\n time = traj.start_time().time()\n timeday = get_timeday(time)\n\n km_list.append(km)\n traveltime_list.append(duration)\n speed_list.append(speed)\n\n traj_per_day[day] += 1\n km_per_day[day] += km\n traveltime_per_day[day] += duration\n max_speed_per_day[day] = max(speed, max_speed_per_day[day]) if day in max_speed_per_day else speed\n min_speed_per_day[day] = min(speed, min_speed_per_day[day]) if day in min_speed_per_day else speed\n\n traj_per_timeday[timeday] += 1\n km_per_timeday[timeday] += km\n traveltime_per_timeday[timeday] += duration\n speed_per_timeday[timeday].append(speed)\n\n features = {\n 'tot_traj': len(trajectories),\n 'tot_km': np.sum(km_list),\n 'tot_traveltime': np.sum(traveltime_list),\n 'avg_km': np.mean(km_list),\n 'avg_traveltime': np.mean(traveltime_list),\n 'avg_speed': np.mean(speed_list),\n 'std_km': np.std(km_list),\n 'std_traveltime': np.std(traveltime_list),\n 'std_speed': np.std(speed_list),\n\n 'avg_traj_per_day': np.mean(list(traj_per_day.values())),\n 'avg_km_per_day': np.mean(list(km_per_day.values())),\n 'avg_traveltime_per_day': np.mean(list(traveltime_per_day.values())),\n 'avg_max_speed_per_day': np.mean(list(max_speed_per_day.values())),\n 'avg_min_speed_per_day': np.mean(list(min_speed_per_day.values())),\n 'std_traj_per_day': np.std(list(traj_per_day.values())),\n 'std_km_per_day': np.std(list(km_per_day.values())),\n 'std_traveltime_per_day': np.std(list(traveltime_per_day.values())),\n 'std_max_speed_per_day': np.std(list(max_speed_per_day.values())),\n 'std_min_speed_per_day': np.std(list(min_speed_per_day.values())),\n\n 'morning_traj': traj_per_timeday.get(0, 0),\n 'afternoon_traj': traj_per_timeday.get(1, 0),\n 'evening_traj': traj_per_timeday.get(2, 0),\n 'night_traj': traj_per_timeday.get(3, 0),\n 'morning_ntraj': traj_per_timeday.get(0, 0) / len(trajectories),\n 'afternoon_ntraj': traj_per_timeday.get(1, 0) / len(trajectories),\n 'evening_ntraj': traj_per_timeday.get(2, 0) / len(trajectories),\n 'night_ntraj': traj_per_timeday.get(3, 0) / len(trajectories),\n\n 'morning_km': km_per_timeday.get(0, 0.0),\n 'afternoon_km': km_per_timeday.get(1, 0.0),\n 'evening_km': km_per_timeday.get(2, 0.0),\n 'night_km': km_per_timeday.get(3, 0.0),\n 'morning_nkm': km_per_timeday.get(0, 0.0) / np.sum(km_list),\n 'afternoon_nkm': km_per_timeday.get(1, 0.0) / np.sum(km_list),\n 'evening_nkm': km_per_timeday.get(2, 0.0) / np.sum(km_list),\n 'night_nkm': km_per_timeday.get(3, 0.0) / np.sum(km_list),\n\n 'morning_ttime': traveltime_per_timeday.get(0, 0.0),\n 'afternoon_ttime': traveltime_per_timeday.get(1, 0.0),\n 'evening_ttime': traveltime_per_timeday.get(2, 0.0),\n 'night_ttime': traveltime_per_timeday.get(3, 0.0),\n 'morning_nttime': traveltime_per_timeday.get(0, 0.0) / np.sum(traveltime_list),\n 'afternoon_nttime': traveltime_per_timeday.get(1, 0.0) / np.sum(traveltime_list),\n 'evening_nttime': traveltime_per_timeday.get(2, 0.0) / np.sum(traveltime_list),\n 'night_nttime': traveltime_per_timeday.get(3, 0.0) / np.sum(traveltime_list),\n\n 'morning_avg_speed': np.mean(speed_per_timeday[0]) if 0 in speed_per_timeday else -1,\n 'afternoon_avg_speed': np.mean(speed_per_timeday[1]) if 1 in speed_per_timeday else -1,\n 'evening_avg_speed': np.mean(speed_per_timeday[2]) if 2 in speed_per_timeday else -1,\n 'night_avg_speed': np.mean(speed_per_timeday[3]) if 3 in speed_per_timeday else -1,\n 'morning_std_speed': np.std(speed_per_timeday[0]) if 0 in speed_per_timeday else -1,\n 'afternoon_std_speed': np.std(speed_per_timeday[1]) if 1 in speed_per_timeday else -1,\n 'evening_std_speed': np.std(speed_per_timeday[2]) if 2 in speed_per_timeday else -1,\n 'night_std_speed': np.std(speed_per_timeday[3]) if 3 in speed_per_timeday else -1,\n\n }\n\n return features\n\n\ndef get_events_features(events):\n nbr_events_per_day = defaultdict(int)\n nbr_events_per_time = defaultdict(int)\n\n nbr_events_type = defaultdict(int)\n nbr_event_type_per_day = defaultdict(lambda: defaultdict(int))\n nbr_event_type_per_time = defaultdict(lambda: defaultdict(int))\n\n nbr_event_location = defaultdict(int)\n nbr_event_location_per_day = defaultdict(lambda: defaultdict(int))\n nbr_event_location_per_time = defaultdict(lambda: defaultdict(int))\n\n nbr_events_type_locations = defaultdict(lambda: defaultdict(int))\n\n durations_list = list()\n durations_event_type = defaultdict(list)\n durations_event_location = defaultdict(list)\n\n avg_acc_list = list()\n avg_acc_event_type = defaultdict(list)\n avg_acc_event_location = defaultdict(list)\n\n max_acc_list = list()\n max_acc_event_type = defaultdict(list)\n max_acc_event_location = defaultdict(list)\n\n angle_list = list()\n angle_event_type = defaultdict(list)\n angle_event_location = defaultdict(list)\n\n for event in events.values():\n day = event['date'].date()\n time = event['date'].time()\n event_type = event['event_type']\n duration = event['duration']\n location = event['location_type']\n avg_acc = event['avg_acc']\n max_acc = event['max_acc']\n angle = event['angle']\n\n nbr_events_per_day[day] += 1\n nbr_events_per_time[time] += 1\n\n nbr_events_type[event_type] += 1\n nbr_event_type_per_day[event_type][day] += 1\n nbr_event_type_per_time[event_type][time] += 1\n\n nbr_event_location[location] += 1\n nbr_event_location_per_day[location][day] += 1\n nbr_event_location_per_time[location][time] += 1\n\n nbr_events_type_locations[event_type][location] += 1\n\n durations_list.append(duration)\n durations_event_type[event_type].append(duration)\n durations_event_location[location].append(duration)\n\n avg_acc_list.append(avg_acc)\n avg_acc_event_type[event_type].append(avg_acc)\n avg_acc_event_location[location].append(avg_acc)\n\n max_acc_list.append(max_acc)\n max_acc_event_type[event_type].append(max_acc)\n max_acc_event_location[location].append(max_acc)\n\n angle_list.append(angle)\n angle_event_type[event_type].append(angle)\n angle_event_location[location].append(angle)\n\n features = {\n 'tot_events': len(events),\n 'avg_events_per_day': np.mean(list(nbr_events_per_day.values())),\n 'std_events_per_day': np.std(list(nbr_events_per_day.values())),\n 'avg_events_per_time': np.mean(list(nbr_events_per_time.values())),\n 'std_events_per_time': np.std(list(nbr_events_per_time.values())),\n 'tot_duration': np.sum(durations_list),\n 'avg_duration': np.mean(durations_list),\n 'std_duration': np.std(durations_list),\n 'avg_avg_acc': np.mean(avg_acc_list),\n 'std_avg_acc': np.std(avg_acc_list),\n 'avg_max_acc': np.mean(max_acc_list),\n 'std_max_acc': np.std(max_acc_list),\n 'avg_angle': np.mean(angle_list),\n 'std_angle': np.std(angle_list),\n }\n\n for event_type in ['Q', 'B', 'A', 'C', 'stop', 'start']:\n features['tot_events_%s' % event_type] = nbr_events_type.get(event_type, 0)\n features['tot_events_p%s' % event_type] = \\\n nbr_events_type.get(event_type, 0) / len(events) if len(events) > 0 else 0.0\n features['avg_events_per_day_%s' % event_type] = np.mean(\n list(nbr_event_type_per_day[event_type].values())) if event_type in nbr_event_type_per_day else -1\n features['avg_events_per_time_%s' % event_type] = np.mean(\n list(nbr_event_type_per_time[event_type].values())) if event_type in nbr_event_type_per_time else -1\n features['std_events_per_day_%s' % event_type] = np.std(\n list(nbr_event_type_per_day[event_type].values())) if event_type in nbr_event_type_per_day else -1\n features['std_events_per_time_%s' % event_type] = np.std(\n list(nbr_event_type_per_time[event_type].values())) if event_type in nbr_event_type_per_time else -1\n\n features['tot_duration_%s' % event_type] = np.sum(\n durations_event_type[event_type]) if event_type in durations_event_type else 0\n features['tot_duration_p%s' % event_type] = np.sum(\n durations_event_type[event_type]) / np.sum(durations_list) if event_type in durations_event_type else 0\n features['avg_duration_%s' % event_type] = np.mean(\n durations_event_type[event_type]) if event_type in durations_event_type else -1\n features['std_duration_%s' % event_type] = np.std(\n durations_event_type[event_type]) if event_type in durations_event_type else -1\n\n features['avg_avg_acc_%s' % event_type] = np.mean(\n avg_acc_event_type[event_type]) if event_type in avg_acc_event_type else -1\n features['std_avg_acc_%s' % event_type] = np.std(\n avg_acc_event_type[event_type]) if event_type in avg_acc_event_type else -1\n\n features['avg_max_acc_%s' % event_type] = np.mean(\n max_acc_event_type[event_type]) if event_type in max_acc_event_type else -1\n features['std_max_acc_%s' % event_type] = np.std(\n max_acc_event_type[event_type]) if event_type in max_acc_event_type else -1\n\n features['avg_angle_%s' % event_type] = np.mean(\n angle_event_type[event_type]) if event_type in angle_event_type else -1\n features['std_angle_%s' % event_type] = np.std(\n angle_event_type[event_type]) if event_type in angle_event_type else -1\n\n for location in ['0', '1', '2']:\n features['tot_events_loc%s' % location] = nbr_event_location.get(location, 0)\n features['tot_events_ploc%s' % location] = \\\n nbr_event_location.get(location, 0) / len(events) if len(events) > 0 else 0.0\n features['avg_events_per_day_loc%s' % location] = np.mean(\n list(nbr_event_location_per_day[location].values())) if location in nbr_event_location_per_day else -1\n features['avg_events_per_time_loc%s' % location] = np.mean(\n list(nbr_event_location_per_time[location].values())) if location in nbr_event_location_per_time else -1\n features['std_events_per_day_loc%s' % location] = np.std(\n list(nbr_event_location_per_day[location].values())) if location in nbr_event_location_per_day else -1\n features['std_events_per_time_loc%s' % location] = np.std(\n list(nbr_event_location_per_time[location].values())) if location in nbr_event_location_per_time else -1\n\n features['tot_duration_loc%s' % location] = np.sum(\n durations_event_location[location]) if location in durations_event_location else 0\n features['tot_duration_ploc%s' % location] = np.sum(\n durations_event_location[location]) / np.sum(durations_list) if location in durations_event_location else 0\n features['avg_duration_loc%s' % location] = np.mean(\n durations_event_location[location]) if location in durations_event_location else -1\n features['std_duration_loc%s' % location] = np.std(\n durations_event_location[location]) if location in durations_event_location else -1\n\n features['avg_avg_acc_loc%s' % location] = np.mean(\n avg_acc_event_location[location]) if location in avg_acc_event_location else -1\n features['std_avg_acc_loc%s' % location] = np.std(\n avg_acc_event_location[location]) if location in avg_acc_event_location else -1\n\n features['avg_max_acc_loc%s' % location] = np.mean(\n max_acc_event_location[location]) if location in max_acc_event_location else -1\n features['std_max_acc_loc%s' % location] = np.std(\n max_acc_event_location[location]) if location in max_acc_event_location else -1\n\n features['avg_angle_loc%s' % location] = np.mean(\n angle_event_location[location]) if location in angle_event_location else -1\n features['std_angle_loc%s' % location] = np.std(\n angle_event_location[location]) if location in angle_event_location else -1\n\n for event_type in ['Q', 'B', 'A', 'C', 'stop', 'start']:\n if event_type in nbr_events_type_locations:\n for location in ['0', '1', '2']:\n if location in nbr_events_type_locations[event_type]:\n features['tot_events_type_%s_loc%s' % (event_type, location)] = \\\n nbr_events_type_locations[event_type][location]\n features['ptot_events_type_%s_loc%s' % (event_type, location)] = \\\n nbr_events_type_locations[event_type][location] / len(events)\n else:\n features['tot_events_type_%s_loc%s' % (event_type, location)] = 0\n features['ptot_events_type_%s_loc%s' % (event_type, location)] = 0.0\n else:\n for location in ['0', '1', '2']:\n features['tot_events_type_%s_loc%s' % (event_type, location)] = 0\n features['ptot_events_type_%s_loc%s' % (event_type, location)] = 0.0\n\n return features\n\n\ndef string2timedelta(s):\n if isinstance(s, str):\n t = datetime.datetime.strptime(s, '%H:%M:%S.%f') if '.' in s else datetime.datetime.strptime(s, '%H:%M:%S')\n return datetime.timedelta(hours=t.hour, minutes=t.minute, seconds=t.second)\n else:\n return s\n\n\ndef get_imn_temporal_features(imn_list, loc_dist_thr=100):\n\n imn_keys_sorted = sorted(list(imn_list.keys()))\n\n imn_key0 = imn_keys_sorted[0]\n imn_key1 = imn_keys_sorted[-1]\n imn0 = imn_list[imn_key0]\n imn1 = imn_list[imn_key1]\n if imn0 is None or imn1 is None:\n features = {\n 'delta_locations': np.nan,\n 'delta_movements': np.nan,\n 'delta_reg_locations': np.nan,\n 'delta_rg': np.nan,\n 'jaccard': np.nan,\n 'cosine': np.nan,\n 'jaccard_mov': np.nan,\n 'cosine_mov': np.nan,\n }\n\n return features\n\n delta_locations = imn1['n_locs'] - imn0['n_locs']\n delta_movements = imn1['n_movs'] - imn0['n_movs']\n delta_reg_locations = imn1['n_reg_locs'] - imn0['n_reg_locs']\n delta_rg = imn1['rg'] - imn0['rg']\n\n imn0loc = np.array([np.array(v) for v in imn0['location_prototype'].values()])\n imn1loc = np.array([np.array(v) for v in imn1['location_prototype'].values()])\n loc0 = np.arange(0, len(imn0loc))\n loc1 = np.arange(0, len(imn1loc)) + max(loc0) + 1\n sup0 = np.array([lf['loc_support'] for lid, lf in imn0['location_features'].items()])\n sup1 = np.array([lf['loc_support'] for lid, lf in imn1['location_features'].items()])\n\n graph0 = imn0['graph']\n graph1 = imn1['graph']\n if not isinstance(graph0, nx.DiGraph):\n graph0 = json_graph.node_link_graph(imn0['graph'], directed=True, multigraph=False,\n attrs={'link': 'edges', 'source': 'from', 'target': 'to'})\n\n if not isinstance(graph1, nx.DiGraph):\n graph1 = json_graph.node_link_graph(imn1['graph'], directed=True, multigraph=False,\n attrs={'link': 'edges', 'source': 'from', 'target': 'to'})\n\n sup0mov = np.array([mf['mov_support'] for mid, mf in imn0['mov_features'].items()])\n sup1mov = np.array([mf['mov_support'] for mid, mf in imn1['mov_features'].items()])\n\n dmatrix = cdist(imn0loc, imn1loc, metric=spherical_distance)\n mapping = dict()\n mapping_loc = dict()\n for j, i in enumerate(np.argmin(dmatrix, axis=0)):\n if dmatrix[i][j] < loc_dist_thr:\n if i not in mapping or dmatrix[i][j] < dmatrix[i][mapping[i]]:\n mapping[i] = j\n mapping_loc[loc0[i]] = loc1[j]\n\n for lid in loc0:\n if lid not in mapping_loc:\n mapping_loc[lid] = lid\n\n loc0to1set = set([mapping_loc[lid] for lid in loc0])\n loc1set = set(loc1)\n jaccard = 1 - len(loc0to1set & loc1set) / len(loc0to1set | loc1set)\n\n sup0remap = np.zeros(max(loc1) + 1)\n sup1remap = np.zeros(max(loc1) + 1)\n for i, v in enumerate(sup0):\n sup0remap[mapping_loc[i]] = v\n for j, v in enumerate(sup1):\n sup1remap[loc1[j]] = v\n\n cosined = cosine(sup0remap, sup1remap)\n graph0to1 = nx.relabel_nodes(graph0, mapping_loc, copy=True)\n mov0to1set = set(graph0to1.edges())\n mov1set = set(graph1.edges())\n jaccard_mov = 1 - len(mov0to1set & mov1set) / len(mov0to1set | mov1set)\n\n sup0mov_remap = np.zeros(len(mov0to1set | mov1set))\n sup1mov_remap = np.zeros(len(mov0to1set | mov1set))\n for lflt in graph0.edges():\n lflt_key = lflt\n if isinstance(list(imn0['location_from_to_movement'].keys())[0], str):\n lflt_key = str(lflt)\n mid = imn0['location_from_to_movement'][lflt_key]\n lflt1 = (mapping_loc[lflt[0]], mapping_loc[lflt[1]])\n index = list(graph0to1.edges()).index(lflt1)\n sup0mov_remap[index] = sup0mov[mid]\n for lflt in graph1.edges():\n lflt_key = lflt\n if isinstance(list(imn1['location_from_to_movement'].keys())[0], str):\n lflt_key = str(lflt)\n mid = imn1['location_from_to_movement'][lflt_key]\n index = list(graph1.edges()).index(lflt)\n sup1mov_remap[index] = sup1mov[mid]\n cosined_mov = cosine(sup0mov_remap, sup1mov_remap)\n\n features = {\n 'delta_locations': delta_locations,\n 'delta_movements': delta_movements,\n 'delta_reg_locations': delta_reg_locations,\n 'delta_rg': delta_rg,\n 'jaccard': jaccard,\n 'cosine': cosined,\n 'jaccard_mov': jaccard_mov,\n 'cosine_mov': cosined_mov,\n }\n\n return features\n\n\ndef get_imn_features(imn_list, event_traj2evntlist):\n\n nbr_locations = list()\n nbr_movements = list()\n nbr_reg_locations = list()\n nbr_reg_movements = list()\n radius_of_gyration = list()\n regular_radius_of_gyration = list()\n entropy = list()\n rentropy = list()\n avg_mov_length = list()\n std_mov_length = list()\n avg_mov_duration = list()\n std_mov_duration = list()\n avg_reg_mov_length = list()\n std_reg_mov_length = list()\n avg_reg_mov_duration = list()\n std_reg_mov_duration = list()\n\n density = list()\n triangles = list()\n clustering_coefficient = list()\n degree = list()\n indegree = list()\n outdegree = list()\n diameter = list()\n eccentricity = list()\n assortativity = list()\n\n l1_count = list()\n l2_count = list()\n l3_count = list()\n l1_indegree = list()\n l2_indegree = list()\n l3_indegree = list()\n l1_outdegree = list()\n l2_outdegree = list()\n l3_outdegree = list()\n l1_dcentrality = list()\n l2_dcentrality = list()\n l3_dcentrality = list()\n l1_bcentrality = list()\n l2_bcentrality = list()\n l3_bcentrality = list()\n l1_events = defaultdict(list)\n l2_events = defaultdict(list)\n l3_events = defaultdict(list)\n\n l1l2_count = list()\n l2l1_count = list()\n l1l3_count = list()\n l3l1_count = list()\n l2l3_count = list()\n l3l2_count = list()\n l1l2_betweenness = list()\n l2l1_betweenness = list()\n l1l3_betweenness = list()\n l3l1_betweenness = list()\n l2l3_betweenness = list()\n l3l2_betweenness = list()\n l1l2_events = defaultdict(list)\n l2l1_events = defaultdict(list)\n l1l3_events = defaultdict(list)\n l3l1_events = defaultdict(list)\n l2l3_events = defaultdict(list)\n l3l2_events = defaultdict(list)\n\n mov_event_entropy = defaultdict(list)\n\n for m0m1, imn in imn_list.items():\n if imn is None:\n continue\n # print(m0m1, imn.keys())\n # print(json.dumps(clear_tuples4json(imn), default=agenda_converter))\n nbr_locations.append(imn['n_locs'])\n nbr_movements.append(imn['n_movs'])\n nbr_reg_locations.append(imn['n_reg_locs'])\n nbr_reg_movements.append(imn['n_reg_movs'])\n radius_of_gyration.append(imn['rg'])\n regular_radius_of_gyration.append(imn['rrg'])\n entropy.append(imn['entropy'])\n rentropy.append(imn['rentropy'])\n avg_mov_length.append(imn['avg_mov_length'])\n std_mov_length.append(imn['std_mov_length'])\n avg_mov_duration.append(string2timedelta(imn['avg_mov_duration']).total_seconds())\n std_mov_duration.append(string2timedelta(imn['std_mov_duration']).total_seconds())\n avg_reg_mov_length.append(imn['avg_reg_mov_length'])\n std_reg_mov_length.append(imn['std_reg_mov_length'])\n avg_reg_mov_duration.append(string2timedelta(imn['avg_reg_mov_duration']).total_seconds())\n std_reg_mov_duration.append(string2timedelta(imn['std_reg_mov_duration']).total_seconds())\n\n graph = imn['graph']\n if not isinstance(graph, nx.DiGraph):\n graph = json_graph.node_link_graph(imn['graph'], directed=True, multigraph=False,\n attrs={'link': 'edges', 'source': 'from', 'target': 'to'})\n density.append(nx.density(graph))\n triangles.append(np.mean(list(nx.triangles(nx.to_undirected(graph)).values())))\n clustering_coefficient.append(nx.average_clustering(graph))\n degree.append(np.mean(list(dict(nx.to_undirected(graph).degree()).values())))\n indegree.append(np.mean(list(dict(graph.in_degree()).values())))\n outdegree.append(np.mean(list(dict(graph.out_degree()).values())))\n if nx.is_connected(nx.to_undirected(graph)):\n diameter.append(nx.diameter(nx.to_undirected(graph)))\n eccentricity.append(np.mean(list(nx.eccentricity(nx.to_undirected(graph)).values())))\n assortativity.append(nx.degree_assortativity_coefficient(nx.to_undirected(graph)))\n else:\n Gc = max(nx.connected_component_subgraphs(nx.to_undirected(graph)), key=len)\n diameter.append(nx.diameter(Gc))\n eccentricity.append(np.mean(list(nx.eccentricity(Gc).values())))\n assortativity.append(nx.degree_assortativity_coefficient(Gc))\n\n # print(imn['location_features'].keys())\n # print(list(imn['location_features'].keys())[0], type(list(imn['location_features'].keys())[0]))\n if isinstance(list(imn['location_features'].keys())[0], int):\n l1, l2, l3 = 0, 1, 2\n else:\n l1, l2, l3 = '0', '1', '2'\n\n l1_count.append(imn['location_features'][l1]['loc_support'])\n l2_count.append(imn['location_features'][l2]['loc_support'])\n if l3 in imn['location_features']:\n l3_count.append(imn['location_features'][l3]['loc_support'])\n in_degree = dict(graph.in_degree())\n out_degree = dict(graph.out_degree())\n dcentrality = nx.degree_centrality(graph)\n bcentrality = nx.betweenness_centrality(graph)\n l1_indegree.append(in_degree[0])\n l2_indegree.append(in_degree[1])\n if 2 in in_degree:\n l3_indegree.append(in_degree[2])\n l1_outdegree.append(out_degree[0])\n l2_outdegree.append(out_degree[1])\n if 2 in out_degree:\n l3_outdegree.append(out_degree[2])\n l1_dcentrality.append(dcentrality[0])\n l2_dcentrality.append(dcentrality[1])\n if 2 in dcentrality:\n l3_dcentrality.append(dcentrality[2])\n l1_bcentrality.append(bcentrality[0])\n l2_bcentrality.append(bcentrality[1])\n if 2 in bcentrality:\n l3_bcentrality.append(bcentrality[2])\n\n l1_nbr_events_type = defaultdict(int)\n l2_nbr_events_type = defaultdict(int)\n l3_nbr_events_type = defaultdict(int)\n\n l1l2_nbr_events_type = defaultdict(int)\n l2l1_nbr_events_type = defaultdict(int)\n l1l3_nbr_events_type = defaultdict(int)\n l3l1_nbr_events_type = defaultdict(int)\n l2l3_nbr_events_type = defaultdict(int)\n l3l2_nbr_events_type = defaultdict(int)\n\n mov_event_count = defaultdict(lambda: defaultdict(int))\n\n for tid in imn['traj_location_from_to']:\n for evnt in event_traj2evntlist[tid]:\n if imn['traj_location_from_to'][tid][1] == 0:\n l1_nbr_events_type[evnt['event_type']] += 1\n elif imn['traj_location_from_to'][tid][1] == 1:\n l2_nbr_events_type[evnt['event_type']] += 1\n elif imn['traj_location_from_to'][tid][1] == 2:\n l3_nbr_events_type[evnt['event_type']] += 1\n\n if imn['traj_location_from_to'][tid][0] == 0 and imn['traj_location_from_to'][tid][1] == 1:\n l1l2_nbr_events_type[evnt['event_type']] += 1\n elif imn['traj_location_from_to'][tid][0] == 1 and imn['traj_location_from_to'][tid][1] == 0:\n l2l1_nbr_events_type[evnt['event_type']] += 1\n elif imn['traj_location_from_to'][tid][0] == 0 and imn['traj_location_from_to'][tid][1] == 2:\n l1l3_nbr_events_type[evnt['event_type']] += 1\n elif imn['traj_location_from_to'][tid][0] == 2 and imn['traj_location_from_to'][tid][1] == 0:\n l3l1_nbr_events_type[evnt['event_type']] += 1\n elif imn['traj_location_from_to'][tid][0] == 1 and imn['traj_location_from_to'][tid][1] == 2:\n l2l3_nbr_events_type[evnt['event_type']] += 1\n elif imn['traj_location_from_to'][tid][0] == 2 and imn['traj_location_from_to'][tid][1] == 1:\n l3l2_nbr_events_type[evnt['event_type']] += 1\n\n lft = imn['traj_location_from_to'][tid][1]\n mov_event_count[evnt['event_type']][lft] += 1\n\n for event_type in ['Q', 'B', 'A', 'C', 'stop', 'start']:\n if event_type in l1_nbr_events_type:\n l1_events[event_type].append(l1_nbr_events_type[event_type])\n else:\n l1_events[event_type].append(0)\n if event_type in l2_nbr_events_type:\n l2_events[event_type].append(l2_nbr_events_type[event_type])\n else:\n l2_events[event_type].append(0)\n if event_type in l3_nbr_events_type:\n l3_events[event_type].append(l3_nbr_events_type[event_type])\n else:\n l3_events[event_type].append(0)\n\n if event_type in l1l2_nbr_events_type:\n l1l2_events[event_type].append(l1l2_nbr_events_type[event_type])\n else:\n l1l2_events[event_type].append(0)\n if event_type in l2l1_nbr_events_type:\n l2l1_events[event_type].append(l2l1_nbr_events_type[event_type])\n else:\n l2l1_events[event_type].append(0)\n\n if event_type in l1l3_nbr_events_type:\n l1l3_events[event_type].append(l1l3_nbr_events_type[event_type])\n else:\n l1l3_events[event_type].append(0)\n if event_type in l3l1_nbr_events_type:\n l3l1_events[event_type].append(l3l1_nbr_events_type[event_type])\n else:\n l3l1_events[event_type].append(0)\n\n if event_type in l2l3_nbr_events_type:\n l2l3_events[event_type].append(l2l3_nbr_events_type[event_type])\n else:\n l2l3_events[event_type].append(0)\n if event_type in l3l1_nbr_events_type:\n l3l2_events[event_type].append(l3l2_nbr_events_type[event_type])\n else:\n l3l2_events[event_type].append(0)\n\n if event_type in mov_event_count:\n vals = list(mov_event_count[event_type].values())\n mov_event_entropy[event_type].append(calculate_entropy(vals, classes=len(vals)))\n else:\n mov_event_entropy[event_type].append(0.0)\n\n l1l2_count.append(imn['location_nextlocs'][l1].get(l2, 0))\n l2l1_count.append(imn['location_nextlocs'][l2].get(l1, 0))\n l1l3_count.append(imn['location_nextlocs'][l1].get(l3, 0))\n if '2' in imn['location_nextlocs']:\n l3l1_count.append(imn['location_nextlocs'][l3].get(l1, 0))\n l2l3_count.append(imn['location_nextlocs'][l2].get(l3, 0))\n l3l2_count.append(imn['location_nextlocs'][l3].get(l2, 0))\n else:\n l3l1_count.append(0)\n l2l3_count.append(0)\n l3l2_count.append(0)\n edge_betweenness = nx.edge_betweenness(graph)\n l1l2_betweenness.append(edge_betweenness.get((0, 1), 0))\n l2l1_betweenness.append(edge_betweenness.get((1, 0), 0))\n l1l3_betweenness.append(edge_betweenness.get((0, 2), 0))\n l3l1_betweenness.append(edge_betweenness.get((2, 0), 0))\n l2l3_betweenness.append(edge_betweenness.get((1, 2), 0))\n l3l2_betweenness.append(edge_betweenness.get((2, 1), 0))\n\n imn_temporal_features = get_imn_temporal_features(imn_list)\n\n features = {\n 'nbr_locations': np.mean(nbr_locations),\n 'nbr_movements': np.mean(nbr_movements),\n 'nbr_reg_locations': np.mean(nbr_reg_locations),\n 'nbr_reg_movements': np.mean(nbr_reg_movements),\n 'radius_of_gyration': np.mean(radius_of_gyration),\n 'regular_radius_of_gyration': np.mean(regular_radius_of_gyration),\n 'entropy': np.mean(entropy),\n 'rentropy': np.mean(rentropy),\n 'avg_mov_length': np.mean(avg_mov_length),\n 'std_mov_length': np.mean(std_mov_length),\n 'avg_mov_duration': np.mean(avg_mov_duration),\n 'std_mov_duration': np.mean(std_mov_duration),\n # 'avg_reg_mov_length': np.mean(avg_reg_mov_length),\n # 'std_reg_mov_length': np.mean(std_reg_mov_length),\n 'avg_reg_mov_duration': np.mean(avg_reg_mov_duration),\n 'std_reg_mov_duration': np.mean(std_reg_mov_duration),\n\n 'density': np.mean(density),\n 'triangles': np.mean(triangles),\n 'clustering_coefficient': np.mean(clustering_coefficient),\n 'avg_degree': np.mean(degree),\n 'avg_indegree': np.mean(indegree),\n 'avg_outdegree': np.mean(outdegree),\n 'diameter': np.mean(diameter),\n 'eccentricity': np.mean(eccentricity),\n 'assortativity': np.mean(assortativity),\n\n 'l1_count': np.mean(l1_count),\n 'l2_count': np.mean(l2_count),\n 'l3_count': np.mean(l3_count),\n 'l1_indegree': np.mean(l1_indegree),\n 'l2_indegree': np.mean(l2_indegree),\n 'l3_indegree': np.mean(l3_indegree),\n 'l1_outdegree': np.mean(l1_outdegree),\n 'l2_outdegree': np.mean(l2_outdegree),\n 'l3_outdegree': np.mean(l3_outdegree),\n 'l1_dcentrality': np.mean(l1_dcentrality),\n 'l2_dcentrality': np.mean(l2_dcentrality),\n 'l3_dcentrality': np.mean(l3_dcentrality),\n 'l1_bcentrality': np.mean(l1_bcentrality),\n 'l2_bcentrality': np.mean(l2_bcentrality),\n 'l3_bcentrality': np.mean(l3_bcentrality),\n\n 'l1l2_count': np.mean(l1l2_count),\n 'l2l1_count': np.mean(l2l1_count),\n 'l1l3_count': np.mean(l1l3_count),\n 'l3l1_count': np.mean(l3l1_count),\n 'l2l3_count': np.mean(l2l3_count),\n 'l3l2_count': np.mean(l3l2_count),\n 'l1l2_betweenness': np.mean(l1l2_betweenness),\n 'l2l1_betweenness': np.mean(l2l1_betweenness),\n 'l1l3_betweenness': np.mean(l1l3_betweenness),\n 'l3l1_betweenness': np.mean(l3l1_betweenness),\n 'l2l3_betweenness': np.mean(l2l3_betweenness),\n 'l3l2_betweenness': np.mean(l3l2_betweenness),\n }\n\n features.update(imn_temporal_features)\n\n for event_type in ['Q', 'B', 'A', 'C', 'stop', 'start']:\n features['l1_%s' % event_type] = np.mean(l1_events[event_type])\n features['l2_%s' % event_type] = np.mean(l2_events[event_type])\n features['l3_%s' % event_type] = np.mean(l3_events[event_type])\n features['l1l2_%s' % event_type] = np.mean(l1l2_events[event_type])\n features['l2l1_%s' % event_type] = np.mean(l2l1_events[event_type])\n features['l1l3_%s' % event_type] = np.mean(l1l3_events[event_type])\n features['l3l1_%s' % event_type] = np.mean(l3l1_events[event_type])\n features['l2l3_%s' % event_type] = np.mean(l2l3_events[event_type])\n features['l3l2_%s' % event_type] = np.mean(l3l2_events[event_type])\n features['mov_entropy_%s' % event_type] = np.mean(mov_event_entropy[event_type])\n\n for k, v in features.items():\n if np.isnan(v):\n features[k] = -1\n\n return features\n\n\ndef path_in_tree(tree, path, max_depth=16):\n idx = 0\n node = tree\n while True:\n if node['is_leaf'] or idx == len(path) or (max_depth is not None and node['depth'] >= max_depth):\n break\n code = path[idx]\n # print(idx, code)\n if code not in node:\n # print('path not found')\n break\n node = node[code]\n idx += 1\n\n return path[:idx]\n\n\ndef get_collective_features(trajectories, imn_list, quadtree, quadtree_features):\n\n features_path_count = defaultdict(int)\n for traj in trajectories.values():\n paths_of_this_traj = set()\n for i, point in enumerate(traj.object):\n lon, lat, _ = point\n path = lon_lat_to_quadtree_path(lon, lat, depth=16)\n if path not in paths_of_this_traj:\n features_path_count[path] += 1\n paths_of_this_traj.add(path)\n\n is_regular_path = {path: False for path in features_path_count}\n for m0m1, imn in imn_list.items():\n if imn is None:\n continue\n if len(imn['regular_locations']) > 0:\n for lid in imn['regular_locations']:\n if isinstance(list(imn['location_prototype'].keys())[0], str):\n lid = str(lid)\n lon, lat = imn['location_prototype'][lid]\n path = lon_lat_to_quadtree_path(lon, lat, depth=16)\n is_regular_path[path] = True\n\n for mid, movement_traj in imn['movement_traj'].items():\n lft = movement_traj[0]\n if lft[0] in imn['regular_locations'] and lft[1] in imn['regular_locations']:\n if isinstance(imn['movement_prototype'][mid], dict):\n traj_object = imn['movement_prototype'][mid]['object']\n else:\n traj_object = imn['movement_prototype'][mid].object\n for i, point in enumerate(traj_object):\n lon, lat, _ = point\n path = lon_lat_to_quadtree_path(lon, lat, depth=16)\n is_regular_path[path] = True\n\n aggregated_quadtree_features_reg = dict()\n aggregated_quadtree_features_irrreg = dict()\n aggregated_quadtree_features_reg_count = dict()\n aggregated_quadtree_features_irrreg_count = dict()\n for path in features_path_count:\n if path not in quadtree_features:\n continue\n apath = path_in_tree(quadtree, path, max_depth=16)\n if is_regular_path[path]:\n if apath not in aggregated_quadtree_features_reg:\n aggregated_quadtree_features_reg[apath] = quadtree_features[path]\n aggregated_quadtree_features_reg_count[apath] = features_path_count[path]\n else:\n for k, v in quadtree_features[path].items():\n aggregated_quadtree_features_reg[apath][k] += v\n aggregated_quadtree_features_reg_count[apath] += features_path_count[path]\n else:\n if apath not in aggregated_quadtree_features_irrreg:\n aggregated_quadtree_features_irrreg[apath] = quadtree_features[path]\n aggregated_quadtree_features_irrreg_count[apath] = features_path_count[path]\n else:\n for k, v in quadtree_features[path].items():\n aggregated_quadtree_features_irrreg[apath][k] += v\n aggregated_quadtree_features_irrreg_count[apath] += features_path_count[path]\n\n aggregated_quadtree_features_reg_comb = dict()\n aggregated_quadtree_features_irrreg_comb = dict()\n for aqf, aqfc in zip([aggregated_quadtree_features_reg, aggregated_quadtree_features_irrreg],\n [aggregated_quadtree_features_reg_comb, aggregated_quadtree_features_irrreg_comb]):\n for path in aqf:\n aqfc[path] = {\n 'nbr_traj_start': aqf[path]['nbr_traj_start'],\n 'nbr_traj_stop': aqf[path]['nbr_traj_stop'],\n 'nbr_traj_move': aqf[path]['nbr_traj_move'],\n 'avg_traj_speed': aqf[path]['traj_speed_count'] / aqf[path]['traj_speed_count']\n if aqf[path]['traj_speed_count'] > 0 else 0,\n 'nbr_evnt_A': aqf[path]['nbr_evnt_A'],\n 'nbr_evnt_B': aqf[path]['nbr_evnt_B'],\n 'nbr_evnt_C': aqf[path]['nbr_evnt_C'],\n 'nbr_evnt_Q': aqf[path]['nbr_evnt_Q'],\n 'nbr_evnt_start': aqf[path]['nbr_evnt_start'],\n 'nbr_evnt_stop': aqf[path]['nbr_evnt_stop'],\n 'avg_speed_A': aqf[path]['speed_A_sum'] / aqf[path]['nbr_evnt_A'] if aqf[path]['nbr_evnt_A'] > 0 else 0,\n 'avg_max_acc_A': aqf[path]['max_acc_A_sum'] / aqf[path]['nbr_evnt_A'] if aqf[path]['nbr_evnt_A'] > 0 else 0,\n 'avg_avg_acc_A': aqf[path]['avg_acc_A_sum'] / aqf[path]['nbr_evnt_A'] if aqf[path]['nbr_evnt_A'] > 0 else 0,\n 'avg_speed_B': aqf[path]['speed_B_sum'] / aqf[path]['nbr_evnt_B'] if aqf[path]['nbr_evnt_B'] > 0 else 0,\n 'avg_max_acc_B': aqf[path]['max_acc_B_sum'] / aqf[path]['nbr_evnt_B'] if aqf[path]['nbr_evnt_B'] > 0 else 0,\n 'avg_avg_acc_B': aqf[path]['avg_acc_B_sum'] / aqf[path]['nbr_evnt_B'] if aqf[path]['nbr_evnt_B'] > 0 else 0,\n 'avg_speed_C': aqf[path]['speed_C_sum'] / aqf[path]['nbr_evnt_C'] if aqf[path]['nbr_evnt_C'] > 0 else 0,\n 'avg_max_acc_C': aqf[path]['max_acc_C_sum'] / aqf[path]['nbr_evnt_C'] if aqf[path]['nbr_evnt_C'] > 0 else 0,\n 'avg_avg_acc_C': aqf[path]['avg_acc_C_sum'] / aqf[path]['nbr_evnt_C'] if aqf[path]['nbr_evnt_C'] > 0 else 0,\n 'avg_speed_Q': aqf[path]['speed_Q_sum'] / aqf[path]['nbr_evnt_Q'] if aqf[path]['nbr_evnt_Q'] > 0 else 0,\n 'avg_max_acc_Q': aqf[path]['max_acc_Q_sum'] / aqf[path]['nbr_evnt_Q'] if aqf[path]['nbr_evnt_Q'] > 0 else 0,\n 'avg_avg_acc_Q': aqf[path]['avg_acc_Q_sum'] / aqf[path]['nbr_evnt_Q'] if aqf[path]['nbr_evnt_Q'] > 0 else 0,\n 'nbr_crash': aqf[path]['nbr_crash'],\n }\n aggregated_quadtree_features_reg = aggregated_quadtree_features_reg_comb\n aggregated_quadtree_features_irrreg = aggregated_quadtree_features_irrreg_comb\n\n features = defaultdict(float)\n total_reg = np.sum(list(aggregated_quadtree_features_reg_count.values()))\n for path, values in aggregated_quadtree_features_reg.items():\n count = aggregated_quadtree_features_reg_count[path]\n for k, v in values.items():\n if total_reg > 0:\n features['reg_%s' % k] += v * count / total_reg\n else:\n features['reg_%s' % k] += 0\n total_occ = np.sum(list(aggregated_quadtree_features_irrreg_count.values()))\n for path, values in aggregated_quadtree_features_irrreg.items():\n count = aggregated_quadtree_features_irrreg_count[path]\n for k, v in values.items():\n if total_occ > 0:\n features['occ_%s' % k] += v * count / total_occ\n else:\n features['occ_%s' % k] += 0\n\n return features\n\n\ndef extract_features_data(uid, data, quadtree, quadtree_features):\n\n features = dict()\n for index, values in data.items():\n trajectories = values['trajectories']\n events = values['events']\n imn_list = values['imns']\n event_traj2evntlist = defaultdict(list)\n for eid, evnt in events.items():\n event_traj2evntlist[evnt['tid']].append(evnt)\n\n if len(trajectories) == 0:\n continue\n\n traj_features = get_trjectory_features(trajectories)\n evnt_features = get_events_features(events)\n imn_features = get_imn_features(imn_list, event_traj2evntlist)\n collective_features = get_collective_features(trajectories, imn_list, quadtree, quadtree_features[index])\n\n features[index] = {\n 'uid': uid,\n 'crash': values['crash'],\n }\n\n features[index].update(traj_features)\n features[index].update(evnt_features)\n features[index].update(imn_features)\n features[index].update(collective_features)\n\n return features\n\n\ndef extract_features(uid, tr_data, ts_data, quadtree, tr_quadtree_features, ts_quadtree_features):\n # print('train')\n training = extract_features_data(uid, tr_data, quadtree, tr_quadtree_features)\n # print('test')\n test = extract_features_data(uid, ts_data, quadtree, ts_quadtree_features)\n return training, test\n\n\ndef store_features(filename, store_obj):\n json_str = '%s\\n' % json.dumps(store_obj, cls=NumpyEncoder)\n json_bytes = json_str.encode('utf-8')\n # print(json_str)\n with gzip.GzipFile(filename, 'a') as fout:\n fout.write(json_bytes)\n\n\n# def main():\n# area = sys.argv[1]\n# country = 'uk' if area == 'london' else 'italy'\n#\n# path = './'\n# path_dataset = path + 'dataset/'\n# path_imn = path + 'imn/'\n# path_crash = path + 'crash/'\n#\n# crash_users_filename = path_dataset + '%s_users_list.csv' % area\n# nocrash_users_filename = path_dataset + '%s_nocrash_users_list.csv' % area\n#\n# crash_users_list = sorted(pd.read_csv(crash_users_filename).values[:, 0].tolist())\n# nocrash_users_list = sorted(pd.read_csv(nocrash_users_filename).values[:, 0].tolist())\n#\n#\n#\n# if __name__ == \"__main__\":\n# main()\n","sub_path":"code/feature_extractor.py","file_name":"feature_extractor.py","file_ext":"py","file_size_in_byte":44140,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"14"} +{"seq_id":"25435205","text":"API_MAPPINGS = {\n\n \"dummy_get\": {\n \"params\": ['p1', 'p2'],\n \"API_method\": ['GET'],\n \"third_party_url\": \"http://0.0.0.0:8080/test_get1\",\n \"output_format\": \"JSON\"\n },\n\n \"dummy_post\": {\n \"params\": [\"p1\", \"p2\"],\n \"API_method\": ['POST'],\n \"third_party_url\": \"http://0.0.0.0:8080/test_post1\",\n \"output_format\": \"JSON\",\n }\n\n}\n","sub_path":"Non Unified URLs/Architecture_one/app/v2/mapping.py","file_name":"mapping.py","file_ext":"py","file_size_in_byte":387,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"14"} +{"seq_id":"312382294","text":"# I decided to beat the exercise from the workbook!\n\nimport requests\nimport pandas as pd\nfrom bs4 import BeautifulSoup\n\npage = requests.get(\"https://forecast.weather.gov/MapClick.php?lat=37.777120000000025&lon=-122.41963999999996#.X9DVpBakolQ\")\nsoup = BeautifulSoup(page.content, 'html.parser')\n\n# first filtering we do over
pip install \" + package_info['name'] + \"\")\n\n official_site = ''\n # check for real links. We can get stuff like 'unknown', '404' in here\n if package_info['home_page'] and re.search(r'www.', package_info['home_page']):\n official_site = '[' + package_info['home_page'] + ' Official site]\\\\\\\\n'\n\n out_file.write('\\t'.join([\n package_info['name'], # Title\n 'A', # Article type\n '', # No redirect\n '', # Other uses (ignored)\n '', # No categories\n '', # References (ignored)\n '', # No related topics\n '', # Further reading (ignored)\n official_site, # External links (ignored)\n '', # Disambiguation (ignored)\n '', # No images\n '